var/home/core/zuul-output/0000755000175000017500000000000015111054341014520 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015111101636015464 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000006441727415111101626017706 0ustar rootrootNov 24 13:12:30 crc systemd[1]: Starting Kubernetes Kubelet... Nov 24 13:12:30 crc restorecon[4755]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:12:30 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:12:31 crc restorecon[4755]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:12:31 crc restorecon[4755]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 24 13:12:32 crc kubenswrapper[4790]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 13:12:32 crc kubenswrapper[4790]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 24 13:12:32 crc kubenswrapper[4790]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 13:12:32 crc kubenswrapper[4790]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 13:12:32 crc kubenswrapper[4790]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 24 13:12:32 crc kubenswrapper[4790]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.041310 4790 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.049815 4790 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.049845 4790 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.049857 4790 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.049866 4790 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.049875 4790 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.049939 4790 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.049950 4790 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.049960 4790 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.049970 4790 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.049978 4790 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.049989 4790 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.049999 4790 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050007 4790 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050017 4790 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050025 4790 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050033 4790 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050041 4790 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050049 4790 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050056 4790 feature_gate.go:330] unrecognized feature gate: Example Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050064 4790 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050072 4790 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050080 4790 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050087 4790 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050095 4790 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050103 4790 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050110 4790 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050118 4790 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050126 4790 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050133 4790 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050141 4790 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050149 4790 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050162 4790 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050171 4790 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050179 4790 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050188 4790 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050197 4790 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050207 4790 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050215 4790 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050224 4790 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050232 4790 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050239 4790 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050247 4790 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050254 4790 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050262 4790 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050273 4790 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050282 4790 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050289 4790 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050297 4790 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050305 4790 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050313 4790 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050320 4790 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050328 4790 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050335 4790 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050343 4790 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050354 4790 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050362 4790 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050370 4790 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050378 4790 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050385 4790 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050393 4790 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050401 4790 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050408 4790 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050416 4790 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050423 4790 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050431 4790 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050439 4790 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050446 4790 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050454 4790 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050462 4790 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050469 4790 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.050476 4790 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050606 4790 flags.go:64] FLAG: --address="0.0.0.0" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050622 4790 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050635 4790 flags.go:64] FLAG: --anonymous-auth="true" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050646 4790 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050657 4790 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050667 4790 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050678 4790 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050688 4790 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050697 4790 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050706 4790 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050716 4790 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050726 4790 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050735 4790 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050744 4790 flags.go:64] FLAG: --cgroup-root="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050753 4790 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050763 4790 flags.go:64] FLAG: --client-ca-file="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050772 4790 flags.go:64] FLAG: --cloud-config="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050780 4790 flags.go:64] FLAG: --cloud-provider="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050789 4790 flags.go:64] FLAG: --cluster-dns="[]" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050802 4790 flags.go:64] FLAG: --cluster-domain="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050811 4790 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050820 4790 flags.go:64] FLAG: --config-dir="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050829 4790 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050839 4790 flags.go:64] FLAG: --container-log-max-files="5" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050849 4790 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050858 4790 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050867 4790 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050877 4790 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050909 4790 flags.go:64] FLAG: --contention-profiling="false" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050918 4790 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050927 4790 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050937 4790 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050945 4790 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050956 4790 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050965 4790 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050974 4790 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050983 4790 flags.go:64] FLAG: --enable-load-reader="false" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.050992 4790 flags.go:64] FLAG: --enable-server="true" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051001 4790 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051012 4790 flags.go:64] FLAG: --event-burst="100" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051021 4790 flags.go:64] FLAG: --event-qps="50" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051030 4790 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051039 4790 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051048 4790 flags.go:64] FLAG: --eviction-hard="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051059 4790 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051068 4790 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051077 4790 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051087 4790 flags.go:64] FLAG: --eviction-soft="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051096 4790 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051105 4790 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051113 4790 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051122 4790 flags.go:64] FLAG: --experimental-mounter-path="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051131 4790 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051140 4790 flags.go:64] FLAG: --fail-swap-on="true" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051149 4790 flags.go:64] FLAG: --feature-gates="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051161 4790 flags.go:64] FLAG: --file-check-frequency="20s" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051170 4790 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051179 4790 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051188 4790 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051197 4790 flags.go:64] FLAG: --healthz-port="10248" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051206 4790 flags.go:64] FLAG: --help="false" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051215 4790 flags.go:64] FLAG: --hostname-override="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051224 4790 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051233 4790 flags.go:64] FLAG: --http-check-frequency="20s" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051242 4790 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051251 4790 flags.go:64] FLAG: --image-credential-provider-config="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051259 4790 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051269 4790 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051278 4790 flags.go:64] FLAG: --image-service-endpoint="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051286 4790 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051295 4790 flags.go:64] FLAG: --kube-api-burst="100" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051304 4790 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051314 4790 flags.go:64] FLAG: --kube-api-qps="50" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051323 4790 flags.go:64] FLAG: --kube-reserved="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051332 4790 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051340 4790 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051349 4790 flags.go:64] FLAG: --kubelet-cgroups="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051358 4790 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051367 4790 flags.go:64] FLAG: --lock-file="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051376 4790 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051385 4790 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051394 4790 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051407 4790 flags.go:64] FLAG: --log-json-split-stream="false" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051416 4790 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051425 4790 flags.go:64] FLAG: --log-text-split-stream="false" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051434 4790 flags.go:64] FLAG: --logging-format="text" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051442 4790 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051452 4790 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051461 4790 flags.go:64] FLAG: --manifest-url="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051470 4790 flags.go:64] FLAG: --manifest-url-header="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051481 4790 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051491 4790 flags.go:64] FLAG: --max-open-files="1000000" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051501 4790 flags.go:64] FLAG: --max-pods="110" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051510 4790 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051520 4790 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051528 4790 flags.go:64] FLAG: --memory-manager-policy="None" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051537 4790 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051546 4790 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051555 4790 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051565 4790 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051583 4790 flags.go:64] FLAG: --node-status-max-images="50" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051592 4790 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051601 4790 flags.go:64] FLAG: --oom-score-adj="-999" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051610 4790 flags.go:64] FLAG: --pod-cidr="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051619 4790 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051631 4790 flags.go:64] FLAG: --pod-manifest-path="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051639 4790 flags.go:64] FLAG: --pod-max-pids="-1" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051648 4790 flags.go:64] FLAG: --pods-per-core="0" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051657 4790 flags.go:64] FLAG: --port="10250" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051674 4790 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051683 4790 flags.go:64] FLAG: --provider-id="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051692 4790 flags.go:64] FLAG: --qos-reserved="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051701 4790 flags.go:64] FLAG: --read-only-port="10255" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051710 4790 flags.go:64] FLAG: --register-node="true" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051719 4790 flags.go:64] FLAG: --register-schedulable="true" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051728 4790 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051742 4790 flags.go:64] FLAG: --registry-burst="10" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051750 4790 flags.go:64] FLAG: --registry-qps="5" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051759 4790 flags.go:64] FLAG: --reserved-cpus="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051768 4790 flags.go:64] FLAG: --reserved-memory="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051778 4790 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051787 4790 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051797 4790 flags.go:64] FLAG: --rotate-certificates="false" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051806 4790 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051815 4790 flags.go:64] FLAG: --runonce="false" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051824 4790 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051833 4790 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051843 4790 flags.go:64] FLAG: --seccomp-default="false" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051852 4790 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051861 4790 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051870 4790 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051902 4790 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051911 4790 flags.go:64] FLAG: --storage-driver-password="root" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051920 4790 flags.go:64] FLAG: --storage-driver-secure="false" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051928 4790 flags.go:64] FLAG: --storage-driver-table="stats" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051937 4790 flags.go:64] FLAG: --storage-driver-user="root" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051946 4790 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051955 4790 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051964 4790 flags.go:64] FLAG: --system-cgroups="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051973 4790 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051986 4790 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.051995 4790 flags.go:64] FLAG: --tls-cert-file="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.052004 4790 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.052015 4790 flags.go:64] FLAG: --tls-min-version="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.052025 4790 flags.go:64] FLAG: --tls-private-key-file="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.052034 4790 flags.go:64] FLAG: --topology-manager-policy="none" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.052043 4790 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.052051 4790 flags.go:64] FLAG: --topology-manager-scope="container" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.052060 4790 flags.go:64] FLAG: --v="2" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.052071 4790 flags.go:64] FLAG: --version="false" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.052083 4790 flags.go:64] FLAG: --vmodule="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.052093 4790 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.052102 4790 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052291 4790 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052303 4790 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052312 4790 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052321 4790 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052328 4790 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052336 4790 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052344 4790 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052351 4790 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052359 4790 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052366 4790 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052375 4790 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052383 4790 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052390 4790 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052398 4790 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052406 4790 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052413 4790 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052421 4790 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052429 4790 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052436 4790 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052444 4790 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052454 4790 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052464 4790 feature_gate.go:330] unrecognized feature gate: Example Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052473 4790 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052482 4790 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052490 4790 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052498 4790 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052506 4790 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052513 4790 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052521 4790 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052529 4790 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052537 4790 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052545 4790 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052553 4790 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052560 4790 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052569 4790 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052576 4790 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052586 4790 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052596 4790 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052605 4790 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052614 4790 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052622 4790 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052631 4790 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052639 4790 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052647 4790 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052655 4790 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052663 4790 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052671 4790 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052679 4790 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052687 4790 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052694 4790 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052703 4790 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052710 4790 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052718 4790 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052726 4790 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052734 4790 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052745 4790 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052753 4790 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052762 4790 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052771 4790 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052785 4790 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052794 4790 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052802 4790 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052811 4790 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052818 4790 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052826 4790 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052834 4790 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052842 4790 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052850 4790 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052857 4790 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052865 4790 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.052872 4790 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.052916 4790 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.067756 4790 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.068092 4790 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068225 4790 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068239 4790 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068251 4790 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068266 4790 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068275 4790 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068285 4790 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068295 4790 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068304 4790 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068313 4790 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068321 4790 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068329 4790 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068337 4790 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068345 4790 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068353 4790 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068361 4790 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068368 4790 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068376 4790 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068383 4790 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068392 4790 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068400 4790 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068408 4790 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068417 4790 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068425 4790 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068433 4790 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068441 4790 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068448 4790 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068456 4790 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068466 4790 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068475 4790 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068484 4790 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068492 4790 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068503 4790 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068513 4790 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068521 4790 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068530 4790 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068538 4790 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068545 4790 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068553 4790 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068561 4790 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068569 4790 feature_gate.go:330] unrecognized feature gate: Example Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068576 4790 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068584 4790 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068592 4790 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068599 4790 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068607 4790 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068615 4790 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068623 4790 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068630 4790 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068638 4790 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068646 4790 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068653 4790 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068661 4790 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068669 4790 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068678 4790 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068686 4790 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068693 4790 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068701 4790 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068709 4790 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068716 4790 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068724 4790 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068732 4790 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068741 4790 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068748 4790 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068758 4790 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068768 4790 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068776 4790 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068784 4790 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068792 4790 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068800 4790 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068807 4790 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.068818 4790 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.068835 4790 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069113 4790 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069132 4790 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069141 4790 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069149 4790 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069159 4790 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069167 4790 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069178 4790 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069190 4790 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069198 4790 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069207 4790 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069216 4790 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069225 4790 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069233 4790 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069243 4790 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069252 4790 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069260 4790 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069268 4790 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069276 4790 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069283 4790 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069291 4790 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069301 4790 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069308 4790 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069318 4790 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069328 4790 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069336 4790 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069345 4790 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069352 4790 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069360 4790 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069370 4790 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069381 4790 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069391 4790 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069401 4790 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069409 4790 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069417 4790 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069425 4790 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069434 4790 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069442 4790 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069452 4790 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069460 4790 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069468 4790 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069477 4790 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069485 4790 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069492 4790 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069500 4790 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069508 4790 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069516 4790 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069524 4790 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069532 4790 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069539 4790 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069546 4790 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069554 4790 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069562 4790 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069570 4790 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069577 4790 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069585 4790 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069592 4790 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069600 4790 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069609 4790 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069616 4790 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069625 4790 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069632 4790 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069640 4790 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069648 4790 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069656 4790 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069663 4790 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069670 4790 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069678 4790 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069686 4790 feature_gate.go:330] unrecognized feature gate: Example Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069694 4790 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069702 4790 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.069709 4790 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.069722 4790 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.070968 4790 server.go:940] "Client rotation is on, will bootstrap in background" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.075504 4790 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.075605 4790 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.077804 4790 server.go:997] "Starting client certificate rotation" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.077834 4790 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.078090 4790 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-14 23:11:24.840010019 +0000 UTC Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.078258 4790 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1233h58m52.761756077s for next certificate rotation Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.101928 4790 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.106789 4790 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.129540 4790 log.go:25] "Validated CRI v1 runtime API" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.164004 4790 log.go:25] "Validated CRI v1 image API" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.166567 4790 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.175120 4790 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-24-13-07-33-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.175166 4790 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.204420 4790 manager.go:217] Machine: {Timestamp:2025-11-24 13:12:32.200803306 +0000 UTC m=+0.580697058 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:3913efcb-c2ad-4861-a3c8-ebb167356dfb BootID:cb4d37b5-209f-4037-867f-c5986b938358 Filesystems:[{Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:cd:7f:8d Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:cd:7f:8d Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:bc:22:32 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:47:d4:f6 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:e5:06:ca Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:ab:42:69 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:2f:c1:fe Speed:-1 Mtu:1496} {Name:eth10 MacAddress:2a:eb:01:08:82:1f Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:8e:70:af:4b:ee:5f Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.204793 4790 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.205027 4790 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.207981 4790 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.208255 4790 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.208303 4790 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.208581 4790 topology_manager.go:138] "Creating topology manager with none policy" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.208593 4790 container_manager_linux.go:303] "Creating device plugin manager" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.209235 4790 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.209276 4790 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.210116 4790 state_mem.go:36] "Initialized new in-memory state store" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.210226 4790 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.214193 4790 kubelet.go:418] "Attempting to sync node with API server" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.214219 4790 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.214276 4790 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.214304 4790 kubelet.go:324] "Adding apiserver pod source" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.214322 4790 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.217973 4790 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.220041 4790 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.221911 4790 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.222201 4790 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.69:6443: connect: connection refused Nov 24 13:12:32 crc kubenswrapper[4790]: E1124 13:12:32.222331 4790 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.69:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.222192 4790 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.69:6443: connect: connection refused Nov 24 13:12:32 crc kubenswrapper[4790]: E1124 13:12:32.222400 4790 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.69:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.225217 4790 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.225240 4790 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.225248 4790 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.225254 4790 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.225265 4790 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.225272 4790 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.225279 4790 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.225290 4790 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.225299 4790 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.225308 4790 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.225319 4790 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.225350 4790 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.226627 4790 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.227295 4790 server.go:1280] "Started kubelet" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.229737 4790 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.69:6443: connect: connection refused Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.229996 4790 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.230005 4790 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 24 13:12:32 crc systemd[1]: Started Kubernetes Kubelet. Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.232586 4790 server.go:460] "Adding debug handlers to kubelet server" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.233870 4790 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.234078 4790 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.234178 4790 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.234179 4790 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 16:42:53.329650836 +0000 UTC Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.234245 4790 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 915h30m21.095410308s for next certificate rotation Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.234793 4790 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.234821 4790 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.234974 4790 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 24 13:12:32 crc kubenswrapper[4790]: E1124 13:12:32.234977 4790 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.242435 4790 factory.go:55] Registering systemd factory Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.242486 4790 factory.go:221] Registration of the systemd container factory successfully Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.243680 4790 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.69:6443: connect: connection refused Nov 24 13:12:32 crc kubenswrapper[4790]: E1124 13:12:32.243961 4790 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.69:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.245199 4790 factory.go:153] Registering CRI-O factory Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.245269 4790 factory.go:221] Registration of the crio container factory successfully Nov 24 13:12:32 crc kubenswrapper[4790]: E1124 13:12:32.245254 4790 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.69:6443: connect: connection refused" interval="200ms" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.245380 4790 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.245411 4790 factory.go:103] Registering Raw factory Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.245444 4790 manager.go:1196] Started watching for new ooms in manager Nov 24 13:12:32 crc kubenswrapper[4790]: E1124 13:12:32.243649 4790 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.129.56.69:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187af383f9dd1cdc default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-24 13:12:32.227253468 +0000 UTC m=+0.607147150,LastTimestamp:2025-11-24 13:12:32.227253468 +0000 UTC m=+0.607147150,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.246470 4790 manager.go:319] Starting recovery of all containers Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.254613 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.254818 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.254952 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.255031 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.255120 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.255204 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.255289 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.255367 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.255444 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.255525 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.255614 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.255702 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.255781 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.255863 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.255965 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.256055 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.256133 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.256212 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.256298 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.256376 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.256452 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.256533 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.256651 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.256730 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.256898 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.256984 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.257076 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.257158 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.257238 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.257322 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.257404 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.257480 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.257593 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.257673 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.257750 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.257832 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.257966 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.258072 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.258151 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.258224 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.258318 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.258409 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.258501 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.258586 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.258670 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.258755 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.258837 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.259030 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.259122 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.259198 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.259275 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.259357 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.259443 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.259528 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.259608 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.259687 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.259770 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.259850 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.259960 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.260040 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.263360 4790 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.263435 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.263478 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.263509 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.263539 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.263567 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.263587 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.263606 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.263626 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.263651 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.263676 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.263700 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.263730 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.263757 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.263783 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.263815 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.263838 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.263916 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.263939 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.263957 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.263975 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.263994 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264013 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264032 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264053 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264072 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264091 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264112 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264133 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264152 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264171 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264194 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264218 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264244 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264272 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264307 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264329 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264348 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264374 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264420 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264492 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264513 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264535 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264554 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264582 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264610 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264635 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264655 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264678 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264700 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264720 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264749 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264774 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264803 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264831 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264852 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264873 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264933 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264952 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264971 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.264989 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265008 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265025 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265045 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265072 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265104 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265124 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265143 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265169 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265188 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265207 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265224 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265244 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265261 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265280 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265306 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265326 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265345 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265364 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265389 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265408 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265426 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265446 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265469 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265490 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265509 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265532 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265559 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265578 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265595 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265614 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265631 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265650 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265674 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265710 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265729 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265749 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265768 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265789 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265813 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265839 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265858 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265877 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265931 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265949 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265973 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.265993 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266012 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266079 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266103 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266122 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266186 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266211 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266230 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266249 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266268 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266286 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266306 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266325 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266342 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266368 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266416 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266436 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266461 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266481 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266497 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266515 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266533 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266551 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266571 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266590 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266607 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266625 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266643 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266666 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266683 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266700 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266720 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266738 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266754 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266773 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266797 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266816 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266835 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266854 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266871 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266923 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.266948 4790 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.267023 4790 reconstruct.go:97] "Volume reconstruction finished" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.267037 4790 reconciler.go:26] "Reconciler: start to sync state" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.270595 4790 manager.go:324] Recovery completed Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.286406 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.288656 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.288702 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.288712 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.293346 4790 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.293372 4790 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.293393 4790 state_mem.go:36] "Initialized new in-memory state store" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.310662 4790 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.313161 4790 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.313242 4790 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.313271 4790 kubelet.go:2335] "Starting kubelet main sync loop" Nov 24 13:12:32 crc kubenswrapper[4790]: E1124 13:12:32.313338 4790 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.315372 4790 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.69:6443: connect: connection refused Nov 24 13:12:32 crc kubenswrapper[4790]: E1124 13:12:32.315546 4790 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.69:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.319257 4790 policy_none.go:49] "None policy: Start" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.320040 4790 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.320071 4790 state_mem.go:35] "Initializing new in-memory state store" Nov 24 13:12:32 crc kubenswrapper[4790]: E1124 13:12:32.336053 4790 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.378910 4790 manager.go:334] "Starting Device Plugin manager" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.379121 4790 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.379137 4790 server.go:79] "Starting device plugin registration server" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.379511 4790 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.379530 4790 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.379705 4790 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.379861 4790 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.379875 4790 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 24 13:12:32 crc kubenswrapper[4790]: E1124 13:12:32.386538 4790 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.413939 4790 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.414059 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.415260 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.415296 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.415308 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.415427 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.415637 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.415681 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.416155 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.416183 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.416193 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.416294 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.416364 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.416387 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.416397 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.416450 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.416473 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.416820 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.416844 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.416856 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.416992 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.417079 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.417110 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.417121 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.417474 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.417503 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.417743 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.417765 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.417776 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.417870 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.418024 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.418073 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.418283 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.418324 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.418344 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.418732 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.418757 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.418768 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.418972 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.418996 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.419004 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.418974 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.419104 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.419701 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.419723 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.419732 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:32 crc kubenswrapper[4790]: E1124 13:12:32.446001 4790 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.69:6443: connect: connection refused" interval="400ms" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.470071 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.470106 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.470125 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.470140 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.470180 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.470200 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.470244 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.470266 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.470283 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.470297 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.470319 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.470369 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.470397 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.470423 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.470437 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.480243 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.481316 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.481345 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.481359 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.481380 4790 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 13:12:32 crc kubenswrapper[4790]: E1124 13:12:32.481705 4790 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.69:6443: connect: connection refused" node="crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.571124 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.571467 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.571504 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.571289 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.571532 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.571568 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.571596 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.571621 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.571625 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.571647 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.571599 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.571655 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.571660 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.571702 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.571686 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.571704 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.571741 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.571770 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.571944 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.572020 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.572044 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.572088 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.572104 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.572113 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.572131 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.572089 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.572101 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.572194 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.572167 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.572205 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.683198 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.684408 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.684443 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.684455 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.684476 4790 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 13:12:32 crc kubenswrapper[4790]: E1124 13:12:32.684824 4790 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.69:6443: connect: connection refused" node="crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.753956 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.760465 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.779511 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.795707 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: I1124 13:12:32.800795 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.805557 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-c7f1ac4024ecbd0dab5fb6acbdb216e3b8433512c1d01b2feac9271b5479c590 WatchSource:0}: Error finding container c7f1ac4024ecbd0dab5fb6acbdb216e3b8433512c1d01b2feac9271b5479c590: Status 404 returned error can't find the container with id c7f1ac4024ecbd0dab5fb6acbdb216e3b8433512c1d01b2feac9271b5479c590 Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.806531 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-4c040ae0e425ff0862a2466f61ea545aeabf1c1930b5de9679aff13c2d012bca WatchSource:0}: Error finding container 4c040ae0e425ff0862a2466f61ea545aeabf1c1930b5de9679aff13c2d012bca: Status 404 returned error can't find the container with id 4c040ae0e425ff0862a2466f61ea545aeabf1c1930b5de9679aff13c2d012bca Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.813338 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-77b9bf1bf2f8a513568b12a96c375e85120815073e0d147f4ff6c9a5cd2f7151 WatchSource:0}: Error finding container 77b9bf1bf2f8a513568b12a96c375e85120815073e0d147f4ff6c9a5cd2f7151: Status 404 returned error can't find the container with id 77b9bf1bf2f8a513568b12a96c375e85120815073e0d147f4ff6c9a5cd2f7151 Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.818162 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-f3cbdaa6306425819744684e840e0ba8068dd555cf5dddd22b05dcc3a56c178e WatchSource:0}: Error finding container f3cbdaa6306425819744684e840e0ba8068dd555cf5dddd22b05dcc3a56c178e: Status 404 returned error can't find the container with id f3cbdaa6306425819744684e840e0ba8068dd555cf5dddd22b05dcc3a56c178e Nov 24 13:12:32 crc kubenswrapper[4790]: W1124 13:12:32.820175 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-e38fa4a4b553db39f6eeed1913358d3443c089c79f1722bbc6eb9060030d5d17 WatchSource:0}: Error finding container e38fa4a4b553db39f6eeed1913358d3443c089c79f1722bbc6eb9060030d5d17: Status 404 returned error can't find the container with id e38fa4a4b553db39f6eeed1913358d3443c089c79f1722bbc6eb9060030d5d17 Nov 24 13:12:32 crc kubenswrapper[4790]: E1124 13:12:32.847510 4790 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.69:6443: connect: connection refused" interval="800ms" Nov 24 13:12:33 crc kubenswrapper[4790]: W1124 13:12:33.061842 4790 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.69:6443: connect: connection refused Nov 24 13:12:33 crc kubenswrapper[4790]: E1124 13:12:33.061950 4790 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.69:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:12:33 crc kubenswrapper[4790]: W1124 13:12:33.064769 4790 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.69:6443: connect: connection refused Nov 24 13:12:33 crc kubenswrapper[4790]: E1124 13:12:33.064825 4790 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.69:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:12:33 crc kubenswrapper[4790]: I1124 13:12:33.084956 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:33 crc kubenswrapper[4790]: I1124 13:12:33.086576 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:33 crc kubenswrapper[4790]: I1124 13:12:33.086617 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:33 crc kubenswrapper[4790]: I1124 13:12:33.086630 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:33 crc kubenswrapper[4790]: I1124 13:12:33.086668 4790 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 13:12:33 crc kubenswrapper[4790]: E1124 13:12:33.087120 4790 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.69:6443: connect: connection refused" node="crc" Nov 24 13:12:33 crc kubenswrapper[4790]: I1124 13:12:33.230433 4790 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.69:6443: connect: connection refused Nov 24 13:12:33 crc kubenswrapper[4790]: W1124 13:12:33.239091 4790 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.69:6443: connect: connection refused Nov 24 13:12:33 crc kubenswrapper[4790]: E1124 13:12:33.239169 4790 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.69:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:12:33 crc kubenswrapper[4790]: I1124 13:12:33.317826 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f3cbdaa6306425819744684e840e0ba8068dd555cf5dddd22b05dcc3a56c178e"} Nov 24 13:12:33 crc kubenswrapper[4790]: I1124 13:12:33.318583 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"77b9bf1bf2f8a513568b12a96c375e85120815073e0d147f4ff6c9a5cd2f7151"} Nov 24 13:12:33 crc kubenswrapper[4790]: I1124 13:12:33.319492 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4c040ae0e425ff0862a2466f61ea545aeabf1c1930b5de9679aff13c2d012bca"} Nov 24 13:12:33 crc kubenswrapper[4790]: I1124 13:12:33.320368 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"c7f1ac4024ecbd0dab5fb6acbdb216e3b8433512c1d01b2feac9271b5479c590"} Nov 24 13:12:33 crc kubenswrapper[4790]: I1124 13:12:33.321203 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e38fa4a4b553db39f6eeed1913358d3443c089c79f1722bbc6eb9060030d5d17"} Nov 24 13:12:33 crc kubenswrapper[4790]: E1124 13:12:33.649000 4790 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.69:6443: connect: connection refused" interval="1.6s" Nov 24 13:12:33 crc kubenswrapper[4790]: W1124 13:12:33.774281 4790 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.69:6443: connect: connection refused Nov 24 13:12:33 crc kubenswrapper[4790]: E1124 13:12:33.774419 4790 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.69:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:12:33 crc kubenswrapper[4790]: I1124 13:12:33.887461 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:33 crc kubenswrapper[4790]: I1124 13:12:33.888494 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:33 crc kubenswrapper[4790]: I1124 13:12:33.888527 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:33 crc kubenswrapper[4790]: I1124 13:12:33.888537 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:33 crc kubenswrapper[4790]: I1124 13:12:33.888560 4790 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 13:12:33 crc kubenswrapper[4790]: E1124 13:12:33.888798 4790 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.69:6443: connect: connection refused" node="crc" Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.230916 4790 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.69:6443: connect: connection refused Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.325582 4790 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="5e9ebe3a1d40785203c03afe7c0234b10973897ab1af0e0afd2b08375cc2f612" exitCode=0 Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.325631 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"5e9ebe3a1d40785203c03afe7c0234b10973897ab1af0e0afd2b08375cc2f612"} Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.325789 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.326691 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.326733 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.326748 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.327166 4790 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="b3febfee771e7c895da4b109e0c8c73fbc1942de721f81d537531bf11cff0f2b" exitCode=0 Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.327221 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"b3febfee771e7c895da4b109e0c8c73fbc1942de721f81d537531bf11cff0f2b"} Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.327238 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.328937 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.328963 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.328972 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.329769 4790 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="a59ab7ed8087587de107062d659abc98e07ce74ae4a8ed5e372bbeee9e6a8908" exitCode=0 Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.329814 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"a59ab7ed8087587de107062d659abc98e07ce74ae4a8ed5e372bbeee9e6a8908"} Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.329865 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.330947 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.330973 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.330984 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.333366 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b"} Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.333386 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4"} Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.333397 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0"} Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.333415 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978"} Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.333419 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.334043 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.334070 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.334082 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.335992 4790 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da" exitCode=0 Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.336029 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da"} Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.336092 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.336836 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.336870 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.336902 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.338337 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.339239 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.339313 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:34 crc kubenswrapper[4790]: I1124 13:12:34.339326 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:34 crc kubenswrapper[4790]: E1124 13:12:34.633525 4790 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.129.56.69:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187af383f9dd1cdc default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-24 13:12:32.227253468 +0000 UTC m=+0.607147150,LastTimestamp:2025-11-24 13:12:32.227253468 +0000 UTC m=+0.607147150,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 24 13:12:35 crc kubenswrapper[4790]: W1124 13:12:35.118707 4790 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.69:6443: connect: connection refused Nov 24 13:12:35 crc kubenswrapper[4790]: E1124 13:12:35.118802 4790 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.69:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:12:35 crc kubenswrapper[4790]: W1124 13:12:35.139096 4790 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.69:6443: connect: connection refused Nov 24 13:12:35 crc kubenswrapper[4790]: E1124 13:12:35.139170 4790 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.69:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.231117 4790 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.69:6443: connect: connection refused Nov 24 13:12:35 crc kubenswrapper[4790]: E1124 13:12:35.250093 4790 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.69:6443: connect: connection refused" interval="3.2s" Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.340612 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"24c1aad9c36a0f829ffd85e2d382255c83dbcad7e885121c6a7a4c7c2e7fc0df"} Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.340664 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64"} Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.340675 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe"} Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.340683 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9"} Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.340691 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14"} Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.340707 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.341735 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.341770 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.341778 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.342061 4790 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="2f3afe21ff61d7784561c5488594523020849d1649bc6688f5f519e682ab5aed" exitCode=0 Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.342136 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.342128 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"2f3afe21ff61d7784561c5488594523020849d1649bc6688f5f519e682ab5aed"} Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.343298 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.343323 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.343332 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.343983 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"4cac56536f98781c80d45847f247b4032416ecb7cbd1f22ff93e17ee86912955"} Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.344031 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.344788 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.344820 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.344832 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.346101 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"103fb4a5a8c9d50f090bd5b73e841e6d6bd3a52817fa60d8681461c3b7658ea8"} Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.346128 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.346151 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.346129 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ee38955b1cc92f846dccf1fb30021559dfa4cefe62e4d095dcf983ac7b0adba5"} Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.346236 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c123986828c19da07e80adf0ff83f2d5858aa1681caf2f097164fa1067e6b247"} Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.346823 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.346849 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.346862 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.346925 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.346943 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.346952 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.489190 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.490388 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.490414 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.490438 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:35 crc kubenswrapper[4790]: I1124 13:12:35.490459 4790 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 13:12:35 crc kubenswrapper[4790]: E1124 13:12:35.491095 4790 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.69:6443: connect: connection refused" node="crc" Nov 24 13:12:35 crc kubenswrapper[4790]: W1124 13:12:35.738729 4790 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.69:6443: connect: connection refused Nov 24 13:12:35 crc kubenswrapper[4790]: E1124 13:12:35.738833 4790 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.69:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:12:36 crc kubenswrapper[4790]: I1124 13:12:36.350311 4790 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="2c2a674281a342da54fb7bea9a9f10a44b2ce647c6e6cadb61f8379d1bab012b" exitCode=0 Nov 24 13:12:36 crc kubenswrapper[4790]: I1124 13:12:36.350433 4790 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 13:12:36 crc kubenswrapper[4790]: I1124 13:12:36.350444 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"2c2a674281a342da54fb7bea9a9f10a44b2ce647c6e6cadb61f8379d1bab012b"} Nov 24 13:12:36 crc kubenswrapper[4790]: I1124 13:12:36.350470 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:36 crc kubenswrapper[4790]: I1124 13:12:36.350486 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:36 crc kubenswrapper[4790]: I1124 13:12:36.350434 4790 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 13:12:36 crc kubenswrapper[4790]: I1124 13:12:36.350534 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:36 crc kubenswrapper[4790]: I1124 13:12:36.350554 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:36 crc kubenswrapper[4790]: I1124 13:12:36.351652 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:36 crc kubenswrapper[4790]: I1124 13:12:36.351691 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:36 crc kubenswrapper[4790]: I1124 13:12:36.351705 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:36 crc kubenswrapper[4790]: I1124 13:12:36.351757 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:36 crc kubenswrapper[4790]: I1124 13:12:36.351779 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:36 crc kubenswrapper[4790]: I1124 13:12:36.351791 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:36 crc kubenswrapper[4790]: I1124 13:12:36.351995 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:36 crc kubenswrapper[4790]: I1124 13:12:36.352024 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:36 crc kubenswrapper[4790]: I1124 13:12:36.352034 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:36 crc kubenswrapper[4790]: I1124 13:12:36.352092 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:36 crc kubenswrapper[4790]: I1124 13:12:36.352100 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:36 crc kubenswrapper[4790]: I1124 13:12:36.352107 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:36 crc kubenswrapper[4790]: I1124 13:12:36.940611 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:12:36 crc kubenswrapper[4790]: I1124 13:12:36.940759 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:36 crc kubenswrapper[4790]: I1124 13:12:36.942089 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:36 crc kubenswrapper[4790]: I1124 13:12:36.942115 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:36 crc kubenswrapper[4790]: I1124 13:12:36.942126 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:37 crc kubenswrapper[4790]: I1124 13:12:37.358400 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"96dfab98b54c0ba0232fd10f69c33d3ebfd791c25afeecca0b93e5399226ba63"} Nov 24 13:12:37 crc kubenswrapper[4790]: I1124 13:12:37.358466 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"0df7b206a25be77094b5af412cd437ff966c069e11bf6e983135169a81c1be9b"} Nov 24 13:12:37 crc kubenswrapper[4790]: I1124 13:12:37.358486 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d1fd5d52d898dad5d4ffc51335493d17ba1b621b7cd11e2b4a03a91de7c8c65a"} Nov 24 13:12:37 crc kubenswrapper[4790]: I1124 13:12:37.358506 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7d51be10ee26b94355358077a8e09a08277ab67219333326f0dfe31b95d7d492"} Nov 24 13:12:37 crc kubenswrapper[4790]: I1124 13:12:37.583962 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:12:37 crc kubenswrapper[4790]: I1124 13:12:37.584105 4790 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 13:12:37 crc kubenswrapper[4790]: I1124 13:12:37.584136 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:37 crc kubenswrapper[4790]: I1124 13:12:37.585637 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:37 crc kubenswrapper[4790]: I1124 13:12:37.585696 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:37 crc kubenswrapper[4790]: I1124 13:12:37.585720 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:37 crc kubenswrapper[4790]: I1124 13:12:37.910637 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:12:38 crc kubenswrapper[4790]: I1124 13:12:38.181955 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:12:38 crc kubenswrapper[4790]: I1124 13:12:38.182112 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:38 crc kubenswrapper[4790]: I1124 13:12:38.183441 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:38 crc kubenswrapper[4790]: I1124 13:12:38.183473 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:38 crc kubenswrapper[4790]: I1124 13:12:38.183482 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:38 crc kubenswrapper[4790]: I1124 13:12:38.363813 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"beb325e20fca2b9d050dcadaf3ff7c08a57d5019c4a0d3f64255c11fdbe24938"} Nov 24 13:12:38 crc kubenswrapper[4790]: I1124 13:12:38.363827 4790 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 13:12:38 crc kubenswrapper[4790]: I1124 13:12:38.363987 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:38 crc kubenswrapper[4790]: I1124 13:12:38.364046 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:38 crc kubenswrapper[4790]: I1124 13:12:38.365536 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:38 crc kubenswrapper[4790]: I1124 13:12:38.365574 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:38 crc kubenswrapper[4790]: I1124 13:12:38.365603 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:38 crc kubenswrapper[4790]: I1124 13:12:38.365647 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:38 crc kubenswrapper[4790]: I1124 13:12:38.365675 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:38 crc kubenswrapper[4790]: I1124 13:12:38.365688 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:38 crc kubenswrapper[4790]: I1124 13:12:38.469390 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 13:12:38 crc kubenswrapper[4790]: I1124 13:12:38.469564 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:38 crc kubenswrapper[4790]: I1124 13:12:38.470648 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:38 crc kubenswrapper[4790]: I1124 13:12:38.470686 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:38 crc kubenswrapper[4790]: I1124 13:12:38.470700 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:38 crc kubenswrapper[4790]: I1124 13:12:38.691802 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:38 crc kubenswrapper[4790]: I1124 13:12:38.692956 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:38 crc kubenswrapper[4790]: I1124 13:12:38.692989 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:38 crc kubenswrapper[4790]: I1124 13:12:38.692999 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:38 crc kubenswrapper[4790]: I1124 13:12:38.693021 4790 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 13:12:38 crc kubenswrapper[4790]: I1124 13:12:38.775484 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:12:38 crc kubenswrapper[4790]: I1124 13:12:38.775676 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:38 crc kubenswrapper[4790]: I1124 13:12:38.776952 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:38 crc kubenswrapper[4790]: I1124 13:12:38.777009 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:38 crc kubenswrapper[4790]: I1124 13:12:38.777045 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:38 crc kubenswrapper[4790]: I1124 13:12:38.839507 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 24 13:12:39 crc kubenswrapper[4790]: I1124 13:12:39.365815 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:39 crc kubenswrapper[4790]: I1124 13:12:39.366830 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:39 crc kubenswrapper[4790]: I1124 13:12:39.366872 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:39 crc kubenswrapper[4790]: I1124 13:12:39.366908 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:39 crc kubenswrapper[4790]: I1124 13:12:39.859374 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:12:39 crc kubenswrapper[4790]: I1124 13:12:39.859609 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:39 crc kubenswrapper[4790]: I1124 13:12:39.861197 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:39 crc kubenswrapper[4790]: I1124 13:12:39.861234 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:39 crc kubenswrapper[4790]: I1124 13:12:39.861246 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:40 crc kubenswrapper[4790]: I1124 13:12:40.194286 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 24 13:12:40 crc kubenswrapper[4790]: I1124 13:12:40.368429 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:40 crc kubenswrapper[4790]: I1124 13:12:40.370036 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:40 crc kubenswrapper[4790]: I1124 13:12:40.370135 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:40 crc kubenswrapper[4790]: I1124 13:12:40.370174 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:41 crc kubenswrapper[4790]: I1124 13:12:41.182722 4790 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 24 13:12:41 crc kubenswrapper[4790]: I1124 13:12:41.182815 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 13:12:41 crc kubenswrapper[4790]: I1124 13:12:41.370588 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:41 crc kubenswrapper[4790]: I1124 13:12:41.371386 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:41 crc kubenswrapper[4790]: I1124 13:12:41.371419 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:41 crc kubenswrapper[4790]: I1124 13:12:41.371432 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:41 crc kubenswrapper[4790]: I1124 13:12:41.883711 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:12:41 crc kubenswrapper[4790]: I1124 13:12:41.883900 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:41 crc kubenswrapper[4790]: I1124 13:12:41.885289 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:41 crc kubenswrapper[4790]: I1124 13:12:41.885346 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:41 crc kubenswrapper[4790]: I1124 13:12:41.885362 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:41 crc kubenswrapper[4790]: I1124 13:12:41.889411 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:12:42 crc kubenswrapper[4790]: I1124 13:12:42.373050 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:42 crc kubenswrapper[4790]: I1124 13:12:42.373781 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:42 crc kubenswrapper[4790]: I1124 13:12:42.373825 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:42 crc kubenswrapper[4790]: I1124 13:12:42.373838 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:42 crc kubenswrapper[4790]: E1124 13:12:42.386661 4790 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 24 13:12:46 crc kubenswrapper[4790]: I1124 13:12:46.006424 4790 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 24 13:12:46 crc kubenswrapper[4790]: I1124 13:12:46.006535 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 24 13:12:46 crc kubenswrapper[4790]: I1124 13:12:46.012218 4790 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 24 13:12:46 crc kubenswrapper[4790]: I1124 13:12:46.012297 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 24 13:12:46 crc kubenswrapper[4790]: I1124 13:12:46.385766 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 24 13:12:46 crc kubenswrapper[4790]: I1124 13:12:46.387831 4790 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="24c1aad9c36a0f829ffd85e2d382255c83dbcad7e885121c6a7a4c7c2e7fc0df" exitCode=255 Nov 24 13:12:46 crc kubenswrapper[4790]: I1124 13:12:46.387931 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"24c1aad9c36a0f829ffd85e2d382255c83dbcad7e885121c6a7a4c7c2e7fc0df"} Nov 24 13:12:46 crc kubenswrapper[4790]: I1124 13:12:46.388161 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:46 crc kubenswrapper[4790]: I1124 13:12:46.389168 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:46 crc kubenswrapper[4790]: I1124 13:12:46.389199 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:46 crc kubenswrapper[4790]: I1124 13:12:46.389211 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:46 crc kubenswrapper[4790]: I1124 13:12:46.389662 4790 scope.go:117] "RemoveContainer" containerID="24c1aad9c36a0f829ffd85e2d382255c83dbcad7e885121c6a7a4c7c2e7fc0df" Nov 24 13:12:47 crc kubenswrapper[4790]: I1124 13:12:47.391745 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 24 13:12:47 crc kubenswrapper[4790]: I1124 13:12:47.393835 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813"} Nov 24 13:12:47 crc kubenswrapper[4790]: I1124 13:12:47.394029 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:47 crc kubenswrapper[4790]: I1124 13:12:47.395353 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:47 crc kubenswrapper[4790]: I1124 13:12:47.395407 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:47 crc kubenswrapper[4790]: I1124 13:12:47.395427 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:47 crc kubenswrapper[4790]: I1124 13:12:47.918340 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:12:48 crc kubenswrapper[4790]: I1124 13:12:48.396477 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:48 crc kubenswrapper[4790]: I1124 13:12:48.396968 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:12:48 crc kubenswrapper[4790]: I1124 13:12:48.397449 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:48 crc kubenswrapper[4790]: I1124 13:12:48.397499 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:48 crc kubenswrapper[4790]: I1124 13:12:48.397512 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:48 crc kubenswrapper[4790]: I1124 13:12:48.403626 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:12:48 crc kubenswrapper[4790]: I1124 13:12:48.783466 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:12:48 crc kubenswrapper[4790]: I1124 13:12:48.783694 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:48 crc kubenswrapper[4790]: I1124 13:12:48.785187 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:48 crc kubenswrapper[4790]: I1124 13:12:48.785233 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:48 crc kubenswrapper[4790]: I1124 13:12:48.785244 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:48 crc kubenswrapper[4790]: I1124 13:12:48.873260 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 24 13:12:48 crc kubenswrapper[4790]: I1124 13:12:48.873441 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:48 crc kubenswrapper[4790]: I1124 13:12:48.874683 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:48 crc kubenswrapper[4790]: I1124 13:12:48.874739 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:48 crc kubenswrapper[4790]: I1124 13:12:48.874749 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:48 crc kubenswrapper[4790]: I1124 13:12:48.888478 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 24 13:12:49 crc kubenswrapper[4790]: I1124 13:12:49.398786 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:49 crc kubenswrapper[4790]: I1124 13:12:49.398908 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:49 crc kubenswrapper[4790]: I1124 13:12:49.404177 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:49 crc kubenswrapper[4790]: I1124 13:12:49.404208 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:49 crc kubenswrapper[4790]: I1124 13:12:49.404239 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:49 crc kubenswrapper[4790]: I1124 13:12:49.404250 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:49 crc kubenswrapper[4790]: I1124 13:12:49.404262 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:49 crc kubenswrapper[4790]: I1124 13:12:49.404265 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:50 crc kubenswrapper[4790]: I1124 13:12:50.401231 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:50 crc kubenswrapper[4790]: I1124 13:12:50.402521 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:50 crc kubenswrapper[4790]: I1124 13:12:50.402568 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:50 crc kubenswrapper[4790]: I1124 13:12:50.402585 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:50 crc kubenswrapper[4790]: E1124 13:12:50.997347 4790 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.000019 4790 trace.go:236] Trace[1309823667]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 13:12:40.696) (total time: 10303ms): Nov 24 13:12:51 crc kubenswrapper[4790]: Trace[1309823667]: ---"Objects listed" error: 10303ms (13:12:50.999) Nov 24 13:12:51 crc kubenswrapper[4790]: Trace[1309823667]: [10.303602778s] [10.303602778s] END Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.000051 4790 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.001489 4790 trace.go:236] Trace[2074072144]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 13:12:36.136) (total time: 14865ms): Nov 24 13:12:51 crc kubenswrapper[4790]: Trace[2074072144]: ---"Objects listed" error: 14865ms (13:12:51.001) Nov 24 13:12:51 crc kubenswrapper[4790]: Trace[2074072144]: [14.865052679s] [14.865052679s] END Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.001513 4790 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.002271 4790 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.002340 4790 trace.go:236] Trace[1109554685]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 13:12:39.741) (total time: 11260ms): Nov 24 13:12:51 crc kubenswrapper[4790]: Trace[1109554685]: ---"Objects listed" error: 11260ms (13:12:51.002) Nov 24 13:12:51 crc kubenswrapper[4790]: Trace[1109554685]: [11.260432091s] [11.260432091s] END Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.002405 4790 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.003125 4790 trace.go:236] Trace[161399013]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 13:12:39.108) (total time: 11894ms): Nov 24 13:12:51 crc kubenswrapper[4790]: Trace[161399013]: ---"Objects listed" error: 11894ms (13:12:51.003) Nov 24 13:12:51 crc kubenswrapper[4790]: Trace[161399013]: [11.894277543s] [11.894277543s] END Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.003256 4790 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 24 13:12:51 crc kubenswrapper[4790]: E1124 13:12:51.004160 4790 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.058454 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.062555 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.224202 4790 apiserver.go:52] "Watching apiserver" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.226159 4790 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.226452 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.226785 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.226894 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.226985 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:12:51 crc kubenswrapper[4790]: E1124 13:12:51.227096 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.227014 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:12:51 crc kubenswrapper[4790]: E1124 13:12:51.227910 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.228049 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.228659 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:12:51 crc kubenswrapper[4790]: E1124 13:12:51.228764 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.229292 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.229297 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.229464 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.230514 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.230528 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.232498 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.232557 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.234312 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.234627 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.235521 4790 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.252986 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.263773 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.273988 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.282401 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.291079 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.301512 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.303633 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.303750 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.303839 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.303942 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.304038 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.304134 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.304226 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.304390 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.304486 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.304553 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.304625 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.304711 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.304783 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.304865 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.304959 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.305072 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.305169 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.305268 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.304182 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.304287 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.304649 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.304783 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.304952 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.304994 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.305073 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.305139 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.305183 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.305326 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.305347 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.305478 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.305602 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.305733 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.305364 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.306091 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.306169 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.306253 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.306348 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.306417 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.306483 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.306614 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.306680 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.306746 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.306820 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.306901 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.306984 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.307063 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.307126 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.307189 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.307259 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.307328 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.307415 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.307481 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.307553 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.307618 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.307686 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.307750 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.307830 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.307943 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.308016 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.308089 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.308157 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.308217 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.308289 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.308358 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.308420 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.308490 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.308582 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.308661 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.308734 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.308796 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.308860 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.308955 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.309028 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.309093 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.309153 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.306021 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.306031 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.306096 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.306122 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.306351 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.306543 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.306551 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.306762 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.306784 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.306948 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.307020 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.307119 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.307255 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.307515 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.307562 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.307723 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.307776 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.308031 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.308081 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.308307 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.308340 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.308501 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.308525 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.308786 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.309196 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.309723 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.309924 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.310837 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.310001 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.310059 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.310250 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.310324 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.310916 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.310340 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.310327 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.310369 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.310517 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.310528 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.310626 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.310718 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.310736 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.310970 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.310779 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.310841 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311055 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.310861 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311136 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311160 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311180 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311202 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311227 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311248 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311267 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311283 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311327 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311091 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311349 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311370 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311389 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311409 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311430 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311448 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311467 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311485 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311504 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311523 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311541 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311520 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311562 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311583 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311606 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311689 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311712 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311731 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311750 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311770 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311792 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311818 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311844 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311864 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311914 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311957 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311978 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311996 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312014 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312036 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312054 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312072 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312093 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312111 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312129 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312149 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312167 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312188 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312207 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312231 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312251 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312275 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312315 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312338 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312359 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312377 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312397 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312414 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312433 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312452 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312470 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312488 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312506 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312526 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312551 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312570 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312588 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312607 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312626 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312647 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312671 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312693 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312716 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312736 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312756 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312775 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312793 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312813 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312835 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312854 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312891 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312911 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312931 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312951 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312972 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312995 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313017 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313039 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313059 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313083 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313130 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313151 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313173 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313193 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313214 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313234 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313254 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313273 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313300 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313319 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313338 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313359 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313382 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313403 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313426 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313446 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313468 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313486 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313507 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313548 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313570 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313591 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313611 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313630 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313653 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313677 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313699 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313718 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313736 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313756 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313776 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313796 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313818 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313838 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313857 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313894 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313917 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313935 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313954 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313973 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313991 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314011 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314037 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314079 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314106 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314129 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314152 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314173 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314210 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311325 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311486 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311610 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311744 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.311894 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312162 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312449 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.312804 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313054 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313169 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313334 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313346 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313789 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.313858 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314213 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314233 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314235 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314403 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314438 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314470 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314497 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314519 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314528 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314561 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314586 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314654 4790 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314681 4790 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314699 4790 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314713 4790 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314726 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314731 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314739 4790 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314768 4790 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314802 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314846 4790 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314860 4790 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314873 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314920 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314910 4790 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314959 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314975 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314986 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.314997 4790 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315008 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315020 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315032 4790 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315042 4790 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315053 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315064 4790 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315134 4790 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315146 4790 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315157 4790 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315166 4790 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315174 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315186 4790 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315198 4790 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315208 4790 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315210 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315219 4790 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315313 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315343 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315365 4790 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315378 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315391 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315404 4790 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315415 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315427 4790 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315438 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315449 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315732 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315748 4790 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315760 4790 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315773 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315789 4790 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315801 4790 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315812 4790 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315813 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315823 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315835 4790 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315847 4790 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315868 4790 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315895 4790 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315906 4790 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315918 4790 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315930 4790 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315941 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315962 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315976 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315987 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315999 4790 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.316019 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.316037 4790 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.316051 4790 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.316062 4790 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.316073 4790 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.316085 4790 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.316098 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.316109 4790 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.316120 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.316131 4790 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.316143 4790 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.316162 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.316183 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.316194 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315578 4790 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.317007 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315385 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315491 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315546 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: E1124 13:12:51.315627 4790 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.317953 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.317964 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315737 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315805 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315905 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.315985 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.316063 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.316079 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.316205 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.316423 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: E1124 13:12:51.316492 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:12:51.816475943 +0000 UTC m=+20.196369685 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.316576 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.316900 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.317082 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.317088 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.317221 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.317439 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.317717 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.317760 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.317753 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.317836 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.326780 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.326925 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.326943 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.327629 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.327799 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.327895 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.328344 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.328488 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.328392 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.328573 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.328624 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.328804 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.328834 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.328862 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.329100 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.329175 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.329191 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.329327 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.329441 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.329469 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.330101 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.329449 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.330185 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.329857 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.330377 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.330412 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: E1124 13:12:51.330762 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:12:51.830517247 +0000 UTC m=+20.210410919 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.330847 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: E1124 13:12:51.330923 4790 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:12:51 crc kubenswrapper[4790]: E1124 13:12:51.330945 4790 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:12:51 crc kubenswrapper[4790]: E1124 13:12:51.330957 4790 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.330971 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.331123 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.331372 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.331390 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.331419 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: E1124 13:12:51.331664 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 13:12:51.831644959 +0000 UTC m=+20.211538621 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.331686 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.331835 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.332089 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.332290 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.332356 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.332371 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.332441 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.331733 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.332940 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.333316 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.333343 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.333680 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.333391 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.334613 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.334688 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.334695 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.334712 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.334958 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.335029 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.335217 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.335306 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: E1124 13:12:51.337306 4790 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:12:51 crc kubenswrapper[4790]: E1124 13:12:51.337479 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:12:51.837404425 +0000 UTC m=+20.217298087 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.346199 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.346617 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.346787 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.347205 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.347404 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.347586 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.348139 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.348268 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.348431 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.348870 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.349858 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.350080 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.350224 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.351377 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.351636 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.351939 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.352211 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.352285 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.352676 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.353530 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.353921 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.354007 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.354275 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.354288 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.354313 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.354339 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.354369 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.354483 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.354635 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.354661 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.354732 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: E1124 13:12:51.354963 4790 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:12:51 crc kubenswrapper[4790]: E1124 13:12:51.354983 4790 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:12:51 crc kubenswrapper[4790]: E1124 13:12:51.355555 4790 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:12:51 crc kubenswrapper[4790]: E1124 13:12:51.355626 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 13:12:51.855609818 +0000 UTC m=+20.235503480 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.355213 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.355015 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.355023 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.355128 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.356409 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.357326 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.357501 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.357503 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.357624 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.357698 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.359582 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.360036 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.372129 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.372669 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.377745 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.380335 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:12:51 crc kubenswrapper[4790]: E1124 13:12:51.408423 4790 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.416767 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.416847 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.416922 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.416916 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.416937 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417000 4790 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417040 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417113 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417131 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417192 4790 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417205 4790 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417215 4790 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417227 4790 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417238 4790 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417254 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417262 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417270 4790 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417279 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417287 4790 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417316 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417325 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417333 4790 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417342 4790 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417350 4790 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417371 4790 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417389 4790 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417401 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417412 4790 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417422 4790 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417434 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417446 4790 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417457 4790 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417468 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417480 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417493 4790 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417534 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417550 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417565 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417576 4790 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417589 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417600 4790 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417611 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417623 4790 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417634 4790 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417645 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417656 4790 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417666 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417677 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417688 4790 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417698 4790 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417728 4790 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417739 4790 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417749 4790 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417760 4790 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417770 4790 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417780 4790 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417791 4790 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417802 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417813 4790 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417823 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417835 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417847 4790 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417859 4790 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417870 4790 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417900 4790 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417913 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417925 4790 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417935 4790 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417948 4790 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417959 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417970 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417983 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.417997 4790 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418008 4790 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418020 4790 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418033 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418044 4790 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418056 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418067 4790 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418078 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418089 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418100 4790 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418111 4790 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418127 4790 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418138 4790 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418151 4790 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418161 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418173 4790 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418186 4790 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418197 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418208 4790 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418219 4790 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418230 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418241 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418253 4790 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418263 4790 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418274 4790 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418285 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418297 4790 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418307 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418331 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418342 4790 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418355 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418366 4790 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418376 4790 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418387 4790 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418398 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418409 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418420 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418430 4790 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418441 4790 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418450 4790 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418461 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418471 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418482 4790 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418492 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418503 4790 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418514 4790 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418525 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418536 4790 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418548 4790 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418559 4790 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.418569 4790 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.540493 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.548347 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 13:12:51 crc kubenswrapper[4790]: W1124 13:12:51.551575 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-a85fb880e625eb112821af1ce56d4701c8cf4c2afa3702611ab20a6f7000dd17 WatchSource:0}: Error finding container a85fb880e625eb112821af1ce56d4701c8cf4c2afa3702611ab20a6f7000dd17: Status 404 returned error can't find the container with id a85fb880e625eb112821af1ce56d4701c8cf4c2afa3702611ab20a6f7000dd17 Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.554492 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 13:12:51 crc kubenswrapper[4790]: W1124 13:12:51.557935 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-661f89039f669f6586c104208706b3534194f0b105a051e61dfd17da33751f2f WatchSource:0}: Error finding container 661f89039f669f6586c104208706b3534194f0b105a051e61dfd17da33751f2f: Status 404 returned error can't find the container with id 661f89039f669f6586c104208706b3534194f0b105a051e61dfd17da33751f2f Nov 24 13:12:51 crc kubenswrapper[4790]: W1124 13:12:51.569947 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-e7f3105976667f06b2d5314febf1ba3584bb091cb431f47ea7f5ffcdca46bef2 WatchSource:0}: Error finding container e7f3105976667f06b2d5314febf1ba3584bb091cb431f47ea7f5ffcdca46bef2: Status 404 returned error can't find the container with id e7f3105976667f06b2d5314febf1ba3584bb091cb431f47ea7f5ffcdca46bef2 Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.820844 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:12:51 crc kubenswrapper[4790]: E1124 13:12:51.821028 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:12:52.821011676 +0000 UTC m=+21.200905338 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.921933 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.921986 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.922013 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:12:51 crc kubenswrapper[4790]: I1124 13:12:51.922036 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:12:51 crc kubenswrapper[4790]: E1124 13:12:51.922098 4790 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:12:51 crc kubenswrapper[4790]: E1124 13:12:51.922110 4790 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:12:51 crc kubenswrapper[4790]: E1124 13:12:51.922137 4790 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:12:51 crc kubenswrapper[4790]: E1124 13:12:51.922160 4790 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:12:51 crc kubenswrapper[4790]: E1124 13:12:51.922165 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:12:52.922148595 +0000 UTC m=+21.302042257 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:12:51 crc kubenswrapper[4790]: E1124 13:12:51.922173 4790 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:12:51 crc kubenswrapper[4790]: E1124 13:12:51.922185 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:12:52.922177336 +0000 UTC m=+21.302070998 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:12:51 crc kubenswrapper[4790]: E1124 13:12:51.922205 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 13:12:52.922195886 +0000 UTC m=+21.302089558 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:12:51 crc kubenswrapper[4790]: E1124 13:12:51.922252 4790 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:12:51 crc kubenswrapper[4790]: E1124 13:12:51.922265 4790 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:12:51 crc kubenswrapper[4790]: E1124 13:12:51.922273 4790 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:12:51 crc kubenswrapper[4790]: E1124 13:12:51.922306 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 13:12:52.922297519 +0000 UTC m=+21.302191181 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.321112 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.321760 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.322810 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.323397 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.324310 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.324831 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.325680 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.326541 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.327101 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.327955 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.328437 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.329472 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.330045 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.335742 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.336518 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.338448 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.339682 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.339759 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.340465 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.342608 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.343954 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.344822 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.346545 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.347035 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.347748 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.348412 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.349045 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.349756 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.350319 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.351919 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.352312 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.352753 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.353512 4790 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.353662 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.355526 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.356367 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.356996 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.358635 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.360651 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.362044 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.364005 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.365301 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.366910 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.367753 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.368232 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.368846 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.370107 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.370551 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.371455 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.372263 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.373374 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.373875 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.374384 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.375203 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.375682 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.376652 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.377238 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.383568 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.400859 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.405607 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"e7f3105976667f06b2d5314febf1ba3584bb091cb431f47ea7f5ffcdca46bef2"} Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.406820 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985"} Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.406862 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"661f89039f669f6586c104208706b3534194f0b105a051e61dfd17da33751f2f"} Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.408337 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8"} Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.408371 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583"} Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.408384 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"a85fb880e625eb112821af1ce56d4701c8cf4c2afa3702611ab20a6f7000dd17"} Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.415097 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.425837 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.435700 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.445313 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.458475 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.469600 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.483261 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.496501 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.508600 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.830203 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:12:52 crc kubenswrapper[4790]: E1124 13:12:52.830470 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:12:54.830418101 +0000 UTC m=+23.210311763 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.932026 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:12:52 crc kubenswrapper[4790]: E1124 13:12:52.932136 4790 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:12:52 crc kubenswrapper[4790]: E1124 13:12:52.932365 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:12:54.932351723 +0000 UTC m=+23.312245385 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:12:52 crc kubenswrapper[4790]: E1124 13:12:52.932484 4790 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:12:52 crc kubenswrapper[4790]: E1124 13:12:52.932504 4790 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:12:52 crc kubenswrapper[4790]: E1124 13:12:52.932517 4790 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:12:52 crc kubenswrapper[4790]: E1124 13:12:52.932572 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 13:12:54.932560909 +0000 UTC m=+23.312454571 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.932315 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.933272 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:12:52 crc kubenswrapper[4790]: I1124 13:12:52.933488 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:12:52 crc kubenswrapper[4790]: E1124 13:12:52.933741 4790 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:12:52 crc kubenswrapper[4790]: E1124 13:12:52.933868 4790 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:12:52 crc kubenswrapper[4790]: E1124 13:12:52.934007 4790 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:12:52 crc kubenswrapper[4790]: E1124 13:12:52.934179 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 13:12:54.934149855 +0000 UTC m=+23.314043527 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:12:52 crc kubenswrapper[4790]: E1124 13:12:52.934399 4790 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:12:52 crc kubenswrapper[4790]: E1124 13:12:52.934555 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:12:54.934535976 +0000 UTC m=+23.314429668 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:12:53 crc kubenswrapper[4790]: I1124 13:12:53.314381 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:12:53 crc kubenswrapper[4790]: I1124 13:12:53.314440 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:12:53 crc kubenswrapper[4790]: I1124 13:12:53.314477 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:12:53 crc kubenswrapper[4790]: E1124 13:12:53.314496 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:12:53 crc kubenswrapper[4790]: E1124 13:12:53.314583 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:12:53 crc kubenswrapper[4790]: E1124 13:12:53.314675 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:12:53 crc kubenswrapper[4790]: I1124 13:12:53.411644 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 24 13:12:53 crc kubenswrapper[4790]: I1124 13:12:53.412128 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 24 13:12:53 crc kubenswrapper[4790]: I1124 13:12:53.413760 4790 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813" exitCode=255 Nov 24 13:12:53 crc kubenswrapper[4790]: I1124 13:12:53.413819 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813"} Nov 24 13:12:53 crc kubenswrapper[4790]: I1124 13:12:53.413961 4790 scope.go:117] "RemoveContainer" containerID="24c1aad9c36a0f829ffd85e2d382255c83dbcad7e885121c6a7a4c7c2e7fc0df" Nov 24 13:12:53 crc kubenswrapper[4790]: I1124 13:12:53.427842 4790 scope.go:117] "RemoveContainer" containerID="85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813" Nov 24 13:12:53 crc kubenswrapper[4790]: E1124 13:12:53.428075 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 24 13:12:53 crc kubenswrapper[4790]: I1124 13:12:53.428653 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 24 13:12:53 crc kubenswrapper[4790]: I1124 13:12:53.432650 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:53 crc kubenswrapper[4790]: I1124 13:12:53.444544 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:53 crc kubenswrapper[4790]: I1124 13:12:53.456049 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:53 crc kubenswrapper[4790]: I1124 13:12:53.467903 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:53 crc kubenswrapper[4790]: I1124 13:12:53.479178 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:53 crc kubenswrapper[4790]: I1124 13:12:53.492724 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:53 crc kubenswrapper[4790]: I1124 13:12:53.505915 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:53 crc kubenswrapper[4790]: I1124 13:12:53.718474 4790 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:12:54 crc kubenswrapper[4790]: I1124 13:12:54.417720 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde"} Nov 24 13:12:54 crc kubenswrapper[4790]: I1124 13:12:54.418981 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 24 13:12:54 crc kubenswrapper[4790]: I1124 13:12:54.421617 4790 scope.go:117] "RemoveContainer" containerID="85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813" Nov 24 13:12:54 crc kubenswrapper[4790]: E1124 13:12:54.421788 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 24 13:12:54 crc kubenswrapper[4790]: I1124 13:12:54.431774 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:54 crc kubenswrapper[4790]: I1124 13:12:54.445140 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:54 crc kubenswrapper[4790]: I1124 13:12:54.462433 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:54 crc kubenswrapper[4790]: I1124 13:12:54.478262 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24c1aad9c36a0f829ffd85e2d382255c83dbcad7e885121c6a7a4c7c2e7fc0df\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:45Z\\\",\\\"message\\\":\\\"W1124 13:12:35.310820 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 13:12:35.312078 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763989955 cert, and key in /tmp/serving-cert-2676678285/serving-signer.crt, /tmp/serving-cert-2676678285/serving-signer.key\\\\nI1124 13:12:35.611458 1 observer_polling.go:159] Starting file observer\\\\nW1124 13:12:35.613911 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 13:12:35.614090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:35.616440 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2676678285/tls.crt::/tmp/serving-cert-2676678285/tls.key\\\\\\\"\\\\nF1124 13:12:45.850150 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:54 crc kubenswrapper[4790]: I1124 13:12:54.493593 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:54 crc kubenswrapper[4790]: I1124 13:12:54.505938 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:54 crc kubenswrapper[4790]: I1124 13:12:54.518340 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:54 crc kubenswrapper[4790]: I1124 13:12:54.528963 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:54 crc kubenswrapper[4790]: I1124 13:12:54.542709 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:54 crc kubenswrapper[4790]: I1124 13:12:54.554078 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:54 crc kubenswrapper[4790]: I1124 13:12:54.568464 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:54 crc kubenswrapper[4790]: I1124 13:12:54.579843 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:54 crc kubenswrapper[4790]: I1124 13:12:54.590048 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:54 crc kubenswrapper[4790]: I1124 13:12:54.604779 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:54 crc kubenswrapper[4790]: I1124 13:12:54.616632 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:54 crc kubenswrapper[4790]: I1124 13:12:54.629967 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:54 crc kubenswrapper[4790]: I1124 13:12:54.851442 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:12:54 crc kubenswrapper[4790]: E1124 13:12:54.851655 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:12:58.851628431 +0000 UTC m=+27.231522093 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:12:54 crc kubenswrapper[4790]: I1124 13:12:54.951944 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:12:54 crc kubenswrapper[4790]: I1124 13:12:54.952003 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:12:54 crc kubenswrapper[4790]: I1124 13:12:54.952032 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:12:54 crc kubenswrapper[4790]: I1124 13:12:54.952086 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:12:54 crc kubenswrapper[4790]: E1124 13:12:54.952119 4790 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:12:54 crc kubenswrapper[4790]: E1124 13:12:54.952181 4790 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:12:54 crc kubenswrapper[4790]: E1124 13:12:54.952190 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:12:58.952175044 +0000 UTC m=+27.332068706 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:12:54 crc kubenswrapper[4790]: E1124 13:12:54.952182 4790 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:12:54 crc kubenswrapper[4790]: E1124 13:12:54.952240 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:12:58.952228695 +0000 UTC m=+27.332122457 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:12:54 crc kubenswrapper[4790]: E1124 13:12:54.952240 4790 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:12:54 crc kubenswrapper[4790]: E1124 13:12:54.952264 4790 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:12:54 crc kubenswrapper[4790]: E1124 13:12:54.952253 4790 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:12:54 crc kubenswrapper[4790]: E1124 13:12:54.952278 4790 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:12:54 crc kubenswrapper[4790]: E1124 13:12:54.952283 4790 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:12:54 crc kubenswrapper[4790]: E1124 13:12:54.952308 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 13:12:58.952302737 +0000 UTC m=+27.332196399 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:12:54 crc kubenswrapper[4790]: E1124 13:12:54.952332 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 13:12:58.952316278 +0000 UTC m=+27.332209950 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:12:55 crc kubenswrapper[4790]: I1124 13:12:55.314058 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:12:55 crc kubenswrapper[4790]: I1124 13:12:55.314092 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:12:55 crc kubenswrapper[4790]: I1124 13:12:55.314082 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:12:55 crc kubenswrapper[4790]: E1124 13:12:55.314175 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:12:55 crc kubenswrapper[4790]: E1124 13:12:55.314253 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:12:55 crc kubenswrapper[4790]: E1124 13:12:55.314324 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:12:55 crc kubenswrapper[4790]: I1124 13:12:55.424322 4790 scope.go:117] "RemoveContainer" containerID="85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813" Nov 24 13:12:55 crc kubenswrapper[4790]: E1124 13:12:55.424443 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.035863 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-2687d"] Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.036106 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-76rkg"] Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.036255 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.036256 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-2687d" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.037970 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.038010 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.038762 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.038764 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.038778 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.039234 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.039314 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.040179 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.055252 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.068703 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.080664 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.090319 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.102723 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.113061 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.127179 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.146862 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.163278 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.169511 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-host-run-netns\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.169558 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-host-var-lib-kubelet\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.169594 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6c0e3499-87ac-481f-a010-708a3a4a730f-cni-binary-copy\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.169613 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-etc-kubernetes\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.169633 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-system-cni-dir\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.169657 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-host-var-lib-cni-bin\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.169676 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-multus-conf-dir\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.169695 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-os-release\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.169716 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-multus-socket-dir-parent\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.169735 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6c0e3499-87ac-481f-a010-708a3a4a730f-multus-daemon-config\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.169755 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27whv\" (UniqueName: \"kubernetes.io/projected/6c0e3499-87ac-481f-a010-708a3a4a730f-kube-api-access-27whv\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.169794 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-multus-cni-dir\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.169816 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-host-var-lib-cni-multus\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.169842 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/206d383c-05e9-47a3-a2e7-1535bf249e6a-hosts-file\") pod \"node-resolver-2687d\" (UID: \"206d383c-05e9-47a3-a2e7-1535bf249e6a\") " pod="openshift-dns/node-resolver-2687d" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.169864 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6xp7\" (UniqueName: \"kubernetes.io/projected/206d383c-05e9-47a3-a2e7-1535bf249e6a-kube-api-access-t6xp7\") pod \"node-resolver-2687d\" (UID: \"206d383c-05e9-47a3-a2e7-1535bf249e6a\") " pod="openshift-dns/node-resolver-2687d" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.169905 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-host-run-k8s-cni-cncf-io\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.169929 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-host-run-multus-certs\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.169961 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-cnibin\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.169982 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-hostroot\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.173408 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.189778 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.204864 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.225736 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.246140 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.264303 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.270953 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-cnibin\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.270991 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-hostroot\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.271014 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-host-run-netns\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.271031 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-host-var-lib-kubelet\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.271054 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6c0e3499-87ac-481f-a010-708a3a4a730f-cni-binary-copy\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.271069 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-etc-kubernetes\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.271084 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-system-cni-dir\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.271100 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-host-var-lib-cni-bin\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.271113 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-multus-conf-dir\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.271127 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-os-release\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.271142 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-multus-socket-dir-parent\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.271139 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-host-run-netns\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.271157 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6c0e3499-87ac-481f-a010-708a3a4a730f-multus-daemon-config\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.271210 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-hostroot\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.271215 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27whv\" (UniqueName: \"kubernetes.io/projected/6c0e3499-87ac-481f-a010-708a3a4a730f-kube-api-access-27whv\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.271360 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-multus-cni-dir\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.271381 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-host-var-lib-cni-multus\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.271414 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/206d383c-05e9-47a3-a2e7-1535bf249e6a-hosts-file\") pod \"node-resolver-2687d\" (UID: \"206d383c-05e9-47a3-a2e7-1535bf249e6a\") " pod="openshift-dns/node-resolver-2687d" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.271433 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6xp7\" (UniqueName: \"kubernetes.io/projected/206d383c-05e9-47a3-a2e7-1535bf249e6a-kube-api-access-t6xp7\") pod \"node-resolver-2687d\" (UID: \"206d383c-05e9-47a3-a2e7-1535bf249e6a\") " pod="openshift-dns/node-resolver-2687d" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.271451 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-host-run-k8s-cni-cncf-io\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.271454 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-host-var-lib-kubelet\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.271467 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-host-run-multus-certs\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.271551 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-host-run-multus-certs\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.271777 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6c0e3499-87ac-481f-a010-708a3a4a730f-multus-daemon-config\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.271803 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-multus-cni-dir\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.271829 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-host-var-lib-cni-bin\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.271844 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-host-var-lib-cni-multus\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.271859 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-multus-conf-dir\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.271161 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-cnibin\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.271907 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/206d383c-05e9-47a3-a2e7-1535bf249e6a-hosts-file\") pod \"node-resolver-2687d\" (UID: \"206d383c-05e9-47a3-a2e7-1535bf249e6a\") " pod="openshift-dns/node-resolver-2687d" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.272017 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6c0e3499-87ac-481f-a010-708a3a4a730f-cni-binary-copy\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.272057 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-etc-kubernetes\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.271434 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-system-cni-dir\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.272097 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-host-run-k8s-cni-cncf-io\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.272132 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-multus-socket-dir-parent\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.272178 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6c0e3499-87ac-481f-a010-708a3a4a730f-os-release\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.286817 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.291987 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6xp7\" (UniqueName: \"kubernetes.io/projected/206d383c-05e9-47a3-a2e7-1535bf249e6a-kube-api-access-t6xp7\") pod \"node-resolver-2687d\" (UID: \"206d383c-05e9-47a3-a2e7-1535bf249e6a\") " pod="openshift-dns/node-resolver-2687d" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.300263 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27whv\" (UniqueName: \"kubernetes.io/projected/6c0e3499-87ac-481f-a010-708a3a4a730f-kube-api-access-27whv\") pod \"multus-76rkg\" (UID: \"6c0e3499-87ac-481f-a010-708a3a4a730f\") " pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.304088 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.313634 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.313677 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.313794 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:12:57 crc kubenswrapper[4790]: E1124 13:12:57.313911 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:12:57 crc kubenswrapper[4790]: E1124 13:12:57.314021 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:12:57 crc kubenswrapper[4790]: E1124 13:12:57.314169 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.316516 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.326973 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.339354 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.349516 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-76rkg" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.354668 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-2687d" Nov 24 13:12:57 crc kubenswrapper[4790]: W1124 13:12:57.363277 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c0e3499_87ac_481f_a010_708a3a4a730f.slice/crio-71131fc7278d255fb19ad7a8c1574e8abd25d3f2ef3e9fb3ef3a0d3e69f8312d WatchSource:0}: Error finding container 71131fc7278d255fb19ad7a8c1574e8abd25d3f2ef3e9fb3ef3a0d3e69f8312d: Status 404 returned error can't find the container with id 71131fc7278d255fb19ad7a8c1574e8abd25d3f2ef3e9fb3ef3a0d3e69f8312d Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.405058 4790 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.408586 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5x94h"] Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.409350 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.412580 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.412726 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.412851 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.412977 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.413574 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.413693 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.413966 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-xz49t"] Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.414187 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-dzbhp"] Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.414634 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.415177 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.415216 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.415226 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.415311 4790 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.415565 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.416141 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.416174 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.418311 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.418439 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.418545 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.418560 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.418634 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.418814 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.423404 4790 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.424067 4790 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.425442 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.425697 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.425742 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.425753 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.425767 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.425780 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:12:57Z","lastTransitionTime":"2025-11-24T13:12:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.433454 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-2687d" event={"ID":"206d383c-05e9-47a3-a2e7-1535bf249e6a","Type":"ContainerStarted","Data":"5e7ae2e93a93cfd124de5d08e160ea6c602be7f48b1c0d7d5c9e272b9dfff194"} Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.435331 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-76rkg" event={"ID":"6c0e3499-87ac-481f-a010-708a3a4a730f","Type":"ContainerStarted","Data":"71131fc7278d255fb19ad7a8c1574e8abd25d3f2ef3e9fb3ef3a0d3e69f8312d"} Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.439305 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: E1124 13:12:57.441523 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.444932 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.444961 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.444971 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.444985 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.444994 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:12:57Z","lastTransitionTime":"2025-11-24T13:12:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.455430 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: E1124 13:12:57.457244 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.461831 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.461867 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.461891 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.461908 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.461917 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:12:57Z","lastTransitionTime":"2025-11-24T13:12:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:12:57 crc kubenswrapper[4790]: E1124 13:12:57.474462 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.475142 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.478134 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.478169 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.478181 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.478198 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.478210 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:12:57Z","lastTransitionTime":"2025-11-24T13:12:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.487776 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: E1124 13:12:57.489364 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.495383 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.495416 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.495424 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.495436 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.495444 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:12:57Z","lastTransitionTime":"2025-11-24T13:12:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.502464 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: E1124 13:12:57.508696 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: E1124 13:12:57.508811 4790 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.511147 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.511191 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.511204 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.511222 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.511235 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:12:57Z","lastTransitionTime":"2025-11-24T13:12:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.513011 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.523247 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.534476 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.546024 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.556624 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.567658 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.574474 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.574610 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/41874b85-3fa0-4d37-b5a0-ced5689c732f-tuning-conf-dir\") pod \"multus-additional-cni-plugins-dzbhp\" (UID: \"41874b85-3fa0-4d37-b5a0-ced5689c732f\") " pod="openshift-multus/multus-additional-cni-plugins-dzbhp" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.574712 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/41874b85-3fa0-4d37-b5a0-ced5689c732f-system-cni-dir\") pod \"multus-additional-cni-plugins-dzbhp\" (UID: \"41874b85-3fa0-4d37-b5a0-ced5689c732f\") " pod="openshift-multus/multus-additional-cni-plugins-dzbhp" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.575088 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0d73b133-48f0-455f-8f6a-742e633f631a-proxy-tls\") pod \"machine-config-daemon-xz49t\" (UID: \"0d73b133-48f0-455f-8f6a-742e633f631a\") " pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.575184 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/41874b85-3fa0-4d37-b5a0-ced5689c732f-cnibin\") pod \"multus-additional-cni-plugins-dzbhp\" (UID: \"41874b85-3fa0-4d37-b5a0-ced5689c732f\") " pod="openshift-multus/multus-additional-cni-plugins-dzbhp" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.575271 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wn84r\" (UniqueName: \"kubernetes.io/projected/0d73b133-48f0-455f-8f6a-742e633f631a-kube-api-access-wn84r\") pod \"machine-config-daemon-xz49t\" (UID: \"0d73b133-48f0-455f-8f6a-742e633f631a\") " pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.575382 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-kubelet\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.575415 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-etc-openvswitch\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.575433 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-node-log\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.575448 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-cni-netd\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.575463 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/0d73b133-48f0-455f-8f6a-742e633f631a-rootfs\") pod \"machine-config-daemon-xz49t\" (UID: \"0d73b133-48f0-455f-8f6a-742e633f631a\") " pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.575482 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0d73b133-48f0-455f-8f6a-742e633f631a-mcd-auth-proxy-config\") pod \"machine-config-daemon-xz49t\" (UID: \"0d73b133-48f0-455f-8f6a-742e633f631a\") " pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.575500 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-run-netns\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.575514 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-cni-bin\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.575529 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/528a4b22-34ba-41b8-8c7d-07d98eebe02b-ovn-node-metrics-cert\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.575561 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5rsh\" (UniqueName: \"kubernetes.io/projected/41874b85-3fa0-4d37-b5a0-ced5689c732f-kube-api-access-s5rsh\") pod \"multus-additional-cni-plugins-dzbhp\" (UID: \"41874b85-3fa0-4d37-b5a0-ced5689c732f\") " pod="openshift-multus/multus-additional-cni-plugins-dzbhp" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.575577 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-run-systemd\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.575592 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-run-ovn\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.575610 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/528a4b22-34ba-41b8-8c7d-07d98eebe02b-env-overrides\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.575688 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-var-lib-openvswitch\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.575720 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-run-ovn-kubernetes\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.575847 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/41874b85-3fa0-4d37-b5a0-ced5689c732f-cni-binary-copy\") pod \"multus-additional-cni-plugins-dzbhp\" (UID: \"41874b85-3fa0-4d37-b5a0-ced5689c732f\") " pod="openshift-multus/multus-additional-cni-plugins-dzbhp" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.575873 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/41874b85-3fa0-4d37-b5a0-ced5689c732f-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-dzbhp\" (UID: \"41874b85-3fa0-4d37-b5a0-ced5689c732f\") " pod="openshift-multus/multus-additional-cni-plugins-dzbhp" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.575930 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-run-openvswitch\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.575965 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-systemd-units\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.575985 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcnzl\" (UniqueName: \"kubernetes.io/projected/528a4b22-34ba-41b8-8c7d-07d98eebe02b-kube-api-access-bcnzl\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.576006 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-slash\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.576028 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-log-socket\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.576048 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/528a4b22-34ba-41b8-8c7d-07d98eebe02b-ovnkube-config\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.576091 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/41874b85-3fa0-4d37-b5a0-ced5689c732f-os-release\") pod \"multus-additional-cni-plugins-dzbhp\" (UID: \"41874b85-3fa0-4d37-b5a0-ced5689c732f\") " pod="openshift-multus/multus-additional-cni-plugins-dzbhp" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.576648 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/528a4b22-34ba-41b8-8c7d-07d98eebe02b-ovnkube-script-lib\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.586417 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.599188 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.612047 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.613557 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.613583 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.613592 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.613606 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.613616 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:12:57Z","lastTransitionTime":"2025-11-24T13:12:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.623359 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.634669 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.650338 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.663040 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.673406 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677256 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/41874b85-3fa0-4d37-b5a0-ced5689c732f-cnibin\") pod \"multus-additional-cni-plugins-dzbhp\" (UID: \"41874b85-3fa0-4d37-b5a0-ced5689c732f\") " pod="openshift-multus/multus-additional-cni-plugins-dzbhp" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677309 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-kubelet\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677333 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-etc-openvswitch\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677356 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wn84r\" (UniqueName: \"kubernetes.io/projected/0d73b133-48f0-455f-8f6a-742e633f631a-kube-api-access-wn84r\") pod \"machine-config-daemon-xz49t\" (UID: \"0d73b133-48f0-455f-8f6a-742e633f631a\") " pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677378 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-node-log\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677378 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/41874b85-3fa0-4d37-b5a0-ced5689c732f-cnibin\") pod \"multus-additional-cni-plugins-dzbhp\" (UID: \"41874b85-3fa0-4d37-b5a0-ced5689c732f\") " pod="openshift-multus/multus-additional-cni-plugins-dzbhp" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677389 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-kubelet\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677400 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-cni-netd\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677437 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/0d73b133-48f0-455f-8f6a-742e633f631a-rootfs\") pod \"machine-config-daemon-xz49t\" (UID: \"0d73b133-48f0-455f-8f6a-742e633f631a\") " pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677440 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-node-log\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677439 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-etc-openvswitch\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677459 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0d73b133-48f0-455f-8f6a-742e633f631a-mcd-auth-proxy-config\") pod \"machine-config-daemon-xz49t\" (UID: \"0d73b133-48f0-455f-8f6a-742e633f631a\") " pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677476 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/0d73b133-48f0-455f-8f6a-742e633f631a-rootfs\") pod \"machine-config-daemon-xz49t\" (UID: \"0d73b133-48f0-455f-8f6a-742e633f631a\") " pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677438 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-cni-netd\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677513 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-cni-bin\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677484 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-cni-bin\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677554 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/528a4b22-34ba-41b8-8c7d-07d98eebe02b-ovn-node-metrics-cert\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677610 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-run-netns\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677633 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-run-systemd\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677657 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-run-ovn\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677664 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-run-netns\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677678 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/528a4b22-34ba-41b8-8c7d-07d98eebe02b-env-overrides\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677725 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5rsh\" (UniqueName: \"kubernetes.io/projected/41874b85-3fa0-4d37-b5a0-ced5689c732f-kube-api-access-s5rsh\") pod \"multus-additional-cni-plugins-dzbhp\" (UID: \"41874b85-3fa0-4d37-b5a0-ced5689c732f\") " pod="openshift-multus/multus-additional-cni-plugins-dzbhp" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677737 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-run-systemd\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677755 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-var-lib-openvswitch\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677760 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-run-ovn\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677776 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-run-ovn-kubernetes\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677800 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-var-lib-openvswitch\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677808 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/41874b85-3fa0-4d37-b5a0-ced5689c732f-cni-binary-copy\") pod \"multus-additional-cni-plugins-dzbhp\" (UID: \"41874b85-3fa0-4d37-b5a0-ced5689c732f\") " pod="openshift-multus/multus-additional-cni-plugins-dzbhp" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677843 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/41874b85-3fa0-4d37-b5a0-ced5689c732f-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-dzbhp\" (UID: \"41874b85-3fa0-4d37-b5a0-ced5689c732f\") " pod="openshift-multus/multus-additional-cni-plugins-dzbhp" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677869 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-run-openvswitch\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677911 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-systemd-units\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677938 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-slash\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677960 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-log-socket\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677962 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-run-openvswitch\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.677981 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcnzl\" (UniqueName: \"kubernetes.io/projected/528a4b22-34ba-41b8-8c7d-07d98eebe02b-kube-api-access-bcnzl\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.678002 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/528a4b22-34ba-41b8-8c7d-07d98eebe02b-ovnkube-config\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.678021 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/41874b85-3fa0-4d37-b5a0-ced5689c732f-os-release\") pod \"multus-additional-cni-plugins-dzbhp\" (UID: \"41874b85-3fa0-4d37-b5a0-ced5689c732f\") " pod="openshift-multus/multus-additional-cni-plugins-dzbhp" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.678084 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/528a4b22-34ba-41b8-8c7d-07d98eebe02b-ovnkube-script-lib\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.678108 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.678131 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/41874b85-3fa0-4d37-b5a0-ced5689c732f-tuning-conf-dir\") pod \"multus-additional-cni-plugins-dzbhp\" (UID: \"41874b85-3fa0-4d37-b5a0-ced5689c732f\") " pod="openshift-multus/multus-additional-cni-plugins-dzbhp" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.678155 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0d73b133-48f0-455f-8f6a-742e633f631a-proxy-tls\") pod \"machine-config-daemon-xz49t\" (UID: \"0d73b133-48f0-455f-8f6a-742e633f631a\") " pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.678172 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0d73b133-48f0-455f-8f6a-742e633f631a-mcd-auth-proxy-config\") pod \"machine-config-daemon-xz49t\" (UID: \"0d73b133-48f0-455f-8f6a-742e633f631a\") " pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.678216 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-systemd-units\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.678217 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/41874b85-3fa0-4d37-b5a0-ced5689c732f-system-cni-dir\") pod \"multus-additional-cni-plugins-dzbhp\" (UID: \"41874b85-3fa0-4d37-b5a0-ced5689c732f\") " pod="openshift-multus/multus-additional-cni-plugins-dzbhp" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.678177 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/41874b85-3fa0-4d37-b5a0-ced5689c732f-system-cni-dir\") pod \"multus-additional-cni-plugins-dzbhp\" (UID: \"41874b85-3fa0-4d37-b5a0-ced5689c732f\") " pod="openshift-multus/multus-additional-cni-plugins-dzbhp" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.678268 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/41874b85-3fa0-4d37-b5a0-ced5689c732f-os-release\") pod \"multus-additional-cni-plugins-dzbhp\" (UID: \"41874b85-3fa0-4d37-b5a0-ced5689c732f\") " pod="openshift-multus/multus-additional-cni-plugins-dzbhp" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.678296 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-log-socket\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.678348 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-slash\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.678393 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/41874b85-3fa0-4d37-b5a0-ced5689c732f-cni-binary-copy\") pod \"multus-additional-cni-plugins-dzbhp\" (UID: \"41874b85-3fa0-4d37-b5a0-ced5689c732f\") " pod="openshift-multus/multus-additional-cni-plugins-dzbhp" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.678429 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-run-ovn-kubernetes\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.678427 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/41874b85-3fa0-4d37-b5a0-ced5689c732f-tuning-conf-dir\") pod \"multus-additional-cni-plugins-dzbhp\" (UID: \"41874b85-3fa0-4d37-b5a0-ced5689c732f\") " pod="openshift-multus/multus-additional-cni-plugins-dzbhp" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.678459 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.678600 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/528a4b22-34ba-41b8-8c7d-07d98eebe02b-env-overrides\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.678639 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/41874b85-3fa0-4d37-b5a0-ced5689c732f-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-dzbhp\" (UID: \"41874b85-3fa0-4d37-b5a0-ced5689c732f\") " pod="openshift-multus/multus-additional-cni-plugins-dzbhp" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.678855 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/528a4b22-34ba-41b8-8c7d-07d98eebe02b-ovnkube-script-lib\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.679042 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/528a4b22-34ba-41b8-8c7d-07d98eebe02b-ovnkube-config\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.685280 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/528a4b22-34ba-41b8-8c7d-07d98eebe02b-ovn-node-metrics-cert\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.685464 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0d73b133-48f0-455f-8f6a-742e633f631a-proxy-tls\") pod \"machine-config-daemon-xz49t\" (UID: \"0d73b133-48f0-455f-8f6a-742e633f631a\") " pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.688504 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.693020 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wn84r\" (UniqueName: \"kubernetes.io/projected/0d73b133-48f0-455f-8f6a-742e633f631a-kube-api-access-wn84r\") pod \"machine-config-daemon-xz49t\" (UID: \"0d73b133-48f0-455f-8f6a-742e633f631a\") " pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.694824 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcnzl\" (UniqueName: \"kubernetes.io/projected/528a4b22-34ba-41b8-8c7d-07d98eebe02b-kube-api-access-bcnzl\") pod \"ovnkube-node-5x94h\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.695313 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5rsh\" (UniqueName: \"kubernetes.io/projected/41874b85-3fa0-4d37-b5a0-ced5689c732f-kube-api-access-s5rsh\") pod \"multus-additional-cni-plugins-dzbhp\" (UID: \"41874b85-3fa0-4d37-b5a0-ced5689c732f\") " pod="openshift-multus/multus-additional-cni-plugins-dzbhp" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.700417 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.714041 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.715659 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.715697 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.715709 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.715725 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.715735 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:12:57Z","lastTransitionTime":"2025-11-24T13:12:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.724488 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.743716 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.752622 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" Nov 24 13:12:57 crc kubenswrapper[4790]: W1124 13:12:57.756045 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod528a4b22_34ba_41b8_8c7d_07d98eebe02b.slice/crio-af11f648dbb79e4be6d1491f965cd72809a6b9117b650da1e064e2270043adaf WatchSource:0}: Error finding container af11f648dbb79e4be6d1491f965cd72809a6b9117b650da1e064e2270043adaf: Status 404 returned error can't find the container with id af11f648dbb79e4be6d1491f965cd72809a6b9117b650da1e064e2270043adaf Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.759211 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 13:12:57 crc kubenswrapper[4790]: W1124 13:12:57.766652 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod41874b85_3fa0_4d37_b5a0_ced5689c732f.slice/crio-ec2baea7a35bc16adb6c68cebebded1b6e0c8a51215c31bc2f7519239c3512c2 WatchSource:0}: Error finding container ec2baea7a35bc16adb6c68cebebded1b6e0c8a51215c31bc2f7519239c3512c2: Status 404 returned error can't find the container with id ec2baea7a35bc16adb6c68cebebded1b6e0c8a51215c31bc2f7519239c3512c2 Nov 24 13:12:57 crc kubenswrapper[4790]: W1124 13:12:57.773516 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0d73b133_48f0_455f_8f6a_742e633f631a.slice/crio-b6236eeae01428a13a2878b66c13682f19ad2e42f6acd4d0be87695fe278a434 WatchSource:0}: Error finding container b6236eeae01428a13a2878b66c13682f19ad2e42f6acd4d0be87695fe278a434: Status 404 returned error can't find the container with id b6236eeae01428a13a2878b66c13682f19ad2e42f6acd4d0be87695fe278a434 Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.819665 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.819697 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.819709 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.819726 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.819737 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:12:57Z","lastTransitionTime":"2025-11-24T13:12:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.923250 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.923283 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.923292 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.923304 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:12:57 crc kubenswrapper[4790]: I1124 13:12:57.923313 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:12:57Z","lastTransitionTime":"2025-11-24T13:12:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.026373 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.026411 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.026422 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.026436 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.026449 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:12:58Z","lastTransitionTime":"2025-11-24T13:12:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.128260 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.128293 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.128304 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.128321 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.128346 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:12:58Z","lastTransitionTime":"2025-11-24T13:12:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.230401 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.230441 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.230452 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.230467 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.230476 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:12:58Z","lastTransitionTime":"2025-11-24T13:12:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.332513 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.332553 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.332563 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.332577 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.332587 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:12:58Z","lastTransitionTime":"2025-11-24T13:12:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.435176 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.435215 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.435225 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.435240 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.435252 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:12:58Z","lastTransitionTime":"2025-11-24T13:12:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.438598 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-2687d" event={"ID":"206d383c-05e9-47a3-a2e7-1535bf249e6a","Type":"ContainerStarted","Data":"6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790"} Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.441084 4790 generic.go:334] "Generic (PLEG): container finished" podID="41874b85-3fa0-4d37-b5a0-ced5689c732f" containerID="2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385" exitCode=0 Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.441170 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" event={"ID":"41874b85-3fa0-4d37-b5a0-ced5689c732f","Type":"ContainerDied","Data":"2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385"} Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.441206 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" event={"ID":"41874b85-3fa0-4d37-b5a0-ced5689c732f","Type":"ContainerStarted","Data":"ec2baea7a35bc16adb6c68cebebded1b6e0c8a51215c31bc2f7519239c3512c2"} Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.442506 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-76rkg" event={"ID":"6c0e3499-87ac-481f-a010-708a3a4a730f","Type":"ContainerStarted","Data":"6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09"} Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.444172 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8"} Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.444199 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989"} Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.444210 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"b6236eeae01428a13a2878b66c13682f19ad2e42f6acd4d0be87695fe278a434"} Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.447826 4790 generic.go:334] "Generic (PLEG): container finished" podID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerID="9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33" exitCode=0 Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.447863 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" event={"ID":"528a4b22-34ba-41b8-8c7d-07d98eebe02b","Type":"ContainerDied","Data":"9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33"} Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.447902 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" event={"ID":"528a4b22-34ba-41b8-8c7d-07d98eebe02b","Type":"ContainerStarted","Data":"af11f648dbb79e4be6d1491f965cd72809a6b9117b650da1e064e2270043adaf"} Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.455154 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.470485 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.494232 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.509331 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.527810 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.538308 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.538342 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.538352 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.538367 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.538378 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:12:58Z","lastTransitionTime":"2025-11-24T13:12:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.545607 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.557395 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.572300 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.589468 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.604246 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.620673 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.635329 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.641828 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.641890 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.642060 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.642084 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.642094 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:12:58Z","lastTransitionTime":"2025-11-24T13:12:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.652135 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.670690 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.690002 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.705335 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.722578 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.739360 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.745507 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.745561 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.745576 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.745598 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.745611 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:12:58Z","lastTransitionTime":"2025-11-24T13:12:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.754455 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.778690 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.798151 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.813060 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.826198 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.840798 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.847993 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.848030 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.848040 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.848055 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.848074 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:12:58Z","lastTransitionTime":"2025-11-24T13:12:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.852319 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.864361 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.891678 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:12:58 crc kubenswrapper[4790]: E1124 13:12:58.891922 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:13:06.891907099 +0000 UTC m=+35.271800751 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.955321 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.955370 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.955380 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.955395 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.955406 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:12:58Z","lastTransitionTime":"2025-11-24T13:12:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.993476 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.993531 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.993557 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:12:58 crc kubenswrapper[4790]: I1124 13:12:58.993586 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:12:58 crc kubenswrapper[4790]: E1124 13:12:58.993929 4790 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:12:58 crc kubenswrapper[4790]: E1124 13:12:58.993952 4790 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:12:58 crc kubenswrapper[4790]: E1124 13:12:58.993967 4790 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:12:58 crc kubenswrapper[4790]: E1124 13:12:58.994024 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 13:13:06.994007366 +0000 UTC m=+35.373901028 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:12:58 crc kubenswrapper[4790]: E1124 13:12:58.994403 4790 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:12:58 crc kubenswrapper[4790]: E1124 13:12:58.994440 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:13:06.994429848 +0000 UTC m=+35.374323510 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:12:58 crc kubenswrapper[4790]: E1124 13:12:58.994496 4790 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:12:58 crc kubenswrapper[4790]: E1124 13:12:58.994509 4790 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:12:58 crc kubenswrapper[4790]: E1124 13:12:58.994519 4790 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:12:58 crc kubenswrapper[4790]: E1124 13:12:58.994546 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 13:13:06.994537511 +0000 UTC m=+35.374431163 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:12:58 crc kubenswrapper[4790]: E1124 13:12:58.994594 4790 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:12:58 crc kubenswrapper[4790]: E1124 13:12:58.994624 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:13:06.994616304 +0000 UTC m=+35.374509966 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.057625 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.057659 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.057670 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.057687 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.057698 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:12:59Z","lastTransitionTime":"2025-11-24T13:12:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.160703 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.161229 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.161240 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.161259 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.161269 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:12:59Z","lastTransitionTime":"2025-11-24T13:12:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.263692 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.263722 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.263732 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.263744 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.263753 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:12:59Z","lastTransitionTime":"2025-11-24T13:12:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.314011 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.314063 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.314088 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:12:59 crc kubenswrapper[4790]: E1124 13:12:59.314125 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:12:59 crc kubenswrapper[4790]: E1124 13:12:59.314207 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:12:59 crc kubenswrapper[4790]: E1124 13:12:59.314301 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.375449 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.375498 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.375603 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.375693 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.375925 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:12:59Z","lastTransitionTime":"2025-11-24T13:12:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.486871 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.486934 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.486945 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.486961 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.486972 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:12:59Z","lastTransitionTime":"2025-11-24T13:12:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.491439 4790 generic.go:334] "Generic (PLEG): container finished" podID="41874b85-3fa0-4d37-b5a0-ced5689c732f" containerID="ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142" exitCode=0 Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.491543 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" event={"ID":"41874b85-3fa0-4d37-b5a0-ced5689c732f","Type":"ContainerDied","Data":"ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142"} Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.502666 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" event={"ID":"528a4b22-34ba-41b8-8c7d-07d98eebe02b","Type":"ContainerStarted","Data":"ad72e45458f88569f2bb77a55dc98e2bad29a7e451e5258d371722616042bebd"} Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.503155 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" event={"ID":"528a4b22-34ba-41b8-8c7d-07d98eebe02b","Type":"ContainerStarted","Data":"1a2bf135bf819ef8ce489337a9f6b7b0bb626ff335b37ec7e6a74e0e13a1ba40"} Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.503169 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" event={"ID":"528a4b22-34ba-41b8-8c7d-07d98eebe02b","Type":"ContainerStarted","Data":"ec06ff65e703be63a81e99901326ec1eaa20c2176ddbd0764cc7cdc0b5761bd5"} Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.503180 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" event={"ID":"528a4b22-34ba-41b8-8c7d-07d98eebe02b","Type":"ContainerStarted","Data":"23e4305b0dc57da3ad6d92e8f04f1f18780c1ebaf63ee19140a8443b99197bdd"} Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.503194 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" event={"ID":"528a4b22-34ba-41b8-8c7d-07d98eebe02b","Type":"ContainerStarted","Data":"908f7697965f4e6d820bfe32f7f59960e917dedea04373ff0e2b8126bb835163"} Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.514282 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.533535 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.548199 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.562068 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.576236 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.589150 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.592761 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.592799 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.592810 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.592829 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.592840 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:12:59Z","lastTransitionTime":"2025-11-24T13:12:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.601014 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.612992 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.626216 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.636078 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.646421 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.661249 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.673979 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:12:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.695519 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.695553 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.695561 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.695574 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.695583 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:12:59Z","lastTransitionTime":"2025-11-24T13:12:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.798277 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.798307 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.798316 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.798330 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.798338 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:12:59Z","lastTransitionTime":"2025-11-24T13:12:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.901102 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.901169 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.901181 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.901198 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:12:59 crc kubenswrapper[4790]: I1124 13:12:59.901210 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:12:59Z","lastTransitionTime":"2025-11-24T13:12:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.004017 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.004102 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.004128 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.004164 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.004203 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:00Z","lastTransitionTime":"2025-11-24T13:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.106784 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.106835 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.106848 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.106864 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.106896 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:00Z","lastTransitionTime":"2025-11-24T13:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.209706 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.209763 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.209776 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.209800 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.209814 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:00Z","lastTransitionTime":"2025-11-24T13:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.313052 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.313112 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.313179 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.313211 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.313233 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:00Z","lastTransitionTime":"2025-11-24T13:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.323837 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-nlnv5"] Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.324293 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-nlnv5" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.327314 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.327443 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.327457 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.327861 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.341089 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.356193 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.368800 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.383088 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.394670 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nlnv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0f7e0af-ed9a-4a5e-be2f-06abac132790\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wxh2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nlnv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.405580 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.415315 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.415348 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.415357 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.415370 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.415381 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:00Z","lastTransitionTime":"2025-11-24T13:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.416853 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.429216 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.442367 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.452685 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.464593 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.476776 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.488478 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.507591 4790 generic.go:334] "Generic (PLEG): container finished" podID="41874b85-3fa0-4d37-b5a0-ced5689c732f" containerID="424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f" exitCode=0 Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.507685 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" event={"ID":"41874b85-3fa0-4d37-b5a0-ced5689c732f","Type":"ContainerDied","Data":"424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f"} Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.508047 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxh2b\" (UniqueName: \"kubernetes.io/projected/e0f7e0af-ed9a-4a5e-be2f-06abac132790-kube-api-access-wxh2b\") pod \"node-ca-nlnv5\" (UID: \"e0f7e0af-ed9a-4a5e-be2f-06abac132790\") " pod="openshift-image-registry/node-ca-nlnv5" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.508116 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e0f7e0af-ed9a-4a5e-be2f-06abac132790-serviceca\") pod \"node-ca-nlnv5\" (UID: \"e0f7e0af-ed9a-4a5e-be2f-06abac132790\") " pod="openshift-image-registry/node-ca-nlnv5" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.508253 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e0f7e0af-ed9a-4a5e-be2f-06abac132790-host\") pod \"node-ca-nlnv5\" (UID: \"e0f7e0af-ed9a-4a5e-be2f-06abac132790\") " pod="openshift-image-registry/node-ca-nlnv5" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.508832 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.511640 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" event={"ID":"528a4b22-34ba-41b8-8c7d-07d98eebe02b","Type":"ContainerStarted","Data":"9f3d5e18b5f456b7627a9154ff8fb00945ec026d4fb8797d4c4f3e2a2fbb1db7"} Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.517359 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.517389 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.517397 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.517410 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.517420 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:00Z","lastTransitionTime":"2025-11-24T13:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.523671 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.544013 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.559086 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.574262 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.586583 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.601486 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.609496 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxh2b\" (UniqueName: \"kubernetes.io/projected/e0f7e0af-ed9a-4a5e-be2f-06abac132790-kube-api-access-wxh2b\") pod \"node-ca-nlnv5\" (UID: \"e0f7e0af-ed9a-4a5e-be2f-06abac132790\") " pod="openshift-image-registry/node-ca-nlnv5" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.609551 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e0f7e0af-ed9a-4a5e-be2f-06abac132790-serviceca\") pod \"node-ca-nlnv5\" (UID: \"e0f7e0af-ed9a-4a5e-be2f-06abac132790\") " pod="openshift-image-registry/node-ca-nlnv5" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.609576 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e0f7e0af-ed9a-4a5e-be2f-06abac132790-host\") pod \"node-ca-nlnv5\" (UID: \"e0f7e0af-ed9a-4a5e-be2f-06abac132790\") " pod="openshift-image-registry/node-ca-nlnv5" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.609987 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e0f7e0af-ed9a-4a5e-be2f-06abac132790-host\") pod \"node-ca-nlnv5\" (UID: \"e0f7e0af-ed9a-4a5e-be2f-06abac132790\") " pod="openshift-image-registry/node-ca-nlnv5" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.611086 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e0f7e0af-ed9a-4a5e-be2f-06abac132790-serviceca\") pod \"node-ca-nlnv5\" (UID: \"e0f7e0af-ed9a-4a5e-be2f-06abac132790\") " pod="openshift-image-registry/node-ca-nlnv5" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.614734 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.619612 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.619804 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.619928 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.620040 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.620129 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:00Z","lastTransitionTime":"2025-11-24T13:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.627657 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.633221 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxh2b\" (UniqueName: \"kubernetes.io/projected/e0f7e0af-ed9a-4a5e-be2f-06abac132790-kube-api-access-wxh2b\") pod \"node-ca-nlnv5\" (UID: \"e0f7e0af-ed9a-4a5e-be2f-06abac132790\") " pod="openshift-image-registry/node-ca-nlnv5" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.636167 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-nlnv5" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.648130 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:00 crc kubenswrapper[4790]: W1124 13:13:00.654309 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0f7e0af_ed9a_4a5e_be2f_06abac132790.slice/crio-4705bf500e209e87489932b2386398de3749b32b622e7d4c989fcf79c8916608 WatchSource:0}: Error finding container 4705bf500e209e87489932b2386398de3749b32b622e7d4c989fcf79c8916608: Status 404 returned error can't find the container with id 4705bf500e209e87489932b2386398de3749b32b622e7d4c989fcf79c8916608 Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.661332 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.672212 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.685975 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.700324 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.714861 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nlnv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0f7e0af-ed9a-4a5e-be2f-06abac132790\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wxh2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nlnv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.724901 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.724933 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.724945 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.724961 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.724972 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:00Z","lastTransitionTime":"2025-11-24T13:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.827473 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.827507 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.827515 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.827528 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.827574 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:00Z","lastTransitionTime":"2025-11-24T13:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.930070 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.930101 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.930110 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.930122 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:00 crc kubenswrapper[4790]: I1124 13:13:00.930131 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:00Z","lastTransitionTime":"2025-11-24T13:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.032283 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.032321 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.032329 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.032344 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.032354 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:01Z","lastTransitionTime":"2025-11-24T13:13:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.134753 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.134789 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.134800 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.134816 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.134827 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:01Z","lastTransitionTime":"2025-11-24T13:13:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.236809 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.236846 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.236855 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.236867 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.236889 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:01Z","lastTransitionTime":"2025-11-24T13:13:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.313969 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.314030 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:13:01 crc kubenswrapper[4790]: E1124 13:13:01.314085 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.313973 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:01 crc kubenswrapper[4790]: E1124 13:13:01.314168 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:13:01 crc kubenswrapper[4790]: E1124 13:13:01.314268 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.339100 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.339131 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.339141 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.339154 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.339163 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:01Z","lastTransitionTime":"2025-11-24T13:13:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.441495 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.441528 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.441539 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.441554 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.441565 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:01Z","lastTransitionTime":"2025-11-24T13:13:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.517993 4790 generic.go:334] "Generic (PLEG): container finished" podID="41874b85-3fa0-4d37-b5a0-ced5689c732f" containerID="b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6" exitCode=0 Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.518080 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" event={"ID":"41874b85-3fa0-4d37-b5a0-ced5689c732f","Type":"ContainerDied","Data":"b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6"} Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.520645 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-nlnv5" event={"ID":"e0f7e0af-ed9a-4a5e-be2f-06abac132790","Type":"ContainerStarted","Data":"48ad38a0b670b452c1c5d9a287fb396d5e9eed494839d3bccac74dd5456ea3d4"} Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.520676 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-nlnv5" event={"ID":"e0f7e0af-ed9a-4a5e-be2f-06abac132790","Type":"ContainerStarted","Data":"4705bf500e209e87489932b2386398de3749b32b622e7d4c989fcf79c8916608"} Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.533337 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.544967 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.545003 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.545012 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.545032 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.545047 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:01Z","lastTransitionTime":"2025-11-24T13:13:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.546451 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.559447 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.570316 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.590294 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.606151 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.616043 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.628629 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.639169 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.647045 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.647081 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.647091 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.647106 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.647116 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:01Z","lastTransitionTime":"2025-11-24T13:13:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.651169 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nlnv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0f7e0af-ed9a-4a5e-be2f-06abac132790\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wxh2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nlnv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.664019 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.675593 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.690063 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.704758 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.717956 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.730043 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.740248 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nlnv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0f7e0af-ed9a-4a5e-be2f-06abac132790\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ad38a0b670b452c1c5d9a287fb396d5e9eed494839d3bccac74dd5456ea3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wxh2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nlnv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.749872 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.749925 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.749934 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.749948 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.749959 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:01Z","lastTransitionTime":"2025-11-24T13:13:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.751553 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.762020 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.771752 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.781819 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.792474 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.807042 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.818495 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.829043 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.845113 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.852647 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.852678 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.852687 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.852699 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.852708 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:01Z","lastTransitionTime":"2025-11-24T13:13:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.859158 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.869894 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.955249 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.955285 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.955294 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.955308 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:01 crc kubenswrapper[4790]: I1124 13:13:01.955318 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:01Z","lastTransitionTime":"2025-11-24T13:13:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.057740 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.057771 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.057779 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.057792 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.057801 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:02Z","lastTransitionTime":"2025-11-24T13:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.160443 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.160513 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.160522 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.160535 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.160545 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:02Z","lastTransitionTime":"2025-11-24T13:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.262129 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.262763 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.262829 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.262905 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.262972 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:02Z","lastTransitionTime":"2025-11-24T13:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.326981 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.339587 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.354452 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.365947 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.365989 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.366000 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.366016 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.366027 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:02Z","lastTransitionTime":"2025-11-24T13:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.370145 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.383544 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.403599 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.415581 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.427028 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.437962 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.449058 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.458246 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nlnv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0f7e0af-ed9a-4a5e-be2f-06abac132790\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ad38a0b670b452c1c5d9a287fb396d5e9eed494839d3bccac74dd5456ea3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wxh2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nlnv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.467928 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.467954 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.467962 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.467974 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.467983 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:02Z","lastTransitionTime":"2025-11-24T13:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.468830 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.479143 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.489775 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.527164 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" event={"ID":"528a4b22-34ba-41b8-8c7d-07d98eebe02b","Type":"ContainerStarted","Data":"b16be7938ff339aa0f92c4fb0d444114ac253439e80bf12ceb55c7e9a1944603"} Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.529596 4790 generic.go:334] "Generic (PLEG): container finished" podID="41874b85-3fa0-4d37-b5a0-ced5689c732f" containerID="ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf" exitCode=0 Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.529642 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" event={"ID":"41874b85-3fa0-4d37-b5a0-ced5689c732f","Type":"ContainerDied","Data":"ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf"} Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.545314 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.558456 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.570415 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.570467 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.570491 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.570504 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.570513 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:02Z","lastTransitionTime":"2025-11-24T13:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.572174 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.582927 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.593373 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.612824 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.625493 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.636624 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.648080 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.657578 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.667217 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nlnv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0f7e0af-ed9a-4a5e-be2f-06abac132790\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ad38a0b670b452c1c5d9a287fb396d5e9eed494839d3bccac74dd5456ea3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wxh2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nlnv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.674724 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.674780 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.674792 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.674809 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.674822 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:02Z","lastTransitionTime":"2025-11-24T13:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.679546 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.691512 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.703471 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.777797 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.777844 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.777853 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.777868 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.777879 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:02Z","lastTransitionTime":"2025-11-24T13:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.880669 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.880724 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.880739 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.880761 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.880773 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:02Z","lastTransitionTime":"2025-11-24T13:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.982824 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.982869 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.982877 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.982908 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:02 crc kubenswrapper[4790]: I1124 13:13:02.982917 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:02Z","lastTransitionTime":"2025-11-24T13:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.085736 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.085854 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.085902 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.085929 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.085944 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:03Z","lastTransitionTime":"2025-11-24T13:13:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.189466 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.189503 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.189511 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.189524 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.189536 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:03Z","lastTransitionTime":"2025-11-24T13:13:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.292701 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.292747 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.292760 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.292777 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.292790 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:03Z","lastTransitionTime":"2025-11-24T13:13:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.313799 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.313837 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.313924 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:13:03 crc kubenswrapper[4790]: E1124 13:13:03.313961 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:13:03 crc kubenswrapper[4790]: E1124 13:13:03.314061 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:13:03 crc kubenswrapper[4790]: E1124 13:13:03.314123 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.395440 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.395485 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.395496 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.395516 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.395526 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:03Z","lastTransitionTime":"2025-11-24T13:13:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.497458 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.497509 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.497522 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.497539 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.497549 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:03Z","lastTransitionTime":"2025-11-24T13:13:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.535570 4790 generic.go:334] "Generic (PLEG): container finished" podID="41874b85-3fa0-4d37-b5a0-ced5689c732f" containerID="8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a" exitCode=0 Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.535613 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" event={"ID":"41874b85-3fa0-4d37-b5a0-ced5689c732f","Type":"ContainerDied","Data":"8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a"} Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.552408 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.567728 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.578711 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.590459 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.599583 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.599808 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.599901 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.600008 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.600088 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:03Z","lastTransitionTime":"2025-11-24T13:13:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.602278 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.620291 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.636773 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.651437 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.668287 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.678092 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.689437 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nlnv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0f7e0af-ed9a-4a5e-be2f-06abac132790\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ad38a0b670b452c1c5d9a287fb396d5e9eed494839d3bccac74dd5456ea3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wxh2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nlnv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.702032 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.703096 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.703129 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.703139 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.703152 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.703161 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:03Z","lastTransitionTime":"2025-11-24T13:13:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.712870 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.724430 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.805939 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.805993 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.806003 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.806019 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.806046 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:03Z","lastTransitionTime":"2025-11-24T13:13:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.908660 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.908718 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.908734 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.908756 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:03 crc kubenswrapper[4790]: I1124 13:13:03.908776 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:03Z","lastTransitionTime":"2025-11-24T13:13:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.010644 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.010690 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.010701 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.010716 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.010726 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:04Z","lastTransitionTime":"2025-11-24T13:13:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.113107 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.113147 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.113159 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.113176 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.113186 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:04Z","lastTransitionTime":"2025-11-24T13:13:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.215697 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.215976 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.216068 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.216172 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.216246 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:04Z","lastTransitionTime":"2025-11-24T13:13:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.319296 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.319331 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.319342 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.319357 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.319369 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:04Z","lastTransitionTime":"2025-11-24T13:13:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.422074 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.422107 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.422116 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.422132 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.422143 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:04Z","lastTransitionTime":"2025-11-24T13:13:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.524231 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.524274 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.524282 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.524295 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.524305 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:04Z","lastTransitionTime":"2025-11-24T13:13:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.541775 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" event={"ID":"41874b85-3fa0-4d37-b5a0-ced5689c732f","Type":"ContainerStarted","Data":"fec678890b719f519b6740a344414413e17705e146440e130765a0f13b8a4407"} Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.546305 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" event={"ID":"528a4b22-34ba-41b8-8c7d-07d98eebe02b","Type":"ContainerStarted","Data":"baccf450bd4cf6f7b3e28c8885aa0db191fa487cfbd9610d92bdca5e0aef838e"} Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.546681 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.546703 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.556140 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.569080 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.579058 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.579139 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.581007 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.592420 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.602634 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nlnv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0f7e0af-ed9a-4a5e-be2f-06abac132790\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ad38a0b670b452c1c5d9a287fb396d5e9eed494839d3bccac74dd5456ea3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wxh2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nlnv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.613805 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.626289 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.626337 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.626698 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.626717 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.626726 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:04Z","lastTransitionTime":"2025-11-24T13:13:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.627318 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.640142 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.651697 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.664745 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.680438 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.694550 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.713956 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.729813 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.729861 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.729873 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.729909 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.729923 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:04Z","lastTransitionTime":"2025-11-24T13:13:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.730460 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fec678890b719f519b6740a344414413e17705e146440e130765a0f13b8a4407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.743590 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.756223 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.768452 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.780479 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.797161 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec06ff65e703be63a81e99901326ec1eaa20c2176ddbd0764cc7cdc0b5761bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a2bf135bf819ef8ce489337a9f6b7b0bb626ff335b37ec7e6a74e0e13a1ba40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f3d5e18b5f456b7627a9154ff8fb00945ec026d4fb8797d4c4f3e2a2fbb1db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad72e45458f88569f2bb77a55dc98e2bad29a7e451e5258d371722616042bebd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23e4305b0dc57da3ad6d92e8f04f1f18780c1ebaf63ee19140a8443b99197bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908f7697965f4e6d820bfe32f7f59960e917dedea04373ff0e2b8126bb835163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://baccf450bd4cf6f7b3e28c8885aa0db191fa487cfbd9610d92bdca5e0aef838e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b16be7938ff339aa0f92c4fb0d444114ac253439e80bf12ceb55c7e9a1944603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.808561 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fec678890b719f519b6740a344414413e17705e146440e130765a0f13b8a4407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.816756 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.829014 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.836797 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.836840 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.836849 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.836863 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.836873 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:04Z","lastTransitionTime":"2025-11-24T13:13:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.845056 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.857070 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nlnv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0f7e0af-ed9a-4a5e-be2f-06abac132790\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ad38a0b670b452c1c5d9a287fb396d5e9eed494839d3bccac74dd5456ea3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wxh2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nlnv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.870548 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.884060 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.895675 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.907540 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.940153 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.940252 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.940267 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.940312 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:04 crc kubenswrapper[4790]: I1124 13:13:04.940328 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:04Z","lastTransitionTime":"2025-11-24T13:13:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.042731 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.042775 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.042787 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.042804 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.042816 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:05Z","lastTransitionTime":"2025-11-24T13:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.146492 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.146539 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.146549 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.146577 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.146587 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:05Z","lastTransitionTime":"2025-11-24T13:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.248822 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.248893 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.248906 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.248923 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.248948 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:05Z","lastTransitionTime":"2025-11-24T13:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.313598 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.313685 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:05 crc kubenswrapper[4790]: E1124 13:13:05.313757 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.313605 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:13:05 crc kubenswrapper[4790]: E1124 13:13:05.313834 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:13:05 crc kubenswrapper[4790]: E1124 13:13:05.313975 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.351353 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.351411 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.351420 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.351432 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.351442 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:05Z","lastTransitionTime":"2025-11-24T13:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.454120 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.454164 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.454179 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.454196 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.454206 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:05Z","lastTransitionTime":"2025-11-24T13:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.549607 4790 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.556104 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.556356 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.556375 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.556394 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.556406 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:05Z","lastTransitionTime":"2025-11-24T13:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.658597 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.658631 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.658640 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.658653 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.658663 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:05Z","lastTransitionTime":"2025-11-24T13:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.761564 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.761623 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.761636 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.761656 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.761668 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:05Z","lastTransitionTime":"2025-11-24T13:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.864254 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.864303 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.864312 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.864329 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.864341 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:05Z","lastTransitionTime":"2025-11-24T13:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.966719 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.966753 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.966762 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.966777 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:05 crc kubenswrapper[4790]: I1124 13:13:05.966788 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:05Z","lastTransitionTime":"2025-11-24T13:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.069385 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.069415 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.069423 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.069468 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.069477 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:06Z","lastTransitionTime":"2025-11-24T13:13:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.172112 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.172156 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.172166 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.172185 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.172229 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:06Z","lastTransitionTime":"2025-11-24T13:13:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.275794 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.275837 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.275847 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.275863 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.275898 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:06Z","lastTransitionTime":"2025-11-24T13:13:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.377960 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.378045 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.378069 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.378134 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.378150 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:06Z","lastTransitionTime":"2025-11-24T13:13:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.481158 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.481198 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.481209 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.481228 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.481239 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:06Z","lastTransitionTime":"2025-11-24T13:13:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.552910 4790 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.585352 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.585426 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.585442 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.585461 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.585511 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:06Z","lastTransitionTime":"2025-11-24T13:13:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.687839 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.687915 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.687927 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.687940 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.687951 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:06Z","lastTransitionTime":"2025-11-24T13:13:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.791399 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.791484 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.791525 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.791546 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.791558 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:06Z","lastTransitionTime":"2025-11-24T13:13:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.894331 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.894374 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.894383 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.894398 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.894407 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:06Z","lastTransitionTime":"2025-11-24T13:13:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.969188 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:13:06 crc kubenswrapper[4790]: E1124 13:13:06.969336 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:13:22.969317874 +0000 UTC m=+51.349211526 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.997046 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.997082 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.997090 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.997105 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:06 crc kubenswrapper[4790]: I1124 13:13:06.997114 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:06Z","lastTransitionTime":"2025-11-24T13:13:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.070300 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.070345 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.070366 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.070389 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:13:07 crc kubenswrapper[4790]: E1124 13:13:07.070440 4790 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:13:07 crc kubenswrapper[4790]: E1124 13:13:07.070500 4790 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:13:07 crc kubenswrapper[4790]: E1124 13:13:07.070515 4790 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:13:07 crc kubenswrapper[4790]: E1124 13:13:07.070525 4790 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:13:07 crc kubenswrapper[4790]: E1124 13:13:07.070535 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:13:23.070508605 +0000 UTC m=+51.450402267 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:13:07 crc kubenswrapper[4790]: E1124 13:13:07.070551 4790 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:13:07 crc kubenswrapper[4790]: E1124 13:13:07.070565 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 13:13:23.070553826 +0000 UTC m=+51.450447488 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:13:07 crc kubenswrapper[4790]: E1124 13:13:07.070674 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:13:23.070654249 +0000 UTC m=+51.450547971 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:13:07 crc kubenswrapper[4790]: E1124 13:13:07.070770 4790 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:13:07 crc kubenswrapper[4790]: E1124 13:13:07.070790 4790 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:13:07 crc kubenswrapper[4790]: E1124 13:13:07.070804 4790 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:13:07 crc kubenswrapper[4790]: E1124 13:13:07.070867 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 13:13:23.070854575 +0000 UTC m=+51.450748237 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.099765 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.099811 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.099821 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.099836 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.099846 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:07Z","lastTransitionTime":"2025-11-24T13:13:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.207004 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.207047 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.207061 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.207076 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.207088 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:07Z","lastTransitionTime":"2025-11-24T13:13:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.309090 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.309132 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.309143 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.309160 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.309172 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:07Z","lastTransitionTime":"2025-11-24T13:13:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.313647 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.313763 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.313800 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:07 crc kubenswrapper[4790]: E1124 13:13:07.313814 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:13:07 crc kubenswrapper[4790]: E1124 13:13:07.313823 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:13:07 crc kubenswrapper[4790]: E1124 13:13:07.313964 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.410929 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.410990 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.411001 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.411016 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.411028 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:07Z","lastTransitionTime":"2025-11-24T13:13:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.513277 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.513314 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.513323 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.513342 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.513352 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:07Z","lastTransitionTime":"2025-11-24T13:13:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.616514 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.616561 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.616574 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.616589 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.616600 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:07Z","lastTransitionTime":"2025-11-24T13:13:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.718420 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.718467 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.718477 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.718492 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.718502 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:07Z","lastTransitionTime":"2025-11-24T13:13:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.818909 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.818948 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.818958 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.818971 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.818981 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:07Z","lastTransitionTime":"2025-11-24T13:13:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:07 crc kubenswrapper[4790]: E1124 13:13:07.830308 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:07Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.834255 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.834317 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.834329 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.834351 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.834363 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:07Z","lastTransitionTime":"2025-11-24T13:13:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:07 crc kubenswrapper[4790]: E1124 13:13:07.846713 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:07Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.850575 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.850617 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.850629 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.850649 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.850659 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:07Z","lastTransitionTime":"2025-11-24T13:13:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:07 crc kubenswrapper[4790]: E1124 13:13:07.863382 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:07Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.867093 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.867137 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.867145 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.867160 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.867168 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:07Z","lastTransitionTime":"2025-11-24T13:13:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:07 crc kubenswrapper[4790]: E1124 13:13:07.878459 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:07Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.881990 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.882025 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.882035 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.882049 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.882058 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:07Z","lastTransitionTime":"2025-11-24T13:13:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:07 crc kubenswrapper[4790]: E1124 13:13:07.894815 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:07Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:07 crc kubenswrapper[4790]: E1124 13:13:07.894977 4790 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.896569 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.896596 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.896605 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.896618 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.896626 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:07Z","lastTransitionTime":"2025-11-24T13:13:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.999358 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.999406 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.999418 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.999436 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:07 crc kubenswrapper[4790]: I1124 13:13:07.999455 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:07Z","lastTransitionTime":"2025-11-24T13:13:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.101543 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.101585 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.101594 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.101607 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.101616 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:08Z","lastTransitionTime":"2025-11-24T13:13:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.204084 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.204124 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.204133 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.204147 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.204157 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:08Z","lastTransitionTime":"2025-11-24T13:13:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.305973 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.306012 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.306021 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.306037 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.306049 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:08Z","lastTransitionTime":"2025-11-24T13:13:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.409571 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.409613 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.409625 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.409641 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.409755 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:08Z","lastTransitionTime":"2025-11-24T13:13:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.512416 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.512453 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.512464 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.512478 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.512489 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:08Z","lastTransitionTime":"2025-11-24T13:13:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.561099 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5x94h_528a4b22-34ba-41b8-8c7d-07d98eebe02b/ovnkube-controller/0.log" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.563734 4790 generic.go:334] "Generic (PLEG): container finished" podID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerID="baccf450bd4cf6f7b3e28c8885aa0db191fa487cfbd9610d92bdca5e0aef838e" exitCode=1 Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.563778 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" event={"ID":"528a4b22-34ba-41b8-8c7d-07d98eebe02b","Type":"ContainerDied","Data":"baccf450bd4cf6f7b3e28c8885aa0db191fa487cfbd9610d92bdca5e0aef838e"} Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.564442 4790 scope.go:117] "RemoveContainer" containerID="baccf450bd4cf6f7b3e28c8885aa0db191fa487cfbd9610d92bdca5e0aef838e" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.577481 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:08Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.594512 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:08Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.608045 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:08Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.616284 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.616316 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.616325 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.616337 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.616346 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:08Z","lastTransitionTime":"2025-11-24T13:13:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.623906 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:08Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.637867 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:08Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.651232 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:08Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.670605 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec06ff65e703be63a81e99901326ec1eaa20c2176ddbd0764cc7cdc0b5761bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a2bf135bf819ef8ce489337a9f6b7b0bb626ff335b37ec7e6a74e0e13a1ba40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f3d5e18b5f456b7627a9154ff8fb00945ec026d4fb8797d4c4f3e2a2fbb1db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad72e45458f88569f2bb77a55dc98e2bad29a7e451e5258d371722616042bebd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23e4305b0dc57da3ad6d92e8f04f1f18780c1ebaf63ee19140a8443b99197bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908f7697965f4e6d820bfe32f7f59960e917dedea04373ff0e2b8126bb835163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://baccf450bd4cf6f7b3e28c8885aa0db191fa487cfbd9610d92bdca5e0aef838e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baccf450bd4cf6f7b3e28c8885aa0db191fa487cfbd9610d92bdca5e0aef838e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"message\\\":\\\"Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:13:07.458859 6106 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:13:07.458938 6106 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 13:13:07.459147 6106 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:13:07.459217 6106 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 13:13:07.459277 6106 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 13:13:07.459314 6106 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 13:13:07.459337 6106 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 13:13:07.459372 6106 factory.go:656] Stopping watch factory\\\\nI1124 13:13:07.459406 6106 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 13:13:07.459432 6106 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 13:13:07.459456 6106 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 13:13:07.459480 6106 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b16be7938ff339aa0f92c4fb0d444114ac253439e80bf12ceb55c7e9a1944603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:08Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.685838 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fec678890b719f519b6740a344414413e17705e146440e130765a0f13b8a4407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:08Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.698993 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:08Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.711550 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:08Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.718382 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.718419 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.718432 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.718447 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.718458 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:08Z","lastTransitionTime":"2025-11-24T13:13:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.725492 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:08Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.735494 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:08Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.746019 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nlnv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0f7e0af-ed9a-4a5e-be2f-06abac132790\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ad38a0b670b452c1c5d9a287fb396d5e9eed494839d3bccac74dd5456ea3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wxh2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nlnv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:08Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.757104 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:08Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.820464 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.820507 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.820516 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.820530 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.820539 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:08Z","lastTransitionTime":"2025-11-24T13:13:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.922996 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.923031 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.923039 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.923055 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:08 crc kubenswrapper[4790]: I1124 13:13:08.923065 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:08Z","lastTransitionTime":"2025-11-24T13:13:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.025982 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.026024 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.026033 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.026048 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.026061 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:09Z","lastTransitionTime":"2025-11-24T13:13:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.127600 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.127638 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.127648 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.127666 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.127675 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:09Z","lastTransitionTime":"2025-11-24T13:13:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.230396 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.230444 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.230457 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.230478 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.230490 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:09Z","lastTransitionTime":"2025-11-24T13:13:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.314075 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.314137 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:13:09 crc kubenswrapper[4790]: E1124 13:13:09.314186 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:13:09 crc kubenswrapper[4790]: E1124 13:13:09.314218 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.314134 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:09 crc kubenswrapper[4790]: E1124 13:13:09.314294 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.332107 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.332137 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.332156 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.332170 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.332179 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:09Z","lastTransitionTime":"2025-11-24T13:13:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.434714 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.434761 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.434773 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.434791 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.434811 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:09Z","lastTransitionTime":"2025-11-24T13:13:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.505851 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn"] Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.506302 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.508023 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.508431 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.519005 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.527909 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nlnv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0f7e0af-ed9a-4a5e-be2f-06abac132790\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ad38a0b670b452c1c5d9a287fb396d5e9eed494839d3bccac74dd5456ea3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wxh2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nlnv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.537269 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.537308 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.537318 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.537332 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.537342 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:09Z","lastTransitionTime":"2025-11-24T13:13:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.539791 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5ttn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.552228 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.563620 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.567698 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5x94h_528a4b22-34ba-41b8-8c7d-07d98eebe02b/ovnkube-controller/0.log" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.569833 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" event={"ID":"528a4b22-34ba-41b8-8c7d-07d98eebe02b","Type":"ContainerStarted","Data":"80a2a718c71e9b762d00ee945b50c2e0c5ec37bea569e7499c17e8a8c5487878"} Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.570002 4790 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.575863 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.587507 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.599969 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.610635 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.621718 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.629479 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ea0741f1-1abe-49ba-a186-5c4a3ebec5b5-env-overrides\") pod \"ovnkube-control-plane-749d76644c-n5ttn\" (UID: \"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.629550 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7vc2\" (UniqueName: \"kubernetes.io/projected/ea0741f1-1abe-49ba-a186-5c4a3ebec5b5-kube-api-access-j7vc2\") pod \"ovnkube-control-plane-749d76644c-n5ttn\" (UID: \"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.629643 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ea0741f1-1abe-49ba-a186-5c4a3ebec5b5-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-n5ttn\" (UID: \"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.629702 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ea0741f1-1abe-49ba-a186-5c4a3ebec5b5-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-n5ttn\" (UID: \"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.635803 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.639173 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.639211 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.639220 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.639233 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.639243 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:09Z","lastTransitionTime":"2025-11-24T13:13:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.653579 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec06ff65e703be63a81e99901326ec1eaa20c2176ddbd0764cc7cdc0b5761bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a2bf135bf819ef8ce489337a9f6b7b0bb626ff335b37ec7e6a74e0e13a1ba40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f3d5e18b5f456b7627a9154ff8fb00945ec026d4fb8797d4c4f3e2a2fbb1db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad72e45458f88569f2bb77a55dc98e2bad29a7e451e5258d371722616042bebd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23e4305b0dc57da3ad6d92e8f04f1f18780c1ebaf63ee19140a8443b99197bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908f7697965f4e6d820bfe32f7f59960e917dedea04373ff0e2b8126bb835163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://baccf450bd4cf6f7b3e28c8885aa0db191fa487cfbd9610d92bdca5e0aef838e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baccf450bd4cf6f7b3e28c8885aa0db191fa487cfbd9610d92bdca5e0aef838e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"message\\\":\\\"Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:13:07.458859 6106 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:13:07.458938 6106 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 13:13:07.459147 6106 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:13:07.459217 6106 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 13:13:07.459277 6106 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 13:13:07.459314 6106 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 13:13:07.459337 6106 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 13:13:07.459372 6106 factory.go:656] Stopping watch factory\\\\nI1124 13:13:07.459406 6106 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 13:13:07.459432 6106 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 13:13:07.459456 6106 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 13:13:07.459480 6106 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b16be7938ff339aa0f92c4fb0d444114ac253439e80bf12ceb55c7e9a1944603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.677365 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fec678890b719f519b6740a344414413e17705e146440e130765a0f13b8a4407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.698761 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.723969 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.731079 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ea0741f1-1abe-49ba-a186-5c4a3ebec5b5-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-n5ttn\" (UID: \"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.731114 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ea0741f1-1abe-49ba-a186-5c4a3ebec5b5-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-n5ttn\" (UID: \"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.731162 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ea0741f1-1abe-49ba-a186-5c4a3ebec5b5-env-overrides\") pod \"ovnkube-control-plane-749d76644c-n5ttn\" (UID: \"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.731185 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7vc2\" (UniqueName: \"kubernetes.io/projected/ea0741f1-1abe-49ba-a186-5c4a3ebec5b5-kube-api-access-j7vc2\") pod \"ovnkube-control-plane-749d76644c-n5ttn\" (UID: \"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.731985 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ea0741f1-1abe-49ba-a186-5c4a3ebec5b5-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-n5ttn\" (UID: \"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.732810 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ea0741f1-1abe-49ba-a186-5c4a3ebec5b5-env-overrides\") pod \"ovnkube-control-plane-749d76644c-n5ttn\" (UID: \"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.736752 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.737127 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ea0741f1-1abe-49ba-a186-5c4a3ebec5b5-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-n5ttn\" (UID: \"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.740693 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.740743 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.740756 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.740775 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.740785 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:09Z","lastTransitionTime":"2025-11-24T13:13:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.746043 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7vc2\" (UniqueName: \"kubernetes.io/projected/ea0741f1-1abe-49ba-a186-5c4a3ebec5b5-kube-api-access-j7vc2\") pod \"ovnkube-control-plane-749d76644c-n5ttn\" (UID: \"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.750619 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.761096 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.770159 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.780480 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nlnv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0f7e0af-ed9a-4a5e-be2f-06abac132790\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ad38a0b670b452c1c5d9a287fb396d5e9eed494839d3bccac74dd5456ea3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wxh2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nlnv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.790416 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5ttn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.803997 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.816454 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.818487 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.830603 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.841000 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.842364 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.842391 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.842401 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.842414 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.842423 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:09Z","lastTransitionTime":"2025-11-24T13:13:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.853175 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.865655 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.877755 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.899554 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec06ff65e703be63a81e99901326ec1eaa20c2176ddbd0764cc7cdc0b5761bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a2bf135bf819ef8ce489337a9f6b7b0bb626ff335b37ec7e6a74e0e13a1ba40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f3d5e18b5f456b7627a9154ff8fb00945ec026d4fb8797d4c4f3e2a2fbb1db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad72e45458f88569f2bb77a55dc98e2bad29a7e451e5258d371722616042bebd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23e4305b0dc57da3ad6d92e8f04f1f18780c1ebaf63ee19140a8443b99197bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908f7697965f4e6d820bfe32f7f59960e917dedea04373ff0e2b8126bb835163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80a2a718c71e9b762d00ee945b50c2e0c5ec37bea569e7499c17e8a8c5487878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baccf450bd4cf6f7b3e28c8885aa0db191fa487cfbd9610d92bdca5e0aef838e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"message\\\":\\\"Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:13:07.458859 6106 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:13:07.458938 6106 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 13:13:07.459147 6106 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:13:07.459217 6106 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 13:13:07.459277 6106 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 13:13:07.459314 6106 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 13:13:07.459337 6106 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 13:13:07.459372 6106 factory.go:656] Stopping watch factory\\\\nI1124 13:13:07.459406 6106 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 13:13:07.459432 6106 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 13:13:07.459456 6106 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 13:13:07.459480 6106 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b16be7938ff339aa0f92c4fb0d444114ac253439e80bf12ceb55c7e9a1944603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.916481 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fec678890b719f519b6740a344414413e17705e146440e130765a0f13b8a4407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.944867 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.944925 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.944934 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.944947 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:09 crc kubenswrapper[4790]: I1124 13:13:09.944956 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:09Z","lastTransitionTime":"2025-11-24T13:13:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.047939 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.047969 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.047978 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.047991 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.047999 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:10Z","lastTransitionTime":"2025-11-24T13:13:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.150460 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.150500 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.150512 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.150530 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.150543 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:10Z","lastTransitionTime":"2025-11-24T13:13:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.252932 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.252964 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.252972 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.252985 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.252993 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:10Z","lastTransitionTime":"2025-11-24T13:13:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.314749 4790 scope.go:117] "RemoveContainer" containerID="85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.354725 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.354763 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.354773 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.354788 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.354799 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:10Z","lastTransitionTime":"2025-11-24T13:13:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.457404 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.457438 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.457449 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.457464 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.457472 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:10Z","lastTransitionTime":"2025-11-24T13:13:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.559653 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.559684 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.559694 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.559710 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.559720 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:10Z","lastTransitionTime":"2025-11-24T13:13:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.574274 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5x94h_528a4b22-34ba-41b8-8c7d-07d98eebe02b/ovnkube-controller/1.log" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.574791 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5x94h_528a4b22-34ba-41b8-8c7d-07d98eebe02b/ovnkube-controller/0.log" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.577259 4790 generic.go:334] "Generic (PLEG): container finished" podID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerID="80a2a718c71e9b762d00ee945b50c2e0c5ec37bea569e7499c17e8a8c5487878" exitCode=1 Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.577319 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" event={"ID":"528a4b22-34ba-41b8-8c7d-07d98eebe02b","Type":"ContainerDied","Data":"80a2a718c71e9b762d00ee945b50c2e0c5ec37bea569e7499c17e8a8c5487878"} Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.577375 4790 scope.go:117] "RemoveContainer" containerID="baccf450bd4cf6f7b3e28c8885aa0db191fa487cfbd9610d92bdca5e0aef838e" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.578147 4790 scope.go:117] "RemoveContainer" containerID="80a2a718c71e9b762d00ee945b50c2e0c5ec37bea569e7499c17e8a8c5487878" Nov 24 13:13:10 crc kubenswrapper[4790]: E1124 13:13:10.578306 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5x94h_openshift-ovn-kubernetes(528a4b22-34ba-41b8-8c7d-07d98eebe02b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.579038 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.580510 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1a3641c5a28dae87d2696ab209ab60e5bfadc8881ee96237f823df0f7ad8e841"} Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.580894 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.582193 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" event={"ID":"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5","Type":"ContainerStarted","Data":"cfdc436c0635036c073c666b7acf24eb7d4d46af0a735122716d469ebf114a8e"} Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.582240 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" event={"ID":"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5","Type":"ContainerStarted","Data":"5898d46ebaf15e8d33d02bea73cdb774bab2ea6001911a6dca0ae33c5d8d69e4"} Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.582257 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" event={"ID":"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5","Type":"ContainerStarted","Data":"f2a0d4bed6fa49f4192697531bd202795c1a711cbdad2942532ddaa7f9ab615b"} Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.591981 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nlnv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0f7e0af-ed9a-4a5e-be2f-06abac132790\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ad38a0b670b452c1c5d9a287fb396d5e9eed494839d3bccac74dd5456ea3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wxh2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nlnv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.595565 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-vdtxl"] Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.595993 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:10 crc kubenswrapper[4790]: E1124 13:13:10.596050 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.603153 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5ttn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.614675 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.626096 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.637801 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.647194 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.656422 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.661902 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.661936 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.661959 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.661972 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.661980 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:10Z","lastTransitionTime":"2025-11-24T13:13:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.669988 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.680406 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.690462 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.708022 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec06ff65e703be63a81e99901326ec1eaa20c2176ddbd0764cc7cdc0b5761bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a2bf135bf819ef8ce489337a9f6b7b0bb626ff335b37ec7e6a74e0e13a1ba40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f3d5e18b5f456b7627a9154ff8fb00945ec026d4fb8797d4c4f3e2a2fbb1db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad72e45458f88569f2bb77a55dc98e2bad29a7e451e5258d371722616042bebd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23e4305b0dc57da3ad6d92e8f04f1f18780c1ebaf63ee19140a8443b99197bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908f7697965f4e6d820bfe32f7f59960e917dedea04373ff0e2b8126bb835163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80a2a718c71e9b762d00ee945b50c2e0c5ec37bea569e7499c17e8a8c5487878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baccf450bd4cf6f7b3e28c8885aa0db191fa487cfbd9610d92bdca5e0aef838e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"message\\\":\\\"Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:13:07.458859 6106 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:13:07.458938 6106 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 13:13:07.459147 6106 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:13:07.459217 6106 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 13:13:07.459277 6106 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 13:13:07.459314 6106 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 13:13:07.459337 6106 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 13:13:07.459372 6106 factory.go:656] Stopping watch factory\\\\nI1124 13:13:07.459406 6106 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 13:13:07.459432 6106 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 13:13:07.459456 6106 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 13:13:07.459480 6106 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80a2a718c71e9b762d00ee945b50c2e0c5ec37bea569e7499c17e8a8c5487878\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"message\\\":\\\" \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}\\\\nI1124 13:13:09.646982 6250 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 666.9µs\\\\nI1124 13:13:09.646986 6250 services_controller.go:356] Processing sync for service openshift-machine-config-operator/machine-config-daemon for network=default\\\\nI1124 13:13:09.646995 6250 services_controller.go:356] Processing sync for service openshift-console-operator/metrics for network=default\\\\nF1124 13:13:09.647013 6250 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z]\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b16be7938ff339aa0f92c4fb0d444114ac253439e80bf12ceb55c7e9a1944603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.721946 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fec678890b719f519b6740a344414413e17705e146440e130765a0f13b8a4407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.733583 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.741060 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qf9zs\" (UniqueName: \"kubernetes.io/projected/bef2f789-12a8-45b6-b875-7307992176ee-kube-api-access-qf9zs\") pod \"network-metrics-daemon-vdtxl\" (UID: \"bef2f789-12a8-45b6-b875-7307992176ee\") " pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.741194 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bef2f789-12a8-45b6-b875-7307992176ee-metrics-certs\") pod \"network-metrics-daemon-vdtxl\" (UID: \"bef2f789-12a8-45b6-b875-7307992176ee\") " pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.749168 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.759970 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.763402 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.763428 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.763435 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.763446 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.763454 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:10Z","lastTransitionTime":"2025-11-24T13:13:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.771848 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.780501 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.791159 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nlnv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0f7e0af-ed9a-4a5e-be2f-06abac132790\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ad38a0b670b452c1c5d9a287fb396d5e9eed494839d3bccac74dd5456ea3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wxh2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nlnv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.803143 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898d46ebaf15e8d33d02bea73cdb774bab2ea6001911a6dca0ae33c5d8d69e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfdc436c0635036c073c666b7acf24eb7d4d46af0a735122716d469ebf114a8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5ttn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.815142 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.828218 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.839859 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.842117 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qf9zs\" (UniqueName: \"kubernetes.io/projected/bef2f789-12a8-45b6-b875-7307992176ee-kube-api-access-qf9zs\") pod \"network-metrics-daemon-vdtxl\" (UID: \"bef2f789-12a8-45b6-b875-7307992176ee\") " pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.842169 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bef2f789-12a8-45b6-b875-7307992176ee-metrics-certs\") pod \"network-metrics-daemon-vdtxl\" (UID: \"bef2f789-12a8-45b6-b875-7307992176ee\") " pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:10 crc kubenswrapper[4790]: E1124 13:13:10.842263 4790 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:13:10 crc kubenswrapper[4790]: E1124 13:13:10.842320 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bef2f789-12a8-45b6-b875-7307992176ee-metrics-certs podName:bef2f789-12a8-45b6-b875-7307992176ee nodeName:}" failed. No retries permitted until 2025-11-24 13:13:11.342306599 +0000 UTC m=+39.722200261 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bef2f789-12a8-45b6-b875-7307992176ee-metrics-certs") pod "network-metrics-daemon-vdtxl" (UID: "bef2f789-12a8-45b6-b875-7307992176ee") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.852579 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.859350 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qf9zs\" (UniqueName: \"kubernetes.io/projected/bef2f789-12a8-45b6-b875-7307992176ee-kube-api-access-qf9zs\") pod \"network-metrics-daemon-vdtxl\" (UID: \"bef2f789-12a8-45b6-b875-7307992176ee\") " pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.865808 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.865963 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.866106 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.866294 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.866491 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:10Z","lastTransitionTime":"2025-11-24T13:13:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.867228 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.879978 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3641c5a28dae87d2696ab209ab60e5bfadc8881ee96237f823df0f7ad8e841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.892617 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.903846 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.923355 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec06ff65e703be63a81e99901326ec1eaa20c2176ddbd0764cc7cdc0b5761bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a2bf135bf819ef8ce489337a9f6b7b0bb626ff335b37ec7e6a74e0e13a1ba40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f3d5e18b5f456b7627a9154ff8fb00945ec026d4fb8797d4c4f3e2a2fbb1db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad72e45458f88569f2bb77a55dc98e2bad29a7e451e5258d371722616042bebd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23e4305b0dc57da3ad6d92e8f04f1f18780c1ebaf63ee19140a8443b99197bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908f7697965f4e6d820bfe32f7f59960e917dedea04373ff0e2b8126bb835163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80a2a718c71e9b762d00ee945b50c2e0c5ec37bea569e7499c17e8a8c5487878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baccf450bd4cf6f7b3e28c8885aa0db191fa487cfbd9610d92bdca5e0aef838e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"message\\\":\\\"Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:13:07.458859 6106 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:13:07.458938 6106 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 13:13:07.459147 6106 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:13:07.459217 6106 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 13:13:07.459277 6106 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 13:13:07.459314 6106 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 13:13:07.459337 6106 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 13:13:07.459372 6106 factory.go:656] Stopping watch factory\\\\nI1124 13:13:07.459406 6106 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 13:13:07.459432 6106 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 13:13:07.459456 6106 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 13:13:07.459480 6106 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80a2a718c71e9b762d00ee945b50c2e0c5ec37bea569e7499c17e8a8c5487878\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"message\\\":\\\" \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}\\\\nI1124 13:13:09.646982 6250 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 666.9µs\\\\nI1124 13:13:09.646986 6250 services_controller.go:356] Processing sync for service openshift-machine-config-operator/machine-config-daemon for network=default\\\\nI1124 13:13:09.646995 6250 services_controller.go:356] Processing sync for service openshift-console-operator/metrics for network=default\\\\nF1124 13:13:09.647013 6250 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z]\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b16be7938ff339aa0f92c4fb0d444114ac253439e80bf12ceb55c7e9a1944603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.938454 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fec678890b719f519b6740a344414413e17705e146440e130765a0f13b8a4407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.949899 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.960571 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdtxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bef2f789-12a8-45b6-b875-7307992176ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdtxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.969449 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.969490 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.969500 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.969517 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:10 crc kubenswrapper[4790]: I1124 13:13:10.969529 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:10Z","lastTransitionTime":"2025-11-24T13:13:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.071666 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.071712 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.071722 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.071735 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.071744 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:11Z","lastTransitionTime":"2025-11-24T13:13:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.174393 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.174431 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.174440 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.174455 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.174465 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:11Z","lastTransitionTime":"2025-11-24T13:13:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.276735 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.276767 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.276776 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.276789 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.276798 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:11Z","lastTransitionTime":"2025-11-24T13:13:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.314154 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.314233 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.314274 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:13:11 crc kubenswrapper[4790]: E1124 13:13:11.314636 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:13:11 crc kubenswrapper[4790]: E1124 13:13:11.314520 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:13:11 crc kubenswrapper[4790]: E1124 13:13:11.314731 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.347988 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bef2f789-12a8-45b6-b875-7307992176ee-metrics-certs\") pod \"network-metrics-daemon-vdtxl\" (UID: \"bef2f789-12a8-45b6-b875-7307992176ee\") " pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:11 crc kubenswrapper[4790]: E1124 13:13:11.348113 4790 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:13:11 crc kubenswrapper[4790]: E1124 13:13:11.348171 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bef2f789-12a8-45b6-b875-7307992176ee-metrics-certs podName:bef2f789-12a8-45b6-b875-7307992176ee nodeName:}" failed. No retries permitted until 2025-11-24 13:13:12.348153119 +0000 UTC m=+40.728046781 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bef2f789-12a8-45b6-b875-7307992176ee-metrics-certs") pod "network-metrics-daemon-vdtxl" (UID: "bef2f789-12a8-45b6-b875-7307992176ee") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.378985 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.379037 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.379047 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.379065 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.379076 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:11Z","lastTransitionTime":"2025-11-24T13:13:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.481659 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.481696 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.481707 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.481720 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.481729 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:11Z","lastTransitionTime":"2025-11-24T13:13:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.584395 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.584433 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.584443 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.584460 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.584471 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:11Z","lastTransitionTime":"2025-11-24T13:13:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.586194 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5x94h_528a4b22-34ba-41b8-8c7d-07d98eebe02b/ovnkube-controller/1.log" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.687188 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.687238 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.687248 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.687265 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.687277 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:11Z","lastTransitionTime":"2025-11-24T13:13:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.789318 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.789367 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.789381 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.789398 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.789409 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:11Z","lastTransitionTime":"2025-11-24T13:13:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.892250 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.892291 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.892303 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.892320 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.892333 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:11Z","lastTransitionTime":"2025-11-24T13:13:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.994509 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.994553 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.994561 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.994576 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:11 crc kubenswrapper[4790]: I1124 13:13:11.994586 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:11Z","lastTransitionTime":"2025-11-24T13:13:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.097160 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.097199 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.097209 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.097226 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.097238 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:12Z","lastTransitionTime":"2025-11-24T13:13:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.200052 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.200094 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.200104 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.200121 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.200132 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:12Z","lastTransitionTime":"2025-11-24T13:13:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.302318 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.302369 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.302383 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.302399 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.302409 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:12Z","lastTransitionTime":"2025-11-24T13:13:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.314331 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:12 crc kubenswrapper[4790]: E1124 13:13:12.314448 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.328961 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:12Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.341280 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:12Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.352786 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:12Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.360860 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bef2f789-12a8-45b6-b875-7307992176ee-metrics-certs\") pod \"network-metrics-daemon-vdtxl\" (UID: \"bef2f789-12a8-45b6-b875-7307992176ee\") " pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:12 crc kubenswrapper[4790]: E1124 13:13:12.361013 4790 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:13:12 crc kubenswrapper[4790]: E1124 13:13:12.361072 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bef2f789-12a8-45b6-b875-7307992176ee-metrics-certs podName:bef2f789-12a8-45b6-b875-7307992176ee nodeName:}" failed. No retries permitted until 2025-11-24 13:13:14.361056895 +0000 UTC m=+42.740950557 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bef2f789-12a8-45b6-b875-7307992176ee-metrics-certs") pod "network-metrics-daemon-vdtxl" (UID: "bef2f789-12a8-45b6-b875-7307992176ee") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.362633 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:12Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.372316 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nlnv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0f7e0af-ed9a-4a5e-be2f-06abac132790\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ad38a0b670b452c1c5d9a287fb396d5e9eed494839d3bccac74dd5456ea3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wxh2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nlnv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:12Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.383371 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898d46ebaf15e8d33d02bea73cdb774bab2ea6001911a6dca0ae33c5d8d69e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfdc436c0635036c073c666b7acf24eb7d4d46af0a735122716d469ebf114a8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5ttn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:12Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.399519 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:12Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.403737 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.403774 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.403784 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.403800 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.403811 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:12Z","lastTransitionTime":"2025-11-24T13:13:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.411360 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:12Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.421919 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:12Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.434224 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fec678890b719f519b6740a344414413e17705e146440e130765a0f13b8a4407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:12Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.443472 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:12Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.456009 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3641c5a28dae87d2696ab209ab60e5bfadc8881ee96237f823df0f7ad8e841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:12Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.469001 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:12Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.481120 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:12Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.497119 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec06ff65e703be63a81e99901326ec1eaa20c2176ddbd0764cc7cdc0b5761bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a2bf135bf819ef8ce489337a9f6b7b0bb626ff335b37ec7e6a74e0e13a1ba40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f3d5e18b5f456b7627a9154ff8fb00945ec026d4fb8797d4c4f3e2a2fbb1db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad72e45458f88569f2bb77a55dc98e2bad29a7e451e5258d371722616042bebd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23e4305b0dc57da3ad6d92e8f04f1f18780c1ebaf63ee19140a8443b99197bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908f7697965f4e6d820bfe32f7f59960e917dedea04373ff0e2b8126bb835163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80a2a718c71e9b762d00ee945b50c2e0c5ec37bea569e7499c17e8a8c5487878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baccf450bd4cf6f7b3e28c8885aa0db191fa487cfbd9610d92bdca5e0aef838e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"message\\\":\\\"Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:13:07.458859 6106 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:13:07.458938 6106 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 13:13:07.459147 6106 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:13:07.459217 6106 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 13:13:07.459277 6106 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 13:13:07.459314 6106 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 13:13:07.459337 6106 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 13:13:07.459372 6106 factory.go:656] Stopping watch factory\\\\nI1124 13:13:07.459406 6106 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 13:13:07.459432 6106 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 13:13:07.459456 6106 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 13:13:07.459480 6106 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80a2a718c71e9b762d00ee945b50c2e0c5ec37bea569e7499c17e8a8c5487878\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"message\\\":\\\" \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}\\\\nI1124 13:13:09.646982 6250 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 666.9µs\\\\nI1124 13:13:09.646986 6250 services_controller.go:356] Processing sync for service openshift-machine-config-operator/machine-config-daemon for network=default\\\\nI1124 13:13:09.646995 6250 services_controller.go:356] Processing sync for service openshift-console-operator/metrics for network=default\\\\nF1124 13:13:09.647013 6250 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z]\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b16be7938ff339aa0f92c4fb0d444114ac253439e80bf12ceb55c7e9a1944603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:12Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.505839 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.505938 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.505954 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.505972 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.505985 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:12Z","lastTransitionTime":"2025-11-24T13:13:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.506952 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdtxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bef2f789-12a8-45b6-b875-7307992176ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdtxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:12Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.607626 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.607662 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.607673 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.607696 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.607707 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:12Z","lastTransitionTime":"2025-11-24T13:13:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.710995 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.711040 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.711052 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.711068 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.711079 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:12Z","lastTransitionTime":"2025-11-24T13:13:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.813521 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.813569 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.813582 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.813600 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.813620 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:12Z","lastTransitionTime":"2025-11-24T13:13:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.916013 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.916066 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.916076 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.916092 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:12 crc kubenswrapper[4790]: I1124 13:13:12.916104 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:12Z","lastTransitionTime":"2025-11-24T13:13:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.018582 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.018624 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.018634 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.018647 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.018656 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:13Z","lastTransitionTime":"2025-11-24T13:13:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.120990 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.121036 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.121045 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.121057 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.121066 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:13Z","lastTransitionTime":"2025-11-24T13:13:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.223116 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.223194 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.223206 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.223227 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.223247 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:13Z","lastTransitionTime":"2025-11-24T13:13:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.314265 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.314405 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:13:13 crc kubenswrapper[4790]: E1124 13:13:13.314434 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.314476 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:13 crc kubenswrapper[4790]: E1124 13:13:13.314577 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:13:13 crc kubenswrapper[4790]: E1124 13:13:13.314661 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.325190 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.325243 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.325259 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.325281 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.325294 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:13Z","lastTransitionTime":"2025-11-24T13:13:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.427203 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.427260 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.427270 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.427286 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.427297 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:13Z","lastTransitionTime":"2025-11-24T13:13:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.529406 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.529438 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.529447 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.529462 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.529472 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:13Z","lastTransitionTime":"2025-11-24T13:13:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.632070 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.632134 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.632151 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.632176 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.632193 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:13Z","lastTransitionTime":"2025-11-24T13:13:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.734261 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.734306 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.734314 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.734328 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.734339 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:13Z","lastTransitionTime":"2025-11-24T13:13:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.836177 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.836232 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.836252 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.836272 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.836288 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:13Z","lastTransitionTime":"2025-11-24T13:13:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.939166 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.939225 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.939237 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.939262 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:13 crc kubenswrapper[4790]: I1124 13:13:13.939274 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:13Z","lastTransitionTime":"2025-11-24T13:13:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.042288 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.042402 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.042417 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.042435 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.042469 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:14Z","lastTransitionTime":"2025-11-24T13:13:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.145372 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.145423 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.145439 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.145456 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.145473 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:14Z","lastTransitionTime":"2025-11-24T13:13:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.248629 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.248676 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.248686 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.248701 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.248710 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:14Z","lastTransitionTime":"2025-11-24T13:13:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.314007 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:14 crc kubenswrapper[4790]: E1124 13:13:14.314143 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.351142 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.351185 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.351194 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.351207 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.351216 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:14Z","lastTransitionTime":"2025-11-24T13:13:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.381953 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bef2f789-12a8-45b6-b875-7307992176ee-metrics-certs\") pod \"network-metrics-daemon-vdtxl\" (UID: \"bef2f789-12a8-45b6-b875-7307992176ee\") " pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:14 crc kubenswrapper[4790]: E1124 13:13:14.382132 4790 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:13:14 crc kubenswrapper[4790]: E1124 13:13:14.382244 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bef2f789-12a8-45b6-b875-7307992176ee-metrics-certs podName:bef2f789-12a8-45b6-b875-7307992176ee nodeName:}" failed. No retries permitted until 2025-11-24 13:13:18.382220194 +0000 UTC m=+46.762113936 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bef2f789-12a8-45b6-b875-7307992176ee-metrics-certs") pod "network-metrics-daemon-vdtxl" (UID: "bef2f789-12a8-45b6-b875-7307992176ee") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.453779 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.453836 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.453850 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.453867 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.453900 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:14Z","lastTransitionTime":"2025-11-24T13:13:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.556862 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.556931 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.556944 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.556962 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.556973 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:14Z","lastTransitionTime":"2025-11-24T13:13:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.658984 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.659024 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.659033 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.659047 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.659058 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:14Z","lastTransitionTime":"2025-11-24T13:13:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.761380 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.761418 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.761428 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.761444 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.761454 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:14Z","lastTransitionTime":"2025-11-24T13:13:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.863707 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.863803 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.863822 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.863841 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.863852 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:14Z","lastTransitionTime":"2025-11-24T13:13:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.966207 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.966244 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.966252 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.966265 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:14 crc kubenswrapper[4790]: I1124 13:13:14.966274 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:14Z","lastTransitionTime":"2025-11-24T13:13:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.068896 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.069984 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.070007 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.070037 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.070054 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:15Z","lastTransitionTime":"2025-11-24T13:13:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.171931 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.171973 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.171983 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.171998 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.172010 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:15Z","lastTransitionTime":"2025-11-24T13:13:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.274744 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.274796 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.274808 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.274853 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.274867 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:15Z","lastTransitionTime":"2025-11-24T13:13:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.314290 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.314330 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.314361 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:15 crc kubenswrapper[4790]: E1124 13:13:15.314434 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:13:15 crc kubenswrapper[4790]: E1124 13:13:15.314543 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:13:15 crc kubenswrapper[4790]: E1124 13:13:15.314625 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.377485 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.377535 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.377544 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.377559 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.377568 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:15Z","lastTransitionTime":"2025-11-24T13:13:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.479857 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.479921 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.479931 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.479946 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.479956 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:15Z","lastTransitionTime":"2025-11-24T13:13:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.582806 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.582869 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.582909 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.582925 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.582938 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:15Z","lastTransitionTime":"2025-11-24T13:13:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.685191 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.685241 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.685250 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.685266 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.685277 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:15Z","lastTransitionTime":"2025-11-24T13:13:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.787527 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.787569 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.787581 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.787596 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.787606 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:15Z","lastTransitionTime":"2025-11-24T13:13:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.890130 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.890173 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.890184 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.890226 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.890239 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:15Z","lastTransitionTime":"2025-11-24T13:13:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.992594 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.992636 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.992648 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.992664 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:15 crc kubenswrapper[4790]: I1124 13:13:15.992676 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:15Z","lastTransitionTime":"2025-11-24T13:13:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.095644 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.095678 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.095688 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.095703 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.095713 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:16Z","lastTransitionTime":"2025-11-24T13:13:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.198861 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.198923 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.198932 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.198948 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.198961 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:16Z","lastTransitionTime":"2025-11-24T13:13:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.301262 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.301294 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.301306 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.301320 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.301330 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:16Z","lastTransitionTime":"2025-11-24T13:13:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.313714 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:16 crc kubenswrapper[4790]: E1124 13:13:16.313851 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.403220 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.403428 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.403543 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.403633 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.403732 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:16Z","lastTransitionTime":"2025-11-24T13:13:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.506390 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.506422 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.506431 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.506446 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.506455 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:16Z","lastTransitionTime":"2025-11-24T13:13:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.608420 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.608667 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.608745 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.608830 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.608962 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:16Z","lastTransitionTime":"2025-11-24T13:13:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.711364 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.711398 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.711406 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.711419 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.711428 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:16Z","lastTransitionTime":"2025-11-24T13:13:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.812902 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.812940 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.812948 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.812961 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.812972 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:16Z","lastTransitionTime":"2025-11-24T13:13:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.915451 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.915491 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.915500 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.915513 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:16 crc kubenswrapper[4790]: I1124 13:13:16.915522 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:16Z","lastTransitionTime":"2025-11-24T13:13:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.018108 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.018142 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.018150 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.018163 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.018173 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:17Z","lastTransitionTime":"2025-11-24T13:13:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.119632 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.119664 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.119677 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.119693 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.119705 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:17Z","lastTransitionTime":"2025-11-24T13:13:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.222179 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.222229 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.222241 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.222261 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.222274 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:17Z","lastTransitionTime":"2025-11-24T13:13:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.314348 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.314437 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.314348 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:13:17 crc kubenswrapper[4790]: E1124 13:13:17.314472 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:13:17 crc kubenswrapper[4790]: E1124 13:13:17.314562 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:13:17 crc kubenswrapper[4790]: E1124 13:13:17.314628 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.324809 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.324845 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.324854 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.324867 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.324897 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:17Z","lastTransitionTime":"2025-11-24T13:13:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.428023 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.428065 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.428074 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.428088 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.428102 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:17Z","lastTransitionTime":"2025-11-24T13:13:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.533247 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.533781 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.533815 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.533832 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.533840 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:17Z","lastTransitionTime":"2025-11-24T13:13:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.636160 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.636189 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.636198 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.636213 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.636223 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:17Z","lastTransitionTime":"2025-11-24T13:13:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.738411 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.738465 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.738474 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.738486 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.738494 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:17Z","lastTransitionTime":"2025-11-24T13:13:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.840462 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.840498 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.840508 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.840521 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.840532 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:17Z","lastTransitionTime":"2025-11-24T13:13:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.942791 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.942822 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.942833 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.942857 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:17 crc kubenswrapper[4790]: I1124 13:13:17.942869 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:17Z","lastTransitionTime":"2025-11-24T13:13:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.037337 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.037370 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.037377 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.037389 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.037398 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:18Z","lastTransitionTime":"2025-11-24T13:13:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:18 crc kubenswrapper[4790]: E1124 13:13:18.049739 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.053004 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.053032 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.053042 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.053054 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.053062 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:18Z","lastTransitionTime":"2025-11-24T13:13:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:18 crc kubenswrapper[4790]: E1124 13:13:18.063512 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.066336 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.066359 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.066369 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.066381 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.066391 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:18Z","lastTransitionTime":"2025-11-24T13:13:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:18 crc kubenswrapper[4790]: E1124 13:13:18.077844 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.081326 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.081361 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.081371 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.081386 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.081396 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:18Z","lastTransitionTime":"2025-11-24T13:13:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:18 crc kubenswrapper[4790]: E1124 13:13:18.092742 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.095937 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.095965 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.095974 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.095990 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.095999 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:18Z","lastTransitionTime":"2025-11-24T13:13:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:18 crc kubenswrapper[4790]: E1124 13:13:18.107125 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:18 crc kubenswrapper[4790]: E1124 13:13:18.107266 4790 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.108571 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.108607 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.108615 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.108630 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.108639 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:18Z","lastTransitionTime":"2025-11-24T13:13:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.211325 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.211374 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.211386 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.211409 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.211425 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:18Z","lastTransitionTime":"2025-11-24T13:13:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.313253 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.313291 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.313302 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.313317 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.313328 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:18Z","lastTransitionTime":"2025-11-24T13:13:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.313525 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:18 crc kubenswrapper[4790]: E1124 13:13:18.313625 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.415537 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.415573 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.415584 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.415600 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.415612 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:18Z","lastTransitionTime":"2025-11-24T13:13:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.420166 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bef2f789-12a8-45b6-b875-7307992176ee-metrics-certs\") pod \"network-metrics-daemon-vdtxl\" (UID: \"bef2f789-12a8-45b6-b875-7307992176ee\") " pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:18 crc kubenswrapper[4790]: E1124 13:13:18.420287 4790 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:13:18 crc kubenswrapper[4790]: E1124 13:13:18.420343 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bef2f789-12a8-45b6-b875-7307992176ee-metrics-certs podName:bef2f789-12a8-45b6-b875-7307992176ee nodeName:}" failed. No retries permitted until 2025-11-24 13:13:26.420328599 +0000 UTC m=+54.800222261 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bef2f789-12a8-45b6-b875-7307992176ee-metrics-certs") pod "network-metrics-daemon-vdtxl" (UID: "bef2f789-12a8-45b6-b875-7307992176ee") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.517630 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.517661 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.517671 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.517690 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.517701 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:18Z","lastTransitionTime":"2025-11-24T13:13:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.619729 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.619778 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.619795 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.619816 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.619834 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:18Z","lastTransitionTime":"2025-11-24T13:13:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.722208 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.722249 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.722266 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.722283 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.722293 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:18Z","lastTransitionTime":"2025-11-24T13:13:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.825004 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.825049 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.825059 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.825075 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.825087 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:18Z","lastTransitionTime":"2025-11-24T13:13:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.926983 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.927029 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.927042 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.927060 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:18 crc kubenswrapper[4790]: I1124 13:13:18.927071 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:18Z","lastTransitionTime":"2025-11-24T13:13:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.029016 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.029051 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.029067 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.029083 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.029093 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:19Z","lastTransitionTime":"2025-11-24T13:13:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.131177 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.131211 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.131234 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.131248 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.131257 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:19Z","lastTransitionTime":"2025-11-24T13:13:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.233568 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.233600 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.233608 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.233620 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.233629 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:19Z","lastTransitionTime":"2025-11-24T13:13:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.314212 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.314253 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.314264 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:13:19 crc kubenswrapper[4790]: E1124 13:13:19.314349 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:13:19 crc kubenswrapper[4790]: E1124 13:13:19.314427 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:13:19 crc kubenswrapper[4790]: E1124 13:13:19.314491 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.335113 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.335150 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.335165 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.335180 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.335190 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:19Z","lastTransitionTime":"2025-11-24T13:13:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.437303 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.437337 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.437353 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.437370 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.437379 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:19Z","lastTransitionTime":"2025-11-24T13:13:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.540467 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.540509 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.540520 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.540534 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.540544 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:19Z","lastTransitionTime":"2025-11-24T13:13:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.643389 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.643423 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.643431 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.643447 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.643456 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:19Z","lastTransitionTime":"2025-11-24T13:13:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.745654 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.745687 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.745694 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.745707 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.745715 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:19Z","lastTransitionTime":"2025-11-24T13:13:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.848207 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.848243 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.848252 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.848264 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.848273 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:19Z","lastTransitionTime":"2025-11-24T13:13:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.950806 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.950845 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.950857 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.950874 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:19 crc kubenswrapper[4790]: I1124 13:13:19.950909 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:19Z","lastTransitionTime":"2025-11-24T13:13:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.052707 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.052736 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.052747 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.052761 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.052769 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:20Z","lastTransitionTime":"2025-11-24T13:13:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.155631 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.155669 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.155680 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.155696 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.155706 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:20Z","lastTransitionTime":"2025-11-24T13:13:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.258238 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.258285 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.258301 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.258324 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.258339 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:20Z","lastTransitionTime":"2025-11-24T13:13:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.314034 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:20 crc kubenswrapper[4790]: E1124 13:13:20.314265 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.361435 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.361501 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.361518 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.361544 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.361560 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:20Z","lastTransitionTime":"2025-11-24T13:13:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.464756 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.464821 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.464843 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.464871 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.464934 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:20Z","lastTransitionTime":"2025-11-24T13:13:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.567865 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.567937 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.567946 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.567962 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.567972 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:20Z","lastTransitionTime":"2025-11-24T13:13:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.670554 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.670598 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.670608 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.670621 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.670630 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:20Z","lastTransitionTime":"2025-11-24T13:13:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.772903 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.773241 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.773328 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.773405 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.773481 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:20Z","lastTransitionTime":"2025-11-24T13:13:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.875634 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.875680 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.875693 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.875709 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.875719 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:20Z","lastTransitionTime":"2025-11-24T13:13:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.977783 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.978035 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.978110 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.978215 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:20 crc kubenswrapper[4790]: I1124 13:13:20.978287 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:20Z","lastTransitionTime":"2025-11-24T13:13:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.081112 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.081386 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.081478 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.081568 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.081656 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:21Z","lastTransitionTime":"2025-11-24T13:13:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.184555 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.184657 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.184682 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.184714 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.184736 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:21Z","lastTransitionTime":"2025-11-24T13:13:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.287237 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.287302 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.287315 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.287333 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.287346 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:21Z","lastTransitionTime":"2025-11-24T13:13:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.314556 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.314564 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:21 crc kubenswrapper[4790]: E1124 13:13:21.314714 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:13:21 crc kubenswrapper[4790]: E1124 13:13:21.314905 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.315064 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:13:21 crc kubenswrapper[4790]: E1124 13:13:21.315246 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.393777 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.393859 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.393899 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.393923 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.393946 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:21Z","lastTransitionTime":"2025-11-24T13:13:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.497001 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.497049 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.497063 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.497079 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.497090 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:21Z","lastTransitionTime":"2025-11-24T13:13:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.599782 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.599820 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.599843 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.599859 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.599870 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:21Z","lastTransitionTime":"2025-11-24T13:13:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.702607 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.702651 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.702661 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.702677 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.702691 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:21Z","lastTransitionTime":"2025-11-24T13:13:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.805248 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.805301 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.805314 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.805330 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.805341 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:21Z","lastTransitionTime":"2025-11-24T13:13:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.907589 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.907629 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.907640 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.907660 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:21 crc kubenswrapper[4790]: I1124 13:13:21.907672 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:21Z","lastTransitionTime":"2025-11-24T13:13:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.010119 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.010170 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.010183 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.010201 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.010213 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:22Z","lastTransitionTime":"2025-11-24T13:13:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.112277 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.112315 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.112326 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.112344 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.112356 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:22Z","lastTransitionTime":"2025-11-24T13:13:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.214484 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.214982 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.215093 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.215193 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.215284 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:22Z","lastTransitionTime":"2025-11-24T13:13:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.313961 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:22 crc kubenswrapper[4790]: E1124 13:13:22.314295 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.317561 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.317597 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.317609 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.317624 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.317635 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:22Z","lastTransitionTime":"2025-11-24T13:13:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.331468 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:22Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.342728 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:22Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.350786 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nlnv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0f7e0af-ed9a-4a5e-be2f-06abac132790\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ad38a0b670b452c1c5d9a287fb396d5e9eed494839d3bccac74dd5456ea3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wxh2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nlnv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:22Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.363133 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898d46ebaf15e8d33d02bea73cdb774bab2ea6001911a6dca0ae33c5d8d69e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfdc436c0635036c073c666b7acf24eb7d4d46af0a735122716d469ebf114a8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5ttn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:22Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.373146 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:22Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.384517 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:22Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.403666 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:22Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.415929 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:22Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.419320 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.419355 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.419363 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.419378 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.419387 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:22Z","lastTransitionTime":"2025-11-24T13:13:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.427942 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:22Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.441099 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3641c5a28dae87d2696ab209ab60e5bfadc8881ee96237f823df0f7ad8e841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:22Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.452591 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:22Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.463526 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:22Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.479842 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec06ff65e703be63a81e99901326ec1eaa20c2176ddbd0764cc7cdc0b5761bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a2bf135bf819ef8ce489337a9f6b7b0bb626ff335b37ec7e6a74e0e13a1ba40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f3d5e18b5f456b7627a9154ff8fb00945ec026d4fb8797d4c4f3e2a2fbb1db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad72e45458f88569f2bb77a55dc98e2bad29a7e451e5258d371722616042bebd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23e4305b0dc57da3ad6d92e8f04f1f18780c1ebaf63ee19140a8443b99197bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908f7697965f4e6d820bfe32f7f59960e917dedea04373ff0e2b8126bb835163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80a2a718c71e9b762d00ee945b50c2e0c5ec37bea569e7499c17e8a8c5487878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baccf450bd4cf6f7b3e28c8885aa0db191fa487cfbd9610d92bdca5e0aef838e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:07Z\\\",\\\"message\\\":\\\"Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:13:07.458859 6106 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:13:07.458938 6106 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 13:13:07.459147 6106 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:13:07.459217 6106 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 13:13:07.459277 6106 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 13:13:07.459314 6106 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 13:13:07.459337 6106 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 13:13:07.459372 6106 factory.go:656] Stopping watch factory\\\\nI1124 13:13:07.459406 6106 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 13:13:07.459432 6106 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 13:13:07.459456 6106 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 13:13:07.459480 6106 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80a2a718c71e9b762d00ee945b50c2e0c5ec37bea569e7499c17e8a8c5487878\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"message\\\":\\\" \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}\\\\nI1124 13:13:09.646982 6250 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 666.9µs\\\\nI1124 13:13:09.646986 6250 services_controller.go:356] Processing sync for service openshift-machine-config-operator/machine-config-daemon for network=default\\\\nI1124 13:13:09.646995 6250 services_controller.go:356] Processing sync for service openshift-console-operator/metrics for network=default\\\\nF1124 13:13:09.647013 6250 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z]\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b16be7938ff339aa0f92c4fb0d444114ac253439e80bf12ceb55c7e9a1944603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:22Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.493248 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fec678890b719f519b6740a344414413e17705e146440e130765a0f13b8a4407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:22Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.505738 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:22Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.516551 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdtxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bef2f789-12a8-45b6-b875-7307992176ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdtxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:22Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.521254 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.521284 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.521292 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.521306 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.521314 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:22Z","lastTransitionTime":"2025-11-24T13:13:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.622776 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.622805 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.622813 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.622826 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.622834 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:22Z","lastTransitionTime":"2025-11-24T13:13:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.724593 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.724629 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.724640 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.724655 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.724667 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:22Z","lastTransitionTime":"2025-11-24T13:13:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.826347 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.826387 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.826400 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.826415 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.826423 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:22Z","lastTransitionTime":"2025-11-24T13:13:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.928977 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.929035 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.929044 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.929060 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:22 crc kubenswrapper[4790]: I1124 13:13:22.929088 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:22Z","lastTransitionTime":"2025-11-24T13:13:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.032030 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.032075 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.032087 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.032104 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.032115 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:23Z","lastTransitionTime":"2025-11-24T13:13:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.063400 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:13:23 crc kubenswrapper[4790]: E1124 13:13:23.063639 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:13:55.06361979 +0000 UTC m=+83.443513452 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.134720 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.134797 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.134814 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.134842 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.134867 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:23Z","lastTransitionTime":"2025-11-24T13:13:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.164564 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.164625 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.164644 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.164666 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:13:23 crc kubenswrapper[4790]: E1124 13:13:23.164695 4790 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:13:23 crc kubenswrapper[4790]: E1124 13:13:23.164791 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:13:55.16477094 +0000 UTC m=+83.544664682 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:13:23 crc kubenswrapper[4790]: E1124 13:13:23.164802 4790 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:13:23 crc kubenswrapper[4790]: E1124 13:13:23.164812 4790 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:13:23 crc kubenswrapper[4790]: E1124 13:13:23.164846 4790 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:13:23 crc kubenswrapper[4790]: E1124 13:13:23.164855 4790 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:13:23 crc kubenswrapper[4790]: E1124 13:13:23.164919 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 13:13:55.164907174 +0000 UTC m=+83.544800836 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:13:23 crc kubenswrapper[4790]: E1124 13:13:23.164822 4790 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:13:23 crc kubenswrapper[4790]: E1124 13:13:23.164940 4790 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:13:23 crc kubenswrapper[4790]: E1124 13:13:23.164959 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 13:13:55.164954005 +0000 UTC m=+83.544847667 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:13:23 crc kubenswrapper[4790]: E1124 13:13:23.164804 4790 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:13:23 crc kubenswrapper[4790]: E1124 13:13:23.165024 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:13:55.165009237 +0000 UTC m=+83.544902989 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.237044 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.237087 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.237097 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.237114 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.237123 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:23Z","lastTransitionTime":"2025-11-24T13:13:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.314106 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.314171 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.314133 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:23 crc kubenswrapper[4790]: E1124 13:13:23.314272 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:13:23 crc kubenswrapper[4790]: E1124 13:13:23.314385 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:13:23 crc kubenswrapper[4790]: E1124 13:13:23.314447 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.314975 4790 scope.go:117] "RemoveContainer" containerID="80a2a718c71e9b762d00ee945b50c2e0c5ec37bea569e7499c17e8a8c5487878" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.328900 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3641c5a28dae87d2696ab209ab60e5bfadc8881ee96237f823df0f7ad8e841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:23Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.340389 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.340668 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.340676 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.340689 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.340699 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:23Z","lastTransitionTime":"2025-11-24T13:13:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.344127 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:23Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.355866 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:23Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.376377 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec06ff65e703be63a81e99901326ec1eaa20c2176ddbd0764cc7cdc0b5761bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a2bf135bf819ef8ce489337a9f6b7b0bb626ff335b37ec7e6a74e0e13a1ba40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f3d5e18b5f456b7627a9154ff8fb00945ec026d4fb8797d4c4f3e2a2fbb1db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad72e45458f88569f2bb77a55dc98e2bad29a7e451e5258d371722616042bebd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23e4305b0dc57da3ad6d92e8f04f1f18780c1ebaf63ee19140a8443b99197bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908f7697965f4e6d820bfe32f7f59960e917dedea04373ff0e2b8126bb835163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80a2a718c71e9b762d00ee945b50c2e0c5ec37bea569e7499c17e8a8c5487878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80a2a718c71e9b762d00ee945b50c2e0c5ec37bea569e7499c17e8a8c5487878\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"message\\\":\\\" \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}\\\\nI1124 13:13:09.646982 6250 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 666.9µs\\\\nI1124 13:13:09.646986 6250 services_controller.go:356] Processing sync for service openshift-machine-config-operator/machine-config-daemon for network=default\\\\nI1124 13:13:09.646995 6250 services_controller.go:356] Processing sync for service openshift-console-operator/metrics for network=default\\\\nF1124 13:13:09.647013 6250 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z]\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5x94h_openshift-ovn-kubernetes(528a4b22-34ba-41b8-8c7d-07d98eebe02b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b16be7938ff339aa0f92c4fb0d444114ac253439e80bf12ceb55c7e9a1944603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:23Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.389600 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fec678890b719f519b6740a344414413e17705e146440e130765a0f13b8a4407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:23Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.399973 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:23Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.409016 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdtxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bef2f789-12a8-45b6-b875-7307992176ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdtxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:23Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.420112 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:23Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.432584 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:23Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.447375 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.447408 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.447419 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.447435 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.447444 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:23Z","lastTransitionTime":"2025-11-24T13:13:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.453051 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:23Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.495284 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:23Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.506311 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nlnv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0f7e0af-ed9a-4a5e-be2f-06abac132790\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ad38a0b670b452c1c5d9a287fb396d5e9eed494839d3bccac74dd5456ea3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wxh2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nlnv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:23Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.515935 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898d46ebaf15e8d33d02bea73cdb774bab2ea6001911a6dca0ae33c5d8d69e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfdc436c0635036c073c666b7acf24eb7d4d46af0a735122716d469ebf114a8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5ttn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:23Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.525894 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:23Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.536551 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:23Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.548529 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:23Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.550019 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.550075 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.550088 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.550104 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.550115 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:23Z","lastTransitionTime":"2025-11-24T13:13:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.620598 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5x94h_528a4b22-34ba-41b8-8c7d-07d98eebe02b/ovnkube-controller/1.log" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.623144 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" event={"ID":"528a4b22-34ba-41b8-8c7d-07d98eebe02b","Type":"ContainerStarted","Data":"07157ab848e1a1ecc3a7d078f861c8e8ceb87e6ec88b7f52513a1b99ba209a15"} Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.623310 4790 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.641531 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:23Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.653032 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.653065 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.653074 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.653088 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.653102 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:23Z","lastTransitionTime":"2025-11-24T13:13:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.663690 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:23Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.682522 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:23Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.700497 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec06ff65e703be63a81e99901326ec1eaa20c2176ddbd0764cc7cdc0b5761bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a2bf135bf819ef8ce489337a9f6b7b0bb626ff335b37ec7e6a74e0e13a1ba40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f3d5e18b5f456b7627a9154ff8fb00945ec026d4fb8797d4c4f3e2a2fbb1db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad72e45458f88569f2bb77a55dc98e2bad29a7e451e5258d371722616042bebd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23e4305b0dc57da3ad6d92e8f04f1f18780c1ebaf63ee19140a8443b99197bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908f7697965f4e6d820bfe32f7f59960e917dedea04373ff0e2b8126bb835163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07157ab848e1a1ecc3a7d078f861c8e8ceb87e6ec88b7f52513a1b99ba209a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80a2a718c71e9b762d00ee945b50c2e0c5ec37bea569e7499c17e8a8c5487878\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"message\\\":\\\" \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}\\\\nI1124 13:13:09.646982 6250 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 666.9µs\\\\nI1124 13:13:09.646986 6250 services_controller.go:356] Processing sync for service openshift-machine-config-operator/machine-config-daemon for network=default\\\\nI1124 13:13:09.646995 6250 services_controller.go:356] Processing sync for service openshift-console-operator/metrics for network=default\\\\nF1124 13:13:09.647013 6250 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z]\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b16be7938ff339aa0f92c4fb0d444114ac253439e80bf12ceb55c7e9a1944603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:23Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.719073 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fec678890b719f519b6740a344414413e17705e146440e130765a0f13b8a4407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:23Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.731770 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:23Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.745349 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3641c5a28dae87d2696ab209ab60e5bfadc8881ee96237f823df0f7ad8e841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:23Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.755901 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.755938 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.755949 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.755966 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.755976 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:23Z","lastTransitionTime":"2025-11-24T13:13:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.759474 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:23Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.771473 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdtxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bef2f789-12a8-45b6-b875-7307992176ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdtxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:23Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.782437 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nlnv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0f7e0af-ed9a-4a5e-be2f-06abac132790\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ad38a0b670b452c1c5d9a287fb396d5e9eed494839d3bccac74dd5456ea3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wxh2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nlnv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:23Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.793045 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898d46ebaf15e8d33d02bea73cdb774bab2ea6001911a6dca0ae33c5d8d69e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfdc436c0635036c073c666b7acf24eb7d4d46af0a735122716d469ebf114a8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5ttn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:23Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.806307 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:23Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.818427 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:23Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.830116 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:23Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.839807 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:23Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.851606 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:23Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.858847 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.858894 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.858903 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.858919 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.858930 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:23Z","lastTransitionTime":"2025-11-24T13:13:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.961221 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.961261 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.961270 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.961284 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:23 crc kubenswrapper[4790]: I1124 13:13:23.961295 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:23Z","lastTransitionTime":"2025-11-24T13:13:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.063873 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.063949 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.063958 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.063998 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.064009 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:24Z","lastTransitionTime":"2025-11-24T13:13:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.166673 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.166707 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.166715 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.166729 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.166743 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:24Z","lastTransitionTime":"2025-11-24T13:13:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.269374 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.269416 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.269426 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.269439 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.269450 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:24Z","lastTransitionTime":"2025-11-24T13:13:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.314183 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:24 crc kubenswrapper[4790]: E1124 13:13:24.314312 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.371216 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.371266 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.371276 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.371294 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.371304 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:24Z","lastTransitionTime":"2025-11-24T13:13:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.473927 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.473979 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.473993 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.474013 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.474025 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:24Z","lastTransitionTime":"2025-11-24T13:13:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.576398 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.576439 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.576448 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.576461 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.576470 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:24Z","lastTransitionTime":"2025-11-24T13:13:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.612974 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.628176 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5x94h_528a4b22-34ba-41b8-8c7d-07d98eebe02b/ovnkube-controller/2.log" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.628964 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5x94h_528a4b22-34ba-41b8-8c7d-07d98eebe02b/ovnkube-controller/1.log" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.631788 4790 generic.go:334] "Generic (PLEG): container finished" podID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerID="07157ab848e1a1ecc3a7d078f861c8e8ceb87e6ec88b7f52513a1b99ba209a15" exitCode=1 Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.631821 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" event={"ID":"528a4b22-34ba-41b8-8c7d-07d98eebe02b","Type":"ContainerDied","Data":"07157ab848e1a1ecc3a7d078f861c8e8ceb87e6ec88b7f52513a1b99ba209a15"} Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.631849 4790 scope.go:117] "RemoveContainer" containerID="80a2a718c71e9b762d00ee945b50c2e0c5ec37bea569e7499c17e8a8c5487878" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.632672 4790 scope.go:117] "RemoveContainer" containerID="07157ab848e1a1ecc3a7d078f861c8e8ceb87e6ec88b7f52513a1b99ba209a15" Nov 24 13:13:24 crc kubenswrapper[4790]: E1124 13:13:24.632923 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5x94h_openshift-ovn-kubernetes(528a4b22-34ba-41b8-8c7d-07d98eebe02b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.653239 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3641c5a28dae87d2696ab209ab60e5bfadc8881ee96237f823df0f7ad8e841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:24Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.667916 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:24Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.679283 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.679344 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.679356 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.679373 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.679385 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:24Z","lastTransitionTime":"2025-11-24T13:13:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.681095 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:24Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.700031 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec06ff65e703be63a81e99901326ec1eaa20c2176ddbd0764cc7cdc0b5761bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a2bf135bf819ef8ce489337a9f6b7b0bb626ff335b37ec7e6a74e0e13a1ba40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f3d5e18b5f456b7627a9154ff8fb00945ec026d4fb8797d4c4f3e2a2fbb1db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad72e45458f88569f2bb77a55dc98e2bad29a7e451e5258d371722616042bebd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23e4305b0dc57da3ad6d92e8f04f1f18780c1ebaf63ee19140a8443b99197bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908f7697965f4e6d820bfe32f7f59960e917dedea04373ff0e2b8126bb835163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07157ab848e1a1ecc3a7d078f861c8e8ceb87e6ec88b7f52513a1b99ba209a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80a2a718c71e9b762d00ee945b50c2e0c5ec37bea569e7499c17e8a8c5487878\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"message\\\":\\\" \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}\\\\nI1124 13:13:09.646982 6250 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 666.9µs\\\\nI1124 13:13:09.646986 6250 services_controller.go:356] Processing sync for service openshift-machine-config-operator/machine-config-daemon for network=default\\\\nI1124 13:13:09.646995 6250 services_controller.go:356] Processing sync for service openshift-console-operator/metrics for network=default\\\\nF1124 13:13:09.647013 6250 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:09Z is after 2025-08-24T17:21:41Z]\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07157ab848e1a1ecc3a7d078f861c8e8ceb87e6ec88b7f52513a1b99ba209a15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:24Z\\\",\\\"message\\\":\\\"\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 13:13:24.030836 6479 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 13:13:24.030868 6479 lb_config.go:1031] Cluster endpoints for openshift-operator-lifecycle-manager/package-server-manager-metrics for network=default are: map[]\\\\nI1124 13:13:24.031184 6479 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 13:13:24.031220 6479 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 13:13:24.031236 6479 services_controller.go:445] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1124 13:13:24.030958 6479 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b16be7938ff339aa0f92c4fb0d444114ac253439e80bf12ceb55c7e9a1944603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:24Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.713265 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fec678890b719f519b6740a344414413e17705e146440e130765a0f13b8a4407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:24Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.726246 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:24Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.736990 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdtxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bef2f789-12a8-45b6-b875-7307992176ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdtxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:24Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.750157 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:24Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.761845 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:24Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.773380 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:24Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.781123 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.781312 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.781380 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.781446 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.781513 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:24Z","lastTransitionTime":"2025-11-24T13:13:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.783006 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:24Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.792979 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nlnv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0f7e0af-ed9a-4a5e-be2f-06abac132790\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ad38a0b670b452c1c5d9a287fb396d5e9eed494839d3bccac74dd5456ea3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wxh2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nlnv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:24Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.803687 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898d46ebaf15e8d33d02bea73cdb774bab2ea6001911a6dca0ae33c5d8d69e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfdc436c0635036c073c666b7acf24eb7d4d46af0a735122716d469ebf114a8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5ttn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:24Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.813252 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:24Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.827389 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:24Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.840139 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:24Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.883988 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.884036 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.884046 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.884064 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.884081 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:24Z","lastTransitionTime":"2025-11-24T13:13:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.986803 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.986842 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.986851 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.986867 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:24 crc kubenswrapper[4790]: I1124 13:13:24.986906 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:24Z","lastTransitionTime":"2025-11-24T13:13:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.088529 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.088579 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.088589 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.088617 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.088626 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:25Z","lastTransitionTime":"2025-11-24T13:13:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.190678 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.190711 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.190728 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.190744 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.190755 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:25Z","lastTransitionTime":"2025-11-24T13:13:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.293239 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.293286 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.293297 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.293312 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.293325 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:25Z","lastTransitionTime":"2025-11-24T13:13:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.313698 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.313752 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.313707 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:13:25 crc kubenswrapper[4790]: E1124 13:13:25.313806 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:13:25 crc kubenswrapper[4790]: E1124 13:13:25.313953 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:13:25 crc kubenswrapper[4790]: E1124 13:13:25.314034 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.395626 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.395674 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.395690 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.395714 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.395730 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:25Z","lastTransitionTime":"2025-11-24T13:13:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.497863 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.497925 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.497936 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.497957 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.497971 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:25Z","lastTransitionTime":"2025-11-24T13:13:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.600735 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.600775 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.600786 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.600803 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.600814 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:25Z","lastTransitionTime":"2025-11-24T13:13:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.636732 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5x94h_528a4b22-34ba-41b8-8c7d-07d98eebe02b/ovnkube-controller/2.log" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.640977 4790 scope.go:117] "RemoveContainer" containerID="07157ab848e1a1ecc3a7d078f861c8e8ceb87e6ec88b7f52513a1b99ba209a15" Nov 24 13:13:25 crc kubenswrapper[4790]: E1124 13:13:25.641286 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5x94h_openshift-ovn-kubernetes(528a4b22-34ba-41b8-8c7d-07d98eebe02b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.652158 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.665429 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.677942 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.690768 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.703286 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.703468 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.703533 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.703602 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.703665 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:25Z","lastTransitionTime":"2025-11-24T13:13:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.709669 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec06ff65e703be63a81e99901326ec1eaa20c2176ddbd0764cc7cdc0b5761bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a2bf135bf819ef8ce489337a9f6b7b0bb626ff335b37ec7e6a74e0e13a1ba40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f3d5e18b5f456b7627a9154ff8fb00945ec026d4fb8797d4c4f3e2a2fbb1db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad72e45458f88569f2bb77a55dc98e2bad29a7e451e5258d371722616042bebd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23e4305b0dc57da3ad6d92e8f04f1f18780c1ebaf63ee19140a8443b99197bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908f7697965f4e6d820bfe32f7f59960e917dedea04373ff0e2b8126bb835163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07157ab848e1a1ecc3a7d078f861c8e8ceb87e6ec88b7f52513a1b99ba209a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07157ab848e1a1ecc3a7d078f861c8e8ceb87e6ec88b7f52513a1b99ba209a15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:24Z\\\",\\\"message\\\":\\\"\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 13:13:24.030836 6479 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 13:13:24.030868 6479 lb_config.go:1031] Cluster endpoints for openshift-operator-lifecycle-manager/package-server-manager-metrics for network=default are: map[]\\\\nI1124 13:13:24.031184 6479 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 13:13:24.031220 6479 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 13:13:24.031236 6479 services_controller.go:445] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1124 13:13:24.030958 6479 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5x94h_openshift-ovn-kubernetes(528a4b22-34ba-41b8-8c7d-07d98eebe02b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b16be7938ff339aa0f92c4fb0d444114ac253439e80bf12ceb55c7e9a1944603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.725041 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fec678890b719f519b6740a344414413e17705e146440e130765a0f13b8a4407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.735207 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.746944 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3641c5a28dae87d2696ab209ab60e5bfadc8881ee96237f823df0f7ad8e841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.757762 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.766471 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdtxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bef2f789-12a8-45b6-b875-7307992176ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdtxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.774163 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nlnv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0f7e0af-ed9a-4a5e-be2f-06abac132790\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ad38a0b670b452c1c5d9a287fb396d5e9eed494839d3bccac74dd5456ea3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wxh2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nlnv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.783494 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898d46ebaf15e8d33d02bea73cdb774bab2ea6001911a6dca0ae33c5d8d69e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfdc436c0635036c073c666b7acf24eb7d4d46af0a735122716d469ebf114a8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5ttn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.796569 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.805850 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.805899 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.805910 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.805927 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.805938 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:25Z","lastTransitionTime":"2025-11-24T13:13:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.808977 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.821949 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.831316 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.908222 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.908316 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.908331 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.908350 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:25 crc kubenswrapper[4790]: I1124 13:13:25.908364 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:25Z","lastTransitionTime":"2025-11-24T13:13:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.010703 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.010756 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.010770 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.010791 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.010809 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:26Z","lastTransitionTime":"2025-11-24T13:13:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.113702 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.113744 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.113754 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.113768 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.113797 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:26Z","lastTransitionTime":"2025-11-24T13:13:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.216236 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.216273 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.216283 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.216298 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.216309 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:26Z","lastTransitionTime":"2025-11-24T13:13:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.314509 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:26 crc kubenswrapper[4790]: E1124 13:13:26.314628 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.318015 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.318186 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.318271 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.318364 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.318449 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:26Z","lastTransitionTime":"2025-11-24T13:13:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.421346 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.421758 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.421850 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.422007 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.422103 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:26Z","lastTransitionTime":"2025-11-24T13:13:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.495005 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bef2f789-12a8-45b6-b875-7307992176ee-metrics-certs\") pod \"network-metrics-daemon-vdtxl\" (UID: \"bef2f789-12a8-45b6-b875-7307992176ee\") " pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:26 crc kubenswrapper[4790]: E1124 13:13:26.495119 4790 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:13:26 crc kubenswrapper[4790]: E1124 13:13:26.495171 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bef2f789-12a8-45b6-b875-7307992176ee-metrics-certs podName:bef2f789-12a8-45b6-b875-7307992176ee nodeName:}" failed. No retries permitted until 2025-11-24 13:13:42.495156888 +0000 UTC m=+70.875050550 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bef2f789-12a8-45b6-b875-7307992176ee-metrics-certs") pod "network-metrics-daemon-vdtxl" (UID: "bef2f789-12a8-45b6-b875-7307992176ee") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.524755 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.525103 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.525294 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.525455 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.525592 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:26Z","lastTransitionTime":"2025-11-24T13:13:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.628727 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.629319 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.629571 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.629867 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.630174 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:26Z","lastTransitionTime":"2025-11-24T13:13:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.733342 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.733397 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.733404 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.733423 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.733435 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:26Z","lastTransitionTime":"2025-11-24T13:13:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.836096 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.836132 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.836144 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.836160 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.836170 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:26Z","lastTransitionTime":"2025-11-24T13:13:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.939475 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.939514 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.939532 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.939559 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:26 crc kubenswrapper[4790]: I1124 13:13:26.939574 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:26Z","lastTransitionTime":"2025-11-24T13:13:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.042079 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.042118 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.042127 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.042142 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.042152 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:27Z","lastTransitionTime":"2025-11-24T13:13:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.144708 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.144766 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.144776 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.144796 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.144807 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:27Z","lastTransitionTime":"2025-11-24T13:13:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.248156 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.248209 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.248219 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.248237 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.248250 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:27Z","lastTransitionTime":"2025-11-24T13:13:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.314296 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:13:27 crc kubenswrapper[4790]: E1124 13:13:27.314497 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.314595 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.314608 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:27 crc kubenswrapper[4790]: E1124 13:13:27.314797 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:13:27 crc kubenswrapper[4790]: E1124 13:13:27.314960 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.351092 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.351145 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.351158 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.351181 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.351201 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:27Z","lastTransitionTime":"2025-11-24T13:13:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.454170 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.454215 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.454227 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.454243 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.454254 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:27Z","lastTransitionTime":"2025-11-24T13:13:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.556934 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.556972 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.556989 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.557009 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.557023 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:27Z","lastTransitionTime":"2025-11-24T13:13:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.659681 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.659727 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.659738 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.659753 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.659763 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:27Z","lastTransitionTime":"2025-11-24T13:13:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.762391 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.762424 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.762434 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.762446 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.762454 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:27Z","lastTransitionTime":"2025-11-24T13:13:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.864646 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.864676 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.864684 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.864698 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.864711 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:27Z","lastTransitionTime":"2025-11-24T13:13:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.967616 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.967987 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.968002 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.968021 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:27 crc kubenswrapper[4790]: I1124 13:13:27.968032 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:27Z","lastTransitionTime":"2025-11-24T13:13:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.070731 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.071481 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.071581 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.071679 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.071778 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:28Z","lastTransitionTime":"2025-11-24T13:13:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.174459 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.174693 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.174778 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.174848 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.174959 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:28Z","lastTransitionTime":"2025-11-24T13:13:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.277080 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.277119 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.277131 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.277147 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.277159 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:28Z","lastTransitionTime":"2025-11-24T13:13:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.314484 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:28 crc kubenswrapper[4790]: E1124 13:13:28.314724 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.380581 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.380652 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.380670 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.380695 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.380713 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:28Z","lastTransitionTime":"2025-11-24T13:13:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.417518 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.417594 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.417603 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.417617 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.417626 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:28Z","lastTransitionTime":"2025-11-24T13:13:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:28 crc kubenswrapper[4790]: E1124 13:13:28.430156 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:28Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.432968 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.433009 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.433018 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.433032 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.433045 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:28Z","lastTransitionTime":"2025-11-24T13:13:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:28 crc kubenswrapper[4790]: E1124 13:13:28.443385 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:28Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.446769 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.446807 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.446817 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.446834 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.446847 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:28Z","lastTransitionTime":"2025-11-24T13:13:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:28 crc kubenswrapper[4790]: E1124 13:13:28.457942 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:28Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.460837 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.460898 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.460912 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.460928 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.460940 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:28Z","lastTransitionTime":"2025-11-24T13:13:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:28 crc kubenswrapper[4790]: E1124 13:13:28.472694 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:28Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.473412 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.477895 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.478123 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.478209 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.478289 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.478359 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:28Z","lastTransitionTime":"2025-11-24T13:13:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.483831 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.486837 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:28Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:28 crc kubenswrapper[4790]: E1124 13:13:28.490521 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:28Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:28 crc kubenswrapper[4790]: E1124 13:13:28.490632 4790 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.492212 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.492245 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.492258 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.492274 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.492285 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:28Z","lastTransitionTime":"2025-11-24T13:13:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.499860 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:28Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.511046 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:28Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.521323 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:28Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.533547 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:28Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.549794 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec06ff65e703be63a81e99901326ec1eaa20c2176ddbd0764cc7cdc0b5761bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a2bf135bf819ef8ce489337a9f6b7b0bb626ff335b37ec7e6a74e0e13a1ba40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f3d5e18b5f456b7627a9154ff8fb00945ec026d4fb8797d4c4f3e2a2fbb1db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad72e45458f88569f2bb77a55dc98e2bad29a7e451e5258d371722616042bebd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23e4305b0dc57da3ad6d92e8f04f1f18780c1ebaf63ee19140a8443b99197bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908f7697965f4e6d820bfe32f7f59960e917dedea04373ff0e2b8126bb835163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07157ab848e1a1ecc3a7d078f861c8e8ceb87e6ec88b7f52513a1b99ba209a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07157ab848e1a1ecc3a7d078f861c8e8ceb87e6ec88b7f52513a1b99ba209a15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:24Z\\\",\\\"message\\\":\\\"\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 13:13:24.030836 6479 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 13:13:24.030868 6479 lb_config.go:1031] Cluster endpoints for openshift-operator-lifecycle-manager/package-server-manager-metrics for network=default are: map[]\\\\nI1124 13:13:24.031184 6479 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 13:13:24.031220 6479 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 13:13:24.031236 6479 services_controller.go:445] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1124 13:13:24.030958 6479 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5x94h_openshift-ovn-kubernetes(528a4b22-34ba-41b8-8c7d-07d98eebe02b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b16be7938ff339aa0f92c4fb0d444114ac253439e80bf12ceb55c7e9a1944603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:28Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.563781 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fec678890b719f519b6740a344414413e17705e146440e130765a0f13b8a4407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:28Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.574120 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:28Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.585532 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3641c5a28dae87d2696ab209ab60e5bfadc8881ee96237f823df0f7ad8e841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:28Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.594776 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdtxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bef2f789-12a8-45b6-b875-7307992176ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdtxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:28Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.595267 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.595307 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.595321 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.595354 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.595365 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:28Z","lastTransitionTime":"2025-11-24T13:13:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.604039 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:28Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.613015 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nlnv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0f7e0af-ed9a-4a5e-be2f-06abac132790\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ad38a0b670b452c1c5d9a287fb396d5e9eed494839d3bccac74dd5456ea3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wxh2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nlnv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:28Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.623992 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898d46ebaf15e8d33d02bea73cdb774bab2ea6001911a6dca0ae33c5d8d69e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfdc436c0635036c073c666b7acf24eb7d4d46af0a735122716d469ebf114a8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5ttn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:28Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.635122 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:28Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.647507 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:28Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.658775 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:28Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.697549 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.697656 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.697680 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.697700 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.697715 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:28Z","lastTransitionTime":"2025-11-24T13:13:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.799704 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.799747 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.799756 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.799770 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.799782 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:28Z","lastTransitionTime":"2025-11-24T13:13:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.902941 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.902996 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.903005 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.903021 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:28 crc kubenswrapper[4790]: I1124 13:13:28.903029 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:28Z","lastTransitionTime":"2025-11-24T13:13:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.005041 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.005082 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.005090 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.005103 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.005113 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:29Z","lastTransitionTime":"2025-11-24T13:13:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.109557 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.109624 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.109641 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.109657 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.109668 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:29Z","lastTransitionTime":"2025-11-24T13:13:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.215549 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.215628 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.215651 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.215681 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.215702 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:29Z","lastTransitionTime":"2025-11-24T13:13:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.313703 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:13:29 crc kubenswrapper[4790]: E1124 13:13:29.313940 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.314003 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:13:29 crc kubenswrapper[4790]: E1124 13:13:29.314201 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.314082 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:29 crc kubenswrapper[4790]: E1124 13:13:29.314345 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.318040 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.318094 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.318106 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.318122 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.318132 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:29Z","lastTransitionTime":"2025-11-24T13:13:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.420340 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.420403 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.420413 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.420426 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.420436 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:29Z","lastTransitionTime":"2025-11-24T13:13:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.523192 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.523226 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.523235 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.523247 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.523256 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:29Z","lastTransitionTime":"2025-11-24T13:13:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.625940 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.625987 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.625999 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.626014 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.626026 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:29Z","lastTransitionTime":"2025-11-24T13:13:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.728520 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.728557 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.728565 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.728577 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.728586 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:29Z","lastTransitionTime":"2025-11-24T13:13:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.830448 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.830480 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.830496 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.830512 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.830527 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:29Z","lastTransitionTime":"2025-11-24T13:13:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.864635 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.878865 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5e86ac0-955f-4455-a55e-db96696dbc80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c123986828c19da07e80adf0ff83f2d5858aa1681caf2f097164fa1067e6b247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee38955b1cc92f846dccf1fb30021559dfa4cefe62e4d095dcf983ac7b0adba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://103fb4a5a8c9d50f090bd5b73e841e6d6bd3a52817fa60d8681461c3b7658ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a59ab7ed8087587de107062d659abc98e07ce74ae4a8ed5e372bbeee9e6a8908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59ab7ed8087587de107062d659abc98e07ce74ae4a8ed5e372bbeee9e6a8908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:29Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.891737 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:29Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.905401 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:29Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.916428 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:29Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.927466 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3641c5a28dae87d2696ab209ab60e5bfadc8881ee96237f823df0f7ad8e841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:29Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.932440 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.932506 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.932522 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.932537 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.932599 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:29Z","lastTransitionTime":"2025-11-24T13:13:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.939478 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:29Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.950467 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:29Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.967264 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec06ff65e703be63a81e99901326ec1eaa20c2176ddbd0764cc7cdc0b5761bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a2bf135bf819ef8ce489337a9f6b7b0bb626ff335b37ec7e6a74e0e13a1ba40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f3d5e18b5f456b7627a9154ff8fb00945ec026d4fb8797d4c4f3e2a2fbb1db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad72e45458f88569f2bb77a55dc98e2bad29a7e451e5258d371722616042bebd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23e4305b0dc57da3ad6d92e8f04f1f18780c1ebaf63ee19140a8443b99197bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908f7697965f4e6d820bfe32f7f59960e917dedea04373ff0e2b8126bb835163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07157ab848e1a1ecc3a7d078f861c8e8ceb87e6ec88b7f52513a1b99ba209a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07157ab848e1a1ecc3a7d078f861c8e8ceb87e6ec88b7f52513a1b99ba209a15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:24Z\\\",\\\"message\\\":\\\"\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 13:13:24.030836 6479 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 13:13:24.030868 6479 lb_config.go:1031] Cluster endpoints for openshift-operator-lifecycle-manager/package-server-manager-metrics for network=default are: map[]\\\\nI1124 13:13:24.031184 6479 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 13:13:24.031220 6479 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 13:13:24.031236 6479 services_controller.go:445] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1124 13:13:24.030958 6479 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5x94h_openshift-ovn-kubernetes(528a4b22-34ba-41b8-8c7d-07d98eebe02b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b16be7938ff339aa0f92c4fb0d444114ac253439e80bf12ceb55c7e9a1944603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:29Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.981150 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fec678890b719f519b6740a344414413e17705e146440e130765a0f13b8a4407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:29Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:29 crc kubenswrapper[4790]: I1124 13:13:29.992157 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:29Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.001811 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdtxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bef2f789-12a8-45b6-b875-7307992176ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdtxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:30Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.013092 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:30Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.025081 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:30Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.035206 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.035304 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.035322 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.035348 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.035367 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:30Z","lastTransitionTime":"2025-11-24T13:13:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.039922 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:30Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.049517 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:30Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.059757 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nlnv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0f7e0af-ed9a-4a5e-be2f-06abac132790\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ad38a0b670b452c1c5d9a287fb396d5e9eed494839d3bccac74dd5456ea3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wxh2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nlnv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:30Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.069701 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898d46ebaf15e8d33d02bea73cdb774bab2ea6001911a6dca0ae33c5d8d69e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfdc436c0635036c073c666b7acf24eb7d4d46af0a735122716d469ebf114a8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5ttn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:30Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.137697 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.137759 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.137770 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.137807 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.137819 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:30Z","lastTransitionTime":"2025-11-24T13:13:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.240109 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.240151 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.240159 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.240173 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.240183 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:30Z","lastTransitionTime":"2025-11-24T13:13:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.314226 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:30 crc kubenswrapper[4790]: E1124 13:13:30.314375 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.343096 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.343165 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.343189 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.343218 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.343235 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:30Z","lastTransitionTime":"2025-11-24T13:13:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.446514 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.446566 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.446581 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.446603 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.446616 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:30Z","lastTransitionTime":"2025-11-24T13:13:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.549405 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.549442 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.549450 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.549462 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.549470 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:30Z","lastTransitionTime":"2025-11-24T13:13:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.653320 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.653457 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.653476 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.653502 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.653558 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:30Z","lastTransitionTime":"2025-11-24T13:13:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.756290 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.756349 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.756361 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.756380 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.756392 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:30Z","lastTransitionTime":"2025-11-24T13:13:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.858844 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.859142 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.859171 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.859201 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.859222 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:30Z","lastTransitionTime":"2025-11-24T13:13:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.961447 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.961488 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.961499 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.961517 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:30 crc kubenswrapper[4790]: I1124 13:13:30.961528 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:30Z","lastTransitionTime":"2025-11-24T13:13:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.064435 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.064480 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.064491 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.064508 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.064520 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:31Z","lastTransitionTime":"2025-11-24T13:13:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.167004 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.167039 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.167048 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.167060 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.167068 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:31Z","lastTransitionTime":"2025-11-24T13:13:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.269733 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.269785 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.269796 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.269810 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.269820 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:31Z","lastTransitionTime":"2025-11-24T13:13:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.314032 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.314064 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.314051 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:13:31 crc kubenswrapper[4790]: E1124 13:13:31.314144 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:13:31 crc kubenswrapper[4790]: E1124 13:13:31.314193 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:13:31 crc kubenswrapper[4790]: E1124 13:13:31.314277 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.372341 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.372376 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.372388 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.372405 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.372414 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:31Z","lastTransitionTime":"2025-11-24T13:13:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.474939 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.474987 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.474998 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.475015 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.475027 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:31Z","lastTransitionTime":"2025-11-24T13:13:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.577326 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.577364 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.577377 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.577392 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.577403 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:31Z","lastTransitionTime":"2025-11-24T13:13:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.679375 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.679426 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.679438 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.679453 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.679463 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:31Z","lastTransitionTime":"2025-11-24T13:13:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.781427 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.781467 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.781478 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.781494 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.781506 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:31Z","lastTransitionTime":"2025-11-24T13:13:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.883639 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.883678 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.883690 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.883706 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.883717 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:31Z","lastTransitionTime":"2025-11-24T13:13:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.985707 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.985751 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.985767 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.985782 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:31 crc kubenswrapper[4790]: I1124 13:13:31.985792 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:31Z","lastTransitionTime":"2025-11-24T13:13:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.088599 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.088653 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.088665 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.088682 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.088694 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:32Z","lastTransitionTime":"2025-11-24T13:13:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.191056 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.191104 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.191114 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.191131 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.191142 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:32Z","lastTransitionTime":"2025-11-24T13:13:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.292777 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.292851 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.292864 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.292900 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.292911 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:32Z","lastTransitionTime":"2025-11-24T13:13:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.313610 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:32 crc kubenswrapper[4790]: E1124 13:13:32.313766 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.325067 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5e86ac0-955f-4455-a55e-db96696dbc80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c123986828c19da07e80adf0ff83f2d5858aa1681caf2f097164fa1067e6b247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee38955b1cc92f846dccf1fb30021559dfa4cefe62e4d095dcf983ac7b0adba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://103fb4a5a8c9d50f090bd5b73e841e6d6bd3a52817fa60d8681461c3b7658ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a59ab7ed8087587de107062d659abc98e07ce74ae4a8ed5e372bbeee9e6a8908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59ab7ed8087587de107062d659abc98e07ce74ae4a8ed5e372bbeee9e6a8908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:32Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.335834 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:32Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.346945 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:32Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.359128 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:32Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.369735 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:32Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.386422 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec06ff65e703be63a81e99901326ec1eaa20c2176ddbd0764cc7cdc0b5761bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a2bf135bf819ef8ce489337a9f6b7b0bb626ff335b37ec7e6a74e0e13a1ba40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f3d5e18b5f456b7627a9154ff8fb00945ec026d4fb8797d4c4f3e2a2fbb1db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad72e45458f88569f2bb77a55dc98e2bad29a7e451e5258d371722616042bebd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23e4305b0dc57da3ad6d92e8f04f1f18780c1ebaf63ee19140a8443b99197bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908f7697965f4e6d820bfe32f7f59960e917dedea04373ff0e2b8126bb835163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07157ab848e1a1ecc3a7d078f861c8e8ceb87e6ec88b7f52513a1b99ba209a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07157ab848e1a1ecc3a7d078f861c8e8ceb87e6ec88b7f52513a1b99ba209a15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:24Z\\\",\\\"message\\\":\\\"\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 13:13:24.030836 6479 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 13:13:24.030868 6479 lb_config.go:1031] Cluster endpoints for openshift-operator-lifecycle-manager/package-server-manager-metrics for network=default are: map[]\\\\nI1124 13:13:24.031184 6479 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 13:13:24.031220 6479 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 13:13:24.031236 6479 services_controller.go:445] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1124 13:13:24.030958 6479 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5x94h_openshift-ovn-kubernetes(528a4b22-34ba-41b8-8c7d-07d98eebe02b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b16be7938ff339aa0f92c4fb0d444114ac253439e80bf12ceb55c7e9a1944603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:32Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.394309 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.394356 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.394366 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.394381 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.394392 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:32Z","lastTransitionTime":"2025-11-24T13:13:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.408177 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fec678890b719f519b6740a344414413e17705e146440e130765a0f13b8a4407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:32Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.419830 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:32Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.434343 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3641c5a28dae87d2696ab209ab60e5bfadc8881ee96237f823df0f7ad8e841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:32Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.444773 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:32Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.454254 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdtxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bef2f789-12a8-45b6-b875-7307992176ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdtxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:32Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.464319 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nlnv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0f7e0af-ed9a-4a5e-be2f-06abac132790\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ad38a0b670b452c1c5d9a287fb396d5e9eed494839d3bccac74dd5456ea3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wxh2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nlnv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:32Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.475519 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898d46ebaf15e8d33d02bea73cdb774bab2ea6001911a6dca0ae33c5d8d69e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfdc436c0635036c073c666b7acf24eb7d4d46af0a735122716d469ebf114a8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5ttn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:32Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.486242 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:32Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.497729 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:32Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.500310 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.500352 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.500361 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.500376 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.500385 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:32Z","lastTransitionTime":"2025-11-24T13:13:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.511201 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:32Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.520954 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:32Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.602498 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.602527 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.602535 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.602550 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.602559 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:32Z","lastTransitionTime":"2025-11-24T13:13:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.705457 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.705494 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.705502 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.705515 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.705524 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:32Z","lastTransitionTime":"2025-11-24T13:13:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.807819 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.807855 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.807865 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.807904 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.807917 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:32Z","lastTransitionTime":"2025-11-24T13:13:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.911032 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.911079 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.911090 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.911108 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:32 crc kubenswrapper[4790]: I1124 13:13:32.911119 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:32Z","lastTransitionTime":"2025-11-24T13:13:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.014441 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.014484 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.014496 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.014512 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.014526 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:33Z","lastTransitionTime":"2025-11-24T13:13:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.120344 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.120384 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.120393 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.120406 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.120416 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:33Z","lastTransitionTime":"2025-11-24T13:13:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.223085 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.223116 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.223124 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.223138 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.223147 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:33Z","lastTransitionTime":"2025-11-24T13:13:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.313791 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.313828 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.313871 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:13:33 crc kubenswrapper[4790]: E1124 13:13:33.314576 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:13:33 crc kubenswrapper[4790]: E1124 13:13:33.314450 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:13:33 crc kubenswrapper[4790]: E1124 13:13:33.314146 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.325378 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.325418 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.325428 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.325441 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.325450 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:33Z","lastTransitionTime":"2025-11-24T13:13:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.427649 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.427694 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.427704 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.427719 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.427729 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:33Z","lastTransitionTime":"2025-11-24T13:13:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.530154 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.530193 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.530202 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.530215 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.530224 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:33Z","lastTransitionTime":"2025-11-24T13:13:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.632579 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.632651 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.632661 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.632679 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.632691 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:33Z","lastTransitionTime":"2025-11-24T13:13:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.736096 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.736163 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.736185 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.736222 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.736245 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:33Z","lastTransitionTime":"2025-11-24T13:13:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.839247 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.839301 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.839321 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.839345 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.839363 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:33Z","lastTransitionTime":"2025-11-24T13:13:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.942615 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.942685 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.942713 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.942748 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:33 crc kubenswrapper[4790]: I1124 13:13:33.942771 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:33Z","lastTransitionTime":"2025-11-24T13:13:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.046131 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.046194 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.046208 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.046226 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.046239 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:34Z","lastTransitionTime":"2025-11-24T13:13:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.149283 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.149354 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.149367 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.149391 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.149408 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:34Z","lastTransitionTime":"2025-11-24T13:13:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.251789 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.251835 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.251848 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.251863 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.251871 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:34Z","lastTransitionTime":"2025-11-24T13:13:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.314085 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:34 crc kubenswrapper[4790]: E1124 13:13:34.314258 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.354567 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.354626 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.354638 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.354663 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.354683 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:34Z","lastTransitionTime":"2025-11-24T13:13:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.457633 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.457670 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.457680 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.457697 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.457708 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:34Z","lastTransitionTime":"2025-11-24T13:13:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.561120 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.561169 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.561180 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.561195 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.561208 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:34Z","lastTransitionTime":"2025-11-24T13:13:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.663784 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.663831 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.663845 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.663862 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.663873 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:34Z","lastTransitionTime":"2025-11-24T13:13:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.766254 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.766299 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.766309 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.766324 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.766335 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:34Z","lastTransitionTime":"2025-11-24T13:13:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.868873 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.868970 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.868981 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.869002 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.869016 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:34Z","lastTransitionTime":"2025-11-24T13:13:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.971616 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.971676 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.971696 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.971717 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:34 crc kubenswrapper[4790]: I1124 13:13:34.971729 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:34Z","lastTransitionTime":"2025-11-24T13:13:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.073673 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.073720 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.073730 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.073747 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.073755 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:35Z","lastTransitionTime":"2025-11-24T13:13:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.176030 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.176071 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.176080 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.176099 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.176113 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:35Z","lastTransitionTime":"2025-11-24T13:13:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.278377 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.278416 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.278430 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.278445 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.278456 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:35Z","lastTransitionTime":"2025-11-24T13:13:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.314268 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.314300 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.314340 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:35 crc kubenswrapper[4790]: E1124 13:13:35.314400 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:13:35 crc kubenswrapper[4790]: E1124 13:13:35.314482 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:13:35 crc kubenswrapper[4790]: E1124 13:13:35.314570 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.381336 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.381388 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.381397 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.381415 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.381426 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:35Z","lastTransitionTime":"2025-11-24T13:13:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.483765 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.483822 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.483842 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.483865 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.483914 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:35Z","lastTransitionTime":"2025-11-24T13:13:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.586605 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.586646 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.586658 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.586674 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.586684 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:35Z","lastTransitionTime":"2025-11-24T13:13:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.688673 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.688731 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.688755 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.688776 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.688790 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:35Z","lastTransitionTime":"2025-11-24T13:13:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.792082 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.792176 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.792204 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.792236 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.792260 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:35Z","lastTransitionTime":"2025-11-24T13:13:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.894534 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.894564 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.894571 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.894584 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.894593 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:35Z","lastTransitionTime":"2025-11-24T13:13:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.997181 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.997228 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.997238 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.997253 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:35 crc kubenswrapper[4790]: I1124 13:13:35.997264 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:35Z","lastTransitionTime":"2025-11-24T13:13:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.099216 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.099256 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.099289 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.099307 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.099319 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:36Z","lastTransitionTime":"2025-11-24T13:13:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.201309 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.201401 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.201411 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.201423 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.201433 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:36Z","lastTransitionTime":"2025-11-24T13:13:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.302964 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.303002 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.303013 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.303031 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.303042 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:36Z","lastTransitionTime":"2025-11-24T13:13:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.315134 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:36 crc kubenswrapper[4790]: E1124 13:13:36.315301 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.405857 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.405975 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.405998 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.406023 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.406043 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:36Z","lastTransitionTime":"2025-11-24T13:13:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.508946 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.509000 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.509018 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.509039 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.509057 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:36Z","lastTransitionTime":"2025-11-24T13:13:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.611523 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.611579 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.611595 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.611614 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.611628 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:36Z","lastTransitionTime":"2025-11-24T13:13:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.713469 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.713506 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.713514 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.713526 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.713535 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:36Z","lastTransitionTime":"2025-11-24T13:13:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.815873 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.815953 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.815964 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.815982 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.815996 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:36Z","lastTransitionTime":"2025-11-24T13:13:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.918491 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.918538 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.918550 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.918564 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:36 crc kubenswrapper[4790]: I1124 13:13:36.918575 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:36Z","lastTransitionTime":"2025-11-24T13:13:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.022696 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.022785 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.022799 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.022820 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.022834 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:37Z","lastTransitionTime":"2025-11-24T13:13:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.125378 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.125422 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.125431 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.125448 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.125462 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:37Z","lastTransitionTime":"2025-11-24T13:13:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.227291 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.227324 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.227333 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.227346 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.227354 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:37Z","lastTransitionTime":"2025-11-24T13:13:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.314256 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:37 crc kubenswrapper[4790]: E1124 13:13:37.314367 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.314267 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:13:37 crc kubenswrapper[4790]: E1124 13:13:37.314451 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.314267 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:13:37 crc kubenswrapper[4790]: E1124 13:13:37.314534 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.329835 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.329870 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.329897 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.329913 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.329926 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:37Z","lastTransitionTime":"2025-11-24T13:13:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.432300 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.432352 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.432361 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.432375 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.432390 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:37Z","lastTransitionTime":"2025-11-24T13:13:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.534780 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.534829 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.534840 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.534856 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.534867 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:37Z","lastTransitionTime":"2025-11-24T13:13:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.637609 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.637643 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.637652 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.637665 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.637674 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:37Z","lastTransitionTime":"2025-11-24T13:13:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.739443 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.739489 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.739502 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.739518 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.739532 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:37Z","lastTransitionTime":"2025-11-24T13:13:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.842171 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.842525 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.842546 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.842565 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.842579 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:37Z","lastTransitionTime":"2025-11-24T13:13:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.945490 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.945538 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.945554 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.945569 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:37 crc kubenswrapper[4790]: I1124 13:13:37.945581 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:37Z","lastTransitionTime":"2025-11-24T13:13:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.048145 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.048184 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.048201 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.048215 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.048329 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:38Z","lastTransitionTime":"2025-11-24T13:13:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.150582 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.150628 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.150637 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.150651 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.150660 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:38Z","lastTransitionTime":"2025-11-24T13:13:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.253396 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.253448 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.253461 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.253484 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.253497 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:38Z","lastTransitionTime":"2025-11-24T13:13:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.314579 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:38 crc kubenswrapper[4790]: E1124 13:13:38.314965 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.315147 4790 scope.go:117] "RemoveContainer" containerID="07157ab848e1a1ecc3a7d078f861c8e8ceb87e6ec88b7f52513a1b99ba209a15" Nov 24 13:13:38 crc kubenswrapper[4790]: E1124 13:13:38.315352 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5x94h_openshift-ovn-kubernetes(528a4b22-34ba-41b8-8c7d-07d98eebe02b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.356132 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.356173 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.356184 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.356199 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.356209 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:38Z","lastTransitionTime":"2025-11-24T13:13:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.458538 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.458580 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.458590 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.458604 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.458613 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:38Z","lastTransitionTime":"2025-11-24T13:13:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.560539 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.560566 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.560574 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.560587 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.560596 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:38Z","lastTransitionTime":"2025-11-24T13:13:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.619202 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.619237 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.619250 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.619266 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.619278 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:38Z","lastTransitionTime":"2025-11-24T13:13:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:38 crc kubenswrapper[4790]: E1124 13:13:38.631134 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:38Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.634853 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.634943 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.634965 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.634990 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.635007 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:38Z","lastTransitionTime":"2025-11-24T13:13:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:38 crc kubenswrapper[4790]: E1124 13:13:38.647085 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:38Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.650912 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.651051 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.651077 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.651094 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.651107 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:38Z","lastTransitionTime":"2025-11-24T13:13:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:38 crc kubenswrapper[4790]: E1124 13:13:38.661795 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:38Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.665630 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.665666 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.665679 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.665695 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.665707 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:38Z","lastTransitionTime":"2025-11-24T13:13:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:38 crc kubenswrapper[4790]: E1124 13:13:38.704753 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:38Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.711694 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.711742 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.711754 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.711769 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.711780 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:38Z","lastTransitionTime":"2025-11-24T13:13:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:38 crc kubenswrapper[4790]: E1124 13:13:38.732533 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:38Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:38 crc kubenswrapper[4790]: E1124 13:13:38.732645 4790 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.733993 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.734062 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.734079 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.734097 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.734110 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:38Z","lastTransitionTime":"2025-11-24T13:13:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.837082 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.837144 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.837164 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.837190 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.837208 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:38Z","lastTransitionTime":"2025-11-24T13:13:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.939705 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.939749 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.939760 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.939775 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:38 crc kubenswrapper[4790]: I1124 13:13:38.939784 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:38Z","lastTransitionTime":"2025-11-24T13:13:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.042507 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.042546 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.042555 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.042569 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.042581 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:39Z","lastTransitionTime":"2025-11-24T13:13:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.144324 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.144358 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.144367 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.144380 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.144388 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:39Z","lastTransitionTime":"2025-11-24T13:13:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.247164 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.247213 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.247222 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.247243 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.247256 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:39Z","lastTransitionTime":"2025-11-24T13:13:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.313747 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.313785 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.313762 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:13:39 crc kubenswrapper[4790]: E1124 13:13:39.313863 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:13:39 crc kubenswrapper[4790]: E1124 13:13:39.314007 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:13:39 crc kubenswrapper[4790]: E1124 13:13:39.314095 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.349663 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.349697 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.349706 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.349720 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.349729 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:39Z","lastTransitionTime":"2025-11-24T13:13:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.451955 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.451994 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.452004 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.452019 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.452028 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:39Z","lastTransitionTime":"2025-11-24T13:13:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.554199 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.554234 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.554241 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.554254 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.554263 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:39Z","lastTransitionTime":"2025-11-24T13:13:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.656466 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.656498 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.656507 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.656518 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.656528 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:39Z","lastTransitionTime":"2025-11-24T13:13:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.758961 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.759001 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.759010 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.759023 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.759034 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:39Z","lastTransitionTime":"2025-11-24T13:13:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.861446 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.861539 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.861557 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.861582 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.861600 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:39Z","lastTransitionTime":"2025-11-24T13:13:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.963543 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.963591 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.963604 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.963620 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:39 crc kubenswrapper[4790]: I1124 13:13:39.963633 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:39Z","lastTransitionTime":"2025-11-24T13:13:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.066014 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.066087 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.066101 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.066117 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.066128 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:40Z","lastTransitionTime":"2025-11-24T13:13:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.169941 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.169980 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.169990 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.170008 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.170019 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:40Z","lastTransitionTime":"2025-11-24T13:13:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.272220 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.272276 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.272307 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.272333 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.272347 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:40Z","lastTransitionTime":"2025-11-24T13:13:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.313811 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:40 crc kubenswrapper[4790]: E1124 13:13:40.313954 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.375634 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.375680 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.375691 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.375706 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.375723 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:40Z","lastTransitionTime":"2025-11-24T13:13:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.478181 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.478218 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.478226 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.478240 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.478250 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:40Z","lastTransitionTime":"2025-11-24T13:13:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.580449 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.580488 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.580498 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.580512 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.580521 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:40Z","lastTransitionTime":"2025-11-24T13:13:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.682686 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.682738 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.682752 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.682768 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.682780 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:40Z","lastTransitionTime":"2025-11-24T13:13:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.784752 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.784792 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.784808 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.784824 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.784834 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:40Z","lastTransitionTime":"2025-11-24T13:13:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.887652 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.887690 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.887701 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.887715 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.887728 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:40Z","lastTransitionTime":"2025-11-24T13:13:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.990506 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.990541 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.990573 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.990592 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:40 crc kubenswrapper[4790]: I1124 13:13:40.990603 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:40Z","lastTransitionTime":"2025-11-24T13:13:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.092736 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.092773 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.092782 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.092796 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.092805 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:41Z","lastTransitionTime":"2025-11-24T13:13:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.195078 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.195105 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.195113 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.195127 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.195136 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:41Z","lastTransitionTime":"2025-11-24T13:13:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.298627 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.299060 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.299233 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.299374 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.299504 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:41Z","lastTransitionTime":"2025-11-24T13:13:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.314165 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.314273 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:41 crc kubenswrapper[4790]: E1124 13:13:41.314326 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:13:41 crc kubenswrapper[4790]: E1124 13:13:41.314419 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.314972 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:13:41 crc kubenswrapper[4790]: E1124 13:13:41.315330 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.402327 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.402411 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.402448 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.402484 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.402512 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:41Z","lastTransitionTime":"2025-11-24T13:13:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.505354 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.505418 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.505429 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.505446 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.505456 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:41Z","lastTransitionTime":"2025-11-24T13:13:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.608390 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.608460 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.608477 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.608502 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.608523 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:41Z","lastTransitionTime":"2025-11-24T13:13:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.711206 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.711253 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.711267 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.711286 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.711303 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:41Z","lastTransitionTime":"2025-11-24T13:13:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.814020 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.814073 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.814083 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.814107 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.814126 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:41Z","lastTransitionTime":"2025-11-24T13:13:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.916593 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.916638 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.916652 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.916673 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:41 crc kubenswrapper[4790]: I1124 13:13:41.916688 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:41Z","lastTransitionTime":"2025-11-24T13:13:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.018848 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.018924 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.018935 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.018953 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.018964 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:42Z","lastTransitionTime":"2025-11-24T13:13:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.122169 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.122204 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.122213 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.122245 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.122254 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:42Z","lastTransitionTime":"2025-11-24T13:13:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.224397 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.224436 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.224452 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.224467 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.224477 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:42Z","lastTransitionTime":"2025-11-24T13:13:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.314580 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:42 crc kubenswrapper[4790]: E1124 13:13:42.314706 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.327386 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.327413 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.327424 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.327439 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.327452 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:42Z","lastTransitionTime":"2025-11-24T13:13:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.328007 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:42Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.339594 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:42Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.352607 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:42Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.365795 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nlnv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0f7e0af-ed9a-4a5e-be2f-06abac132790\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ad38a0b670b452c1c5d9a287fb396d5e9eed494839d3bccac74dd5456ea3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wxh2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nlnv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:42Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.379993 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898d46ebaf15e8d33d02bea73cdb774bab2ea6001911a6dca0ae33c5d8d69e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfdc436c0635036c073c666b7acf24eb7d4d46af0a735122716d469ebf114a8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5ttn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:42Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.395699 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:42Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.408840 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:42Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.424540 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5e86ac0-955f-4455-a55e-db96696dbc80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c123986828c19da07e80adf0ff83f2d5858aa1681caf2f097164fa1067e6b247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee38955b1cc92f846dccf1fb30021559dfa4cefe62e4d095dcf983ac7b0adba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://103fb4a5a8c9d50f090bd5b73e841e6d6bd3a52817fa60d8681461c3b7658ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a59ab7ed8087587de107062d659abc98e07ce74ae4a8ed5e372bbeee9e6a8908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59ab7ed8087587de107062d659abc98e07ce74ae4a8ed5e372bbeee9e6a8908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:42Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.429897 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.429941 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.429950 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.429967 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.430001 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:42Z","lastTransitionTime":"2025-11-24T13:13:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.439564 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:42Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.451995 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:42Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.464858 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3641c5a28dae87d2696ab209ab60e5bfadc8881ee96237f823df0f7ad8e841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:42Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.478396 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:42Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.491835 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:42Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.509388 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec06ff65e703be63a81e99901326ec1eaa20c2176ddbd0764cc7cdc0b5761bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a2bf135bf819ef8ce489337a9f6b7b0bb626ff335b37ec7e6a74e0e13a1ba40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f3d5e18b5f456b7627a9154ff8fb00945ec026d4fb8797d4c4f3e2a2fbb1db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad72e45458f88569f2bb77a55dc98e2bad29a7e451e5258d371722616042bebd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23e4305b0dc57da3ad6d92e8f04f1f18780c1ebaf63ee19140a8443b99197bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908f7697965f4e6d820bfe32f7f59960e917dedea04373ff0e2b8126bb835163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07157ab848e1a1ecc3a7d078f861c8e8ceb87e6ec88b7f52513a1b99ba209a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07157ab848e1a1ecc3a7d078f861c8e8ceb87e6ec88b7f52513a1b99ba209a15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:24Z\\\",\\\"message\\\":\\\"\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 13:13:24.030836 6479 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 13:13:24.030868 6479 lb_config.go:1031] Cluster endpoints for openshift-operator-lifecycle-manager/package-server-manager-metrics for network=default are: map[]\\\\nI1124 13:13:24.031184 6479 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 13:13:24.031220 6479 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 13:13:24.031236 6479 services_controller.go:445] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1124 13:13:24.030958 6479 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5x94h_openshift-ovn-kubernetes(528a4b22-34ba-41b8-8c7d-07d98eebe02b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b16be7938ff339aa0f92c4fb0d444114ac253439e80bf12ceb55c7e9a1944603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:42Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.525556 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fec678890b719f519b6740a344414413e17705e146440e130765a0f13b8a4407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:42Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.533744 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.533794 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.533811 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.533833 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.533846 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:42Z","lastTransitionTime":"2025-11-24T13:13:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.538955 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:42Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.551866 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdtxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bef2f789-12a8-45b6-b875-7307992176ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdtxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:42Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.553502 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bef2f789-12a8-45b6-b875-7307992176ee-metrics-certs\") pod \"network-metrics-daemon-vdtxl\" (UID: \"bef2f789-12a8-45b6-b875-7307992176ee\") " pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:42 crc kubenswrapper[4790]: E1124 13:13:42.553807 4790 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:13:42 crc kubenswrapper[4790]: E1124 13:13:42.554005 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bef2f789-12a8-45b6-b875-7307992176ee-metrics-certs podName:bef2f789-12a8-45b6-b875-7307992176ee nodeName:}" failed. No retries permitted until 2025-11-24 13:14:14.553975167 +0000 UTC m=+102.933868859 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bef2f789-12a8-45b6-b875-7307992176ee-metrics-certs") pod "network-metrics-daemon-vdtxl" (UID: "bef2f789-12a8-45b6-b875-7307992176ee") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.635971 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.636005 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.636014 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.636027 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.636037 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:42Z","lastTransitionTime":"2025-11-24T13:13:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.738196 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.738235 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.738243 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.738257 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.738266 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:42Z","lastTransitionTime":"2025-11-24T13:13:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.840776 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.840825 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.840832 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.840845 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.840853 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:42Z","lastTransitionTime":"2025-11-24T13:13:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.943298 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.943373 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.943386 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.943483 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:42 crc kubenswrapper[4790]: I1124 13:13:42.943501 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:42Z","lastTransitionTime":"2025-11-24T13:13:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.045438 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.045474 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.045484 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.045500 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.045510 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:43Z","lastTransitionTime":"2025-11-24T13:13:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.147306 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.147349 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.147360 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.147376 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.147389 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:43Z","lastTransitionTime":"2025-11-24T13:13:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.249528 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.249561 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.249570 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.249582 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.249590 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:43Z","lastTransitionTime":"2025-11-24T13:13:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.313718 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.313785 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:13:43 crc kubenswrapper[4790]: E1124 13:13:43.313847 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.313967 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:43 crc kubenswrapper[4790]: E1124 13:13:43.314098 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:13:43 crc kubenswrapper[4790]: E1124 13:13:43.314172 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.351647 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.351682 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.351693 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.351708 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.351720 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:43Z","lastTransitionTime":"2025-11-24T13:13:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.453299 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.453342 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.453352 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.453366 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.453375 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:43Z","lastTransitionTime":"2025-11-24T13:13:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.555692 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.555747 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.555758 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.555774 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.555787 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:43Z","lastTransitionTime":"2025-11-24T13:13:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.658439 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.658486 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.658496 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.658512 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.658524 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:43Z","lastTransitionTime":"2025-11-24T13:13:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.761077 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.761138 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.761148 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.761162 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.761172 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:43Z","lastTransitionTime":"2025-11-24T13:13:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.863209 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.863242 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.863253 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.863268 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.863282 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:43Z","lastTransitionTime":"2025-11-24T13:13:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.966343 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.966406 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.966420 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.966437 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:43 crc kubenswrapper[4790]: I1124 13:13:43.966449 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:43Z","lastTransitionTime":"2025-11-24T13:13:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.069009 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.069263 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.069389 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.069505 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.069583 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:44Z","lastTransitionTime":"2025-11-24T13:13:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.171834 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.171865 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.171874 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.171901 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.171910 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:44Z","lastTransitionTime":"2025-11-24T13:13:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.273933 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.274003 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.274016 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.274037 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.274050 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:44Z","lastTransitionTime":"2025-11-24T13:13:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.314102 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:44 crc kubenswrapper[4790]: E1124 13:13:44.314280 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.377516 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.377590 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.377610 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.377639 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.377662 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:44Z","lastTransitionTime":"2025-11-24T13:13:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.480677 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.480732 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.480746 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.480768 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.480781 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:44Z","lastTransitionTime":"2025-11-24T13:13:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.584174 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.584303 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.584318 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.584343 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.584358 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:44Z","lastTransitionTime":"2025-11-24T13:13:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.686310 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.686362 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.686372 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.686389 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.686399 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:44Z","lastTransitionTime":"2025-11-24T13:13:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.705922 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-76rkg_6c0e3499-87ac-481f-a010-708a3a4a730f/kube-multus/0.log" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.705963 4790 generic.go:334] "Generic (PLEG): container finished" podID="6c0e3499-87ac-481f-a010-708a3a4a730f" containerID="6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09" exitCode=1 Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.705990 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-76rkg" event={"ID":"6c0e3499-87ac-481f-a010-708a3a4a730f","Type":"ContainerDied","Data":"6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09"} Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.706333 4790 scope.go:117] "RemoveContainer" containerID="6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.720229 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3641c5a28dae87d2696ab209ab60e5bfadc8881ee96237f823df0f7ad8e841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:44Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.734131 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:44Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.744910 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:44Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.764532 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec06ff65e703be63a81e99901326ec1eaa20c2176ddbd0764cc7cdc0b5761bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a2bf135bf819ef8ce489337a9f6b7b0bb626ff335b37ec7e6a74e0e13a1ba40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f3d5e18b5f456b7627a9154ff8fb00945ec026d4fb8797d4c4f3e2a2fbb1db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad72e45458f88569f2bb77a55dc98e2bad29a7e451e5258d371722616042bebd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23e4305b0dc57da3ad6d92e8f04f1f18780c1ebaf63ee19140a8443b99197bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908f7697965f4e6d820bfe32f7f59960e917dedea04373ff0e2b8126bb835163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07157ab848e1a1ecc3a7d078f861c8e8ceb87e6ec88b7f52513a1b99ba209a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07157ab848e1a1ecc3a7d078f861c8e8ceb87e6ec88b7f52513a1b99ba209a15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:24Z\\\",\\\"message\\\":\\\"\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 13:13:24.030836 6479 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 13:13:24.030868 6479 lb_config.go:1031] Cluster endpoints for openshift-operator-lifecycle-manager/package-server-manager-metrics for network=default are: map[]\\\\nI1124 13:13:24.031184 6479 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 13:13:24.031220 6479 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 13:13:24.031236 6479 services_controller.go:445] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1124 13:13:24.030958 6479 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5x94h_openshift-ovn-kubernetes(528a4b22-34ba-41b8-8c7d-07d98eebe02b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b16be7938ff339aa0f92c4fb0d444114ac253439e80bf12ceb55c7e9a1944603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:44Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.780441 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fec678890b719f519b6740a344414413e17705e146440e130765a0f13b8a4407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:44Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.789404 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.789442 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.789452 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.789467 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.789477 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:44Z","lastTransitionTime":"2025-11-24T13:13:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.790933 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:44Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.800597 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdtxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bef2f789-12a8-45b6-b875-7307992176ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdtxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:44Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.812122 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:44Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.822021 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:44Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.832052 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nlnv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0f7e0af-ed9a-4a5e-be2f-06abac132790\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ad38a0b670b452c1c5d9a287fb396d5e9eed494839d3bccac74dd5456ea3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wxh2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nlnv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:44Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.843944 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898d46ebaf15e8d33d02bea73cdb774bab2ea6001911a6dca0ae33c5d8d69e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfdc436c0635036c073c666b7acf24eb7d4d46af0a735122716d469ebf114a8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5ttn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:44Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.856366 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:44Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.868503 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:44Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.880466 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5e86ac0-955f-4455-a55e-db96696dbc80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c123986828c19da07e80adf0ff83f2d5858aa1681caf2f097164fa1067e6b247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee38955b1cc92f846dccf1fb30021559dfa4cefe62e4d095dcf983ac7b0adba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://103fb4a5a8c9d50f090bd5b73e841e6d6bd3a52817fa60d8681461c3b7658ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a59ab7ed8087587de107062d659abc98e07ce74ae4a8ed5e372bbeee9e6a8908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59ab7ed8087587de107062d659abc98e07ce74ae4a8ed5e372bbeee9e6a8908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:44Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.892353 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.892425 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.892436 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.892477 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.892491 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:44Z","lastTransitionTime":"2025-11-24T13:13:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.894295 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:44Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.909317 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:44Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.924725 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:44Z\\\",\\\"message\\\":\\\"2025-11-24T13:12:58+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_2c499d3c-4900-45ee-b815-27ca113496c9\\\\n2025-11-24T13:12:58+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_2c499d3c-4900-45ee-b815-27ca113496c9 to /host/opt/cni/bin/\\\\n2025-11-24T13:12:59Z [verbose] multus-daemon started\\\\n2025-11-24T13:12:59Z [verbose] Readiness Indicator file check\\\\n2025-11-24T13:13:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:44Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.994709 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.994765 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.994778 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.994793 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:44 crc kubenswrapper[4790]: I1124 13:13:44.994804 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:44Z","lastTransitionTime":"2025-11-24T13:13:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.096929 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.096993 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.097004 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.097019 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.097027 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:45Z","lastTransitionTime":"2025-11-24T13:13:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.199522 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.199568 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.199579 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.199596 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.199608 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:45Z","lastTransitionTime":"2025-11-24T13:13:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.302903 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.302948 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.302958 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.302993 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.303004 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:45Z","lastTransitionTime":"2025-11-24T13:13:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.313737 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:13:45 crc kubenswrapper[4790]: E1124 13:13:45.313868 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.314060 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.314146 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:13:45 crc kubenswrapper[4790]: E1124 13:13:45.314239 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:13:45 crc kubenswrapper[4790]: E1124 13:13:45.314385 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.405213 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.405258 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.405266 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.405280 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.405289 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:45Z","lastTransitionTime":"2025-11-24T13:13:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.507999 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.508040 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.508049 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.508071 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.508088 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:45Z","lastTransitionTime":"2025-11-24T13:13:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.610830 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.610896 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.610906 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.610925 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.610936 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:45Z","lastTransitionTime":"2025-11-24T13:13:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.711632 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-76rkg_6c0e3499-87ac-481f-a010-708a3a4a730f/kube-multus/0.log" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.711708 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-76rkg" event={"ID":"6c0e3499-87ac-481f-a010-708a3a4a730f","Type":"ContainerStarted","Data":"6af0748fbdc734fe38245ac5426022cd063620ec033c5e4285073feb4a376019"} Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.712764 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.712811 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.712829 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.712850 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.712866 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:45Z","lastTransitionTime":"2025-11-24T13:13:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.731972 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.745272 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.758560 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.768442 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.777267 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nlnv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0f7e0af-ed9a-4a5e-be2f-06abac132790\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ad38a0b670b452c1c5d9a287fb396d5e9eed494839d3bccac74dd5456ea3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wxh2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nlnv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.787788 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898d46ebaf15e8d33d02bea73cdb774bab2ea6001911a6dca0ae33c5d8d69e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfdc436c0635036c073c666b7acf24eb7d4d46af0a735122716d469ebf114a8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5ttn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.798246 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5e86ac0-955f-4455-a55e-db96696dbc80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c123986828c19da07e80adf0ff83f2d5858aa1681caf2f097164fa1067e6b247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee38955b1cc92f846dccf1fb30021559dfa4cefe62e4d095dcf983ac7b0adba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://103fb4a5a8c9d50f090bd5b73e841e6d6bd3a52817fa60d8681461c3b7658ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a59ab7ed8087587de107062d659abc98e07ce74ae4a8ed5e372bbeee9e6a8908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59ab7ed8087587de107062d659abc98e07ce74ae4a8ed5e372bbeee9e6a8908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.808008 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.815472 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.815527 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.815540 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.815555 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.815567 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:45Z","lastTransitionTime":"2025-11-24T13:13:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.820393 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.830956 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af0748fbdc734fe38245ac5426022cd063620ec033c5e4285073feb4a376019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:44Z\\\",\\\"message\\\":\\\"2025-11-24T13:12:58+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_2c499d3c-4900-45ee-b815-27ca113496c9\\\\n2025-11-24T13:12:58+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_2c499d3c-4900-45ee-b815-27ca113496c9 to /host/opt/cni/bin/\\\\n2025-11-24T13:12:59Z [verbose] multus-daemon started\\\\n2025-11-24T13:12:59Z [verbose] Readiness Indicator file check\\\\n2025-11-24T13:13:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.844489 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fec678890b719f519b6740a344414413e17705e146440e130765a0f13b8a4407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.853115 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.864096 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3641c5a28dae87d2696ab209ab60e5bfadc8881ee96237f823df0f7ad8e841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.874369 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.884261 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.903410 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec06ff65e703be63a81e99901326ec1eaa20c2176ddbd0764cc7cdc0b5761bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a2bf135bf819ef8ce489337a9f6b7b0bb626ff335b37ec7e6a74e0e13a1ba40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f3d5e18b5f456b7627a9154ff8fb00945ec026d4fb8797d4c4f3e2a2fbb1db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad72e45458f88569f2bb77a55dc98e2bad29a7e451e5258d371722616042bebd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23e4305b0dc57da3ad6d92e8f04f1f18780c1ebaf63ee19140a8443b99197bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908f7697965f4e6d820bfe32f7f59960e917dedea04373ff0e2b8126bb835163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07157ab848e1a1ecc3a7d078f861c8e8ceb87e6ec88b7f52513a1b99ba209a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07157ab848e1a1ecc3a7d078f861c8e8ceb87e6ec88b7f52513a1b99ba209a15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:24Z\\\",\\\"message\\\":\\\"\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 13:13:24.030836 6479 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 13:13:24.030868 6479 lb_config.go:1031] Cluster endpoints for openshift-operator-lifecycle-manager/package-server-manager-metrics for network=default are: map[]\\\\nI1124 13:13:24.031184 6479 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 13:13:24.031220 6479 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 13:13:24.031236 6479 services_controller.go:445] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1124 13:13:24.030958 6479 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5x94h_openshift-ovn-kubernetes(528a4b22-34ba-41b8-8c7d-07d98eebe02b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b16be7938ff339aa0f92c4fb0d444114ac253439e80bf12ceb55c7e9a1944603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.915147 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdtxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bef2f789-12a8-45b6-b875-7307992176ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdtxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.917544 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.917597 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.917611 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.917626 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:45 crc kubenswrapper[4790]: I1124 13:13:45.917638 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:45Z","lastTransitionTime":"2025-11-24T13:13:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.020241 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.020340 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.020358 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.020414 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.020432 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:46Z","lastTransitionTime":"2025-11-24T13:13:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.123365 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.123425 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.123473 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.123495 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.123512 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:46Z","lastTransitionTime":"2025-11-24T13:13:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.226549 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.226606 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.226622 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.226648 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.226666 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:46Z","lastTransitionTime":"2025-11-24T13:13:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.313971 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:46 crc kubenswrapper[4790]: E1124 13:13:46.314196 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.329226 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.329272 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.329283 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.329300 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.329311 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:46Z","lastTransitionTime":"2025-11-24T13:13:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.432214 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.432260 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.432269 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.432284 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.432293 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:46Z","lastTransitionTime":"2025-11-24T13:13:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.535224 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.535260 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.535268 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.535283 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.535294 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:46Z","lastTransitionTime":"2025-11-24T13:13:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.637834 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.637877 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.637916 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.637935 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.637947 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:46Z","lastTransitionTime":"2025-11-24T13:13:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.740469 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.740522 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.740533 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.740546 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.740793 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:46Z","lastTransitionTime":"2025-11-24T13:13:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.844624 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.844735 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.844753 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.844782 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.844800 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:46Z","lastTransitionTime":"2025-11-24T13:13:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.947401 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.947446 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.947459 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.947477 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:46 crc kubenswrapper[4790]: I1124 13:13:46.947488 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:46Z","lastTransitionTime":"2025-11-24T13:13:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.050473 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.050526 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.050536 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.050556 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.050568 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:47Z","lastTransitionTime":"2025-11-24T13:13:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.154046 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.154128 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.154152 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.154180 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.154201 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:47Z","lastTransitionTime":"2025-11-24T13:13:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.256977 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.257030 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.257053 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.257073 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.257087 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:47Z","lastTransitionTime":"2025-11-24T13:13:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.314455 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.314541 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.314544 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:47 crc kubenswrapper[4790]: E1124 13:13:47.314687 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:13:47 crc kubenswrapper[4790]: E1124 13:13:47.314839 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:13:47 crc kubenswrapper[4790]: E1124 13:13:47.315015 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.360012 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.360084 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.360103 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.360131 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.360153 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:47Z","lastTransitionTime":"2025-11-24T13:13:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.463514 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.463583 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.463597 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.463617 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.463630 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:47Z","lastTransitionTime":"2025-11-24T13:13:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.566861 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.566933 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.566949 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.566967 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.566980 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:47Z","lastTransitionTime":"2025-11-24T13:13:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.670330 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.670393 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.670410 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.670433 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.670449 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:47Z","lastTransitionTime":"2025-11-24T13:13:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.774414 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.774479 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.774496 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.774519 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.774536 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:47Z","lastTransitionTime":"2025-11-24T13:13:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.877173 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.877220 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.877231 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.877246 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.877258 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:47Z","lastTransitionTime":"2025-11-24T13:13:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.980184 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.980305 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.980316 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.980329 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:47 crc kubenswrapper[4790]: I1124 13:13:47.980337 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:47Z","lastTransitionTime":"2025-11-24T13:13:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.083074 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.083139 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.083156 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.083182 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.083200 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:48Z","lastTransitionTime":"2025-11-24T13:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.185154 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.185207 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.185219 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.185237 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.185248 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:48Z","lastTransitionTime":"2025-11-24T13:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.288600 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.288681 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.288704 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.288732 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.288753 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:48Z","lastTransitionTime":"2025-11-24T13:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.314515 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:48 crc kubenswrapper[4790]: E1124 13:13:48.314703 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.391709 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.391760 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.391779 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.391799 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.391817 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:48Z","lastTransitionTime":"2025-11-24T13:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.493909 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.493945 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.493956 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.493970 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.493979 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:48Z","lastTransitionTime":"2025-11-24T13:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.596961 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.597001 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.597015 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.597032 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.597042 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:48Z","lastTransitionTime":"2025-11-24T13:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.699729 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.699788 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.699808 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.699837 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.699861 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:48Z","lastTransitionTime":"2025-11-24T13:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.803195 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.803272 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.803285 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.803316 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.803333 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:48Z","lastTransitionTime":"2025-11-24T13:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.905916 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.905956 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.905966 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.905979 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.905987 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:48Z","lastTransitionTime":"2025-11-24T13:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.909718 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.909751 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.909762 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.909775 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.909787 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:48Z","lastTransitionTime":"2025-11-24T13:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:48 crc kubenswrapper[4790]: E1124 13:13:48.922940 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.927191 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.927243 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.927269 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.927291 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.927306 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:48Z","lastTransitionTime":"2025-11-24T13:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:48 crc kubenswrapper[4790]: E1124 13:13:48.942148 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.946128 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.946180 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.946192 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.946207 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.946218 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:48Z","lastTransitionTime":"2025-11-24T13:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:48 crc kubenswrapper[4790]: E1124 13:13:48.961508 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.965200 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.965227 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.965235 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.965248 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.965256 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:48Z","lastTransitionTime":"2025-11-24T13:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:48 crc kubenswrapper[4790]: E1124 13:13:48.978840 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.981706 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.981734 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.981745 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.981760 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:48 crc kubenswrapper[4790]: I1124 13:13:48.981772 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:48Z","lastTransitionTime":"2025-11-24T13:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:48 crc kubenswrapper[4790]: E1124 13:13:48.993721 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:48 crc kubenswrapper[4790]: E1124 13:13:48.993829 4790 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.008279 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.008308 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.008318 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.008329 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.008340 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:49Z","lastTransitionTime":"2025-11-24T13:13:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.110669 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.110708 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.110720 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.110734 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.110744 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:49Z","lastTransitionTime":"2025-11-24T13:13:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.213643 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.213700 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.213716 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.213738 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.213761 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:49Z","lastTransitionTime":"2025-11-24T13:13:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.313874 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.313993 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:13:49 crc kubenswrapper[4790]: E1124 13:13:49.314044 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.314095 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:49 crc kubenswrapper[4790]: E1124 13:13:49.314224 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:13:49 crc kubenswrapper[4790]: E1124 13:13:49.314345 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.316616 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.316667 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.316678 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.316698 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.316710 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:49Z","lastTransitionTime":"2025-11-24T13:13:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.419741 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.419806 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.419820 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.419843 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.419859 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:49Z","lastTransitionTime":"2025-11-24T13:13:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.523509 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.524266 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.524296 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.524335 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.524362 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:49Z","lastTransitionTime":"2025-11-24T13:13:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.627969 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.628057 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.628068 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.628084 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.628095 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:49Z","lastTransitionTime":"2025-11-24T13:13:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.730990 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.731063 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.731073 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.731088 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.731098 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:49Z","lastTransitionTime":"2025-11-24T13:13:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.833990 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.834050 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.834067 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.834090 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.834106 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:49Z","lastTransitionTime":"2025-11-24T13:13:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.937246 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.937293 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.937305 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.937320 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:49 crc kubenswrapper[4790]: I1124 13:13:49.937344 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:49Z","lastTransitionTime":"2025-11-24T13:13:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.040352 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.040412 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.040426 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.040443 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.040455 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:50Z","lastTransitionTime":"2025-11-24T13:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.143688 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.143740 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.143753 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.143774 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.143791 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:50Z","lastTransitionTime":"2025-11-24T13:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.246206 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.246242 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.246252 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.246267 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.246278 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:50Z","lastTransitionTime":"2025-11-24T13:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.314513 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:50 crc kubenswrapper[4790]: E1124 13:13:50.314751 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.349147 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.349209 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.349223 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.349248 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.349263 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:50Z","lastTransitionTime":"2025-11-24T13:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.452440 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.452526 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.452561 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.452593 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.452621 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:50Z","lastTransitionTime":"2025-11-24T13:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.555430 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.555506 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.555524 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.555555 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.555576 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:50Z","lastTransitionTime":"2025-11-24T13:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.658562 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.658623 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.658635 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.658658 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.658673 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:50Z","lastTransitionTime":"2025-11-24T13:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.762753 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.762811 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.762824 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.762841 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.762855 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:50Z","lastTransitionTime":"2025-11-24T13:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.866674 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.866739 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.866758 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.866987 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.867007 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:50Z","lastTransitionTime":"2025-11-24T13:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.970039 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.970123 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.970144 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.970170 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:50 crc kubenswrapper[4790]: I1124 13:13:50.970194 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:50Z","lastTransitionTime":"2025-11-24T13:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.073054 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.073098 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.073132 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.073150 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.073166 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:51Z","lastTransitionTime":"2025-11-24T13:13:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.175620 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.175655 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.175666 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.175682 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.175694 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:51Z","lastTransitionTime":"2025-11-24T13:13:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.278276 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.278323 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.278331 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.278346 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.278355 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:51Z","lastTransitionTime":"2025-11-24T13:13:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.314185 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.314217 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.314447 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:51 crc kubenswrapper[4790]: E1124 13:13:51.314373 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:13:51 crc kubenswrapper[4790]: E1124 13:13:51.314693 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:13:51 crc kubenswrapper[4790]: E1124 13:13:51.314774 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.382237 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.382316 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.382358 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.382378 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.382391 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:51Z","lastTransitionTime":"2025-11-24T13:13:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.486095 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.486148 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.486158 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.486181 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.486196 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:51Z","lastTransitionTime":"2025-11-24T13:13:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.589108 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.589176 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.589189 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.589216 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.589231 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:51Z","lastTransitionTime":"2025-11-24T13:13:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.696587 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.696634 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.696645 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.696660 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.696672 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:51Z","lastTransitionTime":"2025-11-24T13:13:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.800042 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.800084 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.800093 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.800108 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.800117 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:51Z","lastTransitionTime":"2025-11-24T13:13:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.902264 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.902344 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.902384 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.902402 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:51 crc kubenswrapper[4790]: I1124 13:13:51.902413 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:51Z","lastTransitionTime":"2025-11-24T13:13:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.005179 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.005214 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.005223 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.005239 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.005250 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:52Z","lastTransitionTime":"2025-11-24T13:13:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.107858 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.107924 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.107935 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.107952 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.107963 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:52Z","lastTransitionTime":"2025-11-24T13:13:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.211217 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.211255 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.211265 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.211281 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.211290 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:52Z","lastTransitionTime":"2025-11-24T13:13:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.313610 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.313764 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.313791 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.313802 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:52 crc kubenswrapper[4790]: E1124 13:13:52.313796 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.313817 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.313830 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:52Z","lastTransitionTime":"2025-11-24T13:13:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.324437 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5e86ac0-955f-4455-a55e-db96696dbc80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c123986828c19da07e80adf0ff83f2d5858aa1681caf2f097164fa1067e6b247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee38955b1cc92f846dccf1fb30021559dfa4cefe62e4d095dcf983ac7b0adba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://103fb4a5a8c9d50f090bd5b73e841e6d6bd3a52817fa60d8681461c3b7658ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a59ab7ed8087587de107062d659abc98e07ce74ae4a8ed5e372bbeee9e6a8908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59ab7ed8087587de107062d659abc98e07ce74ae4a8ed5e372bbeee9e6a8908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.337144 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.352499 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.365742 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af0748fbdc734fe38245ac5426022cd063620ec033c5e4285073feb4a376019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:44Z\\\",\\\"message\\\":\\\"2025-11-24T13:12:58+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_2c499d3c-4900-45ee-b815-27ca113496c9\\\\n2025-11-24T13:12:58+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_2c499d3c-4900-45ee-b815-27ca113496c9 to /host/opt/cni/bin/\\\\n2025-11-24T13:12:59Z [verbose] multus-daemon started\\\\n2025-11-24T13:12:59Z [verbose] Readiness Indicator file check\\\\n2025-11-24T13:13:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.377215 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.391326 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3641c5a28dae87d2696ab209ab60e5bfadc8881ee96237f823df0f7ad8e841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.408474 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.415835 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.415907 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.415921 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.415936 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.415946 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:52Z","lastTransitionTime":"2025-11-24T13:13:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.420532 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.444403 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec06ff65e703be63a81e99901326ec1eaa20c2176ddbd0764cc7cdc0b5761bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a2bf135bf819ef8ce489337a9f6b7b0bb626ff335b37ec7e6a74e0e13a1ba40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f3d5e18b5f456b7627a9154ff8fb00945ec026d4fb8797d4c4f3e2a2fbb1db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad72e45458f88569f2bb77a55dc98e2bad29a7e451e5258d371722616042bebd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23e4305b0dc57da3ad6d92e8f04f1f18780c1ebaf63ee19140a8443b99197bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908f7697965f4e6d820bfe32f7f59960e917dedea04373ff0e2b8126bb835163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07157ab848e1a1ecc3a7d078f861c8e8ceb87e6ec88b7f52513a1b99ba209a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07157ab848e1a1ecc3a7d078f861c8e8ceb87e6ec88b7f52513a1b99ba209a15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:24Z\\\",\\\"message\\\":\\\"\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 13:13:24.030836 6479 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 13:13:24.030868 6479 lb_config.go:1031] Cluster endpoints for openshift-operator-lifecycle-manager/package-server-manager-metrics for network=default are: map[]\\\\nI1124 13:13:24.031184 6479 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 13:13:24.031220 6479 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 13:13:24.031236 6479 services_controller.go:445] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1124 13:13:24.030958 6479 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5x94h_openshift-ovn-kubernetes(528a4b22-34ba-41b8-8c7d-07d98eebe02b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b16be7938ff339aa0f92c4fb0d444114ac253439e80bf12ceb55c7e9a1944603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.457409 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fec678890b719f519b6740a344414413e17705e146440e130765a0f13b8a4407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.468613 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdtxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bef2f789-12a8-45b6-b875-7307992176ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdtxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.479203 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.492023 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.504222 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.513667 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.518675 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.518723 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.518737 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.518756 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.518769 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:52Z","lastTransitionTime":"2025-11-24T13:13:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.525603 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nlnv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0f7e0af-ed9a-4a5e-be2f-06abac132790\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ad38a0b670b452c1c5d9a287fb396d5e9eed494839d3bccac74dd5456ea3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wxh2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nlnv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.536485 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898d46ebaf15e8d33d02bea73cdb774bab2ea6001911a6dca0ae33c5d8d69e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfdc436c0635036c073c666b7acf24eb7d4d46af0a735122716d469ebf114a8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5ttn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.621983 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.622035 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.622046 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.622063 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.622074 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:52Z","lastTransitionTime":"2025-11-24T13:13:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.724486 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.724536 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.724552 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.724573 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.724588 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:52Z","lastTransitionTime":"2025-11-24T13:13:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.827277 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.827328 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.827339 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.827359 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.827371 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:52Z","lastTransitionTime":"2025-11-24T13:13:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.930077 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.930116 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.930126 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.930141 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:52 crc kubenswrapper[4790]: I1124 13:13:52.930151 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:52Z","lastTransitionTime":"2025-11-24T13:13:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.032689 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.032723 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.032733 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.032746 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.032755 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:53Z","lastTransitionTime":"2025-11-24T13:13:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.135972 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.136003 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.136011 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.136023 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.136032 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:53Z","lastTransitionTime":"2025-11-24T13:13:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.237686 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.237720 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.237728 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.237740 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.237749 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:53Z","lastTransitionTime":"2025-11-24T13:13:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.314103 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.314774 4790 scope.go:117] "RemoveContainer" containerID="07157ab848e1a1ecc3a7d078f861c8e8ceb87e6ec88b7f52513a1b99ba209a15" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.314949 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:13:53 crc kubenswrapper[4790]: E1124 13:13:53.315033 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.315118 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:13:53 crc kubenswrapper[4790]: E1124 13:13:53.315275 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:13:53 crc kubenswrapper[4790]: E1124 13:13:53.315483 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.340803 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.341046 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.341178 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.341314 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.341410 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:53Z","lastTransitionTime":"2025-11-24T13:13:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.447536 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.447627 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.447650 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.447685 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.447710 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:53Z","lastTransitionTime":"2025-11-24T13:13:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.550623 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.550658 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.550667 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.550680 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.550690 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:53Z","lastTransitionTime":"2025-11-24T13:13:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.653180 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.653223 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.653237 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.653254 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.653266 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:53Z","lastTransitionTime":"2025-11-24T13:13:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.737624 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5x94h_528a4b22-34ba-41b8-8c7d-07d98eebe02b/ovnkube-controller/2.log" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.739598 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" event={"ID":"528a4b22-34ba-41b8-8c7d-07d98eebe02b","Type":"ContainerStarted","Data":"92683d18aac71f52361dbb0877043c6ae0cf8452eff0df475873ab76bda9d6a9"} Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.739960 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.752078 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.755660 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.755693 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.755702 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.755716 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.755725 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:53Z","lastTransitionTime":"2025-11-24T13:13:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.764541 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.776848 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.786405 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.795292 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nlnv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0f7e0af-ed9a-4a5e-be2f-06abac132790\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ad38a0b670b452c1c5d9a287fb396d5e9eed494839d3bccac74dd5456ea3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wxh2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nlnv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.807095 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898d46ebaf15e8d33d02bea73cdb774bab2ea6001911a6dca0ae33c5d8d69e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfdc436c0635036c073c666b7acf24eb7d4d46af0a735122716d469ebf114a8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5ttn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.819647 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5e86ac0-955f-4455-a55e-db96696dbc80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c123986828c19da07e80adf0ff83f2d5858aa1681caf2f097164fa1067e6b247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee38955b1cc92f846dccf1fb30021559dfa4cefe62e4d095dcf983ac7b0adba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://103fb4a5a8c9d50f090bd5b73e841e6d6bd3a52817fa60d8681461c3b7658ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a59ab7ed8087587de107062d659abc98e07ce74ae4a8ed5e372bbeee9e6a8908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59ab7ed8087587de107062d659abc98e07ce74ae4a8ed5e372bbeee9e6a8908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.833861 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.849398 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.860859 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.860924 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.860937 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.860969 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.860981 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:53Z","lastTransitionTime":"2025-11-24T13:13:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.869108 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af0748fbdc734fe38245ac5426022cd063620ec033c5e4285073feb4a376019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:44Z\\\",\\\"message\\\":\\\"2025-11-24T13:12:58+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_2c499d3c-4900-45ee-b815-27ca113496c9\\\\n2025-11-24T13:12:58+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_2c499d3c-4900-45ee-b815-27ca113496c9 to /host/opt/cni/bin/\\\\n2025-11-24T13:12:59Z [verbose] multus-daemon started\\\\n2025-11-24T13:12:59Z [verbose] Readiness Indicator file check\\\\n2025-11-24T13:13:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.886513 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fec678890b719f519b6740a344414413e17705e146440e130765a0f13b8a4407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.901719 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.918609 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3641c5a28dae87d2696ab209ab60e5bfadc8881ee96237f823df0f7ad8e841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.932952 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.947101 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.963504 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.963542 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.963555 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.963570 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.963580 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:53Z","lastTransitionTime":"2025-11-24T13:13:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.969576 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec06ff65e703be63a81e99901326ec1eaa20c2176ddbd0764cc7cdc0b5761bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a2bf135bf819ef8ce489337a9f6b7b0bb626ff335b37ec7e6a74e0e13a1ba40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f3d5e18b5f456b7627a9154ff8fb00945ec026d4fb8797d4c4f3e2a2fbb1db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad72e45458f88569f2bb77a55dc98e2bad29a7e451e5258d371722616042bebd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23e4305b0dc57da3ad6d92e8f04f1f18780c1ebaf63ee19140a8443b99197bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908f7697965f4e6d820bfe32f7f59960e917dedea04373ff0e2b8126bb835163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92683d18aac71f52361dbb0877043c6ae0cf8452eff0df475873ab76bda9d6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07157ab848e1a1ecc3a7d078f861c8e8ceb87e6ec88b7f52513a1b99ba209a15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:24Z\\\",\\\"message\\\":\\\"\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 13:13:24.030836 6479 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 13:13:24.030868 6479 lb_config.go:1031] Cluster endpoints for openshift-operator-lifecycle-manager/package-server-manager-metrics for network=default are: map[]\\\\nI1124 13:13:24.031184 6479 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 13:13:24.031220 6479 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 13:13:24.031236 6479 services_controller.go:445] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1124 13:13:24.030958 6479 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b16be7938ff339aa0f92c4fb0d444114ac253439e80bf12ceb55c7e9a1944603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:53 crc kubenswrapper[4790]: I1124 13:13:53.981396 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdtxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bef2f789-12a8-45b6-b875-7307992176ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdtxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:54 crc kubenswrapper[4790]: I1124 13:13:54.066254 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:54 crc kubenswrapper[4790]: I1124 13:13:54.066304 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:54 crc kubenswrapper[4790]: I1124 13:13:54.066321 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:54 crc kubenswrapper[4790]: I1124 13:13:54.066341 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:54 crc kubenswrapper[4790]: I1124 13:13:54.066352 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:54Z","lastTransitionTime":"2025-11-24T13:13:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:54 crc kubenswrapper[4790]: I1124 13:13:54.314549 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:54 crc kubenswrapper[4790]: E1124 13:13:54.314681 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:13:54 crc kubenswrapper[4790]: I1124 13:13:54.733106 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:54 crc kubenswrapper[4790]: I1124 13:13:54.733156 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:54 crc kubenswrapper[4790]: I1124 13:13:54.733167 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:54 crc kubenswrapper[4790]: I1124 13:13:54.733182 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:54 crc kubenswrapper[4790]: I1124 13:13:54.733193 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:54Z","lastTransitionTime":"2025-11-24T13:13:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:54 crc kubenswrapper[4790]: I1124 13:13:54.835806 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:54 crc kubenswrapper[4790]: I1124 13:13:54.835842 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:54 crc kubenswrapper[4790]: I1124 13:13:54.835853 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:54 crc kubenswrapper[4790]: I1124 13:13:54.835870 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:54 crc kubenswrapper[4790]: I1124 13:13:54.835899 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:54Z","lastTransitionTime":"2025-11-24T13:13:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:54 crc kubenswrapper[4790]: I1124 13:13:54.938392 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:54 crc kubenswrapper[4790]: I1124 13:13:54.938432 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:54 crc kubenswrapper[4790]: I1124 13:13:54.938442 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:54 crc kubenswrapper[4790]: I1124 13:13:54.938458 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:54 crc kubenswrapper[4790]: I1124 13:13:54.938468 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:54Z","lastTransitionTime":"2025-11-24T13:13:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.040903 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.040946 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.040959 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.040976 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.040990 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:55Z","lastTransitionTime":"2025-11-24T13:13:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.081984 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:13:55 crc kubenswrapper[4790]: E1124 13:13:55.082192 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:59.082160344 +0000 UTC m=+147.462054016 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.143909 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.143970 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.143986 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.144006 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.144021 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:55Z","lastTransitionTime":"2025-11-24T13:13:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.183221 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.183274 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.183302 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.183331 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:13:55 crc kubenswrapper[4790]: E1124 13:13:55.183446 4790 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:13:55 crc kubenswrapper[4790]: E1124 13:13:55.183465 4790 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:13:55 crc kubenswrapper[4790]: E1124 13:13:55.183477 4790 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:13:55 crc kubenswrapper[4790]: E1124 13:13:55.183525 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 13:14:59.183509679 +0000 UTC m=+147.563403341 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:13:55 crc kubenswrapper[4790]: E1124 13:13:55.183700 4790 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:13:55 crc kubenswrapper[4790]: E1124 13:13:55.183731 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:14:59.183722535 +0000 UTC m=+147.563616197 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:13:55 crc kubenswrapper[4790]: E1124 13:13:55.183869 4790 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:13:55 crc kubenswrapper[4790]: E1124 13:13:55.183900 4790 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:13:55 crc kubenswrapper[4790]: E1124 13:13:55.183910 4790 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:13:55 crc kubenswrapper[4790]: E1124 13:13:55.183942 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 13:14:59.183931541 +0000 UTC m=+147.563825203 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:13:55 crc kubenswrapper[4790]: E1124 13:13:55.184117 4790 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:13:55 crc kubenswrapper[4790]: E1124 13:13:55.184224 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:14:59.184205089 +0000 UTC m=+147.564098801 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.246538 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.246574 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.246584 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.246598 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.246609 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:55Z","lastTransitionTime":"2025-11-24T13:13:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.313657 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.313700 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:13:55 crc kubenswrapper[4790]: E1124 13:13:55.313813 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.313967 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:13:55 crc kubenswrapper[4790]: E1124 13:13:55.313965 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:13:55 crc kubenswrapper[4790]: E1124 13:13:55.314278 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.349371 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.349402 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.349411 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.349424 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.349433 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:55Z","lastTransitionTime":"2025-11-24T13:13:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.453539 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.453607 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.453624 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.453652 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.453670 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:55Z","lastTransitionTime":"2025-11-24T13:13:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.557321 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.557418 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.557438 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.557465 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.557488 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:55Z","lastTransitionTime":"2025-11-24T13:13:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.661504 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.661584 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.661604 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.661640 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.661666 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:55Z","lastTransitionTime":"2025-11-24T13:13:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.749399 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5x94h_528a4b22-34ba-41b8-8c7d-07d98eebe02b/ovnkube-controller/3.log" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.750727 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5x94h_528a4b22-34ba-41b8-8c7d-07d98eebe02b/ovnkube-controller/2.log" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.754751 4790 generic.go:334] "Generic (PLEG): container finished" podID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerID="92683d18aac71f52361dbb0877043c6ae0cf8452eff0df475873ab76bda9d6a9" exitCode=1 Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.754829 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" event={"ID":"528a4b22-34ba-41b8-8c7d-07d98eebe02b","Type":"ContainerDied","Data":"92683d18aac71f52361dbb0877043c6ae0cf8452eff0df475873ab76bda9d6a9"} Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.754918 4790 scope.go:117] "RemoveContainer" containerID="07157ab848e1a1ecc3a7d078f861c8e8ceb87e6ec88b7f52513a1b99ba209a15" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.755931 4790 scope.go:117] "RemoveContainer" containerID="92683d18aac71f52361dbb0877043c6ae0cf8452eff0df475873ab76bda9d6a9" Nov 24 13:13:55 crc kubenswrapper[4790]: E1124 13:13:55.756149 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5x94h_openshift-ovn-kubernetes(528a4b22-34ba-41b8-8c7d-07d98eebe02b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.765734 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.765788 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.765799 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.765823 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.765850 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:55Z","lastTransitionTime":"2025-11-24T13:13:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.783089 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.798301 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.811449 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nlnv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0f7e0af-ed9a-4a5e-be2f-06abac132790\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ad38a0b670b452c1c5d9a287fb396d5e9eed494839d3bccac74dd5456ea3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wxh2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nlnv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.826977 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898d46ebaf15e8d33d02bea73cdb774bab2ea6001911a6dca0ae33c5d8d69e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfdc436c0635036c073c666b7acf24eb7d4d46af0a735122716d469ebf114a8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5ttn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.846661 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.863608 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.868606 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.868676 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.868694 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.868717 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.868729 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:55Z","lastTransitionTime":"2025-11-24T13:13:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.882444 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5e86ac0-955f-4455-a55e-db96696dbc80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c123986828c19da07e80adf0ff83f2d5858aa1681caf2f097164fa1067e6b247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee38955b1cc92f846dccf1fb30021559dfa4cefe62e4d095dcf983ac7b0adba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://103fb4a5a8c9d50f090bd5b73e841e6d6bd3a52817fa60d8681461c3b7658ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a59ab7ed8087587de107062d659abc98e07ce74ae4a8ed5e372bbeee9e6a8908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59ab7ed8087587de107062d659abc98e07ce74ae4a8ed5e372bbeee9e6a8908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.902617 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.919992 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.937122 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af0748fbdc734fe38245ac5426022cd063620ec033c5e4285073feb4a376019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:44Z\\\",\\\"message\\\":\\\"2025-11-24T13:12:58+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_2c499d3c-4900-45ee-b815-27ca113496c9\\\\n2025-11-24T13:12:58+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_2c499d3c-4900-45ee-b815-27ca113496c9 to /host/opt/cni/bin/\\\\n2025-11-24T13:12:59Z [verbose] multus-daemon started\\\\n2025-11-24T13:12:59Z [verbose] Readiness Indicator file check\\\\n2025-11-24T13:13:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.953383 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3641c5a28dae87d2696ab209ab60e5bfadc8881ee96237f823df0f7ad8e841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.969992 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.972015 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.972062 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.972075 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.972097 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.972110 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:55Z","lastTransitionTime":"2025-11-24T13:13:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:55 crc kubenswrapper[4790]: I1124 13:13:55.985923 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.009360 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec06ff65e703be63a81e99901326ec1eaa20c2176ddbd0764cc7cdc0b5761bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a2bf135bf819ef8ce489337a9f6b7b0bb626ff335b37ec7e6a74e0e13a1ba40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f3d5e18b5f456b7627a9154ff8fb00945ec026d4fb8797d4c4f3e2a2fbb1db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad72e45458f88569f2bb77a55dc98e2bad29a7e451e5258d371722616042bebd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23e4305b0dc57da3ad6d92e8f04f1f18780c1ebaf63ee19140a8443b99197bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908f7697965f4e6d820bfe32f7f59960e917dedea04373ff0e2b8126bb835163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92683d18aac71f52361dbb0877043c6ae0cf8452eff0df475873ab76bda9d6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07157ab848e1a1ecc3a7d078f861c8e8ceb87e6ec88b7f52513a1b99ba209a15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:24Z\\\",\\\"message\\\":\\\"\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 13:13:24.030836 6479 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 13:13:24.030868 6479 lb_config.go:1031] Cluster endpoints for openshift-operator-lifecycle-manager/package-server-manager-metrics for network=default are: map[]\\\\nI1124 13:13:24.031184 6479 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 13:13:24.031220 6479 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 13:13:24.031236 6479 services_controller.go:445] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1124 13:13:24.030958 6479 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92683d18aac71f52361dbb0877043c6ae0cf8452eff0df475873ab76bda9d6a9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:54Z\\\",\\\"message\\\":\\\"g(nil)\\\\nI1124 13:13:54.062211 6898 services_controller.go:444] Built service openshift-service-ca-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 13:13:54.062218 6898 services_controller.go:445] Built service openshift-machine-api/machine-api-controllers LB template configs for network=default: []services.lbConfig(nil)\\\\nI1124 13:13:54.062223 6898 services_controller.go:445] Built service openshift-service-ca-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1124 13:13:54.062224 6898 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:54Z is after 2025-08-24T17:21:41Z]\\\\nI1124 13:13:54.062244 6898 services_controller.go:451] Built service openshift-service-ca-oper\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b16be7938ff339aa0f92c4fb0d444114ac253439e80bf12ceb55c7e9a1944603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.038778 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fec678890b719f519b6740a344414413e17705e146440e130765a0f13b8a4407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.053622 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.068106 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdtxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bef2f789-12a8-45b6-b875-7307992176ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdtxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.075385 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.075439 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.075454 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.075479 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.075492 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:56Z","lastTransitionTime":"2025-11-24T13:13:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.177626 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.177662 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.177670 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.177684 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.177693 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:56Z","lastTransitionTime":"2025-11-24T13:13:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.280281 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.280364 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.280378 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.280399 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.280412 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:56Z","lastTransitionTime":"2025-11-24T13:13:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.313775 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:56 crc kubenswrapper[4790]: E1124 13:13:56.313976 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.382486 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.382529 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.382540 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.382555 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.382583 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:56Z","lastTransitionTime":"2025-11-24T13:13:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.485080 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.485393 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.485509 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.485615 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.485704 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:56Z","lastTransitionTime":"2025-11-24T13:13:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.588618 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.588650 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.588659 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.588672 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.588680 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:56Z","lastTransitionTime":"2025-11-24T13:13:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.691845 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.691894 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.691907 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.691922 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.691931 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:56Z","lastTransitionTime":"2025-11-24T13:13:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.758275 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5x94h_528a4b22-34ba-41b8-8c7d-07d98eebe02b/ovnkube-controller/3.log" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.794580 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.794626 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.794635 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.794649 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.794658 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:56Z","lastTransitionTime":"2025-11-24T13:13:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.896816 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.896850 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.896859 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.896871 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.896906 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:56Z","lastTransitionTime":"2025-11-24T13:13:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.999178 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.999221 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.999235 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.999255 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:56 crc kubenswrapper[4790]: I1124 13:13:56.999270 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:56Z","lastTransitionTime":"2025-11-24T13:13:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.101690 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.101954 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.102017 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.102085 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.102146 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:57Z","lastTransitionTime":"2025-11-24T13:13:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.204777 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.204820 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.204829 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.204843 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.204851 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:57Z","lastTransitionTime":"2025-11-24T13:13:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.307382 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.307413 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.307430 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.307448 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.307459 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:57Z","lastTransitionTime":"2025-11-24T13:13:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.313627 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.313702 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:13:57 crc kubenswrapper[4790]: E1124 13:13:57.313744 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.313775 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:57 crc kubenswrapper[4790]: E1124 13:13:57.313814 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:13:57 crc kubenswrapper[4790]: E1124 13:13:57.313925 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.409975 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.410013 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.410025 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.410040 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.410051 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:57Z","lastTransitionTime":"2025-11-24T13:13:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.512838 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.512900 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.512912 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.512928 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.512940 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:57Z","lastTransitionTime":"2025-11-24T13:13:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.616522 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.616574 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.616589 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.616608 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.616621 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:57Z","lastTransitionTime":"2025-11-24T13:13:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.718486 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.718547 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.718564 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.718581 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.718592 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:57Z","lastTransitionTime":"2025-11-24T13:13:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.821262 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.821328 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.821341 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.821358 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.821371 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:57Z","lastTransitionTime":"2025-11-24T13:13:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.923912 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.923959 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.923974 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.923992 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:57 crc kubenswrapper[4790]: I1124 13:13:57.924003 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:57Z","lastTransitionTime":"2025-11-24T13:13:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.025671 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.025719 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.025737 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.025757 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.025768 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:58Z","lastTransitionTime":"2025-11-24T13:13:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.128394 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.128440 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.128449 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.128464 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.128473 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:58Z","lastTransitionTime":"2025-11-24T13:13:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.230833 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.230899 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.230916 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.230933 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.230950 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:58Z","lastTransitionTime":"2025-11-24T13:13:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.314533 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:13:58 crc kubenswrapper[4790]: E1124 13:13:58.314682 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.333532 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.333584 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.333596 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.333611 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.333620 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:58Z","lastTransitionTime":"2025-11-24T13:13:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.435597 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.435638 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.435672 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.435690 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.435701 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:58Z","lastTransitionTime":"2025-11-24T13:13:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.538739 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.538789 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.538803 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.538819 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.538827 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:58Z","lastTransitionTime":"2025-11-24T13:13:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.641643 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.641681 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.641690 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.641704 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.641714 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:58Z","lastTransitionTime":"2025-11-24T13:13:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.743558 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.743588 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.743595 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.743607 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.743615 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:58Z","lastTransitionTime":"2025-11-24T13:13:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.845664 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.845693 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.845701 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.845714 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.845722 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:58Z","lastTransitionTime":"2025-11-24T13:13:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.948176 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.948215 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.948227 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.948242 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:58 crc kubenswrapper[4790]: I1124 13:13:58.948257 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:58Z","lastTransitionTime":"2025-11-24T13:13:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.044974 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.045205 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.045214 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.045230 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.045239 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:59Z","lastTransitionTime":"2025-11-24T13:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:59 crc kubenswrapper[4790]: E1124 13:13:59.056547 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.059424 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.059462 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.059471 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.059485 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.059496 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:59Z","lastTransitionTime":"2025-11-24T13:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:59 crc kubenswrapper[4790]: E1124 13:13:59.071361 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.074383 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.074414 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.074424 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.074442 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.074452 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:59Z","lastTransitionTime":"2025-11-24T13:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:59 crc kubenswrapper[4790]: E1124 13:13:59.088452 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.091738 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.091768 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.091778 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.091792 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.091801 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:59Z","lastTransitionTime":"2025-11-24T13:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:59 crc kubenswrapper[4790]: E1124 13:13:59.102274 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.105187 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.105225 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.105235 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.105251 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.105261 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:59Z","lastTransitionTime":"2025-11-24T13:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:59 crc kubenswrapper[4790]: E1124 13:13:59.119305 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:13:59 crc kubenswrapper[4790]: E1124 13:13:59.119505 4790 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.121904 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.121937 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.121945 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.121959 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.121968 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:59Z","lastTransitionTime":"2025-11-24T13:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.224530 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.224561 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.224569 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.224584 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.224593 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:59Z","lastTransitionTime":"2025-11-24T13:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.313708 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.313760 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.313823 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:13:59 crc kubenswrapper[4790]: E1124 13:13:59.313860 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:13:59 crc kubenswrapper[4790]: E1124 13:13:59.313985 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:13:59 crc kubenswrapper[4790]: E1124 13:13:59.314256 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.327268 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.327304 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.327315 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.327328 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.327337 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:59Z","lastTransitionTime":"2025-11-24T13:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.429848 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.429917 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.429926 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.429961 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.429973 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:59Z","lastTransitionTime":"2025-11-24T13:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.532034 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.532076 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.532085 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.532100 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.532110 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:59Z","lastTransitionTime":"2025-11-24T13:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.634879 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.634946 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.634959 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.634984 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.635008 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:59Z","lastTransitionTime":"2025-11-24T13:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.738575 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.738621 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.738633 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.738648 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.738658 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:59Z","lastTransitionTime":"2025-11-24T13:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.841492 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.841586 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.841612 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.841656 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.841687 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:59Z","lastTransitionTime":"2025-11-24T13:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.945404 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.945483 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.945499 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.945529 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:13:59 crc kubenswrapper[4790]: I1124 13:13:59.945553 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:13:59Z","lastTransitionTime":"2025-11-24T13:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.050391 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.050461 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.050482 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.050514 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.050540 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:00Z","lastTransitionTime":"2025-11-24T13:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.153794 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.153854 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.153873 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.153921 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.153938 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:00Z","lastTransitionTime":"2025-11-24T13:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.255858 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.255905 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.255914 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.255929 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.255937 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:00Z","lastTransitionTime":"2025-11-24T13:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.313874 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:14:00 crc kubenswrapper[4790]: E1124 13:14:00.314083 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.358909 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.358987 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.359001 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.359016 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.359025 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:00Z","lastTransitionTime":"2025-11-24T13:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.461683 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.461735 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.461744 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.461760 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.461774 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:00Z","lastTransitionTime":"2025-11-24T13:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.564313 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.564363 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.564379 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.564404 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.564420 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:00Z","lastTransitionTime":"2025-11-24T13:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.667806 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.667972 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.668000 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.668030 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.668052 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:00Z","lastTransitionTime":"2025-11-24T13:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.771213 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.771256 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.771265 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.771276 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.771284 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:00Z","lastTransitionTime":"2025-11-24T13:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.873960 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.874004 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.874017 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.874033 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.874044 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:00Z","lastTransitionTime":"2025-11-24T13:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.977203 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.977245 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.977256 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.977271 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:00 crc kubenswrapper[4790]: I1124 13:14:00.977279 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:00Z","lastTransitionTime":"2025-11-24T13:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.078994 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.079045 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.079056 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.079071 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.079081 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:01Z","lastTransitionTime":"2025-11-24T13:14:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.181721 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.181792 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.181830 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.181859 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.181934 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:01Z","lastTransitionTime":"2025-11-24T13:14:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.284353 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.284399 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.284409 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.284425 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.284435 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:01Z","lastTransitionTime":"2025-11-24T13:14:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.313990 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.314033 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.314034 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:01 crc kubenswrapper[4790]: E1124 13:14:01.314135 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:01 crc kubenswrapper[4790]: E1124 13:14:01.314243 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:01 crc kubenswrapper[4790]: E1124 13:14:01.314288 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.387333 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.387370 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.387379 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.387393 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.387401 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:01Z","lastTransitionTime":"2025-11-24T13:14:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.490015 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.490063 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.490072 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.490085 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.490095 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:01Z","lastTransitionTime":"2025-11-24T13:14:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.592425 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.592464 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.592490 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.592506 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.592515 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:01Z","lastTransitionTime":"2025-11-24T13:14:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.694730 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.694760 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.694769 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.694781 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.694792 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:01Z","lastTransitionTime":"2025-11-24T13:14:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.797250 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.797340 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.797364 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.797395 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.797419 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:01Z","lastTransitionTime":"2025-11-24T13:14:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.899988 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.900047 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.900078 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.900092 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:01 crc kubenswrapper[4790]: I1124 13:14:01.900101 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:01Z","lastTransitionTime":"2025-11-24T13:14:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.002524 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.002572 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.002580 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.002596 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.002606 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:02Z","lastTransitionTime":"2025-11-24T13:14:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.105241 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.105324 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.105341 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.105365 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.105382 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:02Z","lastTransitionTime":"2025-11-24T13:14:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.207278 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.207333 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.207345 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.207361 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.207372 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:02Z","lastTransitionTime":"2025-11-24T13:14:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.309414 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.309454 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.309464 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.309477 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.309487 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:02Z","lastTransitionTime":"2025-11-24T13:14:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.313969 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:14:02 crc kubenswrapper[4790]: E1124 13:14:02.314119 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.327866 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898d46ebaf15e8d33d02bea73cdb774bab2ea6001911a6dca0ae33c5d8d69e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfdc436c0635036c073c666b7acf24eb7d4d46af0a735122716d469ebf114a8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5ttn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.345166 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.356700 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.369578 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.381782 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.393739 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nlnv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0f7e0af-ed9a-4a5e-be2f-06abac132790\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ad38a0b670b452c1c5d9a287fb396d5e9eed494839d3bccac74dd5456ea3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wxh2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nlnv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.411583 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5e86ac0-955f-4455-a55e-db96696dbc80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c123986828c19da07e80adf0ff83f2d5858aa1681caf2f097164fa1067e6b247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee38955b1cc92f846dccf1fb30021559dfa4cefe62e4d095dcf983ac7b0adba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://103fb4a5a8c9d50f090bd5b73e841e6d6bd3a52817fa60d8681461c3b7658ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a59ab7ed8087587de107062d659abc98e07ce74ae4a8ed5e372bbeee9e6a8908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59ab7ed8087587de107062d659abc98e07ce74ae4a8ed5e372bbeee9e6a8908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.411956 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.411984 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.411995 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.412013 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.412025 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:02Z","lastTransitionTime":"2025-11-24T13:14:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.424627 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.446403 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.457790 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af0748fbdc734fe38245ac5426022cd063620ec033c5e4285073feb4a376019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:44Z\\\",\\\"message\\\":\\\"2025-11-24T13:12:58+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_2c499d3c-4900-45ee-b815-27ca113496c9\\\\n2025-11-24T13:12:58+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_2c499d3c-4900-45ee-b815-27ca113496c9 to /host/opt/cni/bin/\\\\n2025-11-24T13:12:59Z [verbose] multus-daemon started\\\\n2025-11-24T13:12:59Z [verbose] Readiness Indicator file check\\\\n2025-11-24T13:13:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.474013 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec06ff65e703be63a81e99901326ec1eaa20c2176ddbd0764cc7cdc0b5761bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a2bf135bf819ef8ce489337a9f6b7b0bb626ff335b37ec7e6a74e0e13a1ba40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f3d5e18b5f456b7627a9154ff8fb00945ec026d4fb8797d4c4f3e2a2fbb1db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad72e45458f88569f2bb77a55dc98e2bad29a7e451e5258d371722616042bebd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23e4305b0dc57da3ad6d92e8f04f1f18780c1ebaf63ee19140a8443b99197bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908f7697965f4e6d820bfe32f7f59960e917dedea04373ff0e2b8126bb835163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92683d18aac71f52361dbb0877043c6ae0cf8452eff0df475873ab76bda9d6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07157ab848e1a1ecc3a7d078f861c8e8ceb87e6ec88b7f52513a1b99ba209a15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:24Z\\\",\\\"message\\\":\\\"\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 13:13:24.030836 6479 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 13:13:24.030868 6479 lb_config.go:1031] Cluster endpoints for openshift-operator-lifecycle-manager/package-server-manager-metrics for network=default are: map[]\\\\nI1124 13:13:24.031184 6479 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 13:13:24.031220 6479 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 13:13:24.031236 6479 services_controller.go:445] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1124 13:13:24.030958 6479 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92683d18aac71f52361dbb0877043c6ae0cf8452eff0df475873ab76bda9d6a9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:54Z\\\",\\\"message\\\":\\\"g(nil)\\\\nI1124 13:13:54.062211 6898 services_controller.go:444] Built service openshift-service-ca-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 13:13:54.062218 6898 services_controller.go:445] Built service openshift-machine-api/machine-api-controllers LB template configs for network=default: []services.lbConfig(nil)\\\\nI1124 13:13:54.062223 6898 services_controller.go:445] Built service openshift-service-ca-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1124 13:13:54.062224 6898 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:54Z is after 2025-08-24T17:21:41Z]\\\\nI1124 13:13:54.062244 6898 services_controller.go:451] Built service openshift-service-ca-oper\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b16be7938ff339aa0f92c4fb0d444114ac253439e80bf12ceb55c7e9a1944603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.487024 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fec678890b719f519b6740a344414413e17705e146440e130765a0f13b8a4407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.497014 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.509999 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3641c5a28dae87d2696ab209ab60e5bfadc8881ee96237f823df0f7ad8e841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.513705 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.513726 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.513734 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.513747 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.513756 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:02Z","lastTransitionTime":"2025-11-24T13:14:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.522910 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.533589 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.542794 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdtxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bef2f789-12a8-45b6-b875-7307992176ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdtxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.616005 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.616034 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.616042 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.616054 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.616062 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:02Z","lastTransitionTime":"2025-11-24T13:14:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.718997 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.719041 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.719052 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.719067 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.719081 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:02Z","lastTransitionTime":"2025-11-24T13:14:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.822420 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.822495 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.822513 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.822537 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.822563 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:02Z","lastTransitionTime":"2025-11-24T13:14:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.925719 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.925776 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.925792 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.925815 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:02 crc kubenswrapper[4790]: I1124 13:14:02.925832 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:02Z","lastTransitionTime":"2025-11-24T13:14:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.028570 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.028614 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.028625 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.028641 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.028653 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:03Z","lastTransitionTime":"2025-11-24T13:14:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.131155 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.131194 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.131207 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.131224 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.131236 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:03Z","lastTransitionTime":"2025-11-24T13:14:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.234867 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.234923 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.234932 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.234945 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.234954 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:03Z","lastTransitionTime":"2025-11-24T13:14:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.314281 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.314333 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.314314 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:03 crc kubenswrapper[4790]: E1124 13:14:03.314445 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:03 crc kubenswrapper[4790]: E1124 13:14:03.314536 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:03 crc kubenswrapper[4790]: E1124 13:14:03.314638 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.337047 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.337079 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.337089 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.337104 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.337116 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:03Z","lastTransitionTime":"2025-11-24T13:14:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.439642 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.439706 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.439728 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.439758 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.439781 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:03Z","lastTransitionTime":"2025-11-24T13:14:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.541604 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.541646 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.541663 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.541681 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.541697 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:03Z","lastTransitionTime":"2025-11-24T13:14:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.644191 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.644263 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.644288 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.644317 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.644338 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:03Z","lastTransitionTime":"2025-11-24T13:14:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.746835 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.746863 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.746940 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.746956 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.746964 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:03Z","lastTransitionTime":"2025-11-24T13:14:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.849193 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.849238 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.849250 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.849266 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.849277 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:03Z","lastTransitionTime":"2025-11-24T13:14:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.952518 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.952567 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.952578 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.952594 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:03 crc kubenswrapper[4790]: I1124 13:14:03.952608 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:03Z","lastTransitionTime":"2025-11-24T13:14:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.054930 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.054969 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.054978 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.054991 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.055001 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:04Z","lastTransitionTime":"2025-11-24T13:14:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.157426 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.157468 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.157476 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.157491 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.157501 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:04Z","lastTransitionTime":"2025-11-24T13:14:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.259382 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.259431 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.259443 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.259462 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.259473 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:04Z","lastTransitionTime":"2025-11-24T13:14:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.314379 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:14:04 crc kubenswrapper[4790]: E1124 13:14:04.314529 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.361478 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.361514 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.361525 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.361541 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.361551 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:04Z","lastTransitionTime":"2025-11-24T13:14:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.464281 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.464334 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.464344 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.464359 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.464405 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:04Z","lastTransitionTime":"2025-11-24T13:14:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.566654 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.566698 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.566707 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.566721 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.566730 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:04Z","lastTransitionTime":"2025-11-24T13:14:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.669409 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.669449 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.669458 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.669475 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.669486 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:04Z","lastTransitionTime":"2025-11-24T13:14:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.771724 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.771768 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.771779 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.771797 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.771808 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:04Z","lastTransitionTime":"2025-11-24T13:14:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.874210 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.874283 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.874305 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.874327 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.874348 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:04Z","lastTransitionTime":"2025-11-24T13:14:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.976983 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.977013 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.977022 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.977034 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:04 crc kubenswrapper[4790]: I1124 13:14:04.977042 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:04Z","lastTransitionTime":"2025-11-24T13:14:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.080441 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.080496 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.080507 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.080523 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.080536 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:05Z","lastTransitionTime":"2025-11-24T13:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.183370 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.183421 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.183432 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.183451 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.183462 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:05Z","lastTransitionTime":"2025-11-24T13:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.286201 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.286238 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.286249 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.286264 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.286275 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:05Z","lastTransitionTime":"2025-11-24T13:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.313489 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.313599 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:05 crc kubenswrapper[4790]: E1124 13:14:05.313654 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.313701 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:05 crc kubenswrapper[4790]: E1124 13:14:05.313832 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:05 crc kubenswrapper[4790]: E1124 13:14:05.313964 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.325432 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.388691 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.388728 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.388736 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.388750 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.388758 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:05Z","lastTransitionTime":"2025-11-24T13:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.491369 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.491419 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.491429 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.491446 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.491459 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:05Z","lastTransitionTime":"2025-11-24T13:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.594203 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.594245 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.594254 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.594268 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.594277 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:05Z","lastTransitionTime":"2025-11-24T13:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.697051 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.697101 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.697117 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.697133 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.697143 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:05Z","lastTransitionTime":"2025-11-24T13:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.799066 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.799122 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.799130 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.799145 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.799155 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:05Z","lastTransitionTime":"2025-11-24T13:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.903310 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.903352 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.903363 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.903382 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:05 crc kubenswrapper[4790]: I1124 13:14:05.903393 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:05Z","lastTransitionTime":"2025-11-24T13:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.005959 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.006008 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.006019 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.006040 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.006053 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:06Z","lastTransitionTime":"2025-11-24T13:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.108450 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.108506 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.108524 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.108546 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.108562 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:06Z","lastTransitionTime":"2025-11-24T13:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.210915 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.210955 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.210963 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.210978 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.210988 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:06Z","lastTransitionTime":"2025-11-24T13:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.313502 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:14:06 crc kubenswrapper[4790]: E1124 13:14:06.313667 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.313530 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.313751 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.313777 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.313808 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.313836 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:06Z","lastTransitionTime":"2025-11-24T13:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.416595 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.416663 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.416673 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.416710 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.416723 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:06Z","lastTransitionTime":"2025-11-24T13:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.519960 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.520002 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.520012 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.520027 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.520038 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:06Z","lastTransitionTime":"2025-11-24T13:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.623943 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.623996 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.624005 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.624024 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.624034 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:06Z","lastTransitionTime":"2025-11-24T13:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.726262 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.726324 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.726343 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.726367 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.726384 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:06Z","lastTransitionTime":"2025-11-24T13:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.829368 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.829454 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.829475 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.829507 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.829563 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:06Z","lastTransitionTime":"2025-11-24T13:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.932311 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.932401 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.932414 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.932433 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:06 crc kubenswrapper[4790]: I1124 13:14:06.932447 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:06Z","lastTransitionTime":"2025-11-24T13:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.035732 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.035919 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.035947 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.035989 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.036023 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:07Z","lastTransitionTime":"2025-11-24T13:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.138846 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.138941 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.138961 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.138983 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.138996 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:07Z","lastTransitionTime":"2025-11-24T13:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.241221 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.241289 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.241299 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.241317 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.241332 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:07Z","lastTransitionTime":"2025-11-24T13:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.314105 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.314181 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.314104 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:07 crc kubenswrapper[4790]: E1124 13:14:07.314221 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:07 crc kubenswrapper[4790]: E1124 13:14:07.314308 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:07 crc kubenswrapper[4790]: E1124 13:14:07.314458 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.344248 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.344289 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.344299 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.344321 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.344331 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:07Z","lastTransitionTime":"2025-11-24T13:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.447285 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.447573 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.447582 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.447596 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.447606 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:07Z","lastTransitionTime":"2025-11-24T13:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.549366 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.549394 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.549402 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.549417 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.549430 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:07Z","lastTransitionTime":"2025-11-24T13:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.652493 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.652547 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.652558 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.652575 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.652587 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:07Z","lastTransitionTime":"2025-11-24T13:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.754994 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.755043 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.755054 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.755071 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.755085 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:07Z","lastTransitionTime":"2025-11-24T13:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.857758 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.857804 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.857818 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.857835 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.857849 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:07Z","lastTransitionTime":"2025-11-24T13:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.960247 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.960291 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.960302 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.960318 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:07 crc kubenswrapper[4790]: I1124 13:14:07.960330 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:07Z","lastTransitionTime":"2025-11-24T13:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.062935 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.062964 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.062972 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.062985 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.062993 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:08Z","lastTransitionTime":"2025-11-24T13:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.165375 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.165413 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.165421 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.165434 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.165443 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:08Z","lastTransitionTime":"2025-11-24T13:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.267606 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.267642 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.267651 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.267665 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.267675 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:08Z","lastTransitionTime":"2025-11-24T13:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.314368 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:14:08 crc kubenswrapper[4790]: E1124 13:14:08.314507 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.370046 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.370105 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.370120 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.370143 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.370159 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:08Z","lastTransitionTime":"2025-11-24T13:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.472113 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.472155 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.472167 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.472184 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.472195 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:08Z","lastTransitionTime":"2025-11-24T13:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.575048 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.575093 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.575109 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.575128 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.575140 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:08Z","lastTransitionTime":"2025-11-24T13:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.677652 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.677696 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.677709 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.677729 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.677746 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:08Z","lastTransitionTime":"2025-11-24T13:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.780363 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.780419 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.780438 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.780461 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.780478 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:08Z","lastTransitionTime":"2025-11-24T13:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.882809 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.882856 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.882871 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.882909 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.882922 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:08Z","lastTransitionTime":"2025-11-24T13:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.985607 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.985669 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.985691 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.985720 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:08 crc kubenswrapper[4790]: I1124 13:14:08.985737 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:08Z","lastTransitionTime":"2025-11-24T13:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.089081 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.089175 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.089200 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.089230 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.089251 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:09Z","lastTransitionTime":"2025-11-24T13:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.191511 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.191557 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.191572 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.191589 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.191603 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:09Z","lastTransitionTime":"2025-11-24T13:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.294221 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.294423 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.294441 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.294804 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.295034 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:09Z","lastTransitionTime":"2025-11-24T13:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.314536 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.314536 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:09 crc kubenswrapper[4790]: E1124 13:14:09.314723 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.314738 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:09 crc kubenswrapper[4790]: E1124 13:14:09.314876 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:09 crc kubenswrapper[4790]: E1124 13:14:09.314980 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.397632 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.397670 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.397680 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.397692 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.397700 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:09Z","lastTransitionTime":"2025-11-24T13:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.432523 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.432560 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.432572 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.432590 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.432601 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:09Z","lastTransitionTime":"2025-11-24T13:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:09 crc kubenswrapper[4790]: E1124 13:14:09.449233 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.453541 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.453579 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.453590 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.453606 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.453618 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:09Z","lastTransitionTime":"2025-11-24T13:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:09 crc kubenswrapper[4790]: E1124 13:14:09.467616 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.471569 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.471601 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.471610 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.471624 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.471634 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:09Z","lastTransitionTime":"2025-11-24T13:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:09 crc kubenswrapper[4790]: E1124 13:14:09.488483 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.492525 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.492571 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.492586 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.492606 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.492623 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:09Z","lastTransitionTime":"2025-11-24T13:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:09 crc kubenswrapper[4790]: E1124 13:14:09.509788 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.514537 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.514577 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.514589 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.514607 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.514620 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:09Z","lastTransitionTime":"2025-11-24T13:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:09 crc kubenswrapper[4790]: E1124 13:14:09.530937 4790 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb4d37b5-209f-4037-867f-c5986b938358\\\",\\\"systemUUID\\\":\\\"3913efcb-c2ad-4861-a3c8-ebb167356dfb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:09 crc kubenswrapper[4790]: E1124 13:14:09.531067 4790 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.532569 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.532599 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.532608 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.532621 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.532630 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:09Z","lastTransitionTime":"2025-11-24T13:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.635302 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.635346 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.635355 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.635369 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.635378 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:09Z","lastTransitionTime":"2025-11-24T13:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.738747 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.738807 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.738827 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.738856 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.738909 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:09Z","lastTransitionTime":"2025-11-24T13:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.841813 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.841904 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.841922 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.841947 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.841967 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:09Z","lastTransitionTime":"2025-11-24T13:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.945336 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.945408 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.945419 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.945436 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:09 crc kubenswrapper[4790]: I1124 13:14:09.945450 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:09Z","lastTransitionTime":"2025-11-24T13:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.049109 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.049164 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.049175 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.049205 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.049228 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:10Z","lastTransitionTime":"2025-11-24T13:14:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.152433 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.152475 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.152485 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.152502 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.152514 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:10Z","lastTransitionTime":"2025-11-24T13:14:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.256702 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.256773 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.256790 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.256814 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.256831 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:10Z","lastTransitionTime":"2025-11-24T13:14:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.314387 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:14:10 crc kubenswrapper[4790]: E1124 13:14:10.314710 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.315987 4790 scope.go:117] "RemoveContainer" containerID="92683d18aac71f52361dbb0877043c6ae0cf8452eff0df475873ab76bda9d6a9" Nov 24 13:14:10 crc kubenswrapper[4790]: E1124 13:14:10.316322 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5x94h_openshift-ovn-kubernetes(528a4b22-34ba-41b8-8c7d-07d98eebe02b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.331651 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5e86ac0-955f-4455-a55e-db96696dbc80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c123986828c19da07e80adf0ff83f2d5858aa1681caf2f097164fa1067e6b247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee38955b1cc92f846dccf1fb30021559dfa4cefe62e4d095dcf983ac7b0adba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://103fb4a5a8c9d50f090bd5b73e841e6d6bd3a52817fa60d8681461c3b7658ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a59ab7ed8087587de107062d659abc98e07ce74ae4a8ed5e372bbeee9e6a8908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59ab7ed8087587de107062d659abc98e07ce74ae4a8ed5e372bbeee9e6a8908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.346060 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e45e122ae942715678ad8ddb895bc011877005422fc4ef6257e24a273cefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.360687 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56ac3c24f4fd231740a69e0459cfd3c599bfe139078b83606e974fa812686985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.360869 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.360937 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.360956 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.360983 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.361003 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:10Z","lastTransitionTime":"2025-11-24T13:14:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.374286 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-76rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c0e3499-87ac-481f-a010-708a3a4a730f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af0748fbdc734fe38245ac5426022cd063620ec033c5e4285073feb4a376019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:44Z\\\",\\\"message\\\":\\\"2025-11-24T13:12:58+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_2c499d3c-4900-45ee-b815-27ca113496c9\\\\n2025-11-24T13:12:58+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_2c499d3c-4900-45ee-b815-27ca113496c9 to /host/opt/cni/bin/\\\\n2025-11-24T13:12:59Z [verbose] multus-daemon started\\\\n2025-11-24T13:12:59Z [verbose] Readiness Indicator file check\\\\n2025-11-24T13:13:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27whv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-76rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.389118 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a889e59-6d90-4c50-b161-1b9d2371b98d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cac56536f98781c80d45847f247b4032416ecb7cbd1f22ff93e17ee86912955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3febfee771e7c895da4b109e0c8c73fbc1942de721f81d537531bf11cff0f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3febfee771e7c895da4b109e0c8c73fbc1942de721f81d537531bf11cff0f2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.407323 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3641c5a28dae87d2696ab209ab60e5bfadc8881ee96237f823df0f7ad8e841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.427370 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.443394 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.467429 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.467494 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.467513 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.467542 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.467564 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:10Z","lastTransitionTime":"2025-11-24T13:14:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.472784 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec06ff65e703be63a81e99901326ec1eaa20c2176ddbd0764cc7cdc0b5761bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a2bf135bf819ef8ce489337a9f6b7b0bb626ff335b37ec7e6a74e0e13a1ba40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f3d5e18b5f456b7627a9154ff8fb00945ec026d4fb8797d4c4f3e2a2fbb1db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad72e45458f88569f2bb77a55dc98e2bad29a7e451e5258d371722616042bebd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23e4305b0dc57da3ad6d92e8f04f1f18780c1ebaf63ee19140a8443b99197bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908f7697965f4e6d820bfe32f7f59960e917dedea04373ff0e2b8126bb835163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92683d18aac71f52361dbb0877043c6ae0cf8452eff0df475873ab76bda9d6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92683d18aac71f52361dbb0877043c6ae0cf8452eff0df475873ab76bda9d6a9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:54Z\\\",\\\"message\\\":\\\"g(nil)\\\\nI1124 13:13:54.062211 6898 services_controller.go:444] Built service openshift-service-ca-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 13:13:54.062218 6898 services_controller.go:445] Built service openshift-machine-api/machine-api-controllers LB template configs for network=default: []services.lbConfig(nil)\\\\nI1124 13:13:54.062223 6898 services_controller.go:445] Built service openshift-service-ca-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1124 13:13:54.062224 6898 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:54Z is after 2025-08-24T17:21:41Z]\\\\nI1124 13:13:54.062244 6898 services_controller.go:451] Built service openshift-service-ca-oper\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5x94h_openshift-ovn-kubernetes(528a4b22-34ba-41b8-8c7d-07d98eebe02b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b16be7938ff339aa0f92c4fb0d444114ac253439e80bf12ceb55c7e9a1944603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.491316 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fec678890b719f519b6740a344414413e17705e146440e130765a0f13b8a4407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.505236 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.515655 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdtxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bef2f789-12a8-45b6-b875-7307992176ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdtxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.528201 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.541137 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.556476 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.567650 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2687d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"206d383c-05e9-47a3-a2e7-1535bf249e6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b9899145e0c86c643e3e148cd1b375942e8e44cd4c98649d63f701187181790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t6xp7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2687d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.569922 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.569962 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.569977 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.569999 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.570016 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:10Z","lastTransitionTime":"2025-11-24T13:14:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.578155 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nlnv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0f7e0af-ed9a-4a5e-be2f-06abac132790\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ad38a0b670b452c1c5d9a287fb396d5e9eed494839d3bccac74dd5456ea3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wxh2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nlnv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.589758 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898d46ebaf15e8d33d02bea73cdb774bab2ea6001911a6dca0ae33c5d8d69e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfdc436c0635036c073c666b7acf24eb7d4d46af0a735122716d469ebf114a8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5ttn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:10Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.673158 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.673200 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.673209 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.673227 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.673238 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:10Z","lastTransitionTime":"2025-11-24T13:14:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.777420 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.777497 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.777513 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.777539 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.777560 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:10Z","lastTransitionTime":"2025-11-24T13:14:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.880721 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.880795 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.880812 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.880845 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.880869 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:10Z","lastTransitionTime":"2025-11-24T13:14:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.983391 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.983431 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.983439 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.983465 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:10 crc kubenswrapper[4790]: I1124 13:14:10.983476 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:10Z","lastTransitionTime":"2025-11-24T13:14:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.087457 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.087529 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.087543 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.087566 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.087588 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:11Z","lastTransitionTime":"2025-11-24T13:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.190023 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.190111 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.190124 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.190144 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.190157 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:11Z","lastTransitionTime":"2025-11-24T13:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.293084 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.293143 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.293160 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.293183 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.293200 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:11Z","lastTransitionTime":"2025-11-24T13:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.314376 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.314424 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:11 crc kubenswrapper[4790]: E1124 13:14:11.314477 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.314537 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:11 crc kubenswrapper[4790]: E1124 13:14:11.314651 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:11 crc kubenswrapper[4790]: E1124 13:14:11.314791 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.395707 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.395747 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.395755 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.395771 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.395781 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:11Z","lastTransitionTime":"2025-11-24T13:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.497685 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.497734 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.497745 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.497765 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.497777 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:11Z","lastTransitionTime":"2025-11-24T13:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.600391 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.600458 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.600481 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.600509 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.600528 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:11Z","lastTransitionTime":"2025-11-24T13:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.703263 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.703307 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.703317 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.703333 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.703343 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:11Z","lastTransitionTime":"2025-11-24T13:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.804923 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.804966 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.804978 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.804993 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.805005 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:11Z","lastTransitionTime":"2025-11-24T13:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.908259 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.908317 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.908329 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.908347 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:11 crc kubenswrapper[4790]: I1124 13:14:11.908359 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:11Z","lastTransitionTime":"2025-11-24T13:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.011217 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.011287 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.011307 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.011333 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.011353 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:12Z","lastTransitionTime":"2025-11-24T13:14:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.113446 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.113496 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.113519 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.113566 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.113583 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:12Z","lastTransitionTime":"2025-11-24T13:14:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.217411 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.217507 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.217534 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.217573 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.217600 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:12Z","lastTransitionTime":"2025-11-24T13:14:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.314244 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:14:12 crc kubenswrapper[4790]: E1124 13:14:12.315398 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.320523 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.320560 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.320570 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.320584 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.320594 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:12Z","lastTransitionTime":"2025-11-24T13:14:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.332662 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:12Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.355051 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"528a4b22-34ba-41b8-8c7d-07d98eebe02b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec06ff65e703be63a81e99901326ec1eaa20c2176ddbd0764cc7cdc0b5761bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a2bf135bf819ef8ce489337a9f6b7b0bb626ff335b37ec7e6a74e0e13a1ba40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f3d5e18b5f456b7627a9154ff8fb00945ec026d4fb8797d4c4f3e2a2fbb1db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad72e45458f88569f2bb77a55dc98e2bad29a7e451e5258d371722616042bebd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23e4305b0dc57da3ad6d92e8f04f1f18780c1ebaf63ee19140a8443b99197bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908f7697965f4e6d820bfe32f7f59960e917dedea04373ff0e2b8126bb835163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92683d18aac71f52361dbb0877043c6ae0cf8452eff0df475873ab76bda9d6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92683d18aac71f52361dbb0877043c6ae0cf8452eff0df475873ab76bda9d6a9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:13:54Z\\\",\\\"message\\\":\\\"g(nil)\\\\nI1124 13:13:54.062211 6898 services_controller.go:444] Built service openshift-service-ca-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 13:13:54.062218 6898 services_controller.go:445] Built service openshift-machine-api/machine-api-controllers LB template configs for network=default: []services.lbConfig(nil)\\\\nI1124 13:13:54.062223 6898 services_controller.go:445] Built service openshift-service-ca-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1124 13:13:54.062224 6898 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:13:54Z is after 2025-08-24T17:21:41Z]\\\\nI1124 13:13:54.062244 6898 services_controller.go:451] Built service openshift-service-ca-oper\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5x94h_openshift-ovn-kubernetes(528a4b22-34ba-41b8-8c7d-07d98eebe02b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b16be7938ff339aa0f92c4fb0d444114ac253439e80bf12ceb55c7e9a1944603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bcnzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5x94h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:12Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.372093 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41874b85-3fa0-4d37-b5a0-ced5689c732f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fec678890b719f519b6740a344414413e17705e146440e130765a0f13b8a4407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e8c84b52c2d5679abe7eaee3fff29e6277c206a7ef91a3d92c8c96284b12385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae1f6085dc690a8e15a7f754eed9b58474c306d9df9d6822bf964e249277d142\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://424914ff880a5f64d7c13adb9982cc85710b57e6a5cb2a2cc8ca2a2dc10c0d2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7b839ad7aa3e4b01143eaf1a3426840a120a8ded3acd361956f3c4eb0fa30f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae45d5731a377488c79b6ad8ee45cd7dfcafbe06f5dbc020d0ad38b0fa1df7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a2c1d97843e12471047bc31f00c71318eb4366d2910fcab6be6fcbe38b8352a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:13:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:13:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5rsh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dzbhp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:12Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.385081 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d73b133-48f0-455f-8f6a-742e633f631a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98c3515b1c5a10c28bdddeddaf373163aace42e817de372797698160e736f6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wn84r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xz49t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:12Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.395092 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a889e59-6d90-4c50-b161-1b9d2371b98d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cac56536f98781c80d45847f247b4032416ecb7cbd1f22ff93e17ee86912955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3febfee771e7c895da4b109e0c8c73fbc1942de721f81d537531bf11cff0f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3febfee771e7c895da4b109e0c8c73fbc1942de721f81d537531bf11cff0f2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:12Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.408486 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12002956-4437-45d3-b654-edd88efa1c99\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22014c63f8830c00c4897323a9aa5a657d34d195cdf09081dadc553c0251d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://491788dfc55e85f98683bdd5644f5bdfbee75bca91db04755cefcb51e247eefe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c1f53d9446c08e2d57aab170ec714d167aa4726e406e8cd1519f158e191ffc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3641c5a28dae87d2696ab209ab60e5bfadc8881ee96237f823df0f7ad8e841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c359593e2ad4e75c25a9b3de92e639338765cb5a110bfebb7b4bd34d90a813\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"message\\\":\\\"file observer\\\\nW1124 13:12:51.883439 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:12:51.883578 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:12:51.884435 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-520768220/tls.crt::/tmp/serving-cert-520768220/tls.key\\\\\\\"\\\\nI1124 13:12:52.495524 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:12:52.500098 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:12:52.500123 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:12:52.500416 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:12:52.500430 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:12:52.507634 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:12:52.507661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507667 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:12:52.507674 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:12:52.507679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:12:52.507684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:12:52.507689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:12:52.507683 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:12:52.509804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0bf0718508c8db63833d5c3e03813ec887ba446ac037267e82afe8ac39d64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0ed6b38b073595ce10f69dc6f3b96aa19b8cd48b7d68035aad0b3bbbb915da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:12:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:12Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.419484 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d1a53b862b529cccc2551ca95d658d6d08953a3d577e6c430e848296e3ad1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f4a3ac94b289cce59e42b50e1164b990081dc9ca423ff766355894f754b583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:12Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.426141 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.426180 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.426188 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.426203 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.426211 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:12Z","lastTransitionTime":"2025-11-24T13:14:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.429946 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdtxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bef2f789-12a8-45b6-b875-7307992176ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qf9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdtxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:12Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.439327 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nlnv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0f7e0af-ed9a-4a5e-be2f-06abac132790\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ad38a0b670b452c1c5d9a287fb396d5e9eed494839d3bccac74dd5456ea3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wxh2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nlnv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:12Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.449096 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea0741f1-1abe-49ba-a186-5c4a3ebec5b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:13:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898d46ebaf15e8d33d02bea73cdb774bab2ea6001911a6dca0ae33c5d8d69e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfdc436c0635036c073c666b7acf24eb7d4d46af0a735122716d469ebf114a8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:13:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7vc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:13:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5ttn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:12Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.459554 4790 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5b1d2c-ce24-42b7-b0f0-aa18f469612e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b8eca7f94b5ad53b28e3a68c11ee838f72056b389cb42d92ac48f6538fa6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c7698aae1c29b4eb46789b6c5a27725d39c5f818296780650a3377c6b0f978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02c00a15943e6b5a18bff5daa43b5076763ffed9664932bc2d1004d7db5ae9f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a62b6ef704c8900651cbe52908c2e9c3ee0f50dbd972ba6113853a13c5f619b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:12:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:12:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:12Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.523614 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=44.523595051 podStartE2EDuration="44.523595051s" podCreationTimestamp="2025-11-24 13:13:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:12.522987213 +0000 UTC m=+100.902880885" watchObservedRunningTime="2025-11-24 13:14:12.523595051 +0000 UTC m=+100.903488713" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.523749 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-2687d" podStartSLOduration=76.523744615 podStartE2EDuration="1m16.523744615s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:12.502288158 +0000 UTC m=+100.882181820" watchObservedRunningTime="2025-11-24 13:14:12.523744615 +0000 UTC m=+100.903638277" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.528159 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.528197 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.528207 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.528225 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.528237 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:12Z","lastTransitionTime":"2025-11-24T13:14:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.580529 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-76rkg" podStartSLOduration=76.580511328 podStartE2EDuration="1m16.580511328s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:12.579925961 +0000 UTC m=+100.959819633" watchObservedRunningTime="2025-11-24 13:14:12.580511328 +0000 UTC m=+100.960404990" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.630378 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.630406 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.630417 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.630432 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.630441 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:12Z","lastTransitionTime":"2025-11-24T13:14:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.732914 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.732951 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.732959 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.732973 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.732983 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:12Z","lastTransitionTime":"2025-11-24T13:14:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.835704 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.835757 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.835773 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.835795 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.835811 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:12Z","lastTransitionTime":"2025-11-24T13:14:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.938200 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.938241 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.938250 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.938266 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:12 crc kubenswrapper[4790]: I1124 13:14:12.938275 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:12Z","lastTransitionTime":"2025-11-24T13:14:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.046955 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.047233 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.047254 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.047285 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.047307 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:13Z","lastTransitionTime":"2025-11-24T13:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.151153 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.151216 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.151231 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.151250 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.151268 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:13Z","lastTransitionTime":"2025-11-24T13:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.253665 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.253710 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.253721 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.253734 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.253745 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:13Z","lastTransitionTime":"2025-11-24T13:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.313474 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.313541 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.313582 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:13 crc kubenswrapper[4790]: E1124 13:14:13.313609 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:13 crc kubenswrapper[4790]: E1124 13:14:13.313824 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:13 crc kubenswrapper[4790]: E1124 13:14:13.313957 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.332323 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.355541 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.355611 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.355623 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.355636 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.355646 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:13Z","lastTransitionTime":"2025-11-24T13:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.458366 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.458403 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.458413 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.458427 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.458436 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:13Z","lastTransitionTime":"2025-11-24T13:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.560982 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.561022 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.561035 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.561051 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.561061 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:13Z","lastTransitionTime":"2025-11-24T13:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.663768 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.663821 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.663846 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.663866 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.663908 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:13Z","lastTransitionTime":"2025-11-24T13:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.766449 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.766495 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.766526 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.766547 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.766559 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:13Z","lastTransitionTime":"2025-11-24T13:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.869606 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.869658 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.869673 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.869694 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.869707 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:13Z","lastTransitionTime":"2025-11-24T13:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.972466 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.972527 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.972540 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.972556 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:13 crc kubenswrapper[4790]: I1124 13:14:13.972567 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:13Z","lastTransitionTime":"2025-11-24T13:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.074929 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.074976 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.074996 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.075014 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.075027 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:14Z","lastTransitionTime":"2025-11-24T13:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.178661 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.178707 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.178719 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.178736 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.178747 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:14Z","lastTransitionTime":"2025-11-24T13:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.281952 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.282009 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.282026 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.282047 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.282062 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:14Z","lastTransitionTime":"2025-11-24T13:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.315288 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:14:14 crc kubenswrapper[4790]: E1124 13:14:14.315459 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.385048 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.385977 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.386005 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.386033 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.386051 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:14Z","lastTransitionTime":"2025-11-24T13:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.489263 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.489314 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.489330 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.489350 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.489366 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:14Z","lastTransitionTime":"2025-11-24T13:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.575234 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bef2f789-12a8-45b6-b875-7307992176ee-metrics-certs\") pod \"network-metrics-daemon-vdtxl\" (UID: \"bef2f789-12a8-45b6-b875-7307992176ee\") " pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:14:14 crc kubenswrapper[4790]: E1124 13:14:14.575356 4790 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:14:14 crc kubenswrapper[4790]: E1124 13:14:14.575428 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bef2f789-12a8-45b6-b875-7307992176ee-metrics-certs podName:bef2f789-12a8-45b6-b875-7307992176ee nodeName:}" failed. No retries permitted until 2025-11-24 13:15:18.575410994 +0000 UTC m=+166.955304676 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bef2f789-12a8-45b6-b875-7307992176ee-metrics-certs") pod "network-metrics-daemon-vdtxl" (UID: "bef2f789-12a8-45b6-b875-7307992176ee") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.592853 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.593008 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.593038 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.593067 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.593083 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:14Z","lastTransitionTime":"2025-11-24T13:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.695411 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.695455 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.695470 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.695492 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.695507 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:14Z","lastTransitionTime":"2025-11-24T13:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.797727 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.797762 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.797777 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.797796 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.797807 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:14Z","lastTransitionTime":"2025-11-24T13:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.899742 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.899783 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.899802 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.899854 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:14 crc kubenswrapper[4790]: I1124 13:14:14.899865 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:14Z","lastTransitionTime":"2025-11-24T13:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.002701 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.002752 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.002765 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.002781 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.002794 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:15Z","lastTransitionTime":"2025-11-24T13:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.105830 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.106028 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.106049 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.106083 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.106105 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:15Z","lastTransitionTime":"2025-11-24T13:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.217373 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.217462 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.217477 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.217502 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.217519 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:15Z","lastTransitionTime":"2025-11-24T13:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.314029 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.314095 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.314102 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:15 crc kubenswrapper[4790]: E1124 13:14:15.314193 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:15 crc kubenswrapper[4790]: E1124 13:14:15.314298 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:15 crc kubenswrapper[4790]: E1124 13:14:15.314398 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.319407 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.319518 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.319586 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.319663 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.319730 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:15Z","lastTransitionTime":"2025-11-24T13:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.422033 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.422089 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.422106 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.422129 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.422148 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:15Z","lastTransitionTime":"2025-11-24T13:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.525084 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.525138 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.525154 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.525177 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.525195 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:15Z","lastTransitionTime":"2025-11-24T13:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.627417 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.627485 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.627504 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.627528 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.627544 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:15Z","lastTransitionTime":"2025-11-24T13:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.730619 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.730684 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.730704 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.730728 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.730747 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:15Z","lastTransitionTime":"2025-11-24T13:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.833938 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.834002 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.834027 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.834053 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.834071 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:15Z","lastTransitionTime":"2025-11-24T13:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.936810 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.936875 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.936914 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.936934 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:15 crc kubenswrapper[4790]: I1124 13:14:15.936945 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:15Z","lastTransitionTime":"2025-11-24T13:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.038736 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.038800 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.038811 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.038832 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.038848 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:16Z","lastTransitionTime":"2025-11-24T13:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.141595 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.141661 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.141679 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.141706 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.141727 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:16Z","lastTransitionTime":"2025-11-24T13:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.244264 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.244300 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.244308 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.244322 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.244331 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:16Z","lastTransitionTime":"2025-11-24T13:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.314153 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:14:16 crc kubenswrapper[4790]: E1124 13:14:16.314315 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.347125 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.347170 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.347183 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.347201 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.347213 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:16Z","lastTransitionTime":"2025-11-24T13:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.450325 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.450374 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.450388 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.450408 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.450458 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:16Z","lastTransitionTime":"2025-11-24T13:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.553492 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.553560 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.553571 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.553586 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.553598 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:16Z","lastTransitionTime":"2025-11-24T13:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.656397 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.656437 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.656448 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.656464 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.656475 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:16Z","lastTransitionTime":"2025-11-24T13:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.758646 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.758744 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.758776 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.758808 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.758826 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:16Z","lastTransitionTime":"2025-11-24T13:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.861329 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.861444 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.861461 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.861479 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.861489 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:16Z","lastTransitionTime":"2025-11-24T13:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.963529 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.963578 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.963590 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.963609 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:16 crc kubenswrapper[4790]: I1124 13:14:16.963622 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:16Z","lastTransitionTime":"2025-11-24T13:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.065772 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.065819 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.065828 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.065901 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.065918 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:17Z","lastTransitionTime":"2025-11-24T13:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.168449 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.168523 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.168538 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.168556 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.168570 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:17Z","lastTransitionTime":"2025-11-24T13:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.271423 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.271492 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.271510 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.271536 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.271558 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:17Z","lastTransitionTime":"2025-11-24T13:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.314183 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.314256 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:17 crc kubenswrapper[4790]: E1124 13:14:17.314327 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.314275 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:17 crc kubenswrapper[4790]: E1124 13:14:17.314398 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:17 crc kubenswrapper[4790]: E1124 13:14:17.314503 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.374363 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.374422 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.374438 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.374460 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.374476 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:17Z","lastTransitionTime":"2025-11-24T13:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.478086 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.478123 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.478131 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.478143 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.478152 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:17Z","lastTransitionTime":"2025-11-24T13:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.581363 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.581410 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.581421 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.581441 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.581455 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:17Z","lastTransitionTime":"2025-11-24T13:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.684435 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.684471 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.684483 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.684497 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.684507 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:17Z","lastTransitionTime":"2025-11-24T13:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.787977 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.788032 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.788044 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.788061 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.788379 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:17Z","lastTransitionTime":"2025-11-24T13:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.891413 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.891753 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.891867 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.891993 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.892098 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:17Z","lastTransitionTime":"2025-11-24T13:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.995303 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.995335 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.995345 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.995359 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:17 crc kubenswrapper[4790]: I1124 13:14:17.995369 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:17Z","lastTransitionTime":"2025-11-24T13:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.097639 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.097669 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.097680 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.097693 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.097702 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:18Z","lastTransitionTime":"2025-11-24T13:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.200275 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.200307 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.200315 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.200329 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.200338 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:18Z","lastTransitionTime":"2025-11-24T13:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.302956 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.303008 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.303022 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.303043 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.303058 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:18Z","lastTransitionTime":"2025-11-24T13:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.313592 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:14:18 crc kubenswrapper[4790]: E1124 13:14:18.313760 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.405984 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.406306 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.406411 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.406519 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.406627 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:18Z","lastTransitionTime":"2025-11-24T13:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.509836 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.509914 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.509929 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.509948 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.509959 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:18Z","lastTransitionTime":"2025-11-24T13:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.613150 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.613192 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.613202 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.613216 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.613226 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:18Z","lastTransitionTime":"2025-11-24T13:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.715496 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.715750 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.715955 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.716073 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.716178 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:18Z","lastTransitionTime":"2025-11-24T13:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.818593 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.818631 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.818641 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.818656 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.818664 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:18Z","lastTransitionTime":"2025-11-24T13:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.921142 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.921206 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.921220 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.921234 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:18 crc kubenswrapper[4790]: I1124 13:14:18.921244 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:18Z","lastTransitionTime":"2025-11-24T13:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.023674 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.023709 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.023717 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.023731 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.023742 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:19Z","lastTransitionTime":"2025-11-24T13:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.126607 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.126648 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.126659 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.126675 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.126686 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:19Z","lastTransitionTime":"2025-11-24T13:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.230392 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.230452 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.230464 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.230482 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.230493 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:19Z","lastTransitionTime":"2025-11-24T13:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.313461 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.313509 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.313583 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:19 crc kubenswrapper[4790]: E1124 13:14:19.313686 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:19 crc kubenswrapper[4790]: E1124 13:14:19.313954 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:19 crc kubenswrapper[4790]: E1124 13:14:19.314004 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.333111 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.333171 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.333189 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.333212 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.333231 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:19Z","lastTransitionTime":"2025-11-24T13:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.437307 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.437363 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.437381 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.437404 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.437424 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:19Z","lastTransitionTime":"2025-11-24T13:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.539407 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.539454 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.539465 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.539481 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.539492 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:19Z","lastTransitionTime":"2025-11-24T13:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.642867 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.642924 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.642938 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.642955 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.642966 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:19Z","lastTransitionTime":"2025-11-24T13:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.744972 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.745015 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.745027 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.745044 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.745055 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:19Z","lastTransitionTime":"2025-11-24T13:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.763625 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.763673 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.763699 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.763719 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.763737 4790 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:19Z","lastTransitionTime":"2025-11-24T13:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.803806 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-727gb"] Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.804207 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-727gb" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.808275 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.808483 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.810085 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.810513 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.822929 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-nlnv5" podStartSLOduration=83.822909846 podStartE2EDuration="1m23.822909846s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:19.822871554 +0000 UTC m=+108.202765246" watchObservedRunningTime="2025-11-24 13:14:19.822909846 +0000 UTC m=+108.202803518" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.837799 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5ttn" podStartSLOduration=83.837780533 podStartE2EDuration="1m23.837780533s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:19.8373115 +0000 UTC m=+108.217205152" watchObservedRunningTime="2025-11-24 13:14:19.837780533 +0000 UTC m=+108.217674205" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.880601 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=88.880581064 podStartE2EDuration="1m28.880581064s" podCreationTimestamp="2025-11-24 13:12:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:19.864481411 +0000 UTC m=+108.244375083" watchObservedRunningTime="2025-11-24 13:14:19.880581064 +0000 UTC m=+108.260474736" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.896544 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=86.896523302 podStartE2EDuration="1m26.896523302s" podCreationTimestamp="2025-11-24 13:12:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:19.880754579 +0000 UTC m=+108.260648251" watchObservedRunningTime="2025-11-24 13:14:19.896523302 +0000 UTC m=+108.276416974" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.925002 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/4196554d-6688-43fd-aaee-f6128b12c1ad-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-727gb\" (UID: \"4196554d-6688-43fd-aaee-f6128b12c1ad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-727gb" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.925056 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4196554d-6688-43fd-aaee-f6128b12c1ad-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-727gb\" (UID: \"4196554d-6688-43fd-aaee-f6128b12c1ad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-727gb" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.925102 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4196554d-6688-43fd-aaee-f6128b12c1ad-service-ca\") pod \"cluster-version-operator-5c965bbfc6-727gb\" (UID: \"4196554d-6688-43fd-aaee-f6128b12c1ad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-727gb" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.925144 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/4196554d-6688-43fd-aaee-f6128b12c1ad-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-727gb\" (UID: \"4196554d-6688-43fd-aaee-f6128b12c1ad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-727gb" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.925169 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4196554d-6688-43fd-aaee-f6128b12c1ad-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-727gb\" (UID: \"4196554d-6688-43fd-aaee-f6128b12c1ad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-727gb" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.951170 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-dzbhp" podStartSLOduration=83.951153263 podStartE2EDuration="1m23.951153263s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:19.950374411 +0000 UTC m=+108.330268083" watchObservedRunningTime="2025-11-24 13:14:19.951153263 +0000 UTC m=+108.331046925" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.962686 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podStartSLOduration=83.962669745 podStartE2EDuration="1m23.962669745s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:19.962354506 +0000 UTC m=+108.342248158" watchObservedRunningTime="2025-11-24 13:14:19.962669745 +0000 UTC m=+108.342563427" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.973000 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=14.972981741 podStartE2EDuration="14.972981741s" podCreationTimestamp="2025-11-24 13:14:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:19.971825288 +0000 UTC m=+108.351718970" watchObservedRunningTime="2025-11-24 13:14:19.972981741 +0000 UTC m=+108.352875413" Nov 24 13:14:19 crc kubenswrapper[4790]: I1124 13:14:19.996020 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=6.996005033 podStartE2EDuration="6.996005033s" podCreationTimestamp="2025-11-24 13:14:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:19.99486021 +0000 UTC m=+108.374753882" watchObservedRunningTime="2025-11-24 13:14:19.996005033 +0000 UTC m=+108.375898695" Nov 24 13:14:20 crc kubenswrapper[4790]: I1124 13:14:20.025667 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/4196554d-6688-43fd-aaee-f6128b12c1ad-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-727gb\" (UID: \"4196554d-6688-43fd-aaee-f6128b12c1ad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-727gb" Nov 24 13:14:20 crc kubenswrapper[4790]: I1124 13:14:20.025713 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4196554d-6688-43fd-aaee-f6128b12c1ad-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-727gb\" (UID: \"4196554d-6688-43fd-aaee-f6128b12c1ad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-727gb" Nov 24 13:14:20 crc kubenswrapper[4790]: I1124 13:14:20.025749 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/4196554d-6688-43fd-aaee-f6128b12c1ad-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-727gb\" (UID: \"4196554d-6688-43fd-aaee-f6128b12c1ad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-727gb" Nov 24 13:14:20 crc kubenswrapper[4790]: I1124 13:14:20.025773 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4196554d-6688-43fd-aaee-f6128b12c1ad-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-727gb\" (UID: \"4196554d-6688-43fd-aaee-f6128b12c1ad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-727gb" Nov 24 13:14:20 crc kubenswrapper[4790]: I1124 13:14:20.025773 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/4196554d-6688-43fd-aaee-f6128b12c1ad-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-727gb\" (UID: \"4196554d-6688-43fd-aaee-f6128b12c1ad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-727gb" Nov 24 13:14:20 crc kubenswrapper[4790]: I1124 13:14:20.025810 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4196554d-6688-43fd-aaee-f6128b12c1ad-service-ca\") pod \"cluster-version-operator-5c965bbfc6-727gb\" (UID: \"4196554d-6688-43fd-aaee-f6128b12c1ad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-727gb" Nov 24 13:14:20 crc kubenswrapper[4790]: I1124 13:14:20.025851 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/4196554d-6688-43fd-aaee-f6128b12c1ad-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-727gb\" (UID: \"4196554d-6688-43fd-aaee-f6128b12c1ad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-727gb" Nov 24 13:14:20 crc kubenswrapper[4790]: I1124 13:14:20.026662 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4196554d-6688-43fd-aaee-f6128b12c1ad-service-ca\") pod \"cluster-version-operator-5c965bbfc6-727gb\" (UID: \"4196554d-6688-43fd-aaee-f6128b12c1ad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-727gb" Nov 24 13:14:20 crc kubenswrapper[4790]: I1124 13:14:20.039339 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4196554d-6688-43fd-aaee-f6128b12c1ad-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-727gb\" (UID: \"4196554d-6688-43fd-aaee-f6128b12c1ad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-727gb" Nov 24 13:14:20 crc kubenswrapper[4790]: I1124 13:14:20.046638 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4196554d-6688-43fd-aaee-f6128b12c1ad-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-727gb\" (UID: \"4196554d-6688-43fd-aaee-f6128b12c1ad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-727gb" Nov 24 13:14:20 crc kubenswrapper[4790]: I1124 13:14:20.120435 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-727gb" Nov 24 13:14:20 crc kubenswrapper[4790]: W1124 13:14:20.133578 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4196554d_6688_43fd_aaee_f6128b12c1ad.slice/crio-6b4fa19ff85d20ca4f2df5f0944b420ccfda42990f2193ee53f0c94cd12fe8ca WatchSource:0}: Error finding container 6b4fa19ff85d20ca4f2df5f0944b420ccfda42990f2193ee53f0c94cd12fe8ca: Status 404 returned error can't find the container with id 6b4fa19ff85d20ca4f2df5f0944b420ccfda42990f2193ee53f0c94cd12fe8ca Nov 24 13:14:20 crc kubenswrapper[4790]: I1124 13:14:20.314553 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:14:20 crc kubenswrapper[4790]: E1124 13:14:20.314987 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:14:20 crc kubenswrapper[4790]: I1124 13:14:20.831730 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-727gb" event={"ID":"4196554d-6688-43fd-aaee-f6128b12c1ad","Type":"ContainerStarted","Data":"2777560bd09c5f641cd465cdb5d8f4504f34be281183603865a1ae5b2aa0b0a3"} Nov 24 13:14:20 crc kubenswrapper[4790]: I1124 13:14:20.831809 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-727gb" event={"ID":"4196554d-6688-43fd-aaee-f6128b12c1ad","Type":"ContainerStarted","Data":"6b4fa19ff85d20ca4f2df5f0944b420ccfda42990f2193ee53f0c94cd12fe8ca"} Nov 24 13:14:20 crc kubenswrapper[4790]: I1124 13:14:20.846945 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-727gb" podStartSLOduration=84.846927002 podStartE2EDuration="1m24.846927002s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:20.846587292 +0000 UTC m=+109.226480954" watchObservedRunningTime="2025-11-24 13:14:20.846927002 +0000 UTC m=+109.226820664" Nov 24 13:14:21 crc kubenswrapper[4790]: I1124 13:14:21.313550 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:21 crc kubenswrapper[4790]: E1124 13:14:21.313775 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:21 crc kubenswrapper[4790]: I1124 13:14:21.313595 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:21 crc kubenswrapper[4790]: I1124 13:14:21.313563 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:21 crc kubenswrapper[4790]: E1124 13:14:21.313935 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:21 crc kubenswrapper[4790]: E1124 13:14:21.314023 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:22 crc kubenswrapper[4790]: I1124 13:14:22.314275 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:14:22 crc kubenswrapper[4790]: E1124 13:14:22.315090 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:14:23 crc kubenswrapper[4790]: I1124 13:14:23.314362 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:23 crc kubenswrapper[4790]: I1124 13:14:23.314481 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:23 crc kubenswrapper[4790]: I1124 13:14:23.314593 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:23 crc kubenswrapper[4790]: E1124 13:14:23.315033 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:23 crc kubenswrapper[4790]: E1124 13:14:23.315176 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:23 crc kubenswrapper[4790]: E1124 13:14:23.315405 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:23 crc kubenswrapper[4790]: I1124 13:14:23.315600 4790 scope.go:117] "RemoveContainer" containerID="92683d18aac71f52361dbb0877043c6ae0cf8452eff0df475873ab76bda9d6a9" Nov 24 13:14:23 crc kubenswrapper[4790]: E1124 13:14:23.315847 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5x94h_openshift-ovn-kubernetes(528a4b22-34ba-41b8-8c7d-07d98eebe02b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" Nov 24 13:14:24 crc kubenswrapper[4790]: I1124 13:14:24.313529 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:14:24 crc kubenswrapper[4790]: E1124 13:14:24.313644 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:14:25 crc kubenswrapper[4790]: I1124 13:14:25.314173 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:25 crc kubenswrapper[4790]: E1124 13:14:25.314279 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:25 crc kubenswrapper[4790]: I1124 13:14:25.314195 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:25 crc kubenswrapper[4790]: E1124 13:14:25.314337 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:25 crc kubenswrapper[4790]: I1124 13:14:25.314173 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:25 crc kubenswrapper[4790]: E1124 13:14:25.314382 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:26 crc kubenswrapper[4790]: I1124 13:14:26.314350 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:14:26 crc kubenswrapper[4790]: E1124 13:14:26.315187 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:14:27 crc kubenswrapper[4790]: I1124 13:14:27.313902 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:27 crc kubenswrapper[4790]: I1124 13:14:27.313987 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:27 crc kubenswrapper[4790]: E1124 13:14:27.314100 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:27 crc kubenswrapper[4790]: I1124 13:14:27.313912 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:27 crc kubenswrapper[4790]: E1124 13:14:27.314318 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:27 crc kubenswrapper[4790]: E1124 13:14:27.314426 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:28 crc kubenswrapper[4790]: I1124 13:14:28.313719 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:14:28 crc kubenswrapper[4790]: E1124 13:14:28.313901 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:14:29 crc kubenswrapper[4790]: I1124 13:14:29.314003 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:29 crc kubenswrapper[4790]: E1124 13:14:29.314577 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:29 crc kubenswrapper[4790]: I1124 13:14:29.314250 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:29 crc kubenswrapper[4790]: E1124 13:14:29.314768 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:29 crc kubenswrapper[4790]: I1124 13:14:29.314108 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:29 crc kubenswrapper[4790]: E1124 13:14:29.314963 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:30 crc kubenswrapper[4790]: I1124 13:14:30.314517 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:14:30 crc kubenswrapper[4790]: E1124 13:14:30.314707 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:14:30 crc kubenswrapper[4790]: I1124 13:14:30.861548 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-76rkg_6c0e3499-87ac-481f-a010-708a3a4a730f/kube-multus/1.log" Nov 24 13:14:30 crc kubenswrapper[4790]: I1124 13:14:30.862221 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-76rkg_6c0e3499-87ac-481f-a010-708a3a4a730f/kube-multus/0.log" Nov 24 13:14:30 crc kubenswrapper[4790]: I1124 13:14:30.862299 4790 generic.go:334] "Generic (PLEG): container finished" podID="6c0e3499-87ac-481f-a010-708a3a4a730f" containerID="6af0748fbdc734fe38245ac5426022cd063620ec033c5e4285073feb4a376019" exitCode=1 Nov 24 13:14:30 crc kubenswrapper[4790]: I1124 13:14:30.862352 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-76rkg" event={"ID":"6c0e3499-87ac-481f-a010-708a3a4a730f","Type":"ContainerDied","Data":"6af0748fbdc734fe38245ac5426022cd063620ec033c5e4285073feb4a376019"} Nov 24 13:14:30 crc kubenswrapper[4790]: I1124 13:14:30.862403 4790 scope.go:117] "RemoveContainer" containerID="6b30196aac3004609571ca66271468935c3ce39ae06852ef50d9e69a2d6c7d09" Nov 24 13:14:30 crc kubenswrapper[4790]: I1124 13:14:30.864835 4790 scope.go:117] "RemoveContainer" containerID="6af0748fbdc734fe38245ac5426022cd063620ec033c5e4285073feb4a376019" Nov 24 13:14:30 crc kubenswrapper[4790]: E1124 13:14:30.865558 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-76rkg_openshift-multus(6c0e3499-87ac-481f-a010-708a3a4a730f)\"" pod="openshift-multus/multus-76rkg" podUID="6c0e3499-87ac-481f-a010-708a3a4a730f" Nov 24 13:14:31 crc kubenswrapper[4790]: I1124 13:14:31.313954 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:31 crc kubenswrapper[4790]: I1124 13:14:31.313997 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:31 crc kubenswrapper[4790]: E1124 13:14:31.314081 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:31 crc kubenswrapper[4790]: I1124 13:14:31.314153 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:31 crc kubenswrapper[4790]: E1124 13:14:31.314309 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:31 crc kubenswrapper[4790]: E1124 13:14:31.314370 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:31 crc kubenswrapper[4790]: I1124 13:14:31.867265 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-76rkg_6c0e3499-87ac-481f-a010-708a3a4a730f/kube-multus/1.log" Nov 24 13:14:32 crc kubenswrapper[4790]: E1124 13:14:32.301118 4790 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 24 13:14:32 crc kubenswrapper[4790]: I1124 13:14:32.314636 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:14:32 crc kubenswrapper[4790]: E1124 13:14:32.315547 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:14:32 crc kubenswrapper[4790]: E1124 13:14:32.405533 4790 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 13:14:33 crc kubenswrapper[4790]: I1124 13:14:33.313766 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:33 crc kubenswrapper[4790]: I1124 13:14:33.313798 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:33 crc kubenswrapper[4790]: E1124 13:14:33.313943 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:33 crc kubenswrapper[4790]: I1124 13:14:33.313965 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:33 crc kubenswrapper[4790]: E1124 13:14:33.314064 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:33 crc kubenswrapper[4790]: E1124 13:14:33.314145 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:34 crc kubenswrapper[4790]: I1124 13:14:34.313732 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:14:34 crc kubenswrapper[4790]: E1124 13:14:34.313931 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:14:35 crc kubenswrapper[4790]: I1124 13:14:35.313541 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:35 crc kubenswrapper[4790]: E1124 13:14:35.313667 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:35 crc kubenswrapper[4790]: I1124 13:14:35.313560 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:35 crc kubenswrapper[4790]: E1124 13:14:35.313872 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:35 crc kubenswrapper[4790]: I1124 13:14:35.314933 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:35 crc kubenswrapper[4790]: E1124 13:14:35.315311 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:36 crc kubenswrapper[4790]: I1124 13:14:36.314188 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:14:36 crc kubenswrapper[4790]: E1124 13:14:36.314486 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:14:37 crc kubenswrapper[4790]: I1124 13:14:37.314098 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:37 crc kubenswrapper[4790]: I1124 13:14:37.314176 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:37 crc kubenswrapper[4790]: I1124 13:14:37.314246 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:37 crc kubenswrapper[4790]: E1124 13:14:37.314244 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:37 crc kubenswrapper[4790]: E1124 13:14:37.314355 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:37 crc kubenswrapper[4790]: E1124 13:14:37.314402 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:37 crc kubenswrapper[4790]: I1124 13:14:37.315006 4790 scope.go:117] "RemoveContainer" containerID="92683d18aac71f52361dbb0877043c6ae0cf8452eff0df475873ab76bda9d6a9" Nov 24 13:14:37 crc kubenswrapper[4790]: E1124 13:14:37.407114 4790 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 13:14:37 crc kubenswrapper[4790]: I1124 13:14:37.886835 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5x94h_528a4b22-34ba-41b8-8c7d-07d98eebe02b/ovnkube-controller/3.log" Nov 24 13:14:37 crc kubenswrapper[4790]: I1124 13:14:37.889486 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" event={"ID":"528a4b22-34ba-41b8-8c7d-07d98eebe02b","Type":"ContainerStarted","Data":"fbeaf8db5bfb90ff3667f7182b5506b392454f0f0fa6a32223ed0a832e1a86d1"} Nov 24 13:14:37 crc kubenswrapper[4790]: I1124 13:14:37.889841 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:14:37 crc kubenswrapper[4790]: I1124 13:14:37.915932 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" podStartSLOduration=101.915919137 podStartE2EDuration="1m41.915919137s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:37.914571858 +0000 UTC m=+126.294465520" watchObservedRunningTime="2025-11-24 13:14:37.915919137 +0000 UTC m=+126.295812799" Nov 24 13:14:38 crc kubenswrapper[4790]: I1124 13:14:38.012316 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-vdtxl"] Nov 24 13:14:38 crc kubenswrapper[4790]: I1124 13:14:38.012424 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:14:38 crc kubenswrapper[4790]: E1124 13:14:38.012497 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:14:39 crc kubenswrapper[4790]: I1124 13:14:39.313766 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:39 crc kubenswrapper[4790]: I1124 13:14:39.313816 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:14:39 crc kubenswrapper[4790]: I1124 13:14:39.313827 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:39 crc kubenswrapper[4790]: E1124 13:14:39.313949 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:39 crc kubenswrapper[4790]: E1124 13:14:39.314040 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:14:39 crc kubenswrapper[4790]: E1124 13:14:39.314133 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:39 crc kubenswrapper[4790]: I1124 13:14:39.314961 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:39 crc kubenswrapper[4790]: E1124 13:14:39.315056 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:41 crc kubenswrapper[4790]: I1124 13:14:41.313592 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:41 crc kubenswrapper[4790]: E1124 13:14:41.314273 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:41 crc kubenswrapper[4790]: I1124 13:14:41.313641 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:41 crc kubenswrapper[4790]: E1124 13:14:41.314464 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:41 crc kubenswrapper[4790]: I1124 13:14:41.313614 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:14:41 crc kubenswrapper[4790]: E1124 13:14:41.314639 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:14:41 crc kubenswrapper[4790]: I1124 13:14:41.313714 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:41 crc kubenswrapper[4790]: E1124 13:14:41.314893 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:42 crc kubenswrapper[4790]: I1124 13:14:42.315752 4790 scope.go:117] "RemoveContainer" containerID="6af0748fbdc734fe38245ac5426022cd063620ec033c5e4285073feb4a376019" Nov 24 13:14:42 crc kubenswrapper[4790]: E1124 13:14:42.408716 4790 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 13:14:42 crc kubenswrapper[4790]: I1124 13:14:42.910972 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-76rkg_6c0e3499-87ac-481f-a010-708a3a4a730f/kube-multus/1.log" Nov 24 13:14:42 crc kubenswrapper[4790]: I1124 13:14:42.911060 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-76rkg" event={"ID":"6c0e3499-87ac-481f-a010-708a3a4a730f","Type":"ContainerStarted","Data":"24025d94118c5c9d2f93c0792f7c151a99d9f5b9e3b328284faad63d2f3f2a23"} Nov 24 13:14:43 crc kubenswrapper[4790]: I1124 13:14:43.314020 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:43 crc kubenswrapper[4790]: I1124 13:14:43.314173 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:14:43 crc kubenswrapper[4790]: I1124 13:14:43.314426 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:43 crc kubenswrapper[4790]: I1124 13:14:43.314469 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:43 crc kubenswrapper[4790]: E1124 13:14:43.314595 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:43 crc kubenswrapper[4790]: E1124 13:14:43.314687 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:43 crc kubenswrapper[4790]: E1124 13:14:43.314732 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:43 crc kubenswrapper[4790]: E1124 13:14:43.314991 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:14:45 crc kubenswrapper[4790]: I1124 13:14:45.313602 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:45 crc kubenswrapper[4790]: E1124 13:14:45.313731 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:45 crc kubenswrapper[4790]: I1124 13:14:45.313971 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:45 crc kubenswrapper[4790]: I1124 13:14:45.313484 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:14:45 crc kubenswrapper[4790]: E1124 13:14:45.314105 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:45 crc kubenswrapper[4790]: I1124 13:14:45.314331 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:45 crc kubenswrapper[4790]: E1124 13:14:45.314312 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:14:45 crc kubenswrapper[4790]: E1124 13:14:45.314818 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:47 crc kubenswrapper[4790]: I1124 13:14:47.313772 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:14:47 crc kubenswrapper[4790]: I1124 13:14:47.313853 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:47 crc kubenswrapper[4790]: E1124 13:14:47.313913 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdtxl" podUID="bef2f789-12a8-45b6-b875-7307992176ee" Nov 24 13:14:47 crc kubenswrapper[4790]: I1124 13:14:47.314001 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:47 crc kubenswrapper[4790]: I1124 13:14:47.314226 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:47 crc kubenswrapper[4790]: E1124 13:14:47.314232 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:47 crc kubenswrapper[4790]: E1124 13:14:47.314268 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:47 crc kubenswrapper[4790]: E1124 13:14:47.314316 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:49 crc kubenswrapper[4790]: I1124 13:14:49.313926 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:49 crc kubenswrapper[4790]: I1124 13:14:49.313981 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:49 crc kubenswrapper[4790]: I1124 13:14:49.314206 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:49 crc kubenswrapper[4790]: I1124 13:14:49.314278 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:14:49 crc kubenswrapper[4790]: I1124 13:14:49.316849 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 24 13:14:49 crc kubenswrapper[4790]: I1124 13:14:49.316914 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 24 13:14:49 crc kubenswrapper[4790]: I1124 13:14:49.316943 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 24 13:14:49 crc kubenswrapper[4790]: I1124 13:14:49.317303 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 24 13:14:49 crc kubenswrapper[4790]: I1124 13:14:49.317846 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 24 13:14:49 crc kubenswrapper[4790]: I1124 13:14:49.318340 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.577443 4790 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.627914 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-slkpm"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.628561 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-slkpm" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.630851 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.631242 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.631439 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.631618 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.631813 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.631994 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.632131 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-hsvx8"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.632547 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wnlxq"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.632847 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wnlxq" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.633208 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsvx8" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.633494 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f55f4"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.634326 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f55f4" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.637831 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.638357 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.639243 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qwnbb"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.646014 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-cjdhk"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.646851 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-cjdhk" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.647777 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qwnbb" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.648530 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8jkhv"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.662937 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-5hlqm"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.663436 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5hlqm" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.663494 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.664552 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-hkbxm"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.664683 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.664758 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.664832 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.664968 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.665040 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.665159 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.665253 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.665299 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.665344 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-hkbxm" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.665395 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.665448 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.665247 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-t5fjs"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.665632 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.665687 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.665908 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.665991 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.666056 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.666055 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-t5fjs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.666200 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.666286 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.666320 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.666442 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.666486 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.666600 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.666601 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.666671 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.666832 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.669687 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-djs6p"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.670573 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6ljl5"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.671083 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6ljl5" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.672409 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-djs6p" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.672630 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-zvzkq"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.680018 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.680185 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.680308 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.680437 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.680486 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.680495 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.680508 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.680544 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.680549 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.680575 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.680601 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.680601 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.680630 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.680643 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.680658 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.680679 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.680708 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.681694 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.681910 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.683337 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-7d488"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.683730 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-cppvr"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.683926 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-zvzkq" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.683978 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7d488" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.685388 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-96268"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.685686 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ktdmx"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.686133 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qzt4s"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.686472 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qzt4s" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.686776 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-cppvr" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.687066 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-96268" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.687304 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-ktdmx" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.691321 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-z2dld"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.691982 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-z2dld" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.695546 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-4ldk4"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.696106 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-zxkb7"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.696427 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-thd68"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.696588 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.696604 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-4ldk4" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.697725 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.701928 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c4qnk"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.702350 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5lzlv"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.707922 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-xcw79"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.709598 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xcw79" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.710114 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c4qnk" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.714120 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/17718a81-aa70-4f5a-90bd-06e9251f4ef8-images\") pod \"machine-api-operator-5694c8668f-slkpm\" (UID: \"17718a81-aa70-4f5a-90bd-06e9251f4ef8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-slkpm" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.714179 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/69940b0d-4afc-4a7f-8ff7-cb128ce58913-client-ca\") pod \"controller-manager-879f6c89f-cjdhk\" (UID: \"69940b0d-4afc-4a7f-8ff7-cb128ce58913\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cjdhk" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.714218 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/1e6bf7c7-7711-43cb-8817-73b42174b9ee-machine-approver-tls\") pod \"machine-approver-56656f9798-5hlqm\" (UID: \"1e6bf7c7-7711-43cb-8817-73b42174b9ee\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5hlqm" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.714250 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mk45\" (UniqueName: \"kubernetes.io/projected/1e6bf7c7-7711-43cb-8817-73b42174b9ee-kube-api-access-4mk45\") pod \"machine-approver-56656f9798-5hlqm\" (UID: \"1e6bf7c7-7711-43cb-8817-73b42174b9ee\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5hlqm" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.714273 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a2b45ef-e51f-4ad2-a1b7-6d8e7b0c089a-serving-cert\") pod \"openshift-config-operator-7777fb866f-hsvx8\" (UID: \"2a2b45ef-e51f-4ad2-a1b7-6d8e7b0c089a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsvx8" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.714303 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5b226a93-5333-405f-93ad-788f9376faac-encryption-config\") pod \"apiserver-7bbb656c7d-dc2xs\" (UID: \"5b226a93-5333-405f-93ad-788f9376faac\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.714331 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pc7f8\" (UniqueName: \"kubernetes.io/projected/2a2b45ef-e51f-4ad2-a1b7-6d8e7b0c089a-kube-api-access-pc7f8\") pod \"openshift-config-operator-7777fb866f-hsvx8\" (UID: \"2a2b45ef-e51f-4ad2-a1b7-6d8e7b0c089a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsvx8" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.714358 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69940b0d-4afc-4a7f-8ff7-cb128ce58913-config\") pod \"controller-manager-879f6c89f-cjdhk\" (UID: \"69940b0d-4afc-4a7f-8ff7-cb128ce58913\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cjdhk" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.714558 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/66c28684-107f-4208-ae79-4b22a4e485dd-etcd-serving-ca\") pod \"apiserver-76f77b778f-8jkhv\" (UID: \"66c28684-107f-4208-ae79-4b22a4e485dd\") " pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.714611 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/194e75e8-4f01-4862-ba6e-a073a369df5a-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-f55f4\" (UID: \"194e75e8-4f01-4862-ba6e-a073a369df5a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f55f4" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.714648 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f50b116f-b25b-4985-8975-fbb523e1ddf6-config\") pod \"route-controller-manager-6576b87f9c-wnlxq\" (UID: \"f50b116f-b25b-4985-8975-fbb523e1ddf6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wnlxq" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.714688 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/66c28684-107f-4208-ae79-4b22a4e485dd-audit\") pod \"apiserver-76f77b778f-8jkhv\" (UID: \"66c28684-107f-4208-ae79-4b22a4e485dd\") " pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.714732 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5b226a93-5333-405f-93ad-788f9376faac-audit-dir\") pod \"apiserver-7bbb656c7d-dc2xs\" (UID: \"5b226a93-5333-405f-93ad-788f9376faac\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.714769 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dz26\" (UniqueName: \"kubernetes.io/projected/17718a81-aa70-4f5a-90bd-06e9251f4ef8-kube-api-access-9dz26\") pod \"machine-api-operator-5694c8668f-slkpm\" (UID: \"17718a81-aa70-4f5a-90bd-06e9251f4ef8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-slkpm" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.714809 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1e6bf7c7-7711-43cb-8817-73b42174b9ee-auth-proxy-config\") pod \"machine-approver-56656f9798-5hlqm\" (UID: \"1e6bf7c7-7711-43cb-8817-73b42174b9ee\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5hlqm" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.714864 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5lzlv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.718227 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5b226a93-5333-405f-93ad-788f9376faac-serving-cert\") pod \"apiserver-7bbb656c7d-dc2xs\" (UID: \"5b226a93-5333-405f-93ad-788f9376faac\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.718276 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/66c28684-107f-4208-ae79-4b22a4e485dd-image-import-ca\") pod \"apiserver-76f77b778f-8jkhv\" (UID: \"66c28684-107f-4208-ae79-4b22a4e485dd\") " pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.718301 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5b226a93-5333-405f-93ad-788f9376faac-etcd-client\") pod \"apiserver-7bbb656c7d-dc2xs\" (UID: \"5b226a93-5333-405f-93ad-788f9376faac\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.718358 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xflx6\" (UniqueName: \"kubernetes.io/projected/66c28684-107f-4208-ae79-4b22a4e485dd-kube-api-access-xflx6\") pod \"apiserver-76f77b778f-8jkhv\" (UID: \"66c28684-107f-4208-ae79-4b22a4e485dd\") " pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.718379 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e6bf7c7-7711-43cb-8817-73b42174b9ee-config\") pod \"machine-approver-56656f9798-5hlqm\" (UID: \"1e6bf7c7-7711-43cb-8817-73b42174b9ee\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5hlqm" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.718404 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/2a2b45ef-e51f-4ad2-a1b7-6d8e7b0c089a-available-featuregates\") pod \"openshift-config-operator-7777fb866f-hsvx8\" (UID: \"2a2b45ef-e51f-4ad2-a1b7-6d8e7b0c089a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsvx8" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.718434 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpt4f\" (UniqueName: \"kubernetes.io/projected/5b226a93-5333-405f-93ad-788f9376faac-kube-api-access-gpt4f\") pod \"apiserver-7bbb656c7d-dc2xs\" (UID: \"5b226a93-5333-405f-93ad-788f9376faac\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.718475 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17718a81-aa70-4f5a-90bd-06e9251f4ef8-config\") pod \"machine-api-operator-5694c8668f-slkpm\" (UID: \"17718a81-aa70-4f5a-90bd-06e9251f4ef8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-slkpm" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.718521 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f50b116f-b25b-4985-8975-fbb523e1ddf6-client-ca\") pod \"route-controller-manager-6576b87f9c-wnlxq\" (UID: \"f50b116f-b25b-4985-8975-fbb523e1ddf6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wnlxq" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.718529 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.718542 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzgqr\" (UniqueName: \"kubernetes.io/projected/f50b116f-b25b-4985-8975-fbb523e1ddf6-kube-api-access-dzgqr\") pod \"route-controller-manager-6576b87f9c-wnlxq\" (UID: \"f50b116f-b25b-4985-8975-fbb523e1ddf6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wnlxq" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.718573 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgrvw\" (UniqueName: \"kubernetes.io/projected/baa99c88-cee5-42d8-ac79-21551d465f7f-kube-api-access-pgrvw\") pod \"openshift-controller-manager-operator-756b6f6bc6-qwnbb\" (UID: \"baa99c88-cee5-42d8-ac79-21551d465f7f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qwnbb" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.718610 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/66c28684-107f-4208-ae79-4b22a4e485dd-audit-dir\") pod \"apiserver-76f77b778f-8jkhv\" (UID: \"66c28684-107f-4208-ae79-4b22a4e485dd\") " pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.718629 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/69940b0d-4afc-4a7f-8ff7-cb128ce58913-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-cjdhk\" (UID: \"69940b0d-4afc-4a7f-8ff7-cb128ce58913\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cjdhk" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.718668 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5b226a93-5333-405f-93ad-788f9376faac-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-dc2xs\" (UID: \"5b226a93-5333-405f-93ad-788f9376faac\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.718696 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5b226a93-5333-405f-93ad-788f9376faac-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-dc2xs\" (UID: \"5b226a93-5333-405f-93ad-788f9376faac\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.718841 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2x49b\" (UniqueName: \"kubernetes.io/projected/69940b0d-4afc-4a7f-8ff7-cb128ce58913-kube-api-access-2x49b\") pod \"controller-manager-879f6c89f-cjdhk\" (UID: \"69940b0d-4afc-4a7f-8ff7-cb128ce58913\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cjdhk" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.718892 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66c28684-107f-4208-ae79-4b22a4e485dd-config\") pod \"apiserver-76f77b778f-8jkhv\" (UID: \"66c28684-107f-4208-ae79-4b22a4e485dd\") " pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.719592 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/baa99c88-cee5-42d8-ac79-21551d465f7f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-qwnbb\" (UID: \"baa99c88-cee5-42d8-ac79-21551d465f7f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qwnbb" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.719645 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66c28684-107f-4208-ae79-4b22a4e485dd-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8jkhv\" (UID: \"66c28684-107f-4208-ae79-4b22a4e485dd\") " pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.719693 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f50b116f-b25b-4985-8975-fbb523e1ddf6-serving-cert\") pod \"route-controller-manager-6576b87f9c-wnlxq\" (UID: \"f50b116f-b25b-4985-8975-fbb523e1ddf6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wnlxq" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.719744 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/66c28684-107f-4208-ae79-4b22a4e485dd-node-pullsecrets\") pod \"apiserver-76f77b778f-8jkhv\" (UID: \"66c28684-107f-4208-ae79-4b22a4e485dd\") " pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.719776 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66c28684-107f-4208-ae79-4b22a4e485dd-serving-cert\") pod \"apiserver-76f77b778f-8jkhv\" (UID: \"66c28684-107f-4208-ae79-4b22a4e485dd\") " pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.719826 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/baa99c88-cee5-42d8-ac79-21551d465f7f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-qwnbb\" (UID: \"baa99c88-cee5-42d8-ac79-21551d465f7f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qwnbb" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.719866 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/66c28684-107f-4208-ae79-4b22a4e485dd-etcd-client\") pod \"apiserver-76f77b778f-8jkhv\" (UID: \"66c28684-107f-4208-ae79-4b22a4e485dd\") " pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.719956 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/69940b0d-4afc-4a7f-8ff7-cb128ce58913-serving-cert\") pod \"controller-manager-879f6c89f-cjdhk\" (UID: \"69940b0d-4afc-4a7f-8ff7-cb128ce58913\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cjdhk" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.719998 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fb4mb\" (UniqueName: \"kubernetes.io/projected/194e75e8-4f01-4862-ba6e-a073a369df5a-kube-api-access-fb4mb\") pod \"cluster-samples-operator-665b6dd947-f55f4\" (UID: \"194e75e8-4f01-4862-ba6e-a073a369df5a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f55f4" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.720112 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5b226a93-5333-405f-93ad-788f9376faac-audit-policies\") pod \"apiserver-7bbb656c7d-dc2xs\" (UID: \"5b226a93-5333-405f-93ad-788f9376faac\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.720146 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/17718a81-aa70-4f5a-90bd-06e9251f4ef8-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-slkpm\" (UID: \"17718a81-aa70-4f5a-90bd-06e9251f4ef8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-slkpm" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.720174 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/66c28684-107f-4208-ae79-4b22a4e485dd-encryption-config\") pod \"apiserver-76f77b778f-8jkhv\" (UID: \"66c28684-107f-4208-ae79-4b22a4e485dd\") " pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.718855 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.718948 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.720941 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.718963 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.718965 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.719022 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.719089 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.721141 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.721183 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.719133 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.719163 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.719195 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.721332 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.719499 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.719775 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.719994 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.720148 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.720196 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.720302 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.720373 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.721800 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.720473 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.720510 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.720556 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.721984 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.722001 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.722429 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.722644 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.722786 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.723242 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.723323 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.723488 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.723726 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.724034 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.724093 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.724138 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.724255 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.724642 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.724852 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.724982 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.724994 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.725031 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.726127 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.739495 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.746175 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.749319 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xzd7l"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.751741 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.752715 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.757985 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.759484 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.761552 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.762551 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.762761 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.763447 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.763750 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.769146 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-wvnp9"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.769911 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-s7x26"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.770965 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xzd7l" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.771187 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.772154 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-wvnp9" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.772331 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6g5fd"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.772565 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-s7x26" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.772686 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-hx4vr"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.772827 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6g5fd" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.773022 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-6h2jt"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.773379 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hx4vr" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.773444 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6h2jt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.774014 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7x25b"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.774449 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7x25b" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.782947 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-fscrx"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.783977 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fscrx" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.784093 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-54x2b"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.784840 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-54x2b" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.788222 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399820-jr7cd"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.788907 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399820-jr7cd" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.790194 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-79gj2"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.790789 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-79gj2" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.791425 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.791437 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-hsvx8"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.792614 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-slkpm"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.795806 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bfmvc"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.797457 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-t5fjs"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.797581 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bfmvc" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.798851 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qwnbb"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.800900 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wnlxq"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.802749 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8jkhv"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.805029 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-lnmz8"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.805676 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-lnmz8" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.808656 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f55f4"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.811682 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.812615 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-cjdhk"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.814282 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-cppvr"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.819246 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-96268"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.821323 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpt4f\" (UniqueName: \"kubernetes.io/projected/5b226a93-5333-405f-93ad-788f9376faac-kube-api-access-gpt4f\") pod \"apiserver-7bbb656c7d-dc2xs\" (UID: \"5b226a93-5333-405f-93ad-788f9376faac\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.821353 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a86ba968-c6dd-42c8-91fb-207bc839b8ec-audit-policies\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.821372 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c7930b6-15dc-4462-bc47-9137c2d4edc7-config\") pod \"authentication-operator-69f744f599-zvzkq\" (UID: \"3c7930b6-15dc-4462-bc47-9137c2d4edc7\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zvzkq" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.821389 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.821408 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17718a81-aa70-4f5a-90bd-06e9251f4ef8-config\") pod \"machine-api-operator-5694c8668f-slkpm\" (UID: \"17718a81-aa70-4f5a-90bd-06e9251f4ef8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-slkpm" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.821424 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f50b116f-b25b-4985-8975-fbb523e1ddf6-client-ca\") pod \"route-controller-manager-6576b87f9c-wnlxq\" (UID: \"f50b116f-b25b-4985-8975-fbb523e1ddf6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wnlxq" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.821440 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzgqr\" (UniqueName: \"kubernetes.io/projected/f50b116f-b25b-4985-8975-fbb523e1ddf6-kube-api-access-dzgqr\") pod \"route-controller-manager-6576b87f9c-wnlxq\" (UID: \"f50b116f-b25b-4985-8975-fbb523e1ddf6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wnlxq" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.821692 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c42d89b6-6420-4267-93d2-e5dac6e1906b-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6ljl5\" (UID: \"c42d89b6-6420-4267-93d2-e5dac6e1906b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6ljl5" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.821742 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgrvw\" (UniqueName: \"kubernetes.io/projected/baa99c88-cee5-42d8-ac79-21551d465f7f-kube-api-access-pgrvw\") pod \"openshift-controller-manager-operator-756b6f6bc6-qwnbb\" (UID: \"baa99c88-cee5-42d8-ac79-21551d465f7f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qwnbb" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.821768 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.821812 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a2a0fdfc-5465-483d-8961-baa619f6ec0f-trusted-ca\") pod \"console-operator-58897d9998-z2dld\" (UID: \"a2a0fdfc-5465-483d-8961-baa619f6ec0f\") " pod="openshift-console-operator/console-operator-58897d9998-z2dld" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.821839 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/66c28684-107f-4208-ae79-4b22a4e485dd-audit-dir\") pod \"apiserver-76f77b778f-8jkhv\" (UID: \"66c28684-107f-4208-ae79-4b22a4e485dd\") " pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.821856 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.821872 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c36b34a9-5050-4b55-886b-83c0dd80a3df-metrics-certs\") pod \"router-default-5444994796-4ldk4\" (UID: \"c36b34a9-5050-4b55-886b-83c0dd80a3df\") " pod="openshift-ingress/router-default-5444994796-4ldk4" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.821920 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/69940b0d-4afc-4a7f-8ff7-cb128ce58913-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-cjdhk\" (UID: \"69940b0d-4afc-4a7f-8ff7-cb128ce58913\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cjdhk" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.821946 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5b226a93-5333-405f-93ad-788f9376faac-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-dc2xs\" (UID: \"5b226a93-5333-405f-93ad-788f9376faac\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.821969 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5b226a93-5333-405f-93ad-788f9376faac-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-dc2xs\" (UID: \"5b226a93-5333-405f-93ad-788f9376faac\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.821990 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/aef429a6-aa76-4776-a6e2-30096a5e7ccb-service-ca\") pod \"console-f9d7485db-t5fjs\" (UID: \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\") " pod="openshift-console/console-f9d7485db-t5fjs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.822022 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.822039 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7478\" (UniqueName: \"kubernetes.io/projected/c36b34a9-5050-4b55-886b-83c0dd80a3df-kube-api-access-v7478\") pod \"router-default-5444994796-4ldk4\" (UID: \"c36b34a9-5050-4b55-886b-83c0dd80a3df\") " pod="openshift-ingress/router-default-5444994796-4ldk4" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.822057 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2x49b\" (UniqueName: \"kubernetes.io/projected/69940b0d-4afc-4a7f-8ff7-cb128ce58913-kube-api-access-2x49b\") pod \"controller-manager-879f6c89f-cjdhk\" (UID: \"69940b0d-4afc-4a7f-8ff7-cb128ce58913\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cjdhk" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.822074 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66c28684-107f-4208-ae79-4b22a4e485dd-config\") pod \"apiserver-76f77b778f-8jkhv\" (UID: \"66c28684-107f-4208-ae79-4b22a4e485dd\") " pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.822094 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/56177dd2-06b5-4d07-9a9b-395c558e2f13-srv-cert\") pod \"catalog-operator-68c6474976-qzt4s\" (UID: \"56177dd2-06b5-4d07-9a9b-395c558e2f13\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qzt4s" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.822122 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.822140 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/baa99c88-cee5-42d8-ac79-21551d465f7f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-qwnbb\" (UID: \"baa99c88-cee5-42d8-ac79-21551d465f7f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qwnbb" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.822157 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66c28684-107f-4208-ae79-4b22a4e485dd-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8jkhv\" (UID: \"66c28684-107f-4208-ae79-4b22a4e485dd\") " pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.822177 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/692c3c09-be9f-4400-b0fb-e8c8692f0d21-config\") pod \"kube-controller-manager-operator-78b949d7b-5lzlv\" (UID: \"692c3c09-be9f-4400-b0fb-e8c8692f0d21\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5lzlv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.823314 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.823346 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c7930b6-15dc-4462-bc47-9137c2d4edc7-serving-cert\") pod \"authentication-operator-69f744f599-zvzkq\" (UID: \"3c7930b6-15dc-4462-bc47-9137c2d4edc7\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zvzkq" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.823364 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x67fz\" (UniqueName: \"kubernetes.io/projected/3c7930b6-15dc-4462-bc47-9137c2d4edc7-kube-api-access-x67fz\") pod \"authentication-operator-69f744f599-zvzkq\" (UID: \"3c7930b6-15dc-4462-bc47-9137c2d4edc7\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zvzkq" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.823450 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/aef429a6-aa76-4776-a6e2-30096a5e7ccb-console-config\") pod \"console-f9d7485db-t5fjs\" (UID: \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\") " pod="openshift-console/console-f9d7485db-t5fjs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.823456 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66c28684-107f-4208-ae79-4b22a4e485dd-config\") pod \"apiserver-76f77b778f-8jkhv\" (UID: \"66c28684-107f-4208-ae79-4b22a4e485dd\") " pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.823468 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aef429a6-aa76-4776-a6e2-30096a5e7ccb-trusted-ca-bundle\") pod \"console-f9d7485db-t5fjs\" (UID: \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\") " pod="openshift-console/console-f9d7485db-t5fjs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.822295 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17718a81-aa70-4f5a-90bd-06e9251f4ef8-config\") pod \"machine-api-operator-5694c8668f-slkpm\" (UID: \"17718a81-aa70-4f5a-90bd-06e9251f4ef8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-slkpm" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.822672 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5b226a93-5333-405f-93ad-788f9376faac-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-dc2xs\" (UID: \"5b226a93-5333-405f-93ad-788f9376faac\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.822778 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/66c28684-107f-4208-ae79-4b22a4e485dd-audit-dir\") pod \"apiserver-76f77b778f-8jkhv\" (UID: \"66c28684-107f-4208-ae79-4b22a4e485dd\") " pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.823723 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/69940b0d-4afc-4a7f-8ff7-cb128ce58913-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-cjdhk\" (UID: \"69940b0d-4afc-4a7f-8ff7-cb128ce58913\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cjdhk" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.823762 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-z2dld"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.824030 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f50b116f-b25b-4985-8975-fbb523e1ddf6-serving-cert\") pod \"route-controller-manager-6576b87f9c-wnlxq\" (UID: \"f50b116f-b25b-4985-8975-fbb523e1ddf6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wnlxq" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.824075 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/66c28684-107f-4208-ae79-4b22a4e485dd-node-pullsecrets\") pod \"apiserver-76f77b778f-8jkhv\" (UID: \"66c28684-107f-4208-ae79-4b22a4e485dd\") " pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.824133 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlkv2\" (UniqueName: \"kubernetes.io/projected/91336463-df6a-4b98-a990-e6780e417fab-kube-api-access-zlkv2\") pod \"ingress-operator-5b745b69d9-7d488\" (UID: \"91336463-df6a-4b98-a990-e6780e417fab\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7d488" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.824180 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/66c28684-107f-4208-ae79-4b22a4e485dd-node-pullsecrets\") pod \"apiserver-76f77b778f-8jkhv\" (UID: \"66c28684-107f-4208-ae79-4b22a4e485dd\") " pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.824371 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/baa99c88-cee5-42d8-ac79-21551d465f7f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-qwnbb\" (UID: \"baa99c88-cee5-42d8-ac79-21551d465f7f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qwnbb" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.824440 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66c28684-107f-4208-ae79-4b22a4e485dd-serving-cert\") pod \"apiserver-76f77b778f-8jkhv\" (UID: \"66c28684-107f-4208-ae79-4b22a4e485dd\") " pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.824453 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66c28684-107f-4208-ae79-4b22a4e485dd-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8jkhv\" (UID: \"66c28684-107f-4208-ae79-4b22a4e485dd\") " pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.824492 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/56177dd2-06b5-4d07-9a9b-395c558e2f13-profile-collector-cert\") pod \"catalog-operator-68c6474976-qzt4s\" (UID: \"56177dd2-06b5-4d07-9a9b-395c558e2f13\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qzt4s" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.825035 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.825340 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/baa99c88-cee5-42d8-ac79-21551d465f7f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-qwnbb\" (UID: \"baa99c88-cee5-42d8-ac79-21551d465f7f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qwnbb" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.825367 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.825389 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llfmr\" (UniqueName: \"kubernetes.io/projected/a2a0fdfc-5465-483d-8961-baa619f6ec0f-kube-api-access-llfmr\") pod \"console-operator-58897d9998-z2dld\" (UID: \"a2a0fdfc-5465-483d-8961-baa619f6ec0f\") " pod="openshift-console-operator/console-operator-58897d9998-z2dld" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.825409 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/aef429a6-aa76-4776-a6e2-30096a5e7ccb-console-oauth-config\") pod \"console-f9d7485db-t5fjs\" (UID: \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\") " pod="openshift-console/console-f9d7485db-t5fjs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.825430 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2q6p\" (UniqueName: \"kubernetes.io/projected/acb2b70f-4a66-4290-aa8a-a13cb1b4b2a9-kube-api-access-g2q6p\") pod \"control-plane-machine-set-operator-78cbb6b69f-96268\" (UID: \"acb2b70f-4a66-4290-aa8a-a13cb1b4b2a9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-96268" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.825457 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/66c28684-107f-4208-ae79-4b22a4e485dd-etcd-client\") pod \"apiserver-76f77b778f-8jkhv\" (UID: \"66c28684-107f-4208-ae79-4b22a4e485dd\") " pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.825480 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/69940b0d-4afc-4a7f-8ff7-cb128ce58913-serving-cert\") pod \"controller-manager-879f6c89f-cjdhk\" (UID: \"69940b0d-4afc-4a7f-8ff7-cb128ce58913\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cjdhk" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.825502 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3c7930b6-15dc-4462-bc47-9137c2d4edc7-service-ca-bundle\") pod \"authentication-operator-69f744f599-zvzkq\" (UID: \"3c7930b6-15dc-4462-bc47-9137c2d4edc7\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zvzkq" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.825524 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c36b34a9-5050-4b55-886b-83c0dd80a3df-stats-auth\") pod \"router-default-5444994796-4ldk4\" (UID: \"c36b34a9-5050-4b55-886b-83c0dd80a3df\") " pod="openshift-ingress/router-default-5444994796-4ldk4" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.825544 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/692c3c09-be9f-4400-b0fb-e8c8692f0d21-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5lzlv\" (UID: \"692c3c09-be9f-4400-b0fb-e8c8692f0d21\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5lzlv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.825572 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fb4mb\" (UniqueName: \"kubernetes.io/projected/194e75e8-4f01-4862-ba6e-a073a369df5a-kube-api-access-fb4mb\") pod \"cluster-samples-operator-665b6dd947-f55f4\" (UID: \"194e75e8-4f01-4862-ba6e-a073a369df5a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f55f4" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.825593 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5b226a93-5333-405f-93ad-788f9376faac-audit-policies\") pod \"apiserver-7bbb656c7d-dc2xs\" (UID: \"5b226a93-5333-405f-93ad-788f9376faac\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.825614 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c42d89b6-6420-4267-93d2-e5dac6e1906b-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6ljl5\" (UID: \"c42d89b6-6420-4267-93d2-e5dac6e1906b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6ljl5" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.825634 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.825655 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.825679 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/17718a81-aa70-4f5a-90bd-06e9251f4ef8-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-slkpm\" (UID: \"17718a81-aa70-4f5a-90bd-06e9251f4ef8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-slkpm" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.825704 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/66c28684-107f-4208-ae79-4b22a4e485dd-encryption-config\") pod \"apiserver-76f77b778f-8jkhv\" (UID: \"66c28684-107f-4208-ae79-4b22a4e485dd\") " pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.825730 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/aef429a6-aa76-4776-a6e2-30096a5e7ccb-oauth-serving-cert\") pod \"console-f9d7485db-t5fjs\" (UID: \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\") " pod="openshift-console/console-f9d7485db-t5fjs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.825751 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3c7930b6-15dc-4462-bc47-9137c2d4edc7-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-zvzkq\" (UID: \"3c7930b6-15dc-4462-bc47-9137c2d4edc7\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zvzkq" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.825773 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2f77876b-f997-4274-a619-a97e37bc4b9e-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-djs6p\" (UID: \"2f77876b-f997-4274-a619-a97e37bc4b9e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-djs6p" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.825796 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94mzt\" (UniqueName: \"kubernetes.io/projected/2f77876b-f997-4274-a619-a97e37bc4b9e-kube-api-access-94mzt\") pod \"cluster-image-registry-operator-dc59b4c8b-djs6p\" (UID: \"2f77876b-f997-4274-a619-a97e37bc4b9e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-djs6p" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.825824 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/17718a81-aa70-4f5a-90bd-06e9251f4ef8-images\") pod \"machine-api-operator-5694c8668f-slkpm\" (UID: \"17718a81-aa70-4f5a-90bd-06e9251f4ef8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-slkpm" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.825846 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.825871 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/69940b0d-4afc-4a7f-8ff7-cb128ce58913-client-ca\") pod \"controller-manager-879f6c89f-cjdhk\" (UID: \"69940b0d-4afc-4a7f-8ff7-cb128ce58913\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cjdhk" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.825911 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/91336463-df6a-4b98-a990-e6780e417fab-bound-sa-token\") pod \"ingress-operator-5b745b69d9-7d488\" (UID: \"91336463-df6a-4b98-a990-e6780e417fab\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7d488" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.825932 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a86ba968-c6dd-42c8-91fb-207bc839b8ec-audit-dir\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.825977 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/1e6bf7c7-7711-43cb-8817-73b42174b9ee-machine-approver-tls\") pod \"machine-approver-56656f9798-5hlqm\" (UID: \"1e6bf7c7-7711-43cb-8817-73b42174b9ee\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5hlqm" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826000 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mk45\" (UniqueName: \"kubernetes.io/projected/1e6bf7c7-7711-43cb-8817-73b42174b9ee-kube-api-access-4mk45\") pod \"machine-approver-56656f9798-5hlqm\" (UID: \"1e6bf7c7-7711-43cb-8817-73b42174b9ee\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5hlqm" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826024 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c36b34a9-5050-4b55-886b-83c0dd80a3df-service-ca-bundle\") pod \"router-default-5444994796-4ldk4\" (UID: \"c36b34a9-5050-4b55-886b-83c0dd80a3df\") " pod="openshift-ingress/router-default-5444994796-4ldk4" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826050 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a2b45ef-e51f-4ad2-a1b7-6d8e7b0c089a-serving-cert\") pod \"openshift-config-operator-7777fb866f-hsvx8\" (UID: \"2a2b45ef-e51f-4ad2-a1b7-6d8e7b0c089a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsvx8" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826072 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2a0fdfc-5465-483d-8961-baa619f6ec0f-config\") pod \"console-operator-58897d9998-z2dld\" (UID: \"a2a0fdfc-5465-483d-8961-baa619f6ec0f\") " pod="openshift-console-operator/console-operator-58897d9998-z2dld" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826094 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2f77876b-f997-4274-a619-a97e37bc4b9e-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-djs6p\" (UID: \"2f77876b-f997-4274-a619-a97e37bc4b9e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-djs6p" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826135 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5b226a93-5333-405f-93ad-788f9376faac-encryption-config\") pod \"apiserver-7bbb656c7d-dc2xs\" (UID: \"5b226a93-5333-405f-93ad-788f9376faac\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826152 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2a0fdfc-5465-483d-8961-baa619f6ec0f-serving-cert\") pod \"console-operator-58897d9998-z2dld\" (UID: \"a2a0fdfc-5465-483d-8961-baa619f6ec0f\") " pod="openshift-console-operator/console-operator-58897d9998-z2dld" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826171 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/acb2b70f-4a66-4290-aa8a-a13cb1b4b2a9-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-96268\" (UID: \"acb2b70f-4a66-4290-aa8a-a13cb1b4b2a9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-96268" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826191 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pc7f8\" (UniqueName: \"kubernetes.io/projected/2a2b45ef-e51f-4ad2-a1b7-6d8e7b0c089a-kube-api-access-pc7f8\") pod \"openshift-config-operator-7777fb866f-hsvx8\" (UID: \"2a2b45ef-e51f-4ad2-a1b7-6d8e7b0c089a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsvx8" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826208 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfndz\" (UniqueName: \"kubernetes.io/projected/aef429a6-aa76-4776-a6e2-30096a5e7ccb-kube-api-access-sfndz\") pod \"console-f9d7485db-t5fjs\" (UID: \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\") " pod="openshift-console/console-f9d7485db-t5fjs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826233 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69940b0d-4afc-4a7f-8ff7-cb128ce58913-config\") pod \"controller-manager-879f6c89f-cjdhk\" (UID: \"69940b0d-4afc-4a7f-8ff7-cb128ce58913\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cjdhk" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826266 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/66c28684-107f-4208-ae79-4b22a4e485dd-etcd-serving-ca\") pod \"apiserver-76f77b778f-8jkhv\" (UID: \"66c28684-107f-4208-ae79-4b22a4e485dd\") " pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826288 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5zww\" (UniqueName: \"kubernetes.io/projected/a86ba968-c6dd-42c8-91fb-207bc839b8ec-kube-api-access-t5zww\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826311 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2f77876b-f997-4274-a619-a97e37bc4b9e-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-djs6p\" (UID: \"2f77876b-f997-4274-a619-a97e37bc4b9e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-djs6p" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826328 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826333 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/194e75e8-4f01-4862-ba6e-a073a369df5a-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-f55f4\" (UID: \"194e75e8-4f01-4862-ba6e-a073a369df5a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f55f4" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826395 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f50b116f-b25b-4985-8975-fbb523e1ddf6-config\") pod \"route-controller-manager-6576b87f9c-wnlxq\" (UID: \"f50b116f-b25b-4985-8975-fbb523e1ddf6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wnlxq" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826424 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/66c28684-107f-4208-ae79-4b22a4e485dd-audit\") pod \"apiserver-76f77b778f-8jkhv\" (UID: \"66c28684-107f-4208-ae79-4b22a4e485dd\") " pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826445 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c36b34a9-5050-4b55-886b-83c0dd80a3df-default-certificate\") pod \"router-default-5444994796-4ldk4\" (UID: \"c36b34a9-5050-4b55-886b-83c0dd80a3df\") " pod="openshift-ingress/router-default-5444994796-4ldk4" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826469 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5b226a93-5333-405f-93ad-788f9376faac-audit-dir\") pod \"apiserver-7bbb656c7d-dc2xs\" (UID: \"5b226a93-5333-405f-93ad-788f9376faac\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826487 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dz26\" (UniqueName: \"kubernetes.io/projected/17718a81-aa70-4f5a-90bd-06e9251f4ef8-kube-api-access-9dz26\") pod \"machine-api-operator-5694c8668f-slkpm\" (UID: \"17718a81-aa70-4f5a-90bd-06e9251f4ef8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-slkpm" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826503 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1e6bf7c7-7711-43cb-8817-73b42174b9ee-auth-proxy-config\") pod \"machine-approver-56656f9798-5hlqm\" (UID: \"1e6bf7c7-7711-43cb-8817-73b42174b9ee\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5hlqm" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826522 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dp5qr\" (UniqueName: \"kubernetes.io/projected/56177dd2-06b5-4d07-9a9b-395c558e2f13-kube-api-access-dp5qr\") pod \"catalog-operator-68c6474976-qzt4s\" (UID: \"56177dd2-06b5-4d07-9a9b-395c558e2f13\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qzt4s" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826543 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5b226a93-5333-405f-93ad-788f9376faac-serving-cert\") pod \"apiserver-7bbb656c7d-dc2xs\" (UID: \"5b226a93-5333-405f-93ad-788f9376faac\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826562 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/66c28684-107f-4208-ae79-4b22a4e485dd-image-import-ca\") pod \"apiserver-76f77b778f-8jkhv\" (UID: \"66c28684-107f-4208-ae79-4b22a4e485dd\") " pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826582 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5b226a93-5333-405f-93ad-788f9376faac-etcd-client\") pod \"apiserver-7bbb656c7d-dc2xs\" (UID: \"5b226a93-5333-405f-93ad-788f9376faac\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826598 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c42d89b6-6420-4267-93d2-e5dac6e1906b-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6ljl5\" (UID: \"c42d89b6-6420-4267-93d2-e5dac6e1906b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6ljl5" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826613 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/91336463-df6a-4b98-a990-e6780e417fab-trusted-ca\") pod \"ingress-operator-5b745b69d9-7d488\" (UID: \"91336463-df6a-4b98-a990-e6780e417fab\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7d488" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826629 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/692c3c09-be9f-4400-b0fb-e8c8692f0d21-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-5lzlv\" (UID: \"692c3c09-be9f-4400-b0fb-e8c8692f0d21\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5lzlv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826648 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xflx6\" (UniqueName: \"kubernetes.io/projected/66c28684-107f-4208-ae79-4b22a4e485dd-kube-api-access-xflx6\") pod \"apiserver-76f77b778f-8jkhv\" (UID: \"66c28684-107f-4208-ae79-4b22a4e485dd\") " pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826664 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e6bf7c7-7711-43cb-8817-73b42174b9ee-config\") pod \"machine-approver-56656f9798-5hlqm\" (UID: \"1e6bf7c7-7711-43cb-8817-73b42174b9ee\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5hlqm" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826680 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/2a2b45ef-e51f-4ad2-a1b7-6d8e7b0c089a-available-featuregates\") pod \"openshift-config-operator-7777fb866f-hsvx8\" (UID: \"2a2b45ef-e51f-4ad2-a1b7-6d8e7b0c089a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsvx8" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826702 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/aef429a6-aa76-4776-a6e2-30096a5e7ccb-console-serving-cert\") pod \"console-f9d7485db-t5fjs\" (UID: \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\") " pod="openshift-console/console-f9d7485db-t5fjs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.826719 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/91336463-df6a-4b98-a990-e6780e417fab-metrics-tls\") pod \"ingress-operator-5b745b69d9-7d488\" (UID: \"91336463-df6a-4b98-a990-e6780e417fab\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7d488" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.827023 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c4qnk"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.827658 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f50b116f-b25b-4985-8975-fbb523e1ddf6-config\") pod \"route-controller-manager-6576b87f9c-wnlxq\" (UID: \"f50b116f-b25b-4985-8975-fbb523e1ddf6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wnlxq" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.827966 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/66c28684-107f-4208-ae79-4b22a4e485dd-audit\") pod \"apiserver-76f77b778f-8jkhv\" (UID: \"66c28684-107f-4208-ae79-4b22a4e485dd\") " pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.828046 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5b226a93-5333-405f-93ad-788f9376faac-audit-dir\") pod \"apiserver-7bbb656c7d-dc2xs\" (UID: \"5b226a93-5333-405f-93ad-788f9376faac\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.828125 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f50b116f-b25b-4985-8975-fbb523e1ddf6-client-ca\") pod \"route-controller-manager-6576b87f9c-wnlxq\" (UID: \"f50b116f-b25b-4985-8975-fbb523e1ddf6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wnlxq" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.828530 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1e6bf7c7-7711-43cb-8817-73b42174b9ee-auth-proxy-config\") pod \"machine-approver-56656f9798-5hlqm\" (UID: \"1e6bf7c7-7711-43cb-8817-73b42174b9ee\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5hlqm" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.829209 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/69940b0d-4afc-4a7f-8ff7-cb128ce58913-client-ca\") pod \"controller-manager-879f6c89f-cjdhk\" (UID: \"69940b0d-4afc-4a7f-8ff7-cb128ce58913\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cjdhk" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.829936 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/66c28684-107f-4208-ae79-4b22a4e485dd-image-import-ca\") pod \"apiserver-76f77b778f-8jkhv\" (UID: \"66c28684-107f-4208-ae79-4b22a4e485dd\") " pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.830386 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e6bf7c7-7711-43cb-8817-73b42174b9ee-config\") pod \"machine-approver-56656f9798-5hlqm\" (UID: \"1e6bf7c7-7711-43cb-8817-73b42174b9ee\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5hlqm" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.830394 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/66c28684-107f-4208-ae79-4b22a4e485dd-etcd-client\") pod \"apiserver-76f77b778f-8jkhv\" (UID: \"66c28684-107f-4208-ae79-4b22a4e485dd\") " pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.830637 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6ljl5"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.830765 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-thd68"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.830774 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/17718a81-aa70-4f5a-90bd-06e9251f4ef8-images\") pod \"machine-api-operator-5694c8668f-slkpm\" (UID: \"17718a81-aa70-4f5a-90bd-06e9251f4ef8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-slkpm" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.831213 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5b226a93-5333-405f-93ad-788f9376faac-audit-policies\") pod \"apiserver-7bbb656c7d-dc2xs\" (UID: \"5b226a93-5333-405f-93ad-788f9376faac\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.831513 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/66c28684-107f-4208-ae79-4b22a4e485dd-etcd-serving-ca\") pod \"apiserver-76f77b778f-8jkhv\" (UID: \"66c28684-107f-4208-ae79-4b22a4e485dd\") " pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.830737 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66c28684-107f-4208-ae79-4b22a4e485dd-serving-cert\") pod \"apiserver-76f77b778f-8jkhv\" (UID: \"66c28684-107f-4208-ae79-4b22a4e485dd\") " pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.831925 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f50b116f-b25b-4985-8975-fbb523e1ddf6-serving-cert\") pod \"route-controller-manager-6576b87f9c-wnlxq\" (UID: \"f50b116f-b25b-4985-8975-fbb523e1ddf6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wnlxq" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.832278 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/2a2b45ef-e51f-4ad2-a1b7-6d8e7b0c089a-available-featuregates\") pod \"openshift-config-operator-7777fb866f-hsvx8\" (UID: \"2a2b45ef-e51f-4ad2-a1b7-6d8e7b0c089a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsvx8" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.832874 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69940b0d-4afc-4a7f-8ff7-cb128ce58913-config\") pod \"controller-manager-879f6c89f-cjdhk\" (UID: \"69940b0d-4afc-4a7f-8ff7-cb128ce58913\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cjdhk" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.833462 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.833641 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-hkbxm"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.833857 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/194e75e8-4f01-4862-ba6e-a073a369df5a-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-f55f4\" (UID: \"194e75e8-4f01-4862-ba6e-a073a369df5a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f55f4" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.834133 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/1e6bf7c7-7711-43cb-8817-73b42174b9ee-machine-approver-tls\") pod \"machine-approver-56656f9798-5hlqm\" (UID: \"1e6bf7c7-7711-43cb-8817-73b42174b9ee\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5hlqm" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.834512 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5b226a93-5333-405f-93ad-788f9376faac-serving-cert\") pod \"apiserver-7bbb656c7d-dc2xs\" (UID: \"5b226a93-5333-405f-93ad-788f9376faac\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.836240 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/17718a81-aa70-4f5a-90bd-06e9251f4ef8-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-slkpm\" (UID: \"17718a81-aa70-4f5a-90bd-06e9251f4ef8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-slkpm" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.836422 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-54x2b"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.837742 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/66c28684-107f-4208-ae79-4b22a4e485dd-encryption-config\") pod \"apiserver-76f77b778f-8jkhv\" (UID: \"66c28684-107f-4208-ae79-4b22a4e485dd\") " pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.837868 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qzt4s"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.838185 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5b226a93-5333-405f-93ad-788f9376faac-encryption-config\") pod \"apiserver-7bbb656c7d-dc2xs\" (UID: \"5b226a93-5333-405f-93ad-788f9376faac\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.839015 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/69940b0d-4afc-4a7f-8ff7-cb128ce58913-serving-cert\") pod \"controller-manager-879f6c89f-cjdhk\" (UID: \"69940b0d-4afc-4a7f-8ff7-cb128ce58913\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cjdhk" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.839125 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-7d488"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.827002 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5b226a93-5333-405f-93ad-788f9376faac-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-dc2xs\" (UID: \"5b226a93-5333-405f-93ad-788f9376faac\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.840911 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a2b45ef-e51f-4ad2-a1b7-6d8e7b0c089a-serving-cert\") pod \"openshift-config-operator-7777fb866f-hsvx8\" (UID: \"2a2b45ef-e51f-4ad2-a1b7-6d8e7b0c089a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsvx8" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.844997 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-xcw79"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.845079 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xzd7l"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.848293 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5b226a93-5333-405f-93ad-788f9376faac-etcd-client\") pod \"apiserver-7bbb656c7d-dc2xs\" (UID: \"5b226a93-5333-405f-93ad-788f9376faac\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.857903 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-zvzkq"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.858432 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-79gj2"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.859435 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6g5fd"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.861294 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ktdmx"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.862080 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/baa99c88-cee5-42d8-ac79-21551d465f7f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-qwnbb\" (UID: \"baa99c88-cee5-42d8-ac79-21551d465f7f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qwnbb" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.862274 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.866016 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5lzlv"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.867600 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399820-jr7cd"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.869092 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-djs6p"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.870872 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.871076 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-hx4vr"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.872179 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7x25b"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.872738 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-fscrx"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.874023 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-rhnfd"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.875047 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-rhnfd" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.875145 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-6h2jt"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.876327 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-s7x26"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.877469 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-4t9zv"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.889916 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-lnmz8"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.889965 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bfmvc"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.889979 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-4t9zv"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.889993 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-wvnp9"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.890004 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-zxkb7"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.890018 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-csctv"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.890145 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-4t9zv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.891834 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.892358 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-csctv"] Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.892467 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-csctv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.910696 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.927845 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.927895 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llfmr\" (UniqueName: \"kubernetes.io/projected/a2a0fdfc-5465-483d-8961-baa619f6ec0f-kube-api-access-llfmr\") pod \"console-operator-58897d9998-z2dld\" (UID: \"a2a0fdfc-5465-483d-8961-baa619f6ec0f\") " pod="openshift-console-operator/console-operator-58897d9998-z2dld" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.927945 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/aef429a6-aa76-4776-a6e2-30096a5e7ccb-console-oauth-config\") pod \"console-f9d7485db-t5fjs\" (UID: \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\") " pod="openshift-console/console-f9d7485db-t5fjs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.927965 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2q6p\" (UniqueName: \"kubernetes.io/projected/acb2b70f-4a66-4290-aa8a-a13cb1b4b2a9-kube-api-access-g2q6p\") pod \"control-plane-machine-set-operator-78cbb6b69f-96268\" (UID: \"acb2b70f-4a66-4290-aa8a-a13cb1b4b2a9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-96268" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.927987 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3c7930b6-15dc-4462-bc47-9137c2d4edc7-service-ca-bundle\") pod \"authentication-operator-69f744f599-zvzkq\" (UID: \"3c7930b6-15dc-4462-bc47-9137c2d4edc7\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zvzkq" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.928005 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c36b34a9-5050-4b55-886b-83c0dd80a3df-stats-auth\") pod \"router-default-5444994796-4ldk4\" (UID: \"c36b34a9-5050-4b55-886b-83c0dd80a3df\") " pod="openshift-ingress/router-default-5444994796-4ldk4" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.928020 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/692c3c09-be9f-4400-b0fb-e8c8692f0d21-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5lzlv\" (UID: \"692c3c09-be9f-4400-b0fb-e8c8692f0d21\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5lzlv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.928042 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c42d89b6-6420-4267-93d2-e5dac6e1906b-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6ljl5\" (UID: \"c42d89b6-6420-4267-93d2-e5dac6e1906b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6ljl5" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.928057 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.928239 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.928257 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/aef429a6-aa76-4776-a6e2-30096a5e7ccb-oauth-serving-cert\") pod \"console-f9d7485db-t5fjs\" (UID: \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\") " pod="openshift-console/console-f9d7485db-t5fjs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.928293 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3c7930b6-15dc-4462-bc47-9137c2d4edc7-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-zvzkq\" (UID: \"3c7930b6-15dc-4462-bc47-9137c2d4edc7\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zvzkq" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.928310 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2f77876b-f997-4274-a619-a97e37bc4b9e-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-djs6p\" (UID: \"2f77876b-f997-4274-a619-a97e37bc4b9e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-djs6p" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.928327 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94mzt\" (UniqueName: \"kubernetes.io/projected/2f77876b-f997-4274-a619-a97e37bc4b9e-kube-api-access-94mzt\") pod \"cluster-image-registry-operator-dc59b4c8b-djs6p\" (UID: \"2f77876b-f997-4274-a619-a97e37bc4b9e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-djs6p" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.928352 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.929422 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/91336463-df6a-4b98-a990-e6780e417fab-bound-sa-token\") pod \"ingress-operator-5b745b69d9-7d488\" (UID: \"91336463-df6a-4b98-a990-e6780e417fab\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7d488" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.929441 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a86ba968-c6dd-42c8-91fb-207bc839b8ec-audit-dir\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.929439 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/aef429a6-aa76-4776-a6e2-30096a5e7ccb-oauth-serving-cert\") pod \"console-f9d7485db-t5fjs\" (UID: \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\") " pod="openshift-console/console-f9d7485db-t5fjs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.929057 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3c7930b6-15dc-4462-bc47-9137c2d4edc7-service-ca-bundle\") pod \"authentication-operator-69f744f599-zvzkq\" (UID: \"3c7930b6-15dc-4462-bc47-9137c2d4edc7\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zvzkq" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.929326 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3c7930b6-15dc-4462-bc47-9137c2d4edc7-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-zvzkq\" (UID: \"3c7930b6-15dc-4462-bc47-9137c2d4edc7\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zvzkq" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.929537 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c36b34a9-5050-4b55-886b-83c0dd80a3df-service-ca-bundle\") pod \"router-default-5444994796-4ldk4\" (UID: \"c36b34a9-5050-4b55-886b-83c0dd80a3df\") " pod="openshift-ingress/router-default-5444994796-4ldk4" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.929554 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2a0fdfc-5465-483d-8961-baa619f6ec0f-config\") pod \"console-operator-58897d9998-z2dld\" (UID: \"a2a0fdfc-5465-483d-8961-baa619f6ec0f\") " pod="openshift-console-operator/console-operator-58897d9998-z2dld" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.929621 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a86ba968-c6dd-42c8-91fb-207bc839b8ec-audit-dir\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.929644 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2f77876b-f997-4274-a619-a97e37bc4b9e-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-djs6p\" (UID: \"2f77876b-f997-4274-a619-a97e37bc4b9e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-djs6p" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.929763 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2a0fdfc-5465-483d-8961-baa619f6ec0f-serving-cert\") pod \"console-operator-58897d9998-z2dld\" (UID: \"a2a0fdfc-5465-483d-8961-baa619f6ec0f\") " pod="openshift-console-operator/console-operator-58897d9998-z2dld" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.929794 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/acb2b70f-4a66-4290-aa8a-a13cb1b4b2a9-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-96268\" (UID: \"acb2b70f-4a66-4290-aa8a-a13cb1b4b2a9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-96268" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.930280 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfndz\" (UniqueName: \"kubernetes.io/projected/aef429a6-aa76-4776-a6e2-30096a5e7ccb-kube-api-access-sfndz\") pod \"console-f9d7485db-t5fjs\" (UID: \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\") " pod="openshift-console/console-f9d7485db-t5fjs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.930339 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5zww\" (UniqueName: \"kubernetes.io/projected/a86ba968-c6dd-42c8-91fb-207bc839b8ec-kube-api-access-t5zww\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.930360 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2f77876b-f997-4274-a619-a97e37bc4b9e-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-djs6p\" (UID: \"2f77876b-f997-4274-a619-a97e37bc4b9e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-djs6p" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.930400 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c36b34a9-5050-4b55-886b-83c0dd80a3df-default-certificate\") pod \"router-default-5444994796-4ldk4\" (UID: \"c36b34a9-5050-4b55-886b-83c0dd80a3df\") " pod="openshift-ingress/router-default-5444994796-4ldk4" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.930430 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dp5qr\" (UniqueName: \"kubernetes.io/projected/56177dd2-06b5-4d07-9a9b-395c558e2f13-kube-api-access-dp5qr\") pod \"catalog-operator-68c6474976-qzt4s\" (UID: \"56177dd2-06b5-4d07-9a9b-395c558e2f13\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qzt4s" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.930582 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c42d89b6-6420-4267-93d2-e5dac6e1906b-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6ljl5\" (UID: \"c42d89b6-6420-4267-93d2-e5dac6e1906b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6ljl5" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.930601 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/91336463-df6a-4b98-a990-e6780e417fab-trusted-ca\") pod \"ingress-operator-5b745b69d9-7d488\" (UID: \"91336463-df6a-4b98-a990-e6780e417fab\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7d488" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.930617 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/692c3c09-be9f-4400-b0fb-e8c8692f0d21-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-5lzlv\" (UID: \"692c3c09-be9f-4400-b0fb-e8c8692f0d21\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5lzlv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.931854 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/aef429a6-aa76-4776-a6e2-30096a5e7ccb-console-serving-cert\") pod \"console-f9d7485db-t5fjs\" (UID: \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\") " pod="openshift-console/console-f9d7485db-t5fjs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.932224 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/91336463-df6a-4b98-a990-e6780e417fab-metrics-tls\") pod \"ingress-operator-5b745b69d9-7d488\" (UID: \"91336463-df6a-4b98-a990-e6780e417fab\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7d488" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.932261 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a86ba968-c6dd-42c8-91fb-207bc839b8ec-audit-policies\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.932285 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c7930b6-15dc-4462-bc47-9137c2d4edc7-config\") pod \"authentication-operator-69f744f599-zvzkq\" (UID: \"3c7930b6-15dc-4462-bc47-9137c2d4edc7\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zvzkq" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.932303 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.932915 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c42d89b6-6420-4267-93d2-e5dac6e1906b-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6ljl5\" (UID: \"c42d89b6-6420-4267-93d2-e5dac6e1906b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6ljl5" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.932946 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.932962 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a2a0fdfc-5465-483d-8961-baa619f6ec0f-trusted-ca\") pod \"console-operator-58897d9998-z2dld\" (UID: \"a2a0fdfc-5465-483d-8961-baa619f6ec0f\") " pod="openshift-console-operator/console-operator-58897d9998-z2dld" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.933055 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.933072 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c36b34a9-5050-4b55-886b-83c0dd80a3df-metrics-certs\") pod \"router-default-5444994796-4ldk4\" (UID: \"c36b34a9-5050-4b55-886b-83c0dd80a3df\") " pod="openshift-ingress/router-default-5444994796-4ldk4" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.932802 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c7930b6-15dc-4462-bc47-9137c2d4edc7-config\") pod \"authentication-operator-69f744f599-zvzkq\" (UID: \"3c7930b6-15dc-4462-bc47-9137c2d4edc7\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zvzkq" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.931011 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2f77876b-f997-4274-a619-a97e37bc4b9e-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-djs6p\" (UID: \"2f77876b-f997-4274-a619-a97e37bc4b9e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-djs6p" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.932103 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/91336463-df6a-4b98-a990-e6780e417fab-trusted-ca\") pod \"ingress-operator-5b745b69d9-7d488\" (UID: \"91336463-df6a-4b98-a990-e6780e417fab\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7d488" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.930615 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2a0fdfc-5465-483d-8961-baa619f6ec0f-config\") pod \"console-operator-58897d9998-z2dld\" (UID: \"a2a0fdfc-5465-483d-8961-baa619f6ec0f\") " pod="openshift-console-operator/console-operator-58897d9998-z2dld" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.931895 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/aef429a6-aa76-4776-a6e2-30096a5e7ccb-console-oauth-config\") pod \"console-f9d7485db-t5fjs\" (UID: \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\") " pod="openshift-console/console-f9d7485db-t5fjs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.931420 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.933177 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/aef429a6-aa76-4776-a6e2-30096a5e7ccb-service-ca\") pod \"console-f9d7485db-t5fjs\" (UID: \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\") " pod="openshift-console/console-f9d7485db-t5fjs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.933312 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.933336 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7478\" (UniqueName: \"kubernetes.io/projected/c36b34a9-5050-4b55-886b-83c0dd80a3df-kube-api-access-v7478\") pod \"router-default-5444994796-4ldk4\" (UID: \"c36b34a9-5050-4b55-886b-83c0dd80a3df\") " pod="openshift-ingress/router-default-5444994796-4ldk4" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.933366 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/56177dd2-06b5-4d07-9a9b-395c558e2f13-srv-cert\") pod \"catalog-operator-68c6474976-qzt4s\" (UID: \"56177dd2-06b5-4d07-9a9b-395c558e2f13\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qzt4s" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.933399 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.933435 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/692c3c09-be9f-4400-b0fb-e8c8692f0d21-config\") pod \"kube-controller-manager-operator-78b949d7b-5lzlv\" (UID: \"692c3c09-be9f-4400-b0fb-e8c8692f0d21\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5lzlv" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.933471 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.933492 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c7930b6-15dc-4462-bc47-9137c2d4edc7-serving-cert\") pod \"authentication-operator-69f744f599-zvzkq\" (UID: \"3c7930b6-15dc-4462-bc47-9137c2d4edc7\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zvzkq" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.933512 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x67fz\" (UniqueName: \"kubernetes.io/projected/3c7930b6-15dc-4462-bc47-9137c2d4edc7-kube-api-access-x67fz\") pod \"authentication-operator-69f744f599-zvzkq\" (UID: \"3c7930b6-15dc-4462-bc47-9137c2d4edc7\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zvzkq" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.933533 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/aef429a6-aa76-4776-a6e2-30096a5e7ccb-console-config\") pod \"console-f9d7485db-t5fjs\" (UID: \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\") " pod="openshift-console/console-f9d7485db-t5fjs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.933546 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c42d89b6-6420-4267-93d2-e5dac6e1906b-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6ljl5\" (UID: \"c42d89b6-6420-4267-93d2-e5dac6e1906b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6ljl5" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.933554 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aef429a6-aa76-4776-a6e2-30096a5e7ccb-trusted-ca-bundle\") pod \"console-f9d7485db-t5fjs\" (UID: \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\") " pod="openshift-console/console-f9d7485db-t5fjs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.933611 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlkv2\" (UniqueName: \"kubernetes.io/projected/91336463-df6a-4b98-a990-e6780e417fab-kube-api-access-zlkv2\") pod \"ingress-operator-5b745b69d9-7d488\" (UID: \"91336463-df6a-4b98-a990-e6780e417fab\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7d488" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.933674 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/56177dd2-06b5-4d07-9a9b-395c558e2f13-profile-collector-cert\") pod \"catalog-operator-68c6474976-qzt4s\" (UID: \"56177dd2-06b5-4d07-9a9b-395c558e2f13\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qzt4s" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.933705 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.933780 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a2a0fdfc-5465-483d-8961-baa619f6ec0f-trusted-ca\") pod \"console-operator-58897d9998-z2dld\" (UID: \"a2a0fdfc-5465-483d-8961-baa619f6ec0f\") " pod="openshift-console-operator/console-operator-58897d9998-z2dld" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.934425 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aef429a6-aa76-4776-a6e2-30096a5e7ccb-trusted-ca-bundle\") pod \"console-f9d7485db-t5fjs\" (UID: \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\") " pod="openshift-console/console-f9d7485db-t5fjs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.934471 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/aef429a6-aa76-4776-a6e2-30096a5e7ccb-console-config\") pod \"console-f9d7485db-t5fjs\" (UID: \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\") " pod="openshift-console/console-f9d7485db-t5fjs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.935412 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2a0fdfc-5465-483d-8961-baa619f6ec0f-serving-cert\") pod \"console-operator-58897d9998-z2dld\" (UID: \"a2a0fdfc-5465-483d-8961-baa619f6ec0f\") " pod="openshift-console-operator/console-operator-58897d9998-z2dld" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.935421 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2f77876b-f997-4274-a619-a97e37bc4b9e-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-djs6p\" (UID: \"2f77876b-f997-4274-a619-a97e37bc4b9e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-djs6p" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.935586 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/acb2b70f-4a66-4290-aa8a-a13cb1b4b2a9-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-96268\" (UID: \"acb2b70f-4a66-4290-aa8a-a13cb1b4b2a9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-96268" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.935736 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/aef429a6-aa76-4776-a6e2-30096a5e7ccb-service-ca\") pod \"console-f9d7485db-t5fjs\" (UID: \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\") " pod="openshift-console/console-f9d7485db-t5fjs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.935740 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c42d89b6-6420-4267-93d2-e5dac6e1906b-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6ljl5\" (UID: \"c42d89b6-6420-4267-93d2-e5dac6e1906b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6ljl5" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.936849 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/aef429a6-aa76-4776-a6e2-30096a5e7ccb-console-serving-cert\") pod \"console-f9d7485db-t5fjs\" (UID: \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\") " pod="openshift-console/console-f9d7485db-t5fjs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.936983 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.936997 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/91336463-df6a-4b98-a990-e6780e417fab-metrics-tls\") pod \"ingress-operator-5b745b69d9-7d488\" (UID: \"91336463-df6a-4b98-a990-e6780e417fab\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7d488" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.937334 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/56177dd2-06b5-4d07-9a9b-395c558e2f13-profile-collector-cert\") pod \"catalog-operator-68c6474976-qzt4s\" (UID: \"56177dd2-06b5-4d07-9a9b-395c558e2f13\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qzt4s" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.937357 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/56177dd2-06b5-4d07-9a9b-395c558e2f13-srv-cert\") pod \"catalog-operator-68c6474976-qzt4s\" (UID: \"56177dd2-06b5-4d07-9a9b-395c558e2f13\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qzt4s" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.937610 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c7930b6-15dc-4462-bc47-9137c2d4edc7-serving-cert\") pod \"authentication-operator-69f744f599-zvzkq\" (UID: \"3c7930b6-15dc-4462-bc47-9137c2d4edc7\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zvzkq" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.951034 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.956797 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.970840 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 24 13:14:50 crc kubenswrapper[4790]: I1124 13:14:50.977095 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.008599 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.016188 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.016188 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.022115 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.031030 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.050895 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.057988 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.070408 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.073097 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a86ba968-c6dd-42c8-91fb-207bc839b8ec-audit-policies\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.090804 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.102843 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.111812 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.114038 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.135776 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.139625 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.156428 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.164125 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.170170 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.177237 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.190296 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.210667 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.216093 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c36b34a9-5050-4b55-886b-83c0dd80a3df-metrics-certs\") pod \"router-default-5444994796-4ldk4\" (UID: \"c36b34a9-5050-4b55-886b-83c0dd80a3df\") " pod="openshift-ingress/router-default-5444994796-4ldk4" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.231313 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.242991 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c36b34a9-5050-4b55-886b-83c0dd80a3df-stats-auth\") pod \"router-default-5444994796-4ldk4\" (UID: \"c36b34a9-5050-4b55-886b-83c0dd80a3df\") " pod="openshift-ingress/router-default-5444994796-4ldk4" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.250529 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.260709 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c36b34a9-5050-4b55-886b-83c0dd80a3df-service-ca-bundle\") pod \"router-default-5444994796-4ldk4\" (UID: \"c36b34a9-5050-4b55-886b-83c0dd80a3df\") " pod="openshift-ingress/router-default-5444994796-4ldk4" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.270951 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.292687 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.310906 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.324306 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c36b34a9-5050-4b55-886b-83c0dd80a3df-default-certificate\") pod \"router-default-5444994796-4ldk4\" (UID: \"c36b34a9-5050-4b55-886b-83c0dd80a3df\") " pod="openshift-ingress/router-default-5444994796-4ldk4" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.330914 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.351157 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.371143 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.390540 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.411582 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.432836 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.451541 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.470682 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.491246 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.511289 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.531213 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.550829 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.572373 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.582436 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/692c3c09-be9f-4400-b0fb-e8c8692f0d21-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5lzlv\" (UID: \"692c3c09-be9f-4400-b0fb-e8c8692f0d21\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5lzlv" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.591336 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.595118 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/692c3c09-be9f-4400-b0fb-e8c8692f0d21-config\") pod \"kube-controller-manager-operator-78b949d7b-5lzlv\" (UID: \"692c3c09-be9f-4400-b0fb-e8c8692f0d21\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5lzlv" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.651114 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.671192 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.692751 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.710845 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.732025 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.752391 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.771662 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.789792 4790 request.go:700] Waited for 1.017028517s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/secrets?fieldSelector=metadata.name%3Dmarketplace-operator-metrics&limit=500&resourceVersion=0 Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.792157 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.811478 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.842294 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.852548 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.872922 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.892182 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.911153 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.932206 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.951822 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.971174 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 24 13:14:51 crc kubenswrapper[4790]: I1124 13:14:51.991774 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.011810 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.032472 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.051865 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.070807 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.092659 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.111868 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.131980 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.151694 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.171948 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.191714 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.210718 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.231913 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.251693 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.271185 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.291229 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.311514 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.331439 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.351311 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.371750 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.391132 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.410637 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.432618 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.451686 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.471620 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.491434 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.536134 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpt4f\" (UniqueName: \"kubernetes.io/projected/5b226a93-5333-405f-93ad-788f9376faac-kube-api-access-gpt4f\") pod \"apiserver-7bbb656c7d-dc2xs\" (UID: \"5b226a93-5333-405f-93ad-788f9376faac\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.556462 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzgqr\" (UniqueName: \"kubernetes.io/projected/f50b116f-b25b-4985-8975-fbb523e1ddf6-kube-api-access-dzgqr\") pod \"route-controller-manager-6576b87f9c-wnlxq\" (UID: \"f50b116f-b25b-4985-8975-fbb523e1ddf6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wnlxq" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.559064 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.569244 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgrvw\" (UniqueName: \"kubernetes.io/projected/baa99c88-cee5-42d8-ac79-21551d465f7f-kube-api-access-pgrvw\") pod \"openshift-controller-manager-operator-756b6f6bc6-qwnbb\" (UID: \"baa99c88-cee5-42d8-ac79-21551d465f7f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qwnbb" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.580214 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qwnbb" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.587618 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2x49b\" (UniqueName: \"kubernetes.io/projected/69940b0d-4afc-4a7f-8ff7-cb128ce58913-kube-api-access-2x49b\") pod \"controller-manager-879f6c89f-cjdhk\" (UID: \"69940b0d-4afc-4a7f-8ff7-cb128ce58913\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cjdhk" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.605278 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dz26\" (UniqueName: \"kubernetes.io/projected/17718a81-aa70-4f5a-90bd-06e9251f4ef8-kube-api-access-9dz26\") pod \"machine-api-operator-5694c8668f-slkpm\" (UID: \"17718a81-aa70-4f5a-90bd-06e9251f4ef8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-slkpm" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.625822 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fb4mb\" (UniqueName: \"kubernetes.io/projected/194e75e8-4f01-4862-ba6e-a073a369df5a-kube-api-access-fb4mb\") pod \"cluster-samples-operator-665b6dd947-f55f4\" (UID: \"194e75e8-4f01-4862-ba6e-a073a369df5a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f55f4" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.650101 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xflx6\" (UniqueName: \"kubernetes.io/projected/66c28684-107f-4208-ae79-4b22a4e485dd-kube-api-access-xflx6\") pod \"apiserver-76f77b778f-8jkhv\" (UID: \"66c28684-107f-4208-ae79-4b22a4e485dd\") " pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.665583 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pc7f8\" (UniqueName: \"kubernetes.io/projected/2a2b45ef-e51f-4ad2-a1b7-6d8e7b0c089a-kube-api-access-pc7f8\") pod \"openshift-config-operator-7777fb866f-hsvx8\" (UID: \"2a2b45ef-e51f-4ad2-a1b7-6d8e7b0c089a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsvx8" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.685763 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mk45\" (UniqueName: \"kubernetes.io/projected/1e6bf7c7-7711-43cb-8817-73b42174b9ee-kube-api-access-4mk45\") pod \"machine-approver-56656f9798-5hlqm\" (UID: \"1e6bf7c7-7711-43cb-8817-73b42174b9ee\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5hlqm" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.691973 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.718254 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.730984 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.751472 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.766506 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-slkpm" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.771218 4790 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.775227 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qwnbb"] Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.778605 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wnlxq" Nov 24 13:14:52 crc kubenswrapper[4790]: W1124 13:14:52.786431 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbaa99c88_cee5_42d8_ac79_21551d465f7f.slice/crio-ecbe8827803fef53b3732e4621e9d6a0d0328cd6a869e1462b1db19f8ed6bb4d WatchSource:0}: Error finding container ecbe8827803fef53b3732e4621e9d6a0d0328cd6a869e1462b1db19f8ed6bb4d: Status 404 returned error can't find the container with id ecbe8827803fef53b3732e4621e9d6a0d0328cd6a869e1462b1db19f8ed6bb4d Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.790121 4790 request.go:700] Waited for 1.899645515s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/hostpath-provisioner/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.791556 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.799232 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs"] Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.799608 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsvx8" Nov 24 13:14:52 crc kubenswrapper[4790]: W1124 13:14:52.807194 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b226a93_5333_405f_93ad_788f9376faac.slice/crio-6642fbfca2978bee057895207c2854a9584dcd40218b9150b00bf90b773e84c2 WatchSource:0}: Error finding container 6642fbfca2978bee057895207c2854a9584dcd40218b9150b00bf90b773e84c2: Status 404 returned error can't find the container with id 6642fbfca2978bee057895207c2854a9584dcd40218b9150b00bf90b773e84c2 Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.810755 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.826302 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f55f4" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.830615 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.852438 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.869289 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-cjdhk" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.898208 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5hlqm" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.898763 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.899106 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2q6p\" (UniqueName: \"kubernetes.io/projected/acb2b70f-4a66-4290-aa8a-a13cb1b4b2a9-kube-api-access-g2q6p\") pod \"control-plane-machine-set-operator-78cbb6b69f-96268\" (UID: \"acb2b70f-4a66-4290-aa8a-a13cb1b4b2a9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-96268" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.923233 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c42d89b6-6420-4267-93d2-e5dac6e1906b-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6ljl5\" (UID: \"c42d89b6-6420-4267-93d2-e5dac6e1906b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6ljl5" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.932839 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llfmr\" (UniqueName: \"kubernetes.io/projected/a2a0fdfc-5465-483d-8961-baa619f6ec0f-kube-api-access-llfmr\") pod \"console-operator-58897d9998-z2dld\" (UID: \"a2a0fdfc-5465-483d-8961-baa619f6ec0f\") " pod="openshift-console-operator/console-operator-58897d9998-z2dld" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.945871 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5hlqm" event={"ID":"1e6bf7c7-7711-43cb-8817-73b42174b9ee","Type":"ContainerStarted","Data":"cb8995ebb2c1dee94cc200591bd3970d3e033e8604642a6d9ae0715c0160348e"} Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.951407 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2f77876b-f997-4274-a619-a97e37bc4b9e-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-djs6p\" (UID: \"2f77876b-f997-4274-a619-a97e37bc4b9e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-djs6p" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.952690 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" event={"ID":"5b226a93-5333-405f-93ad-788f9376faac","Type":"ContainerStarted","Data":"6642fbfca2978bee057895207c2854a9584dcd40218b9150b00bf90b773e84c2"} Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.955789 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qwnbb" event={"ID":"baa99c88-cee5-42d8-ac79-21551d465f7f","Type":"ContainerStarted","Data":"ecbe8827803fef53b3732e4621e9d6a0d0328cd6a869e1462b1db19f8ed6bb4d"} Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.967283 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94mzt\" (UniqueName: \"kubernetes.io/projected/2f77876b-f997-4274-a619-a97e37bc4b9e-kube-api-access-94mzt\") pod \"cluster-image-registry-operator-dc59b4c8b-djs6p\" (UID: \"2f77876b-f997-4274-a619-a97e37bc4b9e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-djs6p" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.972290 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-djs6p" Nov 24 13:14:52 crc kubenswrapper[4790]: I1124 13:14:52.986719 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/91336463-df6a-4b98-a990-e6780e417fab-bound-sa-token\") pod \"ingress-operator-5b745b69d9-7d488\" (UID: \"91336463-df6a-4b98-a990-e6780e417fab\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7d488" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.013137 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfndz\" (UniqueName: \"kubernetes.io/projected/aef429a6-aa76-4776-a6e2-30096a5e7ccb-kube-api-access-sfndz\") pod \"console-f9d7485db-t5fjs\" (UID: \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\") " pod="openshift-console/console-f9d7485db-t5fjs" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.018951 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-96268" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.030871 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-slkpm"] Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.033538 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5zww\" (UniqueName: \"kubernetes.io/projected/a86ba968-c6dd-42c8-91fb-207bc839b8ec-kube-api-access-t5zww\") pod \"oauth-openshift-558db77b4-zxkb7\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.053727 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dp5qr\" (UniqueName: \"kubernetes.io/projected/56177dd2-06b5-4d07-9a9b-395c558e2f13-kube-api-access-dp5qr\") pod \"catalog-operator-68c6474976-qzt4s\" (UID: \"56177dd2-06b5-4d07-9a9b-395c558e2f13\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qzt4s" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.059951 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-z2dld" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.068702 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.074124 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/692c3c09-be9f-4400-b0fb-e8c8692f0d21-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-5lzlv\" (UID: \"692c3c09-be9f-4400-b0fb-e8c8692f0d21\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5lzlv" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.088599 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlkv2\" (UniqueName: \"kubernetes.io/projected/91336463-df6a-4b98-a990-e6780e417fab-kube-api-access-zlkv2\") pod \"ingress-operator-5b745b69d9-7d488\" (UID: \"91336463-df6a-4b98-a990-e6780e417fab\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7d488" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.110502 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5lzlv" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.122194 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7478\" (UniqueName: \"kubernetes.io/projected/c36b34a9-5050-4b55-886b-83c0dd80a3df-kube-api-access-v7478\") pod \"router-default-5444994796-4ldk4\" (UID: \"c36b34a9-5050-4b55-886b-83c0dd80a3df\") " pod="openshift-ingress/router-default-5444994796-4ldk4" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.142509 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x67fz\" (UniqueName: \"kubernetes.io/projected/3c7930b6-15dc-4462-bc47-9137c2d4edc7-kube-api-access-x67fz\") pod \"authentication-operator-69f744f599-zvzkq\" (UID: \"3c7930b6-15dc-4462-bc47-9137c2d4edc7\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zvzkq" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.166029 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d441620d-6188-465c-8531-888b80826aa3-etcd-ca\") pod \"etcd-operator-b45778765-cppvr\" (UID: \"d441620d-6188-465c-8531-888b80826aa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cppvr" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.166115 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ffdd5f6c-5425-4895-8c2b-6ba6ef3706c9-profile-collector-cert\") pod \"olm-operator-6b444d44fb-xzd7l\" (UID: \"ffdd5f6c-5425-4895-8c2b-6ba6ef3706c9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xzd7l" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.166154 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d441620d-6188-465c-8531-888b80826aa3-etcd-client\") pod \"etcd-operator-b45778765-cppvr\" (UID: \"d441620d-6188-465c-8531-888b80826aa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cppvr" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.166221 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67792\" (UniqueName: \"kubernetes.io/projected/e4177bcc-76f8-43de-83ad-019fe23543e4-kube-api-access-67792\") pod \"multus-admission-controller-857f4d67dd-ktdmx\" (UID: \"e4177bcc-76f8-43de-83ad-019fe23543e4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ktdmx" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.166294 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6381724d-d16b-4d7b-9d9f-613451b93c77-registry-tls\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.166344 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9zzk\" (UniqueName: \"kubernetes.io/projected/6381724d-d16b-4d7b-9d9f-613451b93c77-kube-api-access-t9zzk\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.166390 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d441620d-6188-465c-8531-888b80826aa3-serving-cert\") pod \"etcd-operator-b45778765-cppvr\" (UID: \"d441620d-6188-465c-8531-888b80826aa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cppvr" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.166436 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d441620d-6188-465c-8531-888b80826aa3-etcd-service-ca\") pod \"etcd-operator-b45778765-cppvr\" (UID: \"d441620d-6188-465c-8531-888b80826aa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cppvr" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.166512 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd8b1442-101f-4104-ae72-cf6d2029f2d1-config\") pod \"kube-apiserver-operator-766d6c64bb-c4qnk\" (UID: \"fd8b1442-101f-4104-ae72-cf6d2029f2d1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c4qnk" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.166629 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.166749 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzlxr\" (UniqueName: \"kubernetes.io/projected/ebb47b17-21a5-4ae0-857d-36f913842d4a-kube-api-access-fzlxr\") pod \"migrator-59844c95c7-xcw79\" (UID: \"ebb47b17-21a5-4ae0-857d-36f913842d4a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xcw79" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.166816 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6381724d-d16b-4d7b-9d9f-613451b93c77-ca-trust-extracted\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.166867 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d441620d-6188-465c-8531-888b80826aa3-config\") pod \"etcd-operator-b45778765-cppvr\" (UID: \"d441620d-6188-465c-8531-888b80826aa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cppvr" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.166914 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2bpq\" (UniqueName: \"kubernetes.io/projected/207655ea-f838-43a7-ad87-2bee51686a16-kube-api-access-q2bpq\") pod \"downloads-7954f5f757-hkbxm\" (UID: \"207655ea-f838-43a7-ad87-2bee51686a16\") " pod="openshift-console/downloads-7954f5f757-hkbxm" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.166962 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mh76t\" (UniqueName: \"kubernetes.io/projected/d441620d-6188-465c-8531-888b80826aa3-kube-api-access-mh76t\") pod \"etcd-operator-b45778765-cppvr\" (UID: \"d441620d-6188-465c-8531-888b80826aa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cppvr" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.166989 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd8b1442-101f-4104-ae72-cf6d2029f2d1-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-c4qnk\" (UID: \"fd8b1442-101f-4104-ae72-cf6d2029f2d1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c4qnk" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.167010 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ffdd5f6c-5425-4895-8c2b-6ba6ef3706c9-srv-cert\") pod \"olm-operator-6b444d44fb-xzd7l\" (UID: \"ffdd5f6c-5425-4895-8c2b-6ba6ef3706c9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xzd7l" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.167069 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6381724d-d16b-4d7b-9d9f-613451b93c77-bound-sa-token\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.167097 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6381724d-d16b-4d7b-9d9f-613451b93c77-trusted-ca\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.167119 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e4177bcc-76f8-43de-83ad-019fe23543e4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ktdmx\" (UID: \"e4177bcc-76f8-43de-83ad-019fe23543e4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ktdmx" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.167143 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnr8b\" (UniqueName: \"kubernetes.io/projected/ffdd5f6c-5425-4895-8c2b-6ba6ef3706c9-kube-api-access-lnr8b\") pod \"olm-operator-6b444d44fb-xzd7l\" (UID: \"ffdd5f6c-5425-4895-8c2b-6ba6ef3706c9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xzd7l" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.167165 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fd8b1442-101f-4104-ae72-cf6d2029f2d1-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-c4qnk\" (UID: \"fd8b1442-101f-4104-ae72-cf6d2029f2d1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c4qnk" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.167186 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6381724d-d16b-4d7b-9d9f-613451b93c77-registry-certificates\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.167217 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6381724d-d16b-4d7b-9d9f-613451b93c77-installation-pull-secrets\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:53 crc kubenswrapper[4790]: E1124 13:14:53.168674 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:53.668658988 +0000 UTC m=+142.048552850 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.215172 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-t5fjs" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.219677 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6ljl5" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.267902 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:53 crc kubenswrapper[4790]: E1124 13:14:53.268053 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:53.768019407 +0000 UTC m=+142.147913069 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.268853 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkb7j\" (UniqueName: \"kubernetes.io/projected/d325cb10-86d1-414b-aa0a-282b8d6c1e30-kube-api-access-rkb7j\") pod \"openshift-apiserver-operator-796bbdcf4f-7x25b\" (UID: \"d325cb10-86d1-414b-aa0a-282b8d6c1e30\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7x25b" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.268899 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/7531eb66-63f6-4eef-b30a-c3fb1ab94f7a-certs\") pod \"machine-config-server-rhnfd\" (UID: \"7531eb66-63f6-4eef-b30a-c3fb1ab94f7a\") " pod="openshift-machine-config-operator/machine-config-server-rhnfd" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.269038 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j968g\" (UniqueName: \"kubernetes.io/projected/d059e3f7-ef28-4557-971c-e77be1472c52-kube-api-access-j968g\") pod \"ingress-canary-lnmz8\" (UID: \"d059e3f7-ef28-4557-971c-e77be1472c52\") " pod="openshift-ingress-canary/ingress-canary-lnmz8" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.269076 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6381724d-d16b-4d7b-9d9f-613451b93c77-trusted-ca\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.269095 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e4177bcc-76f8-43de-83ad-019fe23543e4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ktdmx\" (UID: \"e4177bcc-76f8-43de-83ad-019fe23543e4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ktdmx" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.269114 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/951c1cb8-6613-404d-b0ec-ca5c8b66ec30-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-s7x26\" (UID: \"951c1cb8-6613-404d-b0ec-ca5c8b66ec30\") " pod="openshift-marketplace/marketplace-operator-79b997595-s7x26" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.269133 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnr8b\" (UniqueName: \"kubernetes.io/projected/ffdd5f6c-5425-4895-8c2b-6ba6ef3706c9-kube-api-access-lnr8b\") pod \"olm-operator-6b444d44fb-xzd7l\" (UID: \"ffdd5f6c-5425-4895-8c2b-6ba6ef3706c9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xzd7l" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.269777 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7bc385a4-1703-4a76-b27e-8ae71480381d-signing-key\") pod \"service-ca-9c57cc56f-54x2b\" (UID: \"7bc385a4-1703-4a76-b27e-8ae71480381d\") " pod="openshift-service-ca/service-ca-9c57cc56f-54x2b" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.269823 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fd8b1442-101f-4104-ae72-cf6d2029f2d1-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-c4qnk\" (UID: \"fd8b1442-101f-4104-ae72-cf6d2029f2d1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c4qnk" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.269842 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdfzj\" (UniqueName: \"kubernetes.io/projected/7bc385a4-1703-4a76-b27e-8ae71480381d-kube-api-access-zdfzj\") pod \"service-ca-9c57cc56f-54x2b\" (UID: \"7bc385a4-1703-4a76-b27e-8ae71480381d\") " pod="openshift-service-ca/service-ca-9c57cc56f-54x2b" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.269911 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6381724d-d16b-4d7b-9d9f-613451b93c77-registry-certificates\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.269967 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6381724d-d16b-4d7b-9d9f-613451b93c77-installation-pull-secrets\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.269988 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/f6b07b0b-f08b-4ad5-a390-86a8119e96c9-csi-data-dir\") pod \"csi-hostpathplugin-4t9zv\" (UID: \"f6b07b0b-f08b-4ad5-a390-86a8119e96c9\") " pod="hostpath-provisioner/csi-hostpathplugin-4t9zv" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.270009 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ql2q4\" (UniqueName: \"kubernetes.io/projected/e9a8e814-f2c2-46d1-99c9-40d77e0d96da-kube-api-access-ql2q4\") pod \"machine-config-controller-84d6567774-6h2jt\" (UID: \"e9a8e814-f2c2-46d1-99c9-40d77e0d96da\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6h2jt" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.270034 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/f6b07b0b-f08b-4ad5-a390-86a8119e96c9-plugins-dir\") pod \"csi-hostpathplugin-4t9zv\" (UID: \"f6b07b0b-f08b-4ad5-a390-86a8119e96c9\") " pod="hostpath-provisioner/csi-hostpathplugin-4t9zv" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.270122 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d059e3f7-ef28-4557-971c-e77be1472c52-cert\") pod \"ingress-canary-lnmz8\" (UID: \"d059e3f7-ef28-4557-971c-e77be1472c52\") " pod="openshift-ingress-canary/ingress-canary-lnmz8" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.270150 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/f6b07b0b-f08b-4ad5-a390-86a8119e96c9-mountpoint-dir\") pod \"csi-hostpathplugin-4t9zv\" (UID: \"f6b07b0b-f08b-4ad5-a390-86a8119e96c9\") " pod="hostpath-provisioner/csi-hostpathplugin-4t9zv" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.270180 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1d6e992d-904f-4f11-ab29-29583f1dd955-apiservice-cert\") pod \"packageserver-d55dfcdfc-79gj2\" (UID: \"1d6e992d-904f-4f11-ab29-29583f1dd955\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-79gj2" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.270223 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d441620d-6188-465c-8531-888b80826aa3-etcd-ca\") pod \"etcd-operator-b45778765-cppvr\" (UID: \"d441620d-6188-465c-8531-888b80826aa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cppvr" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.270245 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96dzd\" (UniqueName: \"kubernetes.io/projected/163ef77c-3b2f-45ef-81c8-1fc627b1a234-kube-api-access-96dzd\") pod \"dns-default-csctv\" (UID: \"163ef77c-3b2f-45ef-81c8-1fc627b1a234\") " pod="openshift-dns/dns-default-csctv" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.270277 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e68b45a-0978-41b1-ba4c-0b75dc67399e-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-6g5fd\" (UID: \"4e68b45a-0978-41b1-ba4c-0b75dc67399e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6g5fd" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.270304 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ffdd5f6c-5425-4895-8c2b-6ba6ef3706c9-profile-collector-cert\") pod \"olm-operator-6b444d44fb-xzd7l\" (UID: \"ffdd5f6c-5425-4895-8c2b-6ba6ef3706c9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xzd7l" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.270368 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/163ef77c-3b2f-45ef-81c8-1fc627b1a234-metrics-tls\") pod \"dns-default-csctv\" (UID: \"163ef77c-3b2f-45ef-81c8-1fc627b1a234\") " pod="openshift-dns/dns-default-csctv" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.270440 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4e68b45a-0978-41b1-ba4c-0b75dc67399e-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-6g5fd\" (UID: \"4e68b45a-0978-41b1-ba4c-0b75dc67399e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6g5fd" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.270458 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d325cb10-86d1-414b-aa0a-282b8d6c1e30-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7x25b\" (UID: \"d325cb10-86d1-414b-aa0a-282b8d6c1e30\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7x25b" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.270490 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/40f41556-7a45-457f-8f60-5fc87df20363-auth-proxy-config\") pod \"machine-config-operator-74547568cd-fscrx\" (UID: \"40f41556-7a45-457f-8f60-5fc87df20363\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fscrx" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.270510 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d441620d-6188-465c-8531-888b80826aa3-etcd-client\") pod \"etcd-operator-b45778765-cppvr\" (UID: \"d441620d-6188-465c-8531-888b80826aa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cppvr" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.270526 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7bc385a4-1703-4a76-b27e-8ae71480381d-signing-cabundle\") pod \"service-ca-9c57cc56f-54x2b\" (UID: \"7bc385a4-1703-4a76-b27e-8ae71480381d\") " pod="openshift-service-ca/service-ca-9c57cc56f-54x2b" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.270542 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qn6gs\" (UniqueName: \"kubernetes.io/projected/40f41556-7a45-457f-8f60-5fc87df20363-kube-api-access-qn6gs\") pod \"machine-config-operator-74547568cd-fscrx\" (UID: \"40f41556-7a45-457f-8f60-5fc87df20363\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fscrx" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.270599 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/7531eb66-63f6-4eef-b30a-c3fb1ab94f7a-node-bootstrap-token\") pod \"machine-config-server-rhnfd\" (UID: \"7531eb66-63f6-4eef-b30a-c3fb1ab94f7a\") " pod="openshift-machine-config-operator/machine-config-server-rhnfd" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.270620 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67792\" (UniqueName: \"kubernetes.io/projected/e4177bcc-76f8-43de-83ad-019fe23543e4-kube-api-access-67792\") pod \"multus-admission-controller-857f4d67dd-ktdmx\" (UID: \"e4177bcc-76f8-43de-83ad-019fe23543e4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ktdmx" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.270638 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba3cf9ac-3e57-4274-b5ea-727bf77aebf1-config\") pod \"service-ca-operator-777779d784-hx4vr\" (UID: \"ba3cf9ac-3e57-4274-b5ea-727bf77aebf1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hx4vr" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.270657 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9fn8\" (UniqueName: \"kubernetes.io/projected/586ab369-00f4-497e-836f-3aede32089c2-kube-api-access-g9fn8\") pod \"collect-profiles-29399820-jr7cd\" (UID: \"586ab369-00f4-497e-836f-3aede32089c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399820-jr7cd" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.270677 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/1d6e992d-904f-4f11-ab29-29583f1dd955-tmpfs\") pod \"packageserver-d55dfcdfc-79gj2\" (UID: \"1d6e992d-904f-4f11-ab29-29583f1dd955\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-79gj2" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.270706 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6381724d-d16b-4d7b-9d9f-613451b93c77-registry-tls\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.270727 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9zzk\" (UniqueName: \"kubernetes.io/projected/6381724d-d16b-4d7b-9d9f-613451b93c77-kube-api-access-t9zzk\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.270756 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/586ab369-00f4-497e-836f-3aede32089c2-secret-volume\") pod \"collect-profiles-29399820-jr7cd\" (UID: \"586ab369-00f4-497e-836f-3aede32089c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399820-jr7cd" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.271065 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d441620d-6188-465c-8531-888b80826aa3-serving-cert\") pod \"etcd-operator-b45778765-cppvr\" (UID: \"d441620d-6188-465c-8531-888b80826aa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cppvr" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.271087 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d441620d-6188-465c-8531-888b80826aa3-etcd-service-ca\") pod \"etcd-operator-b45778765-cppvr\" (UID: \"d441620d-6188-465c-8531-888b80826aa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cppvr" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.271153 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd8b1442-101f-4104-ae72-cf6d2029f2d1-config\") pod \"kube-apiserver-operator-766d6c64bb-c4qnk\" (UID: \"fd8b1442-101f-4104-ae72-cf6d2029f2d1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c4qnk" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.271193 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6aa4a7b3-bae3-40bb-950d-e907073c0257-metrics-tls\") pod \"dns-operator-744455d44c-wvnp9\" (UID: \"6aa4a7b3-bae3-40bb-950d-e907073c0257\") " pod="openshift-dns-operator/dns-operator-744455d44c-wvnp9" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.271247 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.271318 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/d6761cdb-c9a1-4a58-9023-e1230eab220d-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-bfmvc\" (UID: \"d6761cdb-c9a1-4a58-9023-e1230eab220d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bfmvc" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.271359 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/40f41556-7a45-457f-8f60-5fc87df20363-images\") pod \"machine-config-operator-74547568cd-fscrx\" (UID: \"40f41556-7a45-457f-8f60-5fc87df20363\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fscrx" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.271375 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/40f41556-7a45-457f-8f60-5fc87df20363-proxy-tls\") pod \"machine-config-operator-74547568cd-fscrx\" (UID: \"40f41556-7a45-457f-8f60-5fc87df20363\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fscrx" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.271391 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxdh4\" (UniqueName: \"kubernetes.io/projected/ba3cf9ac-3e57-4274-b5ea-727bf77aebf1-kube-api-access-qxdh4\") pod \"service-ca-operator-777779d784-hx4vr\" (UID: \"ba3cf9ac-3e57-4274-b5ea-727bf77aebf1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hx4vr" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.271463 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzlxr\" (UniqueName: \"kubernetes.io/projected/ebb47b17-21a5-4ae0-857d-36f913842d4a-kube-api-access-fzlxr\") pod \"migrator-59844c95c7-xcw79\" (UID: \"ebb47b17-21a5-4ae0-857d-36f913842d4a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xcw79" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.271488 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhzq4\" (UniqueName: \"kubernetes.io/projected/d6761cdb-c9a1-4a58-9023-e1230eab220d-kube-api-access-qhzq4\") pod \"package-server-manager-789f6589d5-bfmvc\" (UID: \"d6761cdb-c9a1-4a58-9023-e1230eab220d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bfmvc" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.271489 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6381724d-d16b-4d7b-9d9f-613451b93c77-trusted-ca\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.271560 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkvp6\" (UniqueName: \"kubernetes.io/projected/4e68b45a-0978-41b1-ba4c-0b75dc67399e-kube-api-access-hkvp6\") pod \"kube-storage-version-migrator-operator-b67b599dd-6g5fd\" (UID: \"4e68b45a-0978-41b1-ba4c-0b75dc67399e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6g5fd" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.271614 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6381724d-d16b-4d7b-9d9f-613451b93c77-ca-trust-extracted\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.271631 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f6b07b0b-f08b-4ad5-a390-86a8119e96c9-socket-dir\") pod \"csi-hostpathplugin-4t9zv\" (UID: \"f6b07b0b-f08b-4ad5-a390-86a8119e96c9\") " pod="hostpath-provisioner/csi-hostpathplugin-4t9zv" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.271651 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9dkb\" (UniqueName: \"kubernetes.io/projected/f6b07b0b-f08b-4ad5-a390-86a8119e96c9-kube-api-access-d9dkb\") pod \"csi-hostpathplugin-4t9zv\" (UID: \"f6b07b0b-f08b-4ad5-a390-86a8119e96c9\") " pod="hostpath-provisioner/csi-hostpathplugin-4t9zv" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.271702 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba3cf9ac-3e57-4274-b5ea-727bf77aebf1-serving-cert\") pod \"service-ca-operator-777779d784-hx4vr\" (UID: \"ba3cf9ac-3e57-4274-b5ea-727bf77aebf1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hx4vr" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.271738 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d441620d-6188-465c-8531-888b80826aa3-config\") pod \"etcd-operator-b45778765-cppvr\" (UID: \"d441620d-6188-465c-8531-888b80826aa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cppvr" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.271779 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dq57\" (UniqueName: \"kubernetes.io/projected/951c1cb8-6613-404d-b0ec-ca5c8b66ec30-kube-api-access-2dq57\") pod \"marketplace-operator-79b997595-s7x26\" (UID: \"951c1cb8-6613-404d-b0ec-ca5c8b66ec30\") " pod="openshift-marketplace/marketplace-operator-79b997595-s7x26" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.271799 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/586ab369-00f4-497e-836f-3aede32089c2-config-volume\") pod \"collect-profiles-29399820-jr7cd\" (UID: \"586ab369-00f4-497e-836f-3aede32089c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399820-jr7cd" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.271851 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2bpq\" (UniqueName: \"kubernetes.io/projected/207655ea-f838-43a7-ad87-2bee51686a16-kube-api-access-q2bpq\") pod \"downloads-7954f5f757-hkbxm\" (UID: \"207655ea-f838-43a7-ad87-2bee51686a16\") " pod="openshift-console/downloads-7954f5f757-hkbxm" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.271872 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gc5zb\" (UniqueName: \"kubernetes.io/projected/6aa4a7b3-bae3-40bb-950d-e907073c0257-kube-api-access-gc5zb\") pod \"dns-operator-744455d44c-wvnp9\" (UID: \"6aa4a7b3-bae3-40bb-950d-e907073c0257\") " pod="openshift-dns-operator/dns-operator-744455d44c-wvnp9" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.271921 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f6b07b0b-f08b-4ad5-a390-86a8119e96c9-registration-dir\") pod \"csi-hostpathplugin-4t9zv\" (UID: \"f6b07b0b-f08b-4ad5-a390-86a8119e96c9\") " pod="hostpath-provisioner/csi-hostpathplugin-4t9zv" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.271937 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2nxf\" (UniqueName: \"kubernetes.io/projected/7531eb66-63f6-4eef-b30a-c3fb1ab94f7a-kube-api-access-z2nxf\") pod \"machine-config-server-rhnfd\" (UID: \"7531eb66-63f6-4eef-b30a-c3fb1ab94f7a\") " pod="openshift-machine-config-operator/machine-config-server-rhnfd" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.272309 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jq4n4\" (UniqueName: \"kubernetes.io/projected/1d6e992d-904f-4f11-ab29-29583f1dd955-kube-api-access-jq4n4\") pod \"packageserver-d55dfcdfc-79gj2\" (UID: \"1d6e992d-904f-4f11-ab29-29583f1dd955\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-79gj2" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.272344 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1d6e992d-904f-4f11-ab29-29583f1dd955-webhook-cert\") pod \"packageserver-d55dfcdfc-79gj2\" (UID: \"1d6e992d-904f-4f11-ab29-29583f1dd955\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-79gj2" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.272363 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mh76t\" (UniqueName: \"kubernetes.io/projected/d441620d-6188-465c-8531-888b80826aa3-kube-api-access-mh76t\") pod \"etcd-operator-b45778765-cppvr\" (UID: \"d441620d-6188-465c-8531-888b80826aa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cppvr" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.272385 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd8b1442-101f-4104-ae72-cf6d2029f2d1-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-c4qnk\" (UID: \"fd8b1442-101f-4104-ae72-cf6d2029f2d1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c4qnk" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.272408 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ffdd5f6c-5425-4895-8c2b-6ba6ef3706c9-srv-cert\") pod \"olm-operator-6b444d44fb-xzd7l\" (UID: \"ffdd5f6c-5425-4895-8c2b-6ba6ef3706c9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xzd7l" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.272452 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e9a8e814-f2c2-46d1-99c9-40d77e0d96da-proxy-tls\") pod \"machine-config-controller-84d6567774-6h2jt\" (UID: \"e9a8e814-f2c2-46d1-99c9-40d77e0d96da\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6h2jt" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.272475 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/951c1cb8-6613-404d-b0ec-ca5c8b66ec30-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-s7x26\" (UID: \"951c1cb8-6613-404d-b0ec-ca5c8b66ec30\") " pod="openshift-marketplace/marketplace-operator-79b997595-s7x26" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.272495 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e9a8e814-f2c2-46d1-99c9-40d77e0d96da-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-6h2jt\" (UID: \"e9a8e814-f2c2-46d1-99c9-40d77e0d96da\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6h2jt" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.272530 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d325cb10-86d1-414b-aa0a-282b8d6c1e30-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7x25b\" (UID: \"d325cb10-86d1-414b-aa0a-282b8d6c1e30\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7x25b" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.272615 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6381724d-d16b-4d7b-9d9f-613451b93c77-bound-sa-token\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.272646 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/163ef77c-3b2f-45ef-81c8-1fc627b1a234-config-volume\") pod \"dns-default-csctv\" (UID: \"163ef77c-3b2f-45ef-81c8-1fc627b1a234\") " pod="openshift-dns/dns-default-csctv" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.274074 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6381724d-d16b-4d7b-9d9f-613451b93c77-registry-certificates\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.274629 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6381724d-d16b-4d7b-9d9f-613451b93c77-ca-trust-extracted\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.274917 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d441620d-6188-465c-8531-888b80826aa3-etcd-ca\") pod \"etcd-operator-b45778765-cppvr\" (UID: \"d441620d-6188-465c-8531-888b80826aa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cppvr" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.275866 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d441620d-6188-465c-8531-888b80826aa3-config\") pod \"etcd-operator-b45778765-cppvr\" (UID: \"d441620d-6188-465c-8531-888b80826aa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cppvr" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.276748 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd8b1442-101f-4104-ae72-cf6d2029f2d1-config\") pod \"kube-apiserver-operator-766d6c64bb-c4qnk\" (UID: \"fd8b1442-101f-4104-ae72-cf6d2029f2d1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c4qnk" Nov 24 13:14:53 crc kubenswrapper[4790]: E1124 13:14:53.277324 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:53.777308589 +0000 UTC m=+142.157202251 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.279259 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d441620d-6188-465c-8531-888b80826aa3-etcd-service-ca\") pod \"etcd-operator-b45778765-cppvr\" (UID: \"d441620d-6188-465c-8531-888b80826aa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cppvr" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.283069 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-zvzkq" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.284448 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd8b1442-101f-4104-ae72-cf6d2029f2d1-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-c4qnk\" (UID: \"fd8b1442-101f-4104-ae72-cf6d2029f2d1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c4qnk" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.286441 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ffdd5f6c-5425-4895-8c2b-6ba6ef3706c9-profile-collector-cert\") pod \"olm-operator-6b444d44fb-xzd7l\" (UID: \"ffdd5f6c-5425-4895-8c2b-6ba6ef3706c9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xzd7l" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.288188 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wnlxq"] Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.289670 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6381724d-d16b-4d7b-9d9f-613451b93c77-installation-pull-secrets\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.290860 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6381724d-d16b-4d7b-9d9f-613451b93c77-registry-tls\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.291344 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e4177bcc-76f8-43de-83ad-019fe23543e4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ktdmx\" (UID: \"e4177bcc-76f8-43de-83ad-019fe23543e4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ktdmx" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.293453 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7d488" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.293542 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d441620d-6188-465c-8531-888b80826aa3-etcd-client\") pod \"etcd-operator-b45778765-cppvr\" (UID: \"d441620d-6188-465c-8531-888b80826aa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cppvr" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.294164 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d441620d-6188-465c-8531-888b80826aa3-serving-cert\") pod \"etcd-operator-b45778765-cppvr\" (UID: \"d441620d-6188-465c-8531-888b80826aa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cppvr" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.296151 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ffdd5f6c-5425-4895-8c2b-6ba6ef3706c9-srv-cert\") pod \"olm-operator-6b444d44fb-xzd7l\" (UID: \"ffdd5f6c-5425-4895-8c2b-6ba6ef3706c9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xzd7l" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.304689 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qzt4s" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.310318 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnr8b\" (UniqueName: \"kubernetes.io/projected/ffdd5f6c-5425-4895-8c2b-6ba6ef3706c9-kube-api-access-lnr8b\") pod \"olm-operator-6b444d44fb-xzd7l\" (UID: \"ffdd5f6c-5425-4895-8c2b-6ba6ef3706c9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xzd7l" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.331394 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67792\" (UniqueName: \"kubernetes.io/projected/e4177bcc-76f8-43de-83ad-019fe23543e4-kube-api-access-67792\") pod \"multus-admission-controller-857f4d67dd-ktdmx\" (UID: \"e4177bcc-76f8-43de-83ad-019fe23543e4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ktdmx" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.351729 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2bpq\" (UniqueName: \"kubernetes.io/projected/207655ea-f838-43a7-ad87-2bee51686a16-kube-api-access-q2bpq\") pod \"downloads-7954f5f757-hkbxm\" (UID: \"207655ea-f838-43a7-ad87-2bee51686a16\") " pod="openshift-console/downloads-7954f5f757-hkbxm" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.373328 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9zzk\" (UniqueName: \"kubernetes.io/projected/6381724d-d16b-4d7b-9d9f-613451b93c77-kube-api-access-t9zzk\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.373969 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.374117 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkb7j\" (UniqueName: \"kubernetes.io/projected/d325cb10-86d1-414b-aa0a-282b8d6c1e30-kube-api-access-rkb7j\") pod \"openshift-apiserver-operator-796bbdcf4f-7x25b\" (UID: \"d325cb10-86d1-414b-aa0a-282b8d6c1e30\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7x25b" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.374147 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/7531eb66-63f6-4eef-b30a-c3fb1ab94f7a-certs\") pod \"machine-config-server-rhnfd\" (UID: \"7531eb66-63f6-4eef-b30a-c3fb1ab94f7a\") " pod="openshift-machine-config-operator/machine-config-server-rhnfd" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.374171 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j968g\" (UniqueName: \"kubernetes.io/projected/d059e3f7-ef28-4557-971c-e77be1472c52-kube-api-access-j968g\") pod \"ingress-canary-lnmz8\" (UID: \"d059e3f7-ef28-4557-971c-e77be1472c52\") " pod="openshift-ingress-canary/ingress-canary-lnmz8" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.374195 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/951c1cb8-6613-404d-b0ec-ca5c8b66ec30-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-s7x26\" (UID: \"951c1cb8-6613-404d-b0ec-ca5c8b66ec30\") " pod="openshift-marketplace/marketplace-operator-79b997595-s7x26" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.374219 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7bc385a4-1703-4a76-b27e-8ae71480381d-signing-key\") pod \"service-ca-9c57cc56f-54x2b\" (UID: \"7bc385a4-1703-4a76-b27e-8ae71480381d\") " pod="openshift-service-ca/service-ca-9c57cc56f-54x2b" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.374254 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdfzj\" (UniqueName: \"kubernetes.io/projected/7bc385a4-1703-4a76-b27e-8ae71480381d-kube-api-access-zdfzj\") pod \"service-ca-9c57cc56f-54x2b\" (UID: \"7bc385a4-1703-4a76-b27e-8ae71480381d\") " pod="openshift-service-ca/service-ca-9c57cc56f-54x2b" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.374287 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/f6b07b0b-f08b-4ad5-a390-86a8119e96c9-csi-data-dir\") pod \"csi-hostpathplugin-4t9zv\" (UID: \"f6b07b0b-f08b-4ad5-a390-86a8119e96c9\") " pod="hostpath-provisioner/csi-hostpathplugin-4t9zv" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.374307 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ql2q4\" (UniqueName: \"kubernetes.io/projected/e9a8e814-f2c2-46d1-99c9-40d77e0d96da-kube-api-access-ql2q4\") pod \"machine-config-controller-84d6567774-6h2jt\" (UID: \"e9a8e814-f2c2-46d1-99c9-40d77e0d96da\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6h2jt" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.374324 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/f6b07b0b-f08b-4ad5-a390-86a8119e96c9-plugins-dir\") pod \"csi-hostpathplugin-4t9zv\" (UID: \"f6b07b0b-f08b-4ad5-a390-86a8119e96c9\") " pod="hostpath-provisioner/csi-hostpathplugin-4t9zv" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.374342 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d059e3f7-ef28-4557-971c-e77be1472c52-cert\") pod \"ingress-canary-lnmz8\" (UID: \"d059e3f7-ef28-4557-971c-e77be1472c52\") " pod="openshift-ingress-canary/ingress-canary-lnmz8" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.374366 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/f6b07b0b-f08b-4ad5-a390-86a8119e96c9-mountpoint-dir\") pod \"csi-hostpathplugin-4t9zv\" (UID: \"f6b07b0b-f08b-4ad5-a390-86a8119e96c9\") " pod="hostpath-provisioner/csi-hostpathplugin-4t9zv" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.374390 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1d6e992d-904f-4f11-ab29-29583f1dd955-apiservice-cert\") pod \"packageserver-d55dfcdfc-79gj2\" (UID: \"1d6e992d-904f-4f11-ab29-29583f1dd955\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-79gj2" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.374412 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96dzd\" (UniqueName: \"kubernetes.io/projected/163ef77c-3b2f-45ef-81c8-1fc627b1a234-kube-api-access-96dzd\") pod \"dns-default-csctv\" (UID: \"163ef77c-3b2f-45ef-81c8-1fc627b1a234\") " pod="openshift-dns/dns-default-csctv" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.374436 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e68b45a-0978-41b1-ba4c-0b75dc67399e-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-6g5fd\" (UID: \"4e68b45a-0978-41b1-ba4c-0b75dc67399e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6g5fd" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.374458 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/163ef77c-3b2f-45ef-81c8-1fc627b1a234-metrics-tls\") pod \"dns-default-csctv\" (UID: \"163ef77c-3b2f-45ef-81c8-1fc627b1a234\") " pod="openshift-dns/dns-default-csctv" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.374479 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4e68b45a-0978-41b1-ba4c-0b75dc67399e-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-6g5fd\" (UID: \"4e68b45a-0978-41b1-ba4c-0b75dc67399e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6g5fd" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.374502 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d325cb10-86d1-414b-aa0a-282b8d6c1e30-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7x25b\" (UID: \"d325cb10-86d1-414b-aa0a-282b8d6c1e30\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7x25b" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.374526 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/40f41556-7a45-457f-8f60-5fc87df20363-auth-proxy-config\") pod \"machine-config-operator-74547568cd-fscrx\" (UID: \"40f41556-7a45-457f-8f60-5fc87df20363\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fscrx" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.374547 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7bc385a4-1703-4a76-b27e-8ae71480381d-signing-cabundle\") pod \"service-ca-9c57cc56f-54x2b\" (UID: \"7bc385a4-1703-4a76-b27e-8ae71480381d\") " pod="openshift-service-ca/service-ca-9c57cc56f-54x2b" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.374569 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qn6gs\" (UniqueName: \"kubernetes.io/projected/40f41556-7a45-457f-8f60-5fc87df20363-kube-api-access-qn6gs\") pod \"machine-config-operator-74547568cd-fscrx\" (UID: \"40f41556-7a45-457f-8f60-5fc87df20363\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fscrx" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.374601 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/7531eb66-63f6-4eef-b30a-c3fb1ab94f7a-node-bootstrap-token\") pod \"machine-config-server-rhnfd\" (UID: \"7531eb66-63f6-4eef-b30a-c3fb1ab94f7a\") " pod="openshift-machine-config-operator/machine-config-server-rhnfd" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.374626 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba3cf9ac-3e57-4274-b5ea-727bf77aebf1-config\") pod \"service-ca-operator-777779d784-hx4vr\" (UID: \"ba3cf9ac-3e57-4274-b5ea-727bf77aebf1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hx4vr" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.374649 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9fn8\" (UniqueName: \"kubernetes.io/projected/586ab369-00f4-497e-836f-3aede32089c2-kube-api-access-g9fn8\") pod \"collect-profiles-29399820-jr7cd\" (UID: \"586ab369-00f4-497e-836f-3aede32089c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399820-jr7cd" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.374680 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/1d6e992d-904f-4f11-ab29-29583f1dd955-tmpfs\") pod \"packageserver-d55dfcdfc-79gj2\" (UID: \"1d6e992d-904f-4f11-ab29-29583f1dd955\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-79gj2" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.374714 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/586ab369-00f4-497e-836f-3aede32089c2-secret-volume\") pod \"collect-profiles-29399820-jr7cd\" (UID: \"586ab369-00f4-497e-836f-3aede32089c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399820-jr7cd" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.374746 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6aa4a7b3-bae3-40bb-950d-e907073c0257-metrics-tls\") pod \"dns-operator-744455d44c-wvnp9\" (UID: \"6aa4a7b3-bae3-40bb-950d-e907073c0257\") " pod="openshift-dns-operator/dns-operator-744455d44c-wvnp9" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.374786 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/d6761cdb-c9a1-4a58-9023-e1230eab220d-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-bfmvc\" (UID: \"d6761cdb-c9a1-4a58-9023-e1230eab220d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bfmvc" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.374813 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/40f41556-7a45-457f-8f60-5fc87df20363-images\") pod \"machine-config-operator-74547568cd-fscrx\" (UID: \"40f41556-7a45-457f-8f60-5fc87df20363\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fscrx" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.374825 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/f6b07b0b-f08b-4ad5-a390-86a8119e96c9-mountpoint-dir\") pod \"csi-hostpathplugin-4t9zv\" (UID: \"f6b07b0b-f08b-4ad5-a390-86a8119e96c9\") " pod="hostpath-provisioner/csi-hostpathplugin-4t9zv" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.376006 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e68b45a-0978-41b1-ba4c-0b75dc67399e-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-6g5fd\" (UID: \"4e68b45a-0978-41b1-ba4c-0b75dc67399e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6g5fd" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.377615 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-4ldk4" Nov 24 13:14:53 crc kubenswrapper[4790]: E1124 13:14:53.377627 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:53.877610238 +0000 UTC m=+142.257503900 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.374837 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/40f41556-7a45-457f-8f60-5fc87df20363-proxy-tls\") pod \"machine-config-operator-74547568cd-fscrx\" (UID: \"40f41556-7a45-457f-8f60-5fc87df20363\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fscrx" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.379186 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxdh4\" (UniqueName: \"kubernetes.io/projected/ba3cf9ac-3e57-4274-b5ea-727bf77aebf1-kube-api-access-qxdh4\") pod \"service-ca-operator-777779d784-hx4vr\" (UID: \"ba3cf9ac-3e57-4274-b5ea-727bf77aebf1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hx4vr" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.379287 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhzq4\" (UniqueName: \"kubernetes.io/projected/d6761cdb-c9a1-4a58-9023-e1230eab220d-kube-api-access-qhzq4\") pod \"package-server-manager-789f6589d5-bfmvc\" (UID: \"d6761cdb-c9a1-4a58-9023-e1230eab220d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bfmvc" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.379325 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkvp6\" (UniqueName: \"kubernetes.io/projected/4e68b45a-0978-41b1-ba4c-0b75dc67399e-kube-api-access-hkvp6\") pod \"kube-storage-version-migrator-operator-b67b599dd-6g5fd\" (UID: \"4e68b45a-0978-41b1-ba4c-0b75dc67399e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6g5fd" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.379351 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f6b07b0b-f08b-4ad5-a390-86a8119e96c9-socket-dir\") pod \"csi-hostpathplugin-4t9zv\" (UID: \"f6b07b0b-f08b-4ad5-a390-86a8119e96c9\") " pod="hostpath-provisioner/csi-hostpathplugin-4t9zv" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.379369 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9dkb\" (UniqueName: \"kubernetes.io/projected/f6b07b0b-f08b-4ad5-a390-86a8119e96c9-kube-api-access-d9dkb\") pod \"csi-hostpathplugin-4t9zv\" (UID: \"f6b07b0b-f08b-4ad5-a390-86a8119e96c9\") " pod="hostpath-provisioner/csi-hostpathplugin-4t9zv" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.379391 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba3cf9ac-3e57-4274-b5ea-727bf77aebf1-serving-cert\") pod \"service-ca-operator-777779d784-hx4vr\" (UID: \"ba3cf9ac-3e57-4274-b5ea-727bf77aebf1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hx4vr" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.379410 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dq57\" (UniqueName: \"kubernetes.io/projected/951c1cb8-6613-404d-b0ec-ca5c8b66ec30-kube-api-access-2dq57\") pod \"marketplace-operator-79b997595-s7x26\" (UID: \"951c1cb8-6613-404d-b0ec-ca5c8b66ec30\") " pod="openshift-marketplace/marketplace-operator-79b997595-s7x26" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.379425 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/586ab369-00f4-497e-836f-3aede32089c2-config-volume\") pod \"collect-profiles-29399820-jr7cd\" (UID: \"586ab369-00f4-497e-836f-3aede32089c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399820-jr7cd" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.379445 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gc5zb\" (UniqueName: \"kubernetes.io/projected/6aa4a7b3-bae3-40bb-950d-e907073c0257-kube-api-access-gc5zb\") pod \"dns-operator-744455d44c-wvnp9\" (UID: \"6aa4a7b3-bae3-40bb-950d-e907073c0257\") " pod="openshift-dns-operator/dns-operator-744455d44c-wvnp9" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.379465 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f6b07b0b-f08b-4ad5-a390-86a8119e96c9-registration-dir\") pod \"csi-hostpathplugin-4t9zv\" (UID: \"f6b07b0b-f08b-4ad5-a390-86a8119e96c9\") " pod="hostpath-provisioner/csi-hostpathplugin-4t9zv" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.379481 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2nxf\" (UniqueName: \"kubernetes.io/projected/7531eb66-63f6-4eef-b30a-c3fb1ab94f7a-kube-api-access-z2nxf\") pod \"machine-config-server-rhnfd\" (UID: \"7531eb66-63f6-4eef-b30a-c3fb1ab94f7a\") " pod="openshift-machine-config-operator/machine-config-server-rhnfd" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.379499 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jq4n4\" (UniqueName: \"kubernetes.io/projected/1d6e992d-904f-4f11-ab29-29583f1dd955-kube-api-access-jq4n4\") pod \"packageserver-d55dfcdfc-79gj2\" (UID: \"1d6e992d-904f-4f11-ab29-29583f1dd955\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-79gj2" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.379514 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1d6e992d-904f-4f11-ab29-29583f1dd955-webhook-cert\") pod \"packageserver-d55dfcdfc-79gj2\" (UID: \"1d6e992d-904f-4f11-ab29-29583f1dd955\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-79gj2" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.379547 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e9a8e814-f2c2-46d1-99c9-40d77e0d96da-proxy-tls\") pod \"machine-config-controller-84d6567774-6h2jt\" (UID: \"e9a8e814-f2c2-46d1-99c9-40d77e0d96da\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6h2jt" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.379565 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d325cb10-86d1-414b-aa0a-282b8d6c1e30-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7x25b\" (UID: \"d325cb10-86d1-414b-aa0a-282b8d6c1e30\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7x25b" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.379588 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/951c1cb8-6613-404d-b0ec-ca5c8b66ec30-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-s7x26\" (UID: \"951c1cb8-6613-404d-b0ec-ca5c8b66ec30\") " pod="openshift-marketplace/marketplace-operator-79b997595-s7x26" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.379604 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e9a8e814-f2c2-46d1-99c9-40d77e0d96da-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-6h2jt\" (UID: \"e9a8e814-f2c2-46d1-99c9-40d77e0d96da\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6h2jt" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.379630 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/163ef77c-3b2f-45ef-81c8-1fc627b1a234-config-volume\") pod \"dns-default-csctv\" (UID: \"163ef77c-3b2f-45ef-81c8-1fc627b1a234\") " pod="openshift-dns/dns-default-csctv" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.380661 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d325cb10-86d1-414b-aa0a-282b8d6c1e30-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7x25b\" (UID: \"d325cb10-86d1-414b-aa0a-282b8d6c1e30\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7x25b" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.381166 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f6b07b0b-f08b-4ad5-a390-86a8119e96c9-socket-dir\") pod \"csi-hostpathplugin-4t9zv\" (UID: \"f6b07b0b-f08b-4ad5-a390-86a8119e96c9\") " pod="hostpath-provisioner/csi-hostpathplugin-4t9zv" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.381467 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f6b07b0b-f08b-4ad5-a390-86a8119e96c9-registration-dir\") pod \"csi-hostpathplugin-4t9zv\" (UID: \"f6b07b0b-f08b-4ad5-a390-86a8119e96c9\") " pod="hostpath-provisioner/csi-hostpathplugin-4t9zv" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.381915 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/1d6e992d-904f-4f11-ab29-29583f1dd955-tmpfs\") pod \"packageserver-d55dfcdfc-79gj2\" (UID: \"1d6e992d-904f-4f11-ab29-29583f1dd955\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-79gj2" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.382331 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/40f41556-7a45-457f-8f60-5fc87df20363-images\") pod \"machine-config-operator-74547568cd-fscrx\" (UID: \"40f41556-7a45-457f-8f60-5fc87df20363\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fscrx" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.382591 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/f6b07b0b-f08b-4ad5-a390-86a8119e96c9-csi-data-dir\") pod \"csi-hostpathplugin-4t9zv\" (UID: \"f6b07b0b-f08b-4ad5-a390-86a8119e96c9\") " pod="hostpath-provisioner/csi-hostpathplugin-4t9zv" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.384717 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba3cf9ac-3e57-4274-b5ea-727bf77aebf1-config\") pod \"service-ca-operator-777779d784-hx4vr\" (UID: \"ba3cf9ac-3e57-4274-b5ea-727bf77aebf1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hx4vr" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.385640 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e9a8e814-f2c2-46d1-99c9-40d77e0d96da-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-6h2jt\" (UID: \"e9a8e814-f2c2-46d1-99c9-40d77e0d96da\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6h2jt" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.385714 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/586ab369-00f4-497e-836f-3aede32089c2-config-volume\") pod \"collect-profiles-29399820-jr7cd\" (UID: \"586ab369-00f4-497e-836f-3aede32089c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399820-jr7cd" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.388498 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/f6b07b0b-f08b-4ad5-a390-86a8119e96c9-plugins-dir\") pod \"csi-hostpathplugin-4t9zv\" (UID: \"f6b07b0b-f08b-4ad5-a390-86a8119e96c9\") " pod="hostpath-provisioner/csi-hostpathplugin-4t9zv" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.388948 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/40f41556-7a45-457f-8f60-5fc87df20363-auth-proxy-config\") pod \"machine-config-operator-74547568cd-fscrx\" (UID: \"40f41556-7a45-457f-8f60-5fc87df20363\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fscrx" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.389240 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1d6e992d-904f-4f11-ab29-29583f1dd955-apiservice-cert\") pod \"packageserver-d55dfcdfc-79gj2\" (UID: \"1d6e992d-904f-4f11-ab29-29583f1dd955\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-79gj2" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.390732 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/163ef77c-3b2f-45ef-81c8-1fc627b1a234-metrics-tls\") pod \"dns-default-csctv\" (UID: \"163ef77c-3b2f-45ef-81c8-1fc627b1a234\") " pod="openshift-dns/dns-default-csctv" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.392531 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1d6e992d-904f-4f11-ab29-29583f1dd955-webhook-cert\") pod \"packageserver-d55dfcdfc-79gj2\" (UID: \"1d6e992d-904f-4f11-ab29-29583f1dd955\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-79gj2" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.392634 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba3cf9ac-3e57-4274-b5ea-727bf77aebf1-serving-cert\") pod \"service-ca-operator-777779d784-hx4vr\" (UID: \"ba3cf9ac-3e57-4274-b5ea-727bf77aebf1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hx4vr" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.392744 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e9a8e814-f2c2-46d1-99c9-40d77e0d96da-proxy-tls\") pod \"machine-config-controller-84d6567774-6h2jt\" (UID: \"e9a8e814-f2c2-46d1-99c9-40d77e0d96da\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6h2jt" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.392806 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/d6761cdb-c9a1-4a58-9023-e1230eab220d-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-bfmvc\" (UID: \"d6761cdb-c9a1-4a58-9023-e1230eab220d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bfmvc" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.393346 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6aa4a7b3-bae3-40bb-950d-e907073c0257-metrics-tls\") pod \"dns-operator-744455d44c-wvnp9\" (UID: \"6aa4a7b3-bae3-40bb-950d-e907073c0257\") " pod="openshift-dns-operator/dns-operator-744455d44c-wvnp9" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.393752 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/951c1cb8-6613-404d-b0ec-ca5c8b66ec30-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-s7x26\" (UID: \"951c1cb8-6613-404d-b0ec-ca5c8b66ec30\") " pod="openshift-marketplace/marketplace-operator-79b997595-s7x26" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.394034 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d059e3f7-ef28-4557-971c-e77be1472c52-cert\") pod \"ingress-canary-lnmz8\" (UID: \"d059e3f7-ef28-4557-971c-e77be1472c52\") " pod="openshift-ingress-canary/ingress-canary-lnmz8" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.394212 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/7531eb66-63f6-4eef-b30a-c3fb1ab94f7a-node-bootstrap-token\") pod \"machine-config-server-rhnfd\" (UID: \"7531eb66-63f6-4eef-b30a-c3fb1ab94f7a\") " pod="openshift-machine-config-operator/machine-config-server-rhnfd" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.394664 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4e68b45a-0978-41b1-ba4c-0b75dc67399e-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-6g5fd\" (UID: \"4e68b45a-0978-41b1-ba4c-0b75dc67399e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6g5fd" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.395922 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-hsvx8"] Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.396717 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f55f4"] Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.397120 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/586ab369-00f4-497e-836f-3aede32089c2-secret-volume\") pod \"collect-profiles-29399820-jr7cd\" (UID: \"586ab369-00f4-497e-836f-3aede32089c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399820-jr7cd" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.397500 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/163ef77c-3b2f-45ef-81c8-1fc627b1a234-config-volume\") pod \"dns-default-csctv\" (UID: \"163ef77c-3b2f-45ef-81c8-1fc627b1a234\") " pod="openshift-dns/dns-default-csctv" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.397602 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d325cb10-86d1-414b-aa0a-282b8d6c1e30-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7x25b\" (UID: \"d325cb10-86d1-414b-aa0a-282b8d6c1e30\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7x25b" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.397851 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7bc385a4-1703-4a76-b27e-8ae71480381d-signing-cabundle\") pod \"service-ca-9c57cc56f-54x2b\" (UID: \"7bc385a4-1703-4a76-b27e-8ae71480381d\") " pod="openshift-service-ca/service-ca-9c57cc56f-54x2b" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.398671 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/7531eb66-63f6-4eef-b30a-c3fb1ab94f7a-certs\") pod \"machine-config-server-rhnfd\" (UID: \"7531eb66-63f6-4eef-b30a-c3fb1ab94f7a\") " pod="openshift-machine-config-operator/machine-config-server-rhnfd" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.399179 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/40f41556-7a45-457f-8f60-5fc87df20363-proxy-tls\") pod \"machine-config-operator-74547568cd-fscrx\" (UID: \"40f41556-7a45-457f-8f60-5fc87df20363\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fscrx" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.406425 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7bc385a4-1703-4a76-b27e-8ae71480381d-signing-key\") pod \"service-ca-9c57cc56f-54x2b\" (UID: \"7bc385a4-1703-4a76-b27e-8ae71480381d\") " pod="openshift-service-ca/service-ca-9c57cc56f-54x2b" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.409418 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mh76t\" (UniqueName: \"kubernetes.io/projected/d441620d-6188-465c-8531-888b80826aa3-kube-api-access-mh76t\") pod \"etcd-operator-b45778765-cppvr\" (UID: \"d441620d-6188-465c-8531-888b80826aa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cppvr" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.410131 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/951c1cb8-6613-404d-b0ec-ca5c8b66ec30-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-s7x26\" (UID: \"951c1cb8-6613-404d-b0ec-ca5c8b66ec30\") " pod="openshift-marketplace/marketplace-operator-79b997595-s7x26" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.426790 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xzd7l" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.444240 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fd8b1442-101f-4104-ae72-cf6d2029f2d1-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-c4qnk\" (UID: \"fd8b1442-101f-4104-ae72-cf6d2029f2d1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c4qnk" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.446031 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6381724d-d16b-4d7b-9d9f-613451b93c77-bound-sa-token\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.470525 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzlxr\" (UniqueName: \"kubernetes.io/projected/ebb47b17-21a5-4ae0-857d-36f913842d4a-kube-api-access-fzlxr\") pod \"migrator-59844c95c7-xcw79\" (UID: \"ebb47b17-21a5-4ae0-857d-36f913842d4a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xcw79" Nov 24 13:14:53 crc kubenswrapper[4790]: W1124 13:14:53.472644 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc36b34a9_5050_4b55_886b_83c0dd80a3df.slice/crio-501ed4367cd21be30e212d46292c614a7d766e2af4622b2991ee6533d982d78e WatchSource:0}: Error finding container 501ed4367cd21be30e212d46292c614a7d766e2af4622b2991ee6533d982d78e: Status 404 returned error can't find the container with id 501ed4367cd21be30e212d46292c614a7d766e2af4622b2991ee6533d982d78e Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.475799 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-zxkb7"] Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.481423 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:53 crc kubenswrapper[4790]: E1124 13:14:53.481865 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:53.981852825 +0000 UTC m=+142.361746487 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.490212 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qn6gs\" (UniqueName: \"kubernetes.io/projected/40f41556-7a45-457f-8f60-5fc87df20363-kube-api-access-qn6gs\") pod \"machine-config-operator-74547568cd-fscrx\" (UID: \"40f41556-7a45-457f-8f60-5fc87df20363\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fscrx" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.502051 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8jkhv"] Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.502098 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-cjdhk"] Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.503926 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-hkbxm" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.516013 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96dzd\" (UniqueName: \"kubernetes.io/projected/163ef77c-3b2f-45ef-81c8-1fc627b1a234-kube-api-access-96dzd\") pod \"dns-default-csctv\" (UID: \"163ef77c-3b2f-45ef-81c8-1fc627b1a234\") " pod="openshift-dns/dns-default-csctv" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.548238 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxdh4\" (UniqueName: \"kubernetes.io/projected/ba3cf9ac-3e57-4274-b5ea-727bf77aebf1-kube-api-access-qxdh4\") pod \"service-ca-operator-777779d784-hx4vr\" (UID: \"ba3cf9ac-3e57-4274-b5ea-727bf77aebf1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hx4vr" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.573683 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-csctv" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.573766 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhzq4\" (UniqueName: \"kubernetes.io/projected/d6761cdb-c9a1-4a58-9023-e1230eab220d-kube-api-access-qhzq4\") pod \"package-server-manager-789f6589d5-bfmvc\" (UID: \"d6761cdb-c9a1-4a58-9023-e1230eab220d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bfmvc" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.574013 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-96268"] Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.578217 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-djs6p"] Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.582456 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:53 crc kubenswrapper[4790]: E1124 13:14:53.582662 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:54.08263499 +0000 UTC m=+142.462528662 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.582989 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:53 crc kubenswrapper[4790]: E1124 13:14:53.584399 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:54.084383307 +0000 UTC m=+142.464276969 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.598546 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkvp6\" (UniqueName: \"kubernetes.io/projected/4e68b45a-0978-41b1-ba4c-0b75dc67399e-kube-api-access-hkvp6\") pod \"kube-storage-version-migrator-operator-b67b599dd-6g5fd\" (UID: \"4e68b45a-0978-41b1-ba4c-0b75dc67399e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6g5fd" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.608952 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-cppvr" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.610817 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gc5zb\" (UniqueName: \"kubernetes.io/projected/6aa4a7b3-bae3-40bb-950d-e907073c0257-kube-api-access-gc5zb\") pod \"dns-operator-744455d44c-wvnp9\" (UID: \"6aa4a7b3-bae3-40bb-950d-e907073c0257\") " pod="openshift-dns-operator/dns-operator-744455d44c-wvnp9" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.611237 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9dkb\" (UniqueName: \"kubernetes.io/projected/f6b07b0b-f08b-4ad5-a390-86a8119e96c9-kube-api-access-d9dkb\") pod \"csi-hostpathplugin-4t9zv\" (UID: \"f6b07b0b-f08b-4ad5-a390-86a8119e96c9\") " pod="hostpath-provisioner/csi-hostpathplugin-4t9zv" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.626052 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-ktdmx" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.626605 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkb7j\" (UniqueName: \"kubernetes.io/projected/d325cb10-86d1-414b-aa0a-282b8d6c1e30-kube-api-access-rkb7j\") pod \"openshift-apiserver-operator-796bbdcf4f-7x25b\" (UID: \"d325cb10-86d1-414b-aa0a-282b8d6c1e30\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7x25b" Nov 24 13:14:53 crc kubenswrapper[4790]: W1124 13:14:53.633221 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda86ba968_c6dd_42c8_91fb_207bc839b8ec.slice/crio-52aa339943e7758dc3adc470725aa069f384cd6d35be2b78083c3b060213239f WatchSource:0}: Error finding container 52aa339943e7758dc3adc470725aa069f384cd6d35be2b78083c3b060213239f: Status 404 returned error can't find the container with id 52aa339943e7758dc3adc470725aa069f384cd6d35be2b78083c3b060213239f Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.634781 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6ljl5"] Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.648773 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j968g\" (UniqueName: \"kubernetes.io/projected/d059e3f7-ef28-4557-971c-e77be1472c52-kube-api-access-j968g\") pod \"ingress-canary-lnmz8\" (UID: \"d059e3f7-ef28-4557-971c-e77be1472c52\") " pod="openshift-ingress-canary/ingress-canary-lnmz8" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.683595 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:53 crc kubenswrapper[4790]: E1124 13:14:53.683801 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:54.183776526 +0000 UTC m=+142.563670188 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.684463 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:53 crc kubenswrapper[4790]: E1124 13:14:53.684800 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:54.184782929 +0000 UTC m=+142.564676591 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.692388 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2nxf\" (UniqueName: \"kubernetes.io/projected/7531eb66-63f6-4eef-b30a-c3fb1ab94f7a-kube-api-access-z2nxf\") pod \"machine-config-server-rhnfd\" (UID: \"7531eb66-63f6-4eef-b30a-c3fb1ab94f7a\") " pod="openshift-machine-config-operator/machine-config-server-rhnfd" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.700507 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xcw79" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.701047 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c4qnk" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.708630 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dq57\" (UniqueName: \"kubernetes.io/projected/951c1cb8-6613-404d-b0ec-ca5c8b66ec30-kube-api-access-2dq57\") pod \"marketplace-operator-79b997595-s7x26\" (UID: \"951c1cb8-6613-404d-b0ec-ca5c8b66ec30\") " pod="openshift-marketplace/marketplace-operator-79b997595-s7x26" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.711187 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-z2dld"] Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.714016 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5lzlv"] Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.718375 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ql2q4\" (UniqueName: \"kubernetes.io/projected/e9a8e814-f2c2-46d1-99c9-40d77e0d96da-kube-api-access-ql2q4\") pod \"machine-config-controller-84d6567774-6h2jt\" (UID: \"e9a8e814-f2c2-46d1-99c9-40d77e0d96da\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6h2jt" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.726389 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-wvnp9" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.727687 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jq4n4\" (UniqueName: \"kubernetes.io/projected/1d6e992d-904f-4f11-ab29-29583f1dd955-kube-api-access-jq4n4\") pod \"packageserver-d55dfcdfc-79gj2\" (UID: \"1d6e992d-904f-4f11-ab29-29583f1dd955\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-79gj2" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.732494 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-s7x26" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.739087 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xzd7l"] Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.739252 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hx4vr" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.747180 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6h2jt" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.751714 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9fn8\" (UniqueName: \"kubernetes.io/projected/586ab369-00f4-497e-836f-3aede32089c2-kube-api-access-g9fn8\") pod \"collect-profiles-29399820-jr7cd\" (UID: \"586ab369-00f4-497e-836f-3aede32089c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399820-jr7cd" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.759035 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6g5fd" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.768980 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7x25b" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.773189 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fscrx" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.777746 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdfzj\" (UniqueName: \"kubernetes.io/projected/7bc385a4-1703-4a76-b27e-8ae71480381d-kube-api-access-zdfzj\") pod \"service-ca-9c57cc56f-54x2b\" (UID: \"7bc385a4-1703-4a76-b27e-8ae71480381d\") " pod="openshift-service-ca/service-ca-9c57cc56f-54x2b" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.784322 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-54x2b" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.784831 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:53 crc kubenswrapper[4790]: E1124 13:14:53.785218 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:54.285201212 +0000 UTC m=+142.665094874 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.789212 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399820-jr7cd" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.800361 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-79gj2" Nov 24 13:14:53 crc kubenswrapper[4790]: W1124 13:14:53.806921 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2a0fdfc_5465_483d_8961_baa619f6ec0f.slice/crio-b45900a53bb27973fda810c8f75654bde7d23be9c272bfd94d396e55d7952075 WatchSource:0}: Error finding container b45900a53bb27973fda810c8f75654bde7d23be9c272bfd94d396e55d7952075: Status 404 returned error can't find the container with id b45900a53bb27973fda810c8f75654bde7d23be9c272bfd94d396e55d7952075 Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.812181 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bfmvc" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.815413 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-lnmz8" Nov 24 13:14:53 crc kubenswrapper[4790]: W1124 13:14:53.816118 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod692c3c09_be9f_4400_b0fb_e8c8692f0d21.slice/crio-bf6be5682e7e4a26d5a13cc3380ec0464cd62f402d838ebdf49aa30237d4c688 WatchSource:0}: Error finding container bf6be5682e7e4a26d5a13cc3380ec0464cd62f402d838ebdf49aa30237d4c688: Status 404 returned error can't find the container with id bf6be5682e7e4a26d5a13cc3380ec0464cd62f402d838ebdf49aa30237d4c688 Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.826357 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-rhnfd" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.865152 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-4t9zv" Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.890702 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:53 crc kubenswrapper[4790]: E1124 13:14:53.891023 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:54.3910066 +0000 UTC m=+142.770900262 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.891324 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qzt4s"] Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.892263 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-t5fjs"] Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.912971 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-7d488"] Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.940452 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-hkbxm"] Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.944129 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-zvzkq"] Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.966348 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6ljl5" event={"ID":"c42d89b6-6420-4267-93d2-e5dac6e1906b","Type":"ContainerStarted","Data":"1481d38639b78f61917f6aaaf78bba892d41919addd939d0c9923e86e4743abd"} Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.970286 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-96268" event={"ID":"acb2b70f-4a66-4290-aa8a-a13cb1b4b2a9","Type":"ContainerStarted","Data":"f22e7f17f5ce2dffa3cd88179df9060705e913fd234cad24930cf8c133ff4cc7"} Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.971381 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-cjdhk" event={"ID":"69940b0d-4afc-4a7f-8ff7-cb128ce58913","Type":"ContainerStarted","Data":"960cd279ac7a1f420dd9ab80bdc74bdffb985ef736d6e01607277f0782682069"} Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.972451 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-djs6p" event={"ID":"2f77876b-f997-4274-a619-a97e37bc4b9e","Type":"ContainerStarted","Data":"699598aabdb383cfc60ee70889b3d6d1bc6752c4429465359813690e560bc82e"} Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.981747 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ktdmx"] Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.994547 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:53 crc kubenswrapper[4790]: E1124 13:14:53.994670 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:54.494644917 +0000 UTC m=+142.874538579 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:53 crc kubenswrapper[4790]: I1124 13:14:53.995069 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:53 crc kubenswrapper[4790]: E1124 13:14:53.995386 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:54.495374891 +0000 UTC m=+142.875268553 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.014564 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-z2dld" event={"ID":"a2a0fdfc-5465-483d-8961-baa619f6ec0f","Type":"ContainerStarted","Data":"b45900a53bb27973fda810c8f75654bde7d23be9c272bfd94d396e55d7952075"} Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.026129 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wnlxq" event={"ID":"f50b116f-b25b-4985-8975-fbb523e1ddf6","Type":"ContainerStarted","Data":"223274807b4aa4b9eb85d9dbb4d301afa8de029c7481e07a66699b8e1ef76c0f"} Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.026174 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wnlxq" event={"ID":"f50b116f-b25b-4985-8975-fbb523e1ddf6","Type":"ContainerStarted","Data":"7f9d98c4a9b7868ebb84c9e55fd33a58756669f8116a53c60e50931c32ab45b5"} Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.027286 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wnlxq" Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.028268 4790 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-wnlxq container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.028324 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wnlxq" podUID="f50b116f-b25b-4985-8975-fbb523e1ddf6" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.030256 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xzd7l" event={"ID":"ffdd5f6c-5425-4895-8c2b-6ba6ef3706c9","Type":"ContainerStarted","Data":"f6e213985379e8ef7f5b8410e09f1aa63c1a55d39451bef1f30b78a4a10c66a7"} Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.031504 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5lzlv" event={"ID":"692c3c09-be9f-4400-b0fb-e8c8692f0d21","Type":"ContainerStarted","Data":"bf6be5682e7e4a26d5a13cc3380ec0464cd62f402d838ebdf49aa30237d4c688"} Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.032328 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsvx8" event={"ID":"2a2b45ef-e51f-4ad2-a1b7-6d8e7b0c089a","Type":"ContainerStarted","Data":"3eedf4b22d34d2fe45426dd7274e4ecd8d11904931d900029879bb31d8eb291d"} Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.035282 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f55f4" event={"ID":"194e75e8-4f01-4862-ba6e-a073a369df5a","Type":"ContainerStarted","Data":"fd3b8b1ae3e3c80dfa4209ff745c1d405c27925fb8de0b7716806a4ff3bc122a"} Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.045078 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-4ldk4" event={"ID":"c36b34a9-5050-4b55-886b-83c0dd80a3df","Type":"ContainerStarted","Data":"501ed4367cd21be30e212d46292c614a7d766e2af4622b2991ee6533d982d78e"} Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.051981 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qwnbb" event={"ID":"baa99c88-cee5-42d8-ac79-21551d465f7f","Type":"ContainerStarted","Data":"35c4d2417c26d039b8de5dba40204d94ca0c2cfbf24c8cadf5cf3892b9855897"} Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.054769 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-csctv"] Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.056429 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5hlqm" event={"ID":"1e6bf7c7-7711-43cb-8817-73b42174b9ee","Type":"ContainerStarted","Data":"ae021e59ad4ca5464b509d17eb45f69c9864bfbfb942df7f695a840be70171fb"} Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.059179 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" event={"ID":"66c28684-107f-4208-ae79-4b22a4e485dd","Type":"ContainerStarted","Data":"fefbb5924abe648344dfa6129f538c8ffc4c93e50bb0a19fcdaf9e44038cc74f"} Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.068998 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" event={"ID":"a86ba968-c6dd-42c8-91fb-207bc839b8ec","Type":"ContainerStarted","Data":"52aa339943e7758dc3adc470725aa069f384cd6d35be2b78083c3b060213239f"} Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.072254 4790 generic.go:334] "Generic (PLEG): container finished" podID="5b226a93-5333-405f-93ad-788f9376faac" containerID="5702bccadeb2d07bb537580e8ac35d14d2d372b529cddd183f3996f9b87ae87b" exitCode=0 Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.072375 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" event={"ID":"5b226a93-5333-405f-93ad-788f9376faac","Type":"ContainerDied","Data":"5702bccadeb2d07bb537580e8ac35d14d2d372b529cddd183f3996f9b87ae87b"} Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.074010 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-slkpm" event={"ID":"17718a81-aa70-4f5a-90bd-06e9251f4ef8","Type":"ContainerStarted","Data":"4b82b0635e3a6e9a63e05627c2cf948b1bfa216360cf5184154db78f8da6e560"} Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.074056 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-slkpm" event={"ID":"17718a81-aa70-4f5a-90bd-06e9251f4ef8","Type":"ContainerStarted","Data":"8ffdbab8012849ce113037dd3eaefb1a43fcd86c02e85a4d731b851ff90eb265"} Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.097593 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:54 crc kubenswrapper[4790]: E1124 13:14:54.098956 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:54.598926526 +0000 UTC m=+142.978820198 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.114240 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-cppvr"] Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.200308 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:54 crc kubenswrapper[4790]: E1124 13:14:54.201057 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:54.701044924 +0000 UTC m=+143.080938586 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.304873 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.304905 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bfmvc"] Nov 24 13:14:54 crc kubenswrapper[4790]: E1124 13:14:54.305167 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:54.805141496 +0000 UTC m=+143.185035158 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.305277 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:54 crc kubenswrapper[4790]: E1124 13:14:54.305611 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:54.805599751 +0000 UTC m=+143.185493403 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.406253 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:54 crc kubenswrapper[4790]: E1124 13:14:54.407054 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:54.907033157 +0000 UTC m=+143.286926819 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.409628 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-6h2jt"] Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.482678 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-xcw79"] Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.509952 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:54 crc kubenswrapper[4790]: E1124 13:14:54.510407 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:55.010394726 +0000 UTC m=+143.390288388 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.586347 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-lnmz8"] Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.612527 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:54 crc kubenswrapper[4790]: E1124 13:14:54.612892 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:55.112828264 +0000 UTC m=+143.492721926 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.613122 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:54 crc kubenswrapper[4790]: E1124 13:14:54.613461 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:55.113448024 +0000 UTC m=+143.493341686 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:54 crc kubenswrapper[4790]: W1124 13:14:54.626939 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7531eb66_63f6_4eef_b30a_c3fb1ab94f7a.slice/crio-a69a0aa31862a75253644b451c264e9c968dcffdb76c792632ea5fb61ef9bfad WatchSource:0}: Error finding container a69a0aa31862a75253644b451c264e9c968dcffdb76c792632ea5fb61ef9bfad: Status 404 returned error can't find the container with id a69a0aa31862a75253644b451c264e9c968dcffdb76c792632ea5fb61ef9bfad Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.691889 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.715059 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:54 crc kubenswrapper[4790]: E1124 13:14:54.715217 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:55.215203381 +0000 UTC m=+143.595097043 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.715454 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.715041 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wnlxq" podStartSLOduration=118.715023135 podStartE2EDuration="1m58.715023135s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:54.71364581 +0000 UTC m=+143.093539472" watchObservedRunningTime="2025-11-24 13:14:54.715023135 +0000 UTC m=+143.094916797" Nov 24 13:14:54 crc kubenswrapper[4790]: E1124 13:14:54.716488 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:55.216480592 +0000 UTC m=+143.596374254 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.807079 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6g5fd"] Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.816311 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:54 crc kubenswrapper[4790]: E1124 13:14:54.816472 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:55.316446731 +0000 UTC m=+143.696340403 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.816610 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:54 crc kubenswrapper[4790]: E1124 13:14:54.816921 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:55.316913046 +0000 UTC m=+143.696806708 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.901760 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-4t9zv"] Nov 24 13:14:54 crc kubenswrapper[4790]: I1124 13:14:54.918349 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:54 crc kubenswrapper[4790]: E1124 13:14:54.918758 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:55.418738924 +0000 UTC m=+143.798632586 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.023811 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:55 crc kubenswrapper[4790]: E1124 13:14:55.024230 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:55.524212302 +0000 UTC m=+143.904105964 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:55 crc kubenswrapper[4790]: W1124 13:14:55.097482 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e68b45a_0978_41b1_ba4c_0b75dc67399e.slice/crio-2e0616ad67960f956d542631e5fe1d69e6353f82aa8e91a60f5df2023f4ba92f WatchSource:0}: Error finding container 2e0616ad67960f956d542631e5fe1d69e6353f82aa8e91a60f5df2023f4ba92f: Status 404 returned error can't find the container with id 2e0616ad67960f956d542631e5fe1d69e6353f82aa8e91a60f5df2023f4ba92f Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.119961 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" event={"ID":"a86ba968-c6dd-42c8-91fb-207bc839b8ec","Type":"ContainerStarted","Data":"0c06dbf4c49aaeeb09163e81c971e4b532cc334825adcec02f48a8f9eafa867e"} Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.120554 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.124259 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:55 crc kubenswrapper[4790]: E1124 13:14:55.124715 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:55.624694537 +0000 UTC m=+144.004588199 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.124957 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-djs6p" event={"ID":"2f77876b-f997-4274-a619-a97e37bc4b9e","Type":"ContainerStarted","Data":"9415f5d46f46f10c50f1560d435d22a23bf24a8e2c0b53c5bdf19ca76b7d1859"} Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.130309 4790 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-zxkb7 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.41:6443/healthz\": dial tcp 10.217.0.41:6443: connect: connection refused" start-of-body= Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.130361 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" podUID="a86ba968-c6dd-42c8-91fb-207bc839b8ec" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.41:6443/healthz\": dial tcp 10.217.0.41:6443: connect: connection refused" Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.140715 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qzt4s" event={"ID":"56177dd2-06b5-4d07-9a9b-395c558e2f13","Type":"ContainerStarted","Data":"1c8e8fb3fce678f715c608c0a7f56ed4d74ed0e5103cd8d7e09c65c777987485"} Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.145059 4790 generic.go:334] "Generic (PLEG): container finished" podID="2a2b45ef-e51f-4ad2-a1b7-6d8e7b0c089a" containerID="6eca801bc7bd105a40924a59d01a25d3b900e211e2ee2ece4759cbdab1f3a287" exitCode=0 Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.145148 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsvx8" event={"ID":"2a2b45ef-e51f-4ad2-a1b7-6d8e7b0c089a","Type":"ContainerDied","Data":"6eca801bc7bd105a40924a59d01a25d3b900e211e2ee2ece4759cbdab1f3a287"} Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.158423 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-zvzkq" event={"ID":"3c7930b6-15dc-4462-bc47-9137c2d4edc7","Type":"ContainerStarted","Data":"74cf1411872a1019f729f73b9259df33d14683971e5f8e4189e31f57569e0602"} Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.161746 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qwnbb" podStartSLOduration=119.161695699 podStartE2EDuration="1m59.161695699s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:55.147557869 +0000 UTC m=+143.527451531" watchObservedRunningTime="2025-11-24 13:14:55.161695699 +0000 UTC m=+143.541589361" Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.166114 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-rhnfd" event={"ID":"7531eb66-63f6-4eef-b30a-c3fb1ab94f7a","Type":"ContainerStarted","Data":"a69a0aa31862a75253644b451c264e9c968dcffdb76c792632ea5fb61ef9bfad"} Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.167185 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-hkbxm" event={"ID":"207655ea-f838-43a7-ad87-2bee51686a16","Type":"ContainerStarted","Data":"a72e988320878767c277a99203f99f300006acd56f07c8647317ffe90b638a51"} Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.168728 4790 generic.go:334] "Generic (PLEG): container finished" podID="66c28684-107f-4208-ae79-4b22a4e485dd" containerID="2e7e8cabb3998b9faa34680f5dc681975ddc6b2e778cfee4df233e11eb014cc0" exitCode=0 Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.168783 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" event={"ID":"66c28684-107f-4208-ae79-4b22a4e485dd","Type":"ContainerDied","Data":"2e7e8cabb3998b9faa34680f5dc681975ddc6b2e778cfee4df233e11eb014cc0"} Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.176140 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-t5fjs" event={"ID":"aef429a6-aa76-4776-a6e2-30096a5e7ccb","Type":"ContainerStarted","Data":"0fa23a4ec0d51d080fca5722debd6883bf4e5a08d6775739c39c68f81d3d0f8b"} Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.185482 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-4ldk4" event={"ID":"c36b34a9-5050-4b55-886b-83c0dd80a3df","Type":"ContainerStarted","Data":"aad6f89a78a7a68b732316f76b7796414f01788f1e8635504d4bc02ba824282e"} Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.191044 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-96268" event={"ID":"acb2b70f-4a66-4290-aa8a-a13cb1b4b2a9","Type":"ContainerStarted","Data":"a661eb8e232d69086c4ce7504a6a1876ffb0497af52df39e062a4894052fdce3"} Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.198093 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-lnmz8" event={"ID":"d059e3f7-ef28-4557-971c-e77be1472c52","Type":"ContainerStarted","Data":"526aa684ec61f0b9554396b4fe801754347104ed704acbdb1a7298255bdf939d"} Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.199334 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7d488" event={"ID":"91336463-df6a-4b98-a990-e6780e417fab","Type":"ContainerStarted","Data":"88ffbe6a48916c195c6816395ab8d1fb2789706f6de681167ed4454ddf627e58"} Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.201813 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-slkpm" event={"ID":"17718a81-aa70-4f5a-90bd-06e9251f4ef8","Type":"ContainerStarted","Data":"f8490acc6702081977e538fc4de026839d7f674b4faacef229422d20788c3c4f"} Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.204241 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f55f4" event={"ID":"194e75e8-4f01-4862-ba6e-a073a369df5a","Type":"ContainerStarted","Data":"07de105bd4f616d4bff3ec146b3173214d2cca2055cbda2c34d445c7120dc64f"} Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.204999 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xcw79" event={"ID":"ebb47b17-21a5-4ae0-857d-36f913842d4a","Type":"ContainerStarted","Data":"b6d43520e13e104c7c73dfa300d030c07ef8fe020b02c76aa227d854018e3a19"} Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.206079 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ktdmx" event={"ID":"e4177bcc-76f8-43de-83ad-019fe23543e4","Type":"ContainerStarted","Data":"b928d638fee8c0c0a0c1879010eabfed0a999856721bd2a74251df1352cd6e5f"} Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.224486 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-csctv" event={"ID":"163ef77c-3b2f-45ef-81c8-1fc627b1a234","Type":"ContainerStarted","Data":"0a22d960e7caa947e72d208b58ba84ecc5a928eeed6cf12cf3e127b22d8f6043"} Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.225768 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:55 crc kubenswrapper[4790]: E1124 13:14:55.232691 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:55.732672745 +0000 UTC m=+144.112566407 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.233192 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6h2jt" event={"ID":"e9a8e814-f2c2-46d1-99c9-40d77e0d96da","Type":"ContainerStarted","Data":"b6ba98e5682c457dfcc1ef18a0e7548ebdb68e285e23da5070902cd25c2f0c39"} Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.248557 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-cppvr" event={"ID":"d441620d-6188-465c-8531-888b80826aa3","Type":"ContainerStarted","Data":"f3c756bbb440ae87ed84a260c7d55679f8166395c905411640d856071cbf2c3c"} Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.263405 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-cjdhk" event={"ID":"69940b0d-4afc-4a7f-8ff7-cb128ce58913","Type":"ContainerStarted","Data":"675c34e9a7963fc1e18def04b460ba4e931cbb81bbc230859d8cce0fae5bfe06"} Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.264430 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-cjdhk" Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.275103 4790 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-cjdhk container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.275152 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-cjdhk" podUID="69940b0d-4afc-4a7f-8ff7-cb128ce58913" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.284464 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-z2dld" event={"ID":"a2a0fdfc-5465-483d-8961-baa619f6ec0f","Type":"ContainerStarted","Data":"8d8f83ff4fba0c1c24167cefea8c8f4e548d1299594efb510cb32630309c4573"} Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.285341 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-z2dld" Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.287966 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bfmvc" event={"ID":"d6761cdb-c9a1-4a58-9023-e1230eab220d","Type":"ContainerStarted","Data":"5d677d3c8379a18e81457ca2a3e56eec2cddfcf2807d334975b531c43537a702"} Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.290606 4790 patch_prober.go:28] interesting pod/console-operator-58897d9998-z2dld container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/readyz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.290674 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-z2dld" podUID="a2a0fdfc-5465-483d-8961-baa619f6ec0f" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/readyz\": dial tcp 10.217.0.16:8443: connect: connection refused" Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.296381 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wnlxq" Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.326851 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:55 crc kubenswrapper[4790]: E1124 13:14:55.327064 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:55.827028711 +0000 UTC m=+144.206922363 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.327448 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:55 crc kubenswrapper[4790]: E1124 13:14:55.328783 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:55.828764348 +0000 UTC m=+144.208658010 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.383160 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-4ldk4" Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.394148 4790 patch_prober.go:28] interesting pod/router-default-5444994796-4ldk4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:14:55 crc kubenswrapper[4790]: [-]has-synced failed: reason withheld Nov 24 13:14:55 crc kubenswrapper[4790]: [+]process-running ok Nov 24 13:14:55 crc kubenswrapper[4790]: healthz check failed Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.394205 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ldk4" podUID="c36b34a9-5050-4b55-886b-83c0dd80a3df" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.431236 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:55 crc kubenswrapper[4790]: E1124 13:14:55.431787 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:55.931762584 +0000 UTC m=+144.311656246 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.437476 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:55 crc kubenswrapper[4790]: E1124 13:14:55.438576 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:55.938553605 +0000 UTC m=+144.318447267 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.482632 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-slkpm" podStartSLOduration=119.482615367 podStartE2EDuration="1m59.482615367s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:55.435668161 +0000 UTC m=+143.815561823" watchObservedRunningTime="2025-11-24 13:14:55.482615367 +0000 UTC m=+143.862509029" Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.538751 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:55 crc kubenswrapper[4790]: E1124 13:14:55.541459 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:56.041431508 +0000 UTC m=+144.421325170 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.594937 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-96268" podStartSLOduration=119.594916356 podStartE2EDuration="1m59.594916356s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:55.585974745 +0000 UTC m=+143.965868437" watchObservedRunningTime="2025-11-24 13:14:55.594916356 +0000 UTC m=+143.974810028" Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.628021 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-cjdhk" podStartSLOduration=119.628002981 podStartE2EDuration="1m59.628002981s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:55.624869639 +0000 UTC m=+144.004763301" watchObservedRunningTime="2025-11-24 13:14:55.628002981 +0000 UTC m=+144.007896643" Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.642009 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:55 crc kubenswrapper[4790]: E1124 13:14:55.642340 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:56.142328226 +0000 UTC m=+144.522221888 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.651106 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399820-jr7cd"] Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.685085 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-hx4vr"] Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.697287 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c4qnk"] Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.726637 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-fscrx"] Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.728071 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-54x2b"] Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.736729 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-wvnp9"] Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.749095 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:55 crc kubenswrapper[4790]: E1124 13:14:55.749578 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:56.249561961 +0000 UTC m=+144.629455623 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.762627 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7x25b"] Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.776636 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-79gj2"] Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.851712 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:55 crc kubenswrapper[4790]: E1124 13:14:55.852220 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:56.352190706 +0000 UTC m=+144.732084368 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:55 crc kubenswrapper[4790]: W1124 13:14:55.861261 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd325cb10_86d1_414b_aa0a_282b8d6c1e30.slice/crio-04107fc841b91cb3fae419313861d1f36dffb6a042626fd705fc0a3cc57d50eb WatchSource:0}: Error finding container 04107fc841b91cb3fae419313861d1f36dffb6a042626fd705fc0a3cc57d50eb: Status 404 returned error can't find the container with id 04107fc841b91cb3fae419313861d1f36dffb6a042626fd705fc0a3cc57d50eb Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.877726 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-s7x26"] Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.889969 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-djs6p" podStartSLOduration=119.889954343 podStartE2EDuration="1m59.889954343s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:55.888153114 +0000 UTC m=+144.268046776" watchObservedRunningTime="2025-11-24 13:14:55.889954343 +0000 UTC m=+144.269848005" Nov 24 13:14:55 crc kubenswrapper[4790]: W1124 13:14:55.893040 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d6e992d_904f_4f11_ab29_29583f1dd955.slice/crio-75dded4821ea9340a90c731616945134fe9e3d20be8eaf9260cd2cd7eb3a3926 WatchSource:0}: Error finding container 75dded4821ea9340a90c731616945134fe9e3d20be8eaf9260cd2cd7eb3a3926: Status 404 returned error can't find the container with id 75dded4821ea9340a90c731616945134fe9e3d20be8eaf9260cd2cd7eb3a3926 Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.933286 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-z2dld" podStartSLOduration=119.93326891 podStartE2EDuration="1m59.93326891s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:55.932731023 +0000 UTC m=+144.312624685" watchObservedRunningTime="2025-11-24 13:14:55.93326891 +0000 UTC m=+144.313162592" Nov 24 13:14:55 crc kubenswrapper[4790]: I1124 13:14:55.964508 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:55 crc kubenswrapper[4790]: E1124 13:14:55.966118 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:56.466059665 +0000 UTC m=+144.845953327 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.012733 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-4ldk4" podStartSLOduration=120.012706381 podStartE2EDuration="2m0.012706381s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:55.978335924 +0000 UTC m=+144.358229586" watchObservedRunningTime="2025-11-24 13:14:56.012706381 +0000 UTC m=+144.392600043" Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.026289 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" podStartSLOduration=120.026261312 podStartE2EDuration="2m0.026261312s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:56.025399924 +0000 UTC m=+144.405293576" watchObservedRunningTime="2025-11-24 13:14:56.026261312 +0000 UTC m=+144.406154974" Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.066838 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:56 crc kubenswrapper[4790]: E1124 13:14:56.067189 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:56.567179471 +0000 UTC m=+144.947073123 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.169069 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:56 crc kubenswrapper[4790]: E1124 13:14:56.170088 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:56.670059513 +0000 UTC m=+145.049953305 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.271547 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:56 crc kubenswrapper[4790]: E1124 13:14:56.271949 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:56.771937383 +0000 UTC m=+145.151831045 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.350771 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5lzlv" event={"ID":"692c3c09-be9f-4400-b0fb-e8c8692f0d21","Type":"ContainerStarted","Data":"64a80bdb58ae9a28cff0de451ecaed105a1743e7745138a19fd06844e64c6ca3"} Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.368680 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" event={"ID":"5b226a93-5333-405f-93ad-788f9376faac","Type":"ContainerStarted","Data":"39d607cd5f9f230e8624f17b33615534b49abf09a3de3d181b239fdbcf8023f6"} Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.374830 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:56 crc kubenswrapper[4790]: E1124 13:14:56.375169 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:56.875129267 +0000 UTC m=+145.255022929 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.375615 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:56 crc kubenswrapper[4790]: E1124 13:14:56.376441 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:56.876421299 +0000 UTC m=+145.256314961 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.385377 4790 patch_prober.go:28] interesting pod/router-default-5444994796-4ldk4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:14:56 crc kubenswrapper[4790]: [-]has-synced failed: reason withheld Nov 24 13:14:56 crc kubenswrapper[4790]: [+]process-running ok Nov 24 13:14:56 crc kubenswrapper[4790]: healthz check failed Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.385435 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ldk4" podUID="c36b34a9-5050-4b55-886b-83c0dd80a3df" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.432907 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f55f4" event={"ID":"194e75e8-4f01-4862-ba6e-a073a369df5a","Type":"ContainerStarted","Data":"6a534f54d15f05c18e7ca0dfa34f4189660a5e84514cc6a1a6aa06603b367ead"} Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.437896 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5lzlv" podStartSLOduration=120.437845244 podStartE2EDuration="2m0.437845244s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:56.396374427 +0000 UTC m=+144.776268089" watchObservedRunningTime="2025-11-24 13:14:56.437845244 +0000 UTC m=+144.817738906" Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.475911 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qzt4s" event={"ID":"56177dd2-06b5-4d07-9a9b-395c558e2f13","Type":"ContainerStarted","Data":"1bb6c707f30aa5453110f140e59e680e907159f0876be0616f45f604dcdf5f48"} Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.477095 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qzt4s" Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.477392 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:56 crc kubenswrapper[4790]: E1124 13:14:56.478321 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:56.978297469 +0000 UTC m=+145.358191161 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.492471 4790 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-qzt4s container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.492634 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qzt4s" podUID="56177dd2-06b5-4d07-9a9b-395c558e2f13" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.493263 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" podStartSLOduration=120.493223044 podStartE2EDuration="2m0.493223044s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:56.443835339 +0000 UTC m=+144.823729001" watchObservedRunningTime="2025-11-24 13:14:56.493223044 +0000 UTC m=+144.873116706" Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.499408 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c4qnk" event={"ID":"fd8b1442-101f-4104-ae72-cf6d2029f2d1","Type":"ContainerStarted","Data":"120d58eac509e157cd40174a0e3e545be99a38ea6693f72d9dd88079126ad3b8"} Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.517377 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6g5fd" event={"ID":"4e68b45a-0978-41b1-ba4c-0b75dc67399e","Type":"ContainerStarted","Data":"8603779d69c6d8a1ef4ca67c6d4529ee3645768bf3874aac63eeb59ce7fc0dd8"} Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.517449 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6g5fd" event={"ID":"4e68b45a-0978-41b1-ba4c-0b75dc67399e","Type":"ContainerStarted","Data":"2e0616ad67960f956d542631e5fe1d69e6353f82aa8e91a60f5df2023f4ba92f"} Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.521696 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399820-jr7cd" event={"ID":"586ab369-00f4-497e-836f-3aede32089c2","Type":"ContainerStarted","Data":"e4c8bd3bfe5fd6d8aeadbc0586b4af166d665983a75e2f21bdb9f746d5551009"} Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.534514 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6ljl5" event={"ID":"c42d89b6-6420-4267-93d2-e5dac6e1906b","Type":"ContainerStarted","Data":"333b8eb81b113b75d9e3a9f6823d253525ce4c542a7b735a936c9bfab3f79752"} Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.535447 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qzt4s" podStartSLOduration=120.535419445 podStartE2EDuration="2m0.535419445s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:56.534691881 +0000 UTC m=+144.914585543" watchObservedRunningTime="2025-11-24 13:14:56.535419445 +0000 UTC m=+144.915313107" Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.537545 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f55f4" podStartSLOduration=120.537538694 podStartE2EDuration="2m0.537538694s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:56.489378709 +0000 UTC m=+144.869272401" watchObservedRunningTime="2025-11-24 13:14:56.537538694 +0000 UTC m=+144.917432356" Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.553191 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xcw79" event={"ID":"ebb47b17-21a5-4ae0-857d-36f913842d4a","Type":"ContainerStarted","Data":"144956ea8c47e84406f7f45f651ae1150eaadb9fa2b4cd747bfe34741cdf3eb6"} Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.559956 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-cppvr" event={"ID":"d441620d-6188-465c-8531-888b80826aa3","Type":"ContainerStarted","Data":"92f74c93951e6f922e9af9445397f81d0ef07c58bc214f58983e76a33674c5f3"} Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.561318 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-csctv" event={"ID":"163ef77c-3b2f-45ef-81c8-1fc627b1a234","Type":"ContainerStarted","Data":"d6fdd83813da16e532c3a212684a3270354baed33e10ca27d91ba9fbc9325246"} Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.583118 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:56 crc kubenswrapper[4790]: E1124 13:14:56.585063 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:57.085045307 +0000 UTC m=+145.464938969 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.602440 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7d488" event={"ID":"91336463-df6a-4b98-a990-e6780e417fab","Type":"ContainerStarted","Data":"13ee82ac28f0885e62775f7b5ac018ce8b5a256d1bff86216164c6f269b1425d"} Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.617296 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6g5fd" podStartSLOduration=120.617265344 podStartE2EDuration="2m0.617265344s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:56.568262342 +0000 UTC m=+144.948156004" watchObservedRunningTime="2025-11-24 13:14:56.617265344 +0000 UTC m=+144.997159006" Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.617929 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6ljl5" podStartSLOduration=120.617922036 podStartE2EDuration="2m0.617922036s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:56.609176622 +0000 UTC m=+144.989070284" watchObservedRunningTime="2025-11-24 13:14:56.617922036 +0000 UTC m=+144.997815698" Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.626531 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ktdmx" event={"ID":"e4177bcc-76f8-43de-83ad-019fe23543e4","Type":"ContainerStarted","Data":"e19b4d7b9efdf98e3159ff16ec194a99a436012ec4ea9328758c8fa7a908ef59"} Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.630378 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-lnmz8" event={"ID":"d059e3f7-ef28-4557-971c-e77be1472c52","Type":"ContainerStarted","Data":"cd25893073a385901524b552bf1d5c3e78ab816f0ecac2b4cd0a47e454959852"} Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.640352 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7x25b" event={"ID":"d325cb10-86d1-414b-aa0a-282b8d6c1e30","Type":"ContainerStarted","Data":"04107fc841b91cb3fae419313861d1f36dffb6a042626fd705fc0a3cc57d50eb"} Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.644708 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsvx8" event={"ID":"2a2b45ef-e51f-4ad2-a1b7-6d8e7b0c089a","Type":"ContainerStarted","Data":"84a3e740e93ff9a0cd4c7c043f2c5eff9960cd78ba08f68f945b976cb2750aa5"} Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.645538 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsvx8" Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.646959 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-rhnfd" event={"ID":"7531eb66-63f6-4eef-b30a-c3fb1ab94f7a","Type":"ContainerStarted","Data":"e200c15e40bf0ea62d50841d67a012a95e69f3490c52c1ed651ee63689773154"} Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.647935 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-54x2b" event={"ID":"7bc385a4-1703-4a76-b27e-8ae71480381d","Type":"ContainerStarted","Data":"0ecee1e38deaabd7c105a10060d93729fb00330ab7d3ba88e28c778d6ba7e702"} Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.647999 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29399820-jr7cd" podStartSLOduration=120.647974022 podStartE2EDuration="2m0.647974022s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:56.635549989 +0000 UTC m=+145.015443651" watchObservedRunningTime="2025-11-24 13:14:56.647974022 +0000 UTC m=+145.027867684" Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.651294 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-t5fjs" event={"ID":"aef429a6-aa76-4776-a6e2-30096a5e7ccb","Type":"ContainerStarted","Data":"8eaf51275e584017d75742d241cc6c04ee745e5aa9921263e22277a82e17a224"} Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.669495 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-4t9zv" event={"ID":"f6b07b0b-f08b-4ad5-a390-86a8119e96c9","Type":"ContainerStarted","Data":"23f8696a78579d87e43071728bb7ebfbdddaf7832281463127c0b2e75e33bc52"} Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.672050 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xzd7l" event={"ID":"ffdd5f6c-5425-4895-8c2b-6ba6ef3706c9","Type":"ContainerStarted","Data":"95e61aa92c6edb849dcad92f4a6655546e4bdec22f5d644013493ee0a795892c"} Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.674049 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xzd7l" Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.685958 4790 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-xzd7l container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.686040 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xzd7l" podUID="ffdd5f6c-5425-4895-8c2b-6ba6ef3706c9" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.687084 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:56 crc kubenswrapper[4790]: E1124 13:14:56.687330 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:57.18729001 +0000 UTC m=+145.567183672 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.687364 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-cppvr" podStartSLOduration=120.687338851 podStartE2EDuration="2m0.687338851s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:56.660650154 +0000 UTC m=+145.040543816" watchObservedRunningTime="2025-11-24 13:14:56.687338851 +0000 UTC m=+145.067232533" Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.687751 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.688405 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-lnmz8" podStartSLOduration=6.688395656 podStartE2EDuration="6.688395656s" podCreationTimestamp="2025-11-24 13:14:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:56.686315918 +0000 UTC m=+145.066209580" watchObservedRunningTime="2025-11-24 13:14:56.688395656 +0000 UTC m=+145.068289338" Nov 24 13:14:56 crc kubenswrapper[4790]: E1124 13:14:56.698137 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:57.198118602 +0000 UTC m=+145.578012264 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.718723 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6h2jt" event={"ID":"e9a8e814-f2c2-46d1-99c9-40d77e0d96da","Type":"ContainerStarted","Data":"0bebf9f9878fdc90f59ade7c31a344bd24efa6eb38f4cd60dcee3bc26502d6c5"} Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.750695 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hx4vr" event={"ID":"ba3cf9ac-3e57-4274-b5ea-727bf77aebf1","Type":"ContainerStarted","Data":"0d2d0e50d7925d0feb2cd18be859ea5d64e0fc6197cf44c496981195ea6a69e1"} Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.752250 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xzd7l" podStartSLOduration=120.752218699 podStartE2EDuration="2m0.752218699s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:56.738844265 +0000 UTC m=+145.118737937" watchObservedRunningTime="2025-11-24 13:14:56.752218699 +0000 UTC m=+145.132112371" Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.759619 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-zvzkq" event={"ID":"3c7930b6-15dc-4462-bc47-9137c2d4edc7","Type":"ContainerStarted","Data":"bf2c2f79302ecc2839fff356692bcdc482c46b6fcd1b1e89f530a513248ed30b"} Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.761415 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-s7x26" event={"ID":"951c1cb8-6613-404d-b0ec-ca5c8b66ec30","Type":"ContainerStarted","Data":"1cc2d2a338e96c9db5db751721c7e862551d2a108944c87a05f96e2990853495"} Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.766384 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-wvnp9" event={"ID":"6aa4a7b3-bae3-40bb-950d-e907073c0257","Type":"ContainerStarted","Data":"7f777dc83497df7356bd6b29e31be51dfec16da7b1ff08d328f702270093a4c7"} Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.775435 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-79gj2" event={"ID":"1d6e992d-904f-4f11-ab29-29583f1dd955","Type":"ContainerStarted","Data":"75dded4821ea9340a90c731616945134fe9e3d20be8eaf9260cd2cd7eb3a3926"} Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.783130 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-t5fjs" podStartSLOduration=120.783111013 podStartE2EDuration="2m0.783111013s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:56.782341808 +0000 UTC m=+145.162235480" watchObservedRunningTime="2025-11-24 13:14:56.783111013 +0000 UTC m=+145.163004675" Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.800509 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.810058 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5hlqm" event={"ID":"1e6bf7c7-7711-43cb-8817-73b42174b9ee","Type":"ContainerStarted","Data":"6b4432f52651a3983e77c007e6e0500e7bdb93533b99161c682eea220bc3ff2d"} Nov 24 13:14:56 crc kubenswrapper[4790]: E1124 13:14:56.845348 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:57.345296434 +0000 UTC m=+145.725190106 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.864034 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsvx8" podStartSLOduration=120.864003642 podStartE2EDuration="2m0.864003642s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:56.856458787 +0000 UTC m=+145.236352449" watchObservedRunningTime="2025-11-24 13:14:56.864003642 +0000 UTC m=+145.243897304" Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.895713 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" event={"ID":"66c28684-107f-4208-ae79-4b22a4e485dd","Type":"ContainerStarted","Data":"8c6b81b7064006f5acf9345d6b742a17924411d3a34cc70c8de7754405c80799"} Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.902538 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:56 crc kubenswrapper[4790]: E1124 13:14:56.904501 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:57.404476687 +0000 UTC m=+145.784370529 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.924748 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fscrx" event={"ID":"40f41556-7a45-457f-8f60-5fc87df20363","Type":"ContainerStarted","Data":"4f53ff3af9cc4e621ff5141394def6f4d578718afea9967ccc054dbea51bacca"} Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.931500 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-rhnfd" podStartSLOduration=6.931472024 podStartE2EDuration="6.931472024s" podCreationTimestamp="2025-11-24 13:14:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:56.890851124 +0000 UTC m=+145.270744796" watchObservedRunningTime="2025-11-24 13:14:56.931472024 +0000 UTC m=+145.311365686" Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.932430 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-zvzkq" podStartSLOduration=120.932424575 podStartE2EDuration="2m0.932424575s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:56.931076591 +0000 UTC m=+145.310970273" watchObservedRunningTime="2025-11-24 13:14:56.932424575 +0000 UTC m=+145.312318237" Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.941943 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bfmvc" event={"ID":"d6761cdb-c9a1-4a58-9023-e1230eab220d","Type":"ContainerStarted","Data":"7c05850e54f3e8013859f4040e8141055759d0b874564924be9dc07391bca78b"} Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.942191 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bfmvc" Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.955091 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-hkbxm" event={"ID":"207655ea-f838-43a7-ad87-2bee51686a16","Type":"ContainerStarted","Data":"7cba0252a09fb4d60e4415a41cfc7945fe4a25ad23887371c0009f52decff33d"} Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.955141 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-hkbxm" Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.968856 4790 patch_prober.go:28] interesting pod/downloads-7954f5f757-hkbxm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.968928 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-hkbxm" podUID="207655ea-f838-43a7-ad87-2bee51686a16" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.971004 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5hlqm" podStartSLOduration=120.970980768 podStartE2EDuration="2m0.970980768s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:56.970710879 +0000 UTC m=+145.350604541" watchObservedRunningTime="2025-11-24 13:14:56.970980768 +0000 UTC m=+145.350874430" Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.976525 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:14:56 crc kubenswrapper[4790]: I1124 13:14:56.989970 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-cjdhk" Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.006078 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:57 crc kubenswrapper[4790]: E1124 13:14:57.008999 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:57.508968092 +0000 UTC m=+145.888861754 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.059413 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6h2jt" podStartSLOduration=121.059397751 podStartE2EDuration="2m1.059397751s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:57.058305805 +0000 UTC m=+145.438199467" watchObservedRunningTime="2025-11-24 13:14:57.059397751 +0000 UTC m=+145.439291413" Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.070489 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hx4vr" podStartSLOduration=121.0704535 podStartE2EDuration="2m1.0704535s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:57.017175699 +0000 UTC m=+145.397069391" watchObservedRunningTime="2025-11-24 13:14:57.0704535 +0000 UTC m=+145.450347162" Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.108464 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:57 crc kubenswrapper[4790]: E1124 13:14:57.108951 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:57.60893248 +0000 UTC m=+145.988826142 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.188180 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-hkbxm" podStartSLOduration=121.188148064 podStartE2EDuration="2m1.188148064s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:57.187986709 +0000 UTC m=+145.567880371" watchObservedRunningTime="2025-11-24 13:14:57.188148064 +0000 UTC m=+145.568041726" Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.188483 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bfmvc" podStartSLOduration=121.188478375 podStartE2EDuration="2m1.188478375s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:57.154733879 +0000 UTC m=+145.534627541" watchObservedRunningTime="2025-11-24 13:14:57.188478375 +0000 UTC m=+145.568372037" Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.209500 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:57 crc kubenswrapper[4790]: E1124 13:14:57.209942 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:57.709917112 +0000 UTC m=+146.089810774 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.313133 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:57 crc kubenswrapper[4790]: E1124 13:14:57.313723 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:57.813696344 +0000 UTC m=+146.193590006 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.387282 4790 patch_prober.go:28] interesting pod/router-default-5444994796-4ldk4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:14:57 crc kubenswrapper[4790]: [-]has-synced failed: reason withheld Nov 24 13:14:57 crc kubenswrapper[4790]: [+]process-running ok Nov 24 13:14:57 crc kubenswrapper[4790]: healthz check failed Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.387329 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ldk4" podUID="c36b34a9-5050-4b55-886b-83c0dd80a3df" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.418310 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:57 crc kubenswrapper[4790]: E1124 13:14:57.418675 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:57.918659374 +0000 UTC m=+146.298553036 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.520306 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:57 crc kubenswrapper[4790]: E1124 13:14:57.520790 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:58.020774852 +0000 UTC m=+146.400668504 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.560794 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.560842 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.564507 4790 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-dc2xs container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.18:8443/livez\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.564566 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" podUID="5b226a93-5333-405f-93ad-788f9376faac" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.18:8443/livez\": dial tcp 10.217.0.18:8443: connect: connection refused" Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.622332 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:57 crc kubenswrapper[4790]: E1124 13:14:57.622638 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:58.122622112 +0000 UTC m=+146.502515774 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.724419 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:57 crc kubenswrapper[4790]: E1124 13:14:57.725094 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:58.225063671 +0000 UTC m=+146.604957333 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.824908 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:57 crc kubenswrapper[4790]: E1124 13:14:57.825334 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:58.325316878 +0000 UTC m=+146.705210540 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.926608 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:57 crc kubenswrapper[4790]: E1124 13:14:57.927253 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:58.42724131 +0000 UTC m=+146.807134972 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.959607 4790 patch_prober.go:28] interesting pod/console-operator-58897d9998-z2dld container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.959651 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-z2dld" podUID="a2a0fdfc-5465-483d-8961-baa619f6ec0f" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.961049 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-4t9zv" event={"ID":"f6b07b0b-f08b-4ad5-a390-86a8119e96c9","Type":"ContainerStarted","Data":"d95aa08c4c15e02a3170c6185c58de430537c1aeb2db3f64a88a722c2396104a"} Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.963190 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" event={"ID":"66c28684-107f-4208-ae79-4b22a4e485dd","Type":"ContainerStarted","Data":"a8d9a413453f52bdd60fb536653a10f1a31d20f0ea3706ab1c71cdeed3629b29"} Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.965579 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fscrx" event={"ID":"40f41556-7a45-457f-8f60-5fc87df20363","Type":"ContainerStarted","Data":"b32a616f3e10a26919fa9c1bcf71bf5f497e574a9d49f9d9ec48f3ba60d85d18"} Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.965609 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fscrx" event={"ID":"40f41556-7a45-457f-8f60-5fc87df20363","Type":"ContainerStarted","Data":"0654e78f36894d290ca53331f6d898c5342d6cb2e5a6611866d30d1e2d65987a"} Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.967222 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bfmvc" event={"ID":"d6761cdb-c9a1-4a58-9023-e1230eab220d","Type":"ContainerStarted","Data":"076da6cbf6fdf69f859211631b6f6001623922368cbde68315e750f07ff56638"} Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.968909 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-csctv" event={"ID":"163ef77c-3b2f-45ef-81c8-1fc627b1a234","Type":"ContainerStarted","Data":"e57033dab03d743e6ba7a89d271073e733f0d44524120508bd38743a17e6456a"} Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.969266 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-csctv" Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.970577 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-wvnp9" event={"ID":"6aa4a7b3-bae3-40bb-950d-e907073c0257","Type":"ContainerStarted","Data":"89fd88a782e1745d19ecc7e9249d75531ae4358b4d000c1ffe7ed89191f819a6"} Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.970598 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-wvnp9" event={"ID":"6aa4a7b3-bae3-40bb-950d-e907073c0257","Type":"ContainerStarted","Data":"dfe6df663bab8fa3e37c5051b1d1ddf34949c2472cfa2a5c9b722007cf83e57b"} Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.972767 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xcw79" event={"ID":"ebb47b17-21a5-4ae0-857d-36f913842d4a","Type":"ContainerStarted","Data":"d248ae1774c0dbed94e84aab4f32aff4c52ddeb3043f2fbed0093bd06eb80a7c"} Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.974628 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c4qnk" event={"ID":"fd8b1442-101f-4104-ae72-cf6d2029f2d1","Type":"ContainerStarted","Data":"0fa2469e245c213a2c4e773652f7c3083f5b9aab09627137a72316a3c00f4e89"} Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.976504 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7d488" event={"ID":"91336463-df6a-4b98-a990-e6780e417fab","Type":"ContainerStarted","Data":"b0f5903508d1b61160e7752b64fbdb01e1dcdd553b8377fa7229f0d2b8ee7bf5"} Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.978080 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399820-jr7cd" event={"ID":"586ab369-00f4-497e-836f-3aede32089c2","Type":"ContainerStarted","Data":"2619928eaae473e7bc12331d14016eb1b26b13485f0b39df900f3172af893b75"} Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.979753 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-54x2b" event={"ID":"7bc385a4-1703-4a76-b27e-8ae71480381d","Type":"ContainerStarted","Data":"c5fb513d9765012db61de34c4ecf37b28f8093a27cdcfde0ddfe276428e81488"} Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.981667 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-s7x26" event={"ID":"951c1cb8-6613-404d-b0ec-ca5c8b66ec30","Type":"ContainerStarted","Data":"2c08f3df7115b36a938fa7e6de536a3abb3482f5ed97fb396fd826fbdd252083"} Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.982447 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-s7x26" Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.983770 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-79gj2" event={"ID":"1d6e992d-904f-4f11-ab29-29583f1dd955","Type":"ContainerStarted","Data":"7926e91966c8532eafdbe044202620600e318540a073b8048d99e86af1ab9ee0"} Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.984355 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-79gj2" Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.984419 4790 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-s7x26 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.984444 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-s7x26" podUID="951c1cb8-6613-404d-b0ec-ca5c8b66ec30" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.985360 4790 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-79gj2 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.31:5443/healthz\": dial tcp 10.217.0.31:5443: connect: connection refused" start-of-body= Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.985395 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-79gj2" podUID="1d6e992d-904f-4f11-ab29-29583f1dd955" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.31:5443/healthz\": dial tcp 10.217.0.31:5443: connect: connection refused" Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.986225 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ktdmx" event={"ID":"e4177bcc-76f8-43de-83ad-019fe23543e4","Type":"ContainerStarted","Data":"5a308dd20f990d4ba772d94e58fac0c4520fc594b1b25bf512026cefbd8fd45c"} Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.987782 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hx4vr" event={"ID":"ba3cf9ac-3e57-4274-b5ea-727bf77aebf1","Type":"ContainerStarted","Data":"71ddc6ae279802dd42070715db74ebadd965059634f5df1fb905fc24ab16817f"} Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.989414 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7x25b" event={"ID":"d325cb10-86d1-414b-aa0a-282b8d6c1e30","Type":"ContainerStarted","Data":"b749b0af1f2083861c96cc8b1c1ce0e7818db5febafd80b863fa058f9a980844"} Nov 24 13:14:57 crc kubenswrapper[4790]: I1124 13:14:57.991677 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6h2jt" event={"ID":"e9a8e814-f2c2-46d1-99c9-40d77e0d96da","Type":"ContainerStarted","Data":"c8e9b378f05d064a80f8972e1c16170010799cf1794d918cdc65beba8d5692bd"} Nov 24 13:14:58 crc kubenswrapper[4790]: I1124 13:14:58.001496 4790 patch_prober.go:28] interesting pod/downloads-7954f5f757-hkbxm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Nov 24 13:14:58 crc kubenswrapper[4790]: I1124 13:14:58.001555 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-hkbxm" podUID="207655ea-f838-43a7-ad87-2bee51686a16" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" Nov 24 13:14:58 crc kubenswrapper[4790]: I1124 13:14:58.001701 4790 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-xzd7l container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Nov 24 13:14:58 crc kubenswrapper[4790]: I1124 13:14:58.001720 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xzd7l" podUID="ffdd5f6c-5425-4895-8c2b-6ba6ef3706c9" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" Nov 24 13:14:58 crc kubenswrapper[4790]: I1124 13:14:58.002445 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qzt4s" Nov 24 13:14:58 crc kubenswrapper[4790]: I1124 13:14:58.029394 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:58 crc kubenswrapper[4790]: E1124 13:14:58.032364 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:58.532344545 +0000 UTC m=+146.912238207 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:58 crc kubenswrapper[4790]: I1124 13:14:58.132395 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:58 crc kubenswrapper[4790]: E1124 13:14:58.134814 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:58.634795214 +0000 UTC m=+147.014688876 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:58 crc kubenswrapper[4790]: I1124 13:14:58.175404 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" podStartSLOduration=122.175385113 podStartE2EDuration="2m2.175385113s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:58.076769709 +0000 UTC m=+146.456663381" watchObservedRunningTime="2025-11-24 13:14:58.175385113 +0000 UTC m=+146.555278775" Nov 24 13:14:58 crc kubenswrapper[4790]: I1124 13:14:58.229657 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-z2dld" Nov 24 13:14:58 crc kubenswrapper[4790]: I1124 13:14:58.237364 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:58 crc kubenswrapper[4790]: E1124 13:14:58.237852 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:58.737833842 +0000 UTC m=+147.117727504 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:58 crc kubenswrapper[4790]: I1124 13:14:58.263772 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-ktdmx" podStartSLOduration=122.263755155 podStartE2EDuration="2m2.263755155s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:58.254165153 +0000 UTC m=+146.634058815" watchObservedRunningTime="2025-11-24 13:14:58.263755155 +0000 UTC m=+146.643648817" Nov 24 13:14:58 crc kubenswrapper[4790]: I1124 13:14:58.304012 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7x25b" podStartSLOduration=122.303995162 podStartE2EDuration="2m2.303995162s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:58.302593637 +0000 UTC m=+146.682487299" watchObservedRunningTime="2025-11-24 13:14:58.303995162 +0000 UTC m=+146.683888824" Nov 24 13:14:58 crc kubenswrapper[4790]: I1124 13:14:58.339196 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:58 crc kubenswrapper[4790]: E1124 13:14:58.339610 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:58.839593889 +0000 UTC m=+147.219487551 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:58 crc kubenswrapper[4790]: I1124 13:14:58.384553 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-54x2b" podStartSLOduration=122.384537459 podStartE2EDuration="2m2.384537459s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:58.382276756 +0000 UTC m=+146.762170418" watchObservedRunningTime="2025-11-24 13:14:58.384537459 +0000 UTC m=+146.764431121" Nov 24 13:14:58 crc kubenswrapper[4790]: I1124 13:14:58.386601 4790 patch_prober.go:28] interesting pod/router-default-5444994796-4ldk4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:14:58 crc kubenswrapper[4790]: [-]has-synced failed: reason withheld Nov 24 13:14:58 crc kubenswrapper[4790]: [+]process-running ok Nov 24 13:14:58 crc kubenswrapper[4790]: healthz check failed Nov 24 13:14:58 crc kubenswrapper[4790]: I1124 13:14:58.386656 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ldk4" podUID="c36b34a9-5050-4b55-886b-83c0dd80a3df" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:14:58 crc kubenswrapper[4790]: I1124 13:14:58.414593 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xcw79" podStartSLOduration=122.414567895 podStartE2EDuration="2m2.414567895s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:58.410031018 +0000 UTC m=+146.789924680" watchObservedRunningTime="2025-11-24 13:14:58.414567895 +0000 UTC m=+146.794461557" Nov 24 13:14:58 crc kubenswrapper[4790]: I1124 13:14:58.428199 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fscrx" podStartSLOduration=122.428179407 podStartE2EDuration="2m2.428179407s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:58.427978811 +0000 UTC m=+146.807872473" watchObservedRunningTime="2025-11-24 13:14:58.428179407 +0000 UTC m=+146.808073069" Nov 24 13:14:58 crc kubenswrapper[4790]: I1124 13:14:58.443417 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:58 crc kubenswrapper[4790]: E1124 13:14:58.443732 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:58.943717242 +0000 UTC m=+147.323610904 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:58 crc kubenswrapper[4790]: I1124 13:14:58.500674 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c4qnk" podStartSLOduration=122.500660542 podStartE2EDuration="2m2.500660542s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:58.497333994 +0000 UTC m=+146.877227656" watchObservedRunningTime="2025-11-24 13:14:58.500660542 +0000 UTC m=+146.880554204" Nov 24 13:14:58 crc kubenswrapper[4790]: I1124 13:14:58.501229 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-s7x26" podStartSLOduration=122.501223331 podStartE2EDuration="2m2.501223331s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:58.45473925 +0000 UTC m=+146.834632932" watchObservedRunningTime="2025-11-24 13:14:58.501223331 +0000 UTC m=+146.881116993" Nov 24 13:14:58 crc kubenswrapper[4790]: I1124 13:14:58.545565 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:58 crc kubenswrapper[4790]: E1124 13:14:58.545859 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:59.045847491 +0000 UTC m=+147.425741153 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:58 crc kubenswrapper[4790]: I1124 13:14:58.548363 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-wvnp9" podStartSLOduration=122.548351272 podStartE2EDuration="2m2.548351272s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:58.546232543 +0000 UTC m=+146.926126205" watchObservedRunningTime="2025-11-24 13:14:58.548351272 +0000 UTC m=+146.928244934" Nov 24 13:14:58 crc kubenswrapper[4790]: I1124 13:14:58.595762 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-csctv" podStartSLOduration=8.595742992 podStartE2EDuration="8.595742992s" podCreationTimestamp="2025-11-24 13:14:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:58.593702736 +0000 UTC m=+146.973596398" watchObservedRunningTime="2025-11-24 13:14:58.595742992 +0000 UTC m=+146.975636654" Nov 24 13:14:58 crc kubenswrapper[4790]: I1124 13:14:58.621291 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-79gj2" podStartSLOduration=122.621277312 podStartE2EDuration="2m2.621277312s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:58.619004358 +0000 UTC m=+146.998898020" watchObservedRunningTime="2025-11-24 13:14:58.621277312 +0000 UTC m=+147.001170974" Nov 24 13:14:58 crc kubenswrapper[4790]: I1124 13:14:58.648184 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:58 crc kubenswrapper[4790]: E1124 13:14:58.648499 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:59.148483586 +0000 UTC m=+147.528377248 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:58 crc kubenswrapper[4790]: I1124 13:14:58.663292 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7d488" podStartSLOduration=122.663272586 podStartE2EDuration="2m2.663272586s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:14:58.659432892 +0000 UTC m=+147.039326554" watchObservedRunningTime="2025-11-24 13:14:58.663272586 +0000 UTC m=+147.043166248" Nov 24 13:14:58 crc kubenswrapper[4790]: I1124 13:14:58.751681 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:58 crc kubenswrapper[4790]: E1124 13:14:58.751986 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:59.251973239 +0000 UTC m=+147.631866901 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:58 crc kubenswrapper[4790]: I1124 13:14:58.853178 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:58 crc kubenswrapper[4790]: E1124 13:14:58.853951 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:59.353930972 +0000 UTC m=+147.733824634 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:58 crc kubenswrapper[4790]: I1124 13:14:58.954542 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:58 crc kubenswrapper[4790]: E1124 13:14:58.954962 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:59.454951014 +0000 UTC m=+147.834844676 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.043798 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-4t9zv" event={"ID":"f6b07b0b-f08b-4ad5-a390-86a8119e96c9","Type":"ContainerStarted","Data":"2ad88a410505b51ec72d318f7225cc5f7ff4a67339f15d8bcc0ec5fc711edb43"} Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.047430 4790 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-s7x26 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.047465 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-s7x26" podUID="951c1cb8-6613-404d-b0ec-ca5c8b66ec30" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.063620 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:59 crc kubenswrapper[4790]: E1124 13:14:59.063842 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:59.563799621 +0000 UTC m=+147.943693283 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.074581 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:59 crc kubenswrapper[4790]: E1124 13:14:59.075012 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:59.575000935 +0000 UTC m=+147.954894597 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.113141 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xzd7l" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.179420 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:59 crc kubenswrapper[4790]: E1124 13:14:59.181180 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:59.681155434 +0000 UTC m=+148.061049096 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.276903 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsvx8" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.280978 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.281028 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.281063 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.281082 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.281126 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:59 crc kubenswrapper[4790]: E1124 13:14:59.281401 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:59.781378901 +0000 UTC m=+148.161272563 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.281868 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.288447 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.288473 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.292565 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.382240 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:59 crc kubenswrapper[4790]: E1124 13:14:59.382426 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:59.882400333 +0000 UTC m=+148.262293985 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.383023 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:59 crc kubenswrapper[4790]: E1124 13:14:59.383357 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:59.883349564 +0000 UTC m=+148.263243226 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.384339 4790 patch_prober.go:28] interesting pod/router-default-5444994796-4ldk4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:14:59 crc kubenswrapper[4790]: [-]has-synced failed: reason withheld Nov 24 13:14:59 crc kubenswrapper[4790]: [+]process-running ok Nov 24 13:14:59 crc kubenswrapper[4790]: healthz check failed Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.384409 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ldk4" podUID="c36b34a9-5050-4b55-886b-83c0dd80a3df" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.395566 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qxxm7"] Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.396495 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qxxm7" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.415597 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.484210 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:59 crc kubenswrapper[4790]: E1124 13:14:59.484433 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:59.984403558 +0000 UTC m=+148.364297230 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.484508 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/594c7e16-73fa-4653-aa63-9f93214efd12-utilities\") pod \"certified-operators-qxxm7\" (UID: \"594c7e16-73fa-4653-aa63-9f93214efd12\") " pod="openshift-marketplace/certified-operators-qxxm7" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.484658 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvdlx\" (UniqueName: \"kubernetes.io/projected/594c7e16-73fa-4653-aa63-9f93214efd12-kube-api-access-vvdlx\") pod \"certified-operators-qxxm7\" (UID: \"594c7e16-73fa-4653-aa63-9f93214efd12\") " pod="openshift-marketplace/certified-operators-qxxm7" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.484753 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.484798 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/594c7e16-73fa-4653-aa63-9f93214efd12-catalog-content\") pod \"certified-operators-qxxm7\" (UID: \"594c7e16-73fa-4653-aa63-9f93214efd12\") " pod="openshift-marketplace/certified-operators-qxxm7" Nov 24 13:14:59 crc kubenswrapper[4790]: E1124 13:14:59.485211 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:14:59.985201074 +0000 UTC m=+148.365094756 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.490368 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qxxm7"] Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.530543 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.543174 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.549661 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-g5nsf"] Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.550664 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g5nsf" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.556146 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.558372 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.586799 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.587092 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/594c7e16-73fa-4653-aa63-9f93214efd12-catalog-content\") pod \"certified-operators-qxxm7\" (UID: \"594c7e16-73fa-4653-aa63-9f93214efd12\") " pod="openshift-marketplace/certified-operators-qxxm7" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.587130 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c3ca964-4656-4248-bbcd-cd29957699cf-catalog-content\") pod \"community-operators-g5nsf\" (UID: \"3c3ca964-4656-4248-bbcd-cd29957699cf\") " pod="openshift-marketplace/community-operators-g5nsf" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.587167 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/594c7e16-73fa-4653-aa63-9f93214efd12-utilities\") pod \"certified-operators-qxxm7\" (UID: \"594c7e16-73fa-4653-aa63-9f93214efd12\") " pod="openshift-marketplace/certified-operators-qxxm7" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.587215 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvdlx\" (UniqueName: \"kubernetes.io/projected/594c7e16-73fa-4653-aa63-9f93214efd12-kube-api-access-vvdlx\") pod \"certified-operators-qxxm7\" (UID: \"594c7e16-73fa-4653-aa63-9f93214efd12\") " pod="openshift-marketplace/certified-operators-qxxm7" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.587239 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgvhh\" (UniqueName: \"kubernetes.io/projected/3c3ca964-4656-4248-bbcd-cd29957699cf-kube-api-access-mgvhh\") pod \"community-operators-g5nsf\" (UID: \"3c3ca964-4656-4248-bbcd-cd29957699cf\") " pod="openshift-marketplace/community-operators-g5nsf" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.587262 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c3ca964-4656-4248-bbcd-cd29957699cf-utilities\") pod \"community-operators-g5nsf\" (UID: \"3c3ca964-4656-4248-bbcd-cd29957699cf\") " pod="openshift-marketplace/community-operators-g5nsf" Nov 24 13:14:59 crc kubenswrapper[4790]: E1124 13:14:59.587363 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:15:00.087345053 +0000 UTC m=+148.467238715 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.588288 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/594c7e16-73fa-4653-aa63-9f93214efd12-catalog-content\") pod \"certified-operators-qxxm7\" (UID: \"594c7e16-73fa-4653-aa63-9f93214efd12\") " pod="openshift-marketplace/certified-operators-qxxm7" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.588553 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/594c7e16-73fa-4653-aa63-9f93214efd12-utilities\") pod \"certified-operators-qxxm7\" (UID: \"594c7e16-73fa-4653-aa63-9f93214efd12\") " pod="openshift-marketplace/certified-operators-qxxm7" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.609391 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g5nsf"] Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.650704 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvdlx\" (UniqueName: \"kubernetes.io/projected/594c7e16-73fa-4653-aa63-9f93214efd12-kube-api-access-vvdlx\") pod \"certified-operators-qxxm7\" (UID: \"594c7e16-73fa-4653-aa63-9f93214efd12\") " pod="openshift-marketplace/certified-operators-qxxm7" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.690746 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgvhh\" (UniqueName: \"kubernetes.io/projected/3c3ca964-4656-4248-bbcd-cd29957699cf-kube-api-access-mgvhh\") pod \"community-operators-g5nsf\" (UID: \"3c3ca964-4656-4248-bbcd-cd29957699cf\") " pod="openshift-marketplace/community-operators-g5nsf" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.690811 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c3ca964-4656-4248-bbcd-cd29957699cf-utilities\") pod \"community-operators-g5nsf\" (UID: \"3c3ca964-4656-4248-bbcd-cd29957699cf\") " pod="openshift-marketplace/community-operators-g5nsf" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.690846 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.690894 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c3ca964-4656-4248-bbcd-cd29957699cf-catalog-content\") pod \"community-operators-g5nsf\" (UID: \"3c3ca964-4656-4248-bbcd-cd29957699cf\") " pod="openshift-marketplace/community-operators-g5nsf" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.691472 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c3ca964-4656-4248-bbcd-cd29957699cf-catalog-content\") pod \"community-operators-g5nsf\" (UID: \"3c3ca964-4656-4248-bbcd-cd29957699cf\") " pod="openshift-marketplace/community-operators-g5nsf" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.692240 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c3ca964-4656-4248-bbcd-cd29957699cf-utilities\") pod \"community-operators-g5nsf\" (UID: \"3c3ca964-4656-4248-bbcd-cd29957699cf\") " pod="openshift-marketplace/community-operators-g5nsf" Nov 24 13:14:59 crc kubenswrapper[4790]: E1124 13:14:59.692592 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:15:00.192574512 +0000 UTC m=+148.572468174 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.716179 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qxxm7" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.724729 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgvhh\" (UniqueName: \"kubernetes.io/projected/3c3ca964-4656-4248-bbcd-cd29957699cf-kube-api-access-mgvhh\") pod \"community-operators-g5nsf\" (UID: \"3c3ca964-4656-4248-bbcd-cd29957699cf\") " pod="openshift-marketplace/community-operators-g5nsf" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.728322 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7mwgd"] Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.729187 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7mwgd" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.747857 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7mwgd"] Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.795777 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.816466 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/036855a4-3747-4525-bcda-03c03daf917c-utilities\") pod \"certified-operators-7mwgd\" (UID: \"036855a4-3747-4525-bcda-03c03daf917c\") " pod="openshift-marketplace/certified-operators-7mwgd" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.816660 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5wmq\" (UniqueName: \"kubernetes.io/projected/036855a4-3747-4525-bcda-03c03daf917c-kube-api-access-p5wmq\") pod \"certified-operators-7mwgd\" (UID: \"036855a4-3747-4525-bcda-03c03daf917c\") " pod="openshift-marketplace/certified-operators-7mwgd" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.816727 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/036855a4-3747-4525-bcda-03c03daf917c-catalog-content\") pod \"certified-operators-7mwgd\" (UID: \"036855a4-3747-4525-bcda-03c03daf917c\") " pod="openshift-marketplace/certified-operators-7mwgd" Nov 24 13:14:59 crc kubenswrapper[4790]: E1124 13:14:59.829383 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:15:00.329350035 +0000 UTC m=+148.709243697 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.920268 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5wmq\" (UniqueName: \"kubernetes.io/projected/036855a4-3747-4525-bcda-03c03daf917c-kube-api-access-p5wmq\") pod \"certified-operators-7mwgd\" (UID: \"036855a4-3747-4525-bcda-03c03daf917c\") " pod="openshift-marketplace/certified-operators-7mwgd" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.920306 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.920329 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/036855a4-3747-4525-bcda-03c03daf917c-catalog-content\") pod \"certified-operators-7mwgd\" (UID: \"036855a4-3747-4525-bcda-03c03daf917c\") " pod="openshift-marketplace/certified-operators-7mwgd" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.920366 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/036855a4-3747-4525-bcda-03c03daf917c-utilities\") pod \"certified-operators-7mwgd\" (UID: \"036855a4-3747-4525-bcda-03c03daf917c\") " pod="openshift-marketplace/certified-operators-7mwgd" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.920768 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/036855a4-3747-4525-bcda-03c03daf917c-utilities\") pod \"certified-operators-7mwgd\" (UID: \"036855a4-3747-4525-bcda-03c03daf917c\") " pod="openshift-marketplace/certified-operators-7mwgd" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.921256 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/036855a4-3747-4525-bcda-03c03daf917c-catalog-content\") pod \"certified-operators-7mwgd\" (UID: \"036855a4-3747-4525-bcda-03c03daf917c\") " pod="openshift-marketplace/certified-operators-7mwgd" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.922951 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4gq7f"] Nov 24 13:14:59 crc kubenswrapper[4790]: E1124 13:14:59.927280 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:15:00.427257697 +0000 UTC m=+148.807151529 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.932727 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4gq7f" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.955190 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g5nsf" Nov 24 13:14:59 crc kubenswrapper[4790]: I1124 13:14:59.964825 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4gq7f"] Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.026385 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:15:00 crc kubenswrapper[4790]: E1124 13:15:00.026863 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:15:00.526849613 +0000 UTC m=+148.906743275 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.029263 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5wmq\" (UniqueName: \"kubernetes.io/projected/036855a4-3747-4525-bcda-03c03daf917c-kube-api-access-p5wmq\") pod \"certified-operators-7mwgd\" (UID: \"036855a4-3747-4525-bcda-03c03daf917c\") " pod="openshift-marketplace/certified-operators-7mwgd" Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.057322 4790 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-79gj2 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.31:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.057359 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-79gj2" podUID="1d6e992d-904f-4f11-ab29-29583f1dd955" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.31:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.058108 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7mwgd" Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.142330 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6551fecc-df79-44f6-8600-fd016469a1e9-catalog-content\") pod \"community-operators-4gq7f\" (UID: \"6551fecc-df79-44f6-8600-fd016469a1e9\") " pod="openshift-marketplace/community-operators-4gq7f" Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.142418 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6551fecc-df79-44f6-8600-fd016469a1e9-utilities\") pod \"community-operators-4gq7f\" (UID: \"6551fecc-df79-44f6-8600-fd016469a1e9\") " pod="openshift-marketplace/community-operators-4gq7f" Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.142454 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zdnf\" (UniqueName: \"kubernetes.io/projected/6551fecc-df79-44f6-8600-fd016469a1e9-kube-api-access-9zdnf\") pod \"community-operators-4gq7f\" (UID: \"6551fecc-df79-44f6-8600-fd016469a1e9\") " pod="openshift-marketplace/community-operators-4gq7f" Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.142512 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:15:00 crc kubenswrapper[4790]: E1124 13:15:00.142814 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:15:00.64280003 +0000 UTC m=+149.022693702 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.176577 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399820-jr7cd"] Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.183642 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-operator-lifecycle-manager/collect-profiles-29399820-jr7cd" podUID="586ab369-00f4-497e-836f-3aede32089c2" containerName="collect-profiles" containerID="cri-o://2619928eaae473e7bc12331d14016eb1b26b13485f0b39df900f3172af893b75" gracePeriod=30 Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.183809 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-4t9zv" event={"ID":"f6b07b0b-f08b-4ad5-a390-86a8119e96c9","Type":"ContainerStarted","Data":"83d0975dc456c80fd1e6f067a2a8e61fcc5b0e3938de759c043171f1d7da54f6"} Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.185825 4790 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-s7x26 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.185903 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-s7x26" podUID="951c1cb8-6613-404d-b0ec-ca5c8b66ec30" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.203006 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399835-8lrlt"] Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.216233 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-8lrlt" Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.218458 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399835-8lrlt"] Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.243974 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.244129 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6551fecc-df79-44f6-8600-fd016469a1e9-catalog-content\") pod \"community-operators-4gq7f\" (UID: \"6551fecc-df79-44f6-8600-fd016469a1e9\") " pod="openshift-marketplace/community-operators-4gq7f" Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.244174 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6551fecc-df79-44f6-8600-fd016469a1e9-utilities\") pod \"community-operators-4gq7f\" (UID: \"6551fecc-df79-44f6-8600-fd016469a1e9\") " pod="openshift-marketplace/community-operators-4gq7f" Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.244201 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zdnf\" (UniqueName: \"kubernetes.io/projected/6551fecc-df79-44f6-8600-fd016469a1e9-kube-api-access-9zdnf\") pod \"community-operators-4gq7f\" (UID: \"6551fecc-df79-44f6-8600-fd016469a1e9\") " pod="openshift-marketplace/community-operators-4gq7f" Nov 24 13:15:00 crc kubenswrapper[4790]: E1124 13:15:00.250325 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:15:00.750304414 +0000 UTC m=+149.130198076 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.251136 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6551fecc-df79-44f6-8600-fd016469a1e9-catalog-content\") pod \"community-operators-4gq7f\" (UID: \"6551fecc-df79-44f6-8600-fd016469a1e9\") " pod="openshift-marketplace/community-operators-4gq7f" Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.251411 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6551fecc-df79-44f6-8600-fd016469a1e9-utilities\") pod \"community-operators-4gq7f\" (UID: \"6551fecc-df79-44f6-8600-fd016469a1e9\") " pod="openshift-marketplace/community-operators-4gq7f" Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.343926 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zdnf\" (UniqueName: \"kubernetes.io/projected/6551fecc-df79-44f6-8600-fd016469a1e9-kube-api-access-9zdnf\") pod \"community-operators-4gq7f\" (UID: \"6551fecc-df79-44f6-8600-fd016469a1e9\") " pod="openshift-marketplace/community-operators-4gq7f" Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.344917 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kn6v\" (UniqueName: \"kubernetes.io/projected/620bb83f-d131-4cce-b059-63ac16d078db-kube-api-access-2kn6v\") pod \"collect-profiles-29399835-8lrlt\" (UID: \"620bb83f-d131-4cce-b059-63ac16d078db\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-8lrlt" Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.344949 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/620bb83f-d131-4cce-b059-63ac16d078db-secret-volume\") pod \"collect-profiles-29399835-8lrlt\" (UID: \"620bb83f-d131-4cce-b059-63ac16d078db\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-8lrlt" Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.345032 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/620bb83f-d131-4cce-b059-63ac16d078db-config-volume\") pod \"collect-profiles-29399835-8lrlt\" (UID: \"620bb83f-d131-4cce-b059-63ac16d078db\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-8lrlt" Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.345065 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:15:00 crc kubenswrapper[4790]: E1124 13:15:00.347488 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:15:00.847473091 +0000 UTC m=+149.227366753 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.365457 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4gq7f" Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.387180 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-79gj2" Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.388028 4790 patch_prober.go:28] interesting pod/router-default-5444994796-4ldk4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:15:00 crc kubenswrapper[4790]: [-]has-synced failed: reason withheld Nov 24 13:15:00 crc kubenswrapper[4790]: [+]process-running ok Nov 24 13:15:00 crc kubenswrapper[4790]: healthz check failed Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.388224 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ldk4" podUID="c36b34a9-5050-4b55-886b-83c0dd80a3df" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.446055 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.446540 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kn6v\" (UniqueName: \"kubernetes.io/projected/620bb83f-d131-4cce-b059-63ac16d078db-kube-api-access-2kn6v\") pod \"collect-profiles-29399835-8lrlt\" (UID: \"620bb83f-d131-4cce-b059-63ac16d078db\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-8lrlt" Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.446565 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/620bb83f-d131-4cce-b059-63ac16d078db-secret-volume\") pod \"collect-profiles-29399835-8lrlt\" (UID: \"620bb83f-d131-4cce-b059-63ac16d078db\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-8lrlt" Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.446606 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/620bb83f-d131-4cce-b059-63ac16d078db-config-volume\") pod \"collect-profiles-29399835-8lrlt\" (UID: \"620bb83f-d131-4cce-b059-63ac16d078db\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-8lrlt" Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.450727 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/620bb83f-d131-4cce-b059-63ac16d078db-config-volume\") pod \"collect-profiles-29399835-8lrlt\" (UID: \"620bb83f-d131-4cce-b059-63ac16d078db\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-8lrlt" Nov 24 13:15:00 crc kubenswrapper[4790]: E1124 13:15:00.450799 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:15:00.950785768 +0000 UTC m=+149.330679430 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.479812 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/620bb83f-d131-4cce-b059-63ac16d078db-secret-volume\") pod \"collect-profiles-29399835-8lrlt\" (UID: \"620bb83f-d131-4cce-b059-63ac16d078db\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-8lrlt" Nov 24 13:15:00 crc kubenswrapper[4790]: W1124 13:15:00.489509 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-1a6d591f93c6901bdf994750ee67716c444357568a2ce37c4821287d4434d937 WatchSource:0}: Error finding container 1a6d591f93c6901bdf994750ee67716c444357568a2ce37c4821287d4434d937: Status 404 returned error can't find the container with id 1a6d591f93c6901bdf994750ee67716c444357568a2ce37c4821287d4434d937 Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.512567 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kn6v\" (UniqueName: \"kubernetes.io/projected/620bb83f-d131-4cce-b059-63ac16d078db-kube-api-access-2kn6v\") pod \"collect-profiles-29399835-8lrlt\" (UID: \"620bb83f-d131-4cce-b059-63ac16d078db\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-8lrlt" Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.548091 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:15:00 crc kubenswrapper[4790]: E1124 13:15:00.548612 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:15:01.048594836 +0000 UTC m=+149.428488498 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.622587 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qxxm7"] Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.623129 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-8lrlt" Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.651424 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:15:00 crc kubenswrapper[4790]: E1124 13:15:00.651802 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:15:01.151785339 +0000 UTC m=+149.531679001 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.753035 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:15:00 crc kubenswrapper[4790]: E1124 13:15:00.753372 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:15:01.25335939 +0000 UTC m=+149.633253052 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.854312 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:15:00 crc kubenswrapper[4790]: E1124 13:15:00.854645 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:15:01.3546276 +0000 UTC m=+149.734521262 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.870149 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g5nsf"] Nov 24 13:15:00 crc kubenswrapper[4790]: I1124 13:15:00.955869 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:15:00 crc kubenswrapper[4790]: E1124 13:15:00.957033 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:15:01.457019037 +0000 UTC m=+149.836912699 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.045316 4790 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.058293 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:15:01 crc kubenswrapper[4790]: E1124 13:15:01.058753 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:15:01.558737112 +0000 UTC m=+149.938630774 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.106987 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7mwgd"] Nov 24 13:15:01 crc kubenswrapper[4790]: W1124 13:15:01.157333 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod036855a4_3747_4525_bcda_03c03daf917c.slice/crio-4e97071d7f696f2e9a0d01f6f27d8a5c57fa969055c10d4accb3a5b7b6838462 WatchSource:0}: Error finding container 4e97071d7f696f2e9a0d01f6f27d8a5c57fa969055c10d4accb3a5b7b6838462: Status 404 returned error can't find the container with id 4e97071d7f696f2e9a0d01f6f27d8a5c57fa969055c10d4accb3a5b7b6838462 Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.159771 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:15:01 crc kubenswrapper[4790]: E1124 13:15:01.160249 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:15:01.66023536 +0000 UTC m=+150.040129022 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.186530 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operator-lifecycle-manager_collect-profiles-29399820-jr7cd_586ab369-00f4-497e-836f-3aede32089c2/collect-profiles/0.log" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.186624 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399820-jr7cd" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.220218 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mwgd" event={"ID":"036855a4-3747-4525-bcda-03c03daf917c","Type":"ContainerStarted","Data":"4e97071d7f696f2e9a0d01f6f27d8a5c57fa969055c10d4accb3a5b7b6838462"} Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.237753 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"bb98dff451bdd7534e2a5fee1053604a91c719290f43daad663913a6c7b78a24"} Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.261024 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9fn8\" (UniqueName: \"kubernetes.io/projected/586ab369-00f4-497e-836f-3aede32089c2-kube-api-access-g9fn8\") pod \"586ab369-00f4-497e-836f-3aede32089c2\" (UID: \"586ab369-00f4-497e-836f-3aede32089c2\") " Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.261848 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.261940 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/586ab369-00f4-497e-836f-3aede32089c2-secret-volume\") pod \"586ab369-00f4-497e-836f-3aede32089c2\" (UID: \"586ab369-00f4-497e-836f-3aede32089c2\") " Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.261985 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/586ab369-00f4-497e-836f-3aede32089c2-config-volume\") pod \"586ab369-00f4-497e-836f-3aede32089c2\" (UID: \"586ab369-00f4-497e-836f-3aede32089c2\") " Nov 24 13:15:01 crc kubenswrapper[4790]: E1124 13:15:01.262128 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:15:01.76209496 +0000 UTC m=+150.141988622 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.262263 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:15:01 crc kubenswrapper[4790]: E1124 13:15:01.262764 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:15:01.762747121 +0000 UTC m=+150.142640783 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.265350 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/586ab369-00f4-497e-836f-3aede32089c2-config-volume" (OuterVolumeSpecName: "config-volume") pod "586ab369-00f4-497e-836f-3aede32089c2" (UID: "586ab369-00f4-497e-836f-3aede32089c2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.268097 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g5nsf" event={"ID":"3c3ca964-4656-4248-bbcd-cd29957699cf","Type":"ContainerStarted","Data":"c6da93a593b6a7c000003a0430d78bbc1166d5f9239729cda1cf6e7e5e08ddb3"} Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.273273 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/586ab369-00f4-497e-836f-3aede32089c2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "586ab369-00f4-497e-836f-3aede32089c2" (UID: "586ab369-00f4-497e-836f-3aede32089c2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.275695 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"3f59dfbd9378494895fd16e3ceacd56e3a6d5da439a188d959e744c37c94b9af"} Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.277937 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/586ab369-00f4-497e-836f-3aede32089c2-kube-api-access-g9fn8" (OuterVolumeSpecName: "kube-api-access-g9fn8") pod "586ab369-00f4-497e-836f-3aede32089c2" (UID: "586ab369-00f4-497e-836f-3aede32089c2"). InnerVolumeSpecName "kube-api-access-g9fn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.278934 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-4t9zv" event={"ID":"f6b07b0b-f08b-4ad5-a390-86a8119e96c9","Type":"ContainerStarted","Data":"b3f0b41fd846bdac2a3c29a11ef0665d319a72f655d584831317a8602c28fe91"} Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.283769 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qxxm7" event={"ID":"594c7e16-73fa-4653-aa63-9f93214efd12","Type":"ContainerStarted","Data":"87897b4e662cb478bcf88adf746a1f56cc38ebb43e296b983da5ae30160b21f7"} Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.286541 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operator-lifecycle-manager_collect-profiles-29399820-jr7cd_586ab369-00f4-497e-836f-3aede32089c2/collect-profiles/0.log" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.286774 4790 generic.go:334] "Generic (PLEG): container finished" podID="586ab369-00f4-497e-836f-3aede32089c2" containerID="2619928eaae473e7bc12331d14016eb1b26b13485f0b39df900f3172af893b75" exitCode=2 Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.286921 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399820-jr7cd" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.287916 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399820-jr7cd" event={"ID":"586ab369-00f4-497e-836f-3aede32089c2","Type":"ContainerDied","Data":"2619928eaae473e7bc12331d14016eb1b26b13485f0b39df900f3172af893b75"} Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.287984 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399820-jr7cd" event={"ID":"586ab369-00f4-497e-836f-3aede32089c2","Type":"ContainerDied","Data":"e4c8bd3bfe5fd6d8aeadbc0586b4af166d665983a75e2f21bdb9f746d5551009"} Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.288008 4790 scope.go:117] "RemoveContainer" containerID="2619928eaae473e7bc12331d14016eb1b26b13485f0b39df900f3172af893b75" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.299230 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7sjpz"] Nov 24 13:15:01 crc kubenswrapper[4790]: E1124 13:15:01.299569 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="586ab369-00f4-497e-836f-3aede32089c2" containerName="collect-profiles" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.299584 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="586ab369-00f4-497e-836f-3aede32089c2" containerName="collect-profiles" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.299714 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="586ab369-00f4-497e-836f-3aede32089c2" containerName="collect-profiles" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.308729 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"4bf5a3c796a4680fe333bd7e4ece3524f49feb524f55ab667efcfae619fde7a4"} Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.309107 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"1a6d591f93c6901bdf994750ee67716c444357568a2ce37c4821287d4434d937"} Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.310312 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7sjpz" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.315122 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.338543 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-4t9zv" podStartSLOduration=11.338522943 podStartE2EDuration="11.338522943s" podCreationTimestamp="2025-11-24 13:14:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:15:01.319386712 +0000 UTC m=+149.699280374" watchObservedRunningTime="2025-11-24 13:15:01.338522943 +0000 UTC m=+149.718416605" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.338611 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7sjpz"] Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.363805 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.364456 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.364644 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.364955 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9420c18f-997b-4d89-b537-3cb99fe22682-utilities\") pod \"redhat-marketplace-7sjpz\" (UID: \"9420c18f-997b-4d89-b537-3cb99fe22682\") " pod="openshift-marketplace/redhat-marketplace-7sjpz" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.365007 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrnmj\" (UniqueName: \"kubernetes.io/projected/9420c18f-997b-4d89-b537-3cb99fe22682-kube-api-access-qrnmj\") pod \"redhat-marketplace-7sjpz\" (UID: \"9420c18f-997b-4d89-b537-3cb99fe22682\") " pod="openshift-marketplace/redhat-marketplace-7sjpz" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.365044 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9420c18f-997b-4d89-b537-3cb99fe22682-catalog-content\") pod \"redhat-marketplace-7sjpz\" (UID: \"9420c18f-997b-4d89-b537-3cb99fe22682\") " pod="openshift-marketplace/redhat-marketplace-7sjpz" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.365128 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9fn8\" (UniqueName: \"kubernetes.io/projected/586ab369-00f4-497e-836f-3aede32089c2-kube-api-access-g9fn8\") on node \"crc\" DevicePath \"\"" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.365141 4790 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/586ab369-00f4-497e-836f-3aede32089c2-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.365150 4790 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/586ab369-00f4-497e-836f-3aede32089c2-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 13:15:01 crc kubenswrapper[4790]: E1124 13:15:01.365296 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:15:01.865280443 +0000 UTC m=+150.245174105 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.369290 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.369460 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.369667 4790 scope.go:117] "RemoveContainer" containerID="2619928eaae473e7bc12331d14016eb1b26b13485f0b39df900f3172af893b75" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.374021 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399820-jr7cd"] Nov 24 13:15:01 crc kubenswrapper[4790]: E1124 13:15:01.379383 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2619928eaae473e7bc12331d14016eb1b26b13485f0b39df900f3172af893b75\": container with ID starting with 2619928eaae473e7bc12331d14016eb1b26b13485f0b39df900f3172af893b75 not found: ID does not exist" containerID="2619928eaae473e7bc12331d14016eb1b26b13485f0b39df900f3172af893b75" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.379445 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2619928eaae473e7bc12331d14016eb1b26b13485f0b39df900f3172af893b75"} err="failed to get container status \"2619928eaae473e7bc12331d14016eb1b26b13485f0b39df900f3172af893b75\": rpc error: code = NotFound desc = could not find container \"2619928eaae473e7bc12331d14016eb1b26b13485f0b39df900f3172af893b75\": container with ID starting with 2619928eaae473e7bc12331d14016eb1b26b13485f0b39df900f3172af893b75 not found: ID does not exist" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.387130 4790 patch_prober.go:28] interesting pod/router-default-5444994796-4ldk4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:15:01 crc kubenswrapper[4790]: [-]has-synced failed: reason withheld Nov 24 13:15:01 crc kubenswrapper[4790]: [+]process-running ok Nov 24 13:15:01 crc kubenswrapper[4790]: healthz check failed Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.387303 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ldk4" podUID="c36b34a9-5050-4b55-886b-83c0dd80a3df" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.410942 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399820-jr7cd"] Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.413969 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.418811 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4gq7f"] Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.462987 4790 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-24T13:15:01.045355737Z","Handler":null,"Name":""} Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.468614 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9420c18f-997b-4d89-b537-3cb99fe22682-utilities\") pod \"redhat-marketplace-7sjpz\" (UID: \"9420c18f-997b-4d89-b537-3cb99fe22682\") " pod="openshift-marketplace/redhat-marketplace-7sjpz" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.468665 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/be665b83-d54a-46e0-9a52-6f74c116aa61-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"be665b83-d54a-46e0-9a52-6f74c116aa61\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.468713 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrnmj\" (UniqueName: \"kubernetes.io/projected/9420c18f-997b-4d89-b537-3cb99fe22682-kube-api-access-qrnmj\") pod \"redhat-marketplace-7sjpz\" (UID: \"9420c18f-997b-4d89-b537-3cb99fe22682\") " pod="openshift-marketplace/redhat-marketplace-7sjpz" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.468748 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9420c18f-997b-4d89-b537-3cb99fe22682-catalog-content\") pod \"redhat-marketplace-7sjpz\" (UID: \"9420c18f-997b-4d89-b537-3cb99fe22682\") " pod="openshift-marketplace/redhat-marketplace-7sjpz" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.468780 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.468811 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/be665b83-d54a-46e0-9a52-6f74c116aa61-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"be665b83-d54a-46e0-9a52-6f74c116aa61\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.469341 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9420c18f-997b-4d89-b537-3cb99fe22682-utilities\") pod \"redhat-marketplace-7sjpz\" (UID: \"9420c18f-997b-4d89-b537-3cb99fe22682\") " pod="openshift-marketplace/redhat-marketplace-7sjpz" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.469975 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9420c18f-997b-4d89-b537-3cb99fe22682-catalog-content\") pod \"redhat-marketplace-7sjpz\" (UID: \"9420c18f-997b-4d89-b537-3cb99fe22682\") " pod="openshift-marketplace/redhat-marketplace-7sjpz" Nov 24 13:15:01 crc kubenswrapper[4790]: E1124 13:15:01.470221 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:15:01.970209952 +0000 UTC m=+150.350103614 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-thd68" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.499079 4790 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.499116 4790 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.521117 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrnmj\" (UniqueName: \"kubernetes.io/projected/9420c18f-997b-4d89-b537-3cb99fe22682-kube-api-access-qrnmj\") pod \"redhat-marketplace-7sjpz\" (UID: \"9420c18f-997b-4d89-b537-3cb99fe22682\") " pod="openshift-marketplace/redhat-marketplace-7sjpz" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.557287 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399835-8lrlt"] Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.570344 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.570720 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/be665b83-d54a-46e0-9a52-6f74c116aa61-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"be665b83-d54a-46e0-9a52-6f74c116aa61\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.570814 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/be665b83-d54a-46e0-9a52-6f74c116aa61-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"be665b83-d54a-46e0-9a52-6f74c116aa61\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.571283 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/be665b83-d54a-46e0-9a52-6f74c116aa61-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"be665b83-d54a-46e0-9a52-6f74c116aa61\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.575100 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.602581 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/be665b83-d54a-46e0-9a52-6f74c116aa61-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"be665b83-d54a-46e0-9a52-6f74c116aa61\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.670603 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7sjpz" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.671571 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.678006 4790 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.678064 4790 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.699553 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.708614 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4r724"] Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.709717 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4r724" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.727317 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-thd68\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.739679 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4r724"] Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.774636 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d53d9ecd-c244-4a4a-9b0d-f21530ed721e-utilities\") pod \"redhat-marketplace-4r724\" (UID: \"d53d9ecd-c244-4a4a-9b0d-f21530ed721e\") " pod="openshift-marketplace/redhat-marketplace-4r724" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.774719 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d53d9ecd-c244-4a4a-9b0d-f21530ed721e-catalog-content\") pod \"redhat-marketplace-4r724\" (UID: \"d53d9ecd-c244-4a4a-9b0d-f21530ed721e\") " pod="openshift-marketplace/redhat-marketplace-4r724" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.774815 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vb77f\" (UniqueName: \"kubernetes.io/projected/d53d9ecd-c244-4a4a-9b0d-f21530ed721e-kube-api-access-vb77f\") pod \"redhat-marketplace-4r724\" (UID: \"d53d9ecd-c244-4a4a-9b0d-f21530ed721e\") " pod="openshift-marketplace/redhat-marketplace-4r724" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.784540 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.875928 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d53d9ecd-c244-4a4a-9b0d-f21530ed721e-utilities\") pod \"redhat-marketplace-4r724\" (UID: \"d53d9ecd-c244-4a4a-9b0d-f21530ed721e\") " pod="openshift-marketplace/redhat-marketplace-4r724" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.876257 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d53d9ecd-c244-4a4a-9b0d-f21530ed721e-catalog-content\") pod \"redhat-marketplace-4r724\" (UID: \"d53d9ecd-c244-4a4a-9b0d-f21530ed721e\") " pod="openshift-marketplace/redhat-marketplace-4r724" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.876317 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vb77f\" (UniqueName: \"kubernetes.io/projected/d53d9ecd-c244-4a4a-9b0d-f21530ed721e-kube-api-access-vb77f\") pod \"redhat-marketplace-4r724\" (UID: \"d53d9ecd-c244-4a4a-9b0d-f21530ed721e\") " pod="openshift-marketplace/redhat-marketplace-4r724" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.876864 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d53d9ecd-c244-4a4a-9b0d-f21530ed721e-utilities\") pod \"redhat-marketplace-4r724\" (UID: \"d53d9ecd-c244-4a4a-9b0d-f21530ed721e\") " pod="openshift-marketplace/redhat-marketplace-4r724" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.877185 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d53d9ecd-c244-4a4a-9b0d-f21530ed721e-catalog-content\") pod \"redhat-marketplace-4r724\" (UID: \"d53d9ecd-c244-4a4a-9b0d-f21530ed721e\") " pod="openshift-marketplace/redhat-marketplace-4r724" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.896779 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vb77f\" (UniqueName: \"kubernetes.io/projected/d53d9ecd-c244-4a4a-9b0d-f21530ed721e-kube-api-access-vb77f\") pod \"redhat-marketplace-4r724\" (UID: \"d53d9ecd-c244-4a4a-9b0d-f21530ed721e\") " pod="openshift-marketplace/redhat-marketplace-4r724" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.897464 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4r724" Nov 24 13:15:01 crc kubenswrapper[4790]: I1124 13:15:01.966803 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7sjpz"] Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.020806 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 13:15:02 crc kubenswrapper[4790]: W1124 13:15:02.037054 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podbe665b83_d54a_46e0_9a52_6f74c116aa61.slice/crio-ba909656c441255ec02e6f55857101af5c593ae88ed3ce10bbaf96e5e9e9688c WatchSource:0}: Error finding container ba909656c441255ec02e6f55857101af5c593ae88ed3ce10bbaf96e5e9e9688c: Status 404 returned error can't find the container with id ba909656c441255ec02e6f55857101af5c593ae88ed3ce10bbaf96e5e9e9688c Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.234019 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4r724"] Nov 24 13:15:02 crc kubenswrapper[4790]: W1124 13:15:02.254593 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd53d9ecd_c244_4a4a_9b0d_f21530ed721e.slice/crio-b43ae547dc9e75f6eace74a9633b475e05ba96aa4fb5286159e996401dadfbc3 WatchSource:0}: Error finding container b43ae547dc9e75f6eace74a9633b475e05ba96aa4fb5286159e996401dadfbc3: Status 404 returned error can't find the container with id b43ae547dc9e75f6eace74a9633b475e05ba96aa4fb5286159e996401dadfbc3 Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.310443 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-thd68"] Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.312833 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"be665b83-d54a-46e0-9a52-6f74c116aa61","Type":"ContainerStarted","Data":"ba909656c441255ec02e6f55857101af5c593ae88ed3ce10bbaf96e5e9e9688c"} Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.319780 4790 generic.go:334] "Generic (PLEG): container finished" podID="594c7e16-73fa-4653-aa63-9f93214efd12" containerID="eabdb0f6a84865cf987b267782d98e2708385c7918d7c2b55bdc40888793a9b0" exitCode=0 Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.331870 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="586ab369-00f4-497e-836f-3aede32089c2" path="/var/lib/kubelet/pods/586ab369-00f4-497e-836f-3aede32089c2/volumes" Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.332962 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.333447 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qxxm7" event={"ID":"594c7e16-73fa-4653-aa63-9f93214efd12","Type":"ContainerDied","Data":"eabdb0f6a84865cf987b267782d98e2708385c7918d7c2b55bdc40888793a9b0"} Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.333481 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4r724" event={"ID":"d53d9ecd-c244-4a4a-9b0d-f21530ed721e","Type":"ContainerStarted","Data":"b43ae547dc9e75f6eace74a9633b475e05ba96aa4fb5286159e996401dadfbc3"} Nov 24 13:15:02 crc kubenswrapper[4790]: W1124 13:15:02.335211 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6381724d_d16b_4d7b_9d9f_613451b93c77.slice/crio-b3f06e4ad38a76ed234e9f3fb96db72680ff8656cb1b928307a804e0fc9a8fe3 WatchSource:0}: Error finding container b3f06e4ad38a76ed234e9f3fb96db72680ff8656cb1b928307a804e0fc9a8fe3: Status 404 returned error can't find the container with id b3f06e4ad38a76ed234e9f3fb96db72680ff8656cb1b928307a804e0fc9a8fe3 Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.335318 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-8lrlt" event={"ID":"620bb83f-d131-4cce-b059-63ac16d078db","Type":"ContainerStarted","Data":"35193b2e078bbc0405057f2ab163fa82c6b85b36d9d0b00ae061fc71843346c1"} Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.335339 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-8lrlt" event={"ID":"620bb83f-d131-4cce-b059-63ac16d078db","Type":"ContainerStarted","Data":"e4dcc9d6f7c37ca4866c3463a7f23ed766c02c704e704e53a5c8925d0aff97eb"} Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.337450 4790 generic.go:334] "Generic (PLEG): container finished" podID="036855a4-3747-4525-bcda-03c03daf917c" containerID="9d341d978351dc4ba171434b3fb3a4ef21bb8d0db3793155383bd465c04c080a" exitCode=0 Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.337494 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mwgd" event={"ID":"036855a4-3747-4525-bcda-03c03daf917c","Type":"ContainerDied","Data":"9d341d978351dc4ba171434b3fb3a4ef21bb8d0db3793155383bd465c04c080a"} Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.340848 4790 generic.go:334] "Generic (PLEG): container finished" podID="3c3ca964-4656-4248-bbcd-cd29957699cf" containerID="382422e39267f8e3cffa06519a5c49d19a2912c3172f2013f441180e59b7dcac" exitCode=0 Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.340923 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g5nsf" event={"ID":"3c3ca964-4656-4248-bbcd-cd29957699cf","Type":"ContainerDied","Data":"382422e39267f8e3cffa06519a5c49d19a2912c3172f2013f441180e59b7dcac"} Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.342940 4790 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.343697 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"7da2f3e23a606bfa80366c42bf6d5fcb48718021536eb65c75f8aab656055a82"} Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.350846 4790 generic.go:334] "Generic (PLEG): container finished" podID="6551fecc-df79-44f6-8600-fd016469a1e9" containerID="bbdb78fe3c3610827c9f7ae7a89b7acaca4a30ab557ba170224ed4500313107d" exitCode=0 Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.350945 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4gq7f" event={"ID":"6551fecc-df79-44f6-8600-fd016469a1e9","Type":"ContainerDied","Data":"bbdb78fe3c3610827c9f7ae7a89b7acaca4a30ab557ba170224ed4500313107d"} Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.350969 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4gq7f" event={"ID":"6551fecc-df79-44f6-8600-fd016469a1e9","Type":"ContainerStarted","Data":"76bcc228cfd641d37790e8e140007306401c42e33ea3545b4181f5f22205cfe0"} Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.352900 4790 generic.go:334] "Generic (PLEG): container finished" podID="9420c18f-997b-4d89-b537-3cb99fe22682" containerID="ee0a266eda880b7d7eb3ee729adeb053d6151082a6b64edfd8f7690905f84bbf" exitCode=0 Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.353519 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sjpz" event={"ID":"9420c18f-997b-4d89-b537-3cb99fe22682","Type":"ContainerDied","Data":"ee0a266eda880b7d7eb3ee729adeb053d6151082a6b64edfd8f7690905f84bbf"} Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.353573 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sjpz" event={"ID":"9420c18f-997b-4d89-b537-3cb99fe22682","Type":"ContainerStarted","Data":"de788473233b1593d640a745ebd3401285b49f955f81e4a48d3e1b6b7c8cfe56"} Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.358609 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"76da3ab727b8579fea35e79fb749a5a0e7cb0be549978e4ecc5d858c47de0f45"} Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.359080 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.389370 4790 patch_prober.go:28] interesting pod/router-default-5444994796-4ldk4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:15:02 crc kubenswrapper[4790]: [-]has-synced failed: reason withheld Nov 24 13:15:02 crc kubenswrapper[4790]: [+]process-running ok Nov 24 13:15:02 crc kubenswrapper[4790]: healthz check failed Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.389439 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ldk4" podUID="c36b34a9-5050-4b55-886b-83c0dd80a3df" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.566031 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.575816 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dc2xs" Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.694177 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-8lrlt" podStartSLOduration=2.694163633 podStartE2EDuration="2.694163633s" podCreationTimestamp="2025-11-24 13:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:15:02.691094573 +0000 UTC m=+151.070988235" watchObservedRunningTime="2025-11-24 13:15:02.694163633 +0000 UTC m=+151.074057295" Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.705342 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-w2gmf"] Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.706295 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w2gmf" Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.708471 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.718979 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w2gmf"] Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.787960 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37ec51fd-7e4c-4217-9a70-389efca8e855-utilities\") pod \"redhat-operators-w2gmf\" (UID: \"37ec51fd-7e4c-4217-9a70-389efca8e855\") " pod="openshift-marketplace/redhat-operators-w2gmf" Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.788043 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37ec51fd-7e4c-4217-9a70-389efca8e855-catalog-content\") pod \"redhat-operators-w2gmf\" (UID: \"37ec51fd-7e4c-4217-9a70-389efca8e855\") " pod="openshift-marketplace/redhat-operators-w2gmf" Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.788112 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9n2j\" (UniqueName: \"kubernetes.io/projected/37ec51fd-7e4c-4217-9a70-389efca8e855-kube-api-access-b9n2j\") pod \"redhat-operators-w2gmf\" (UID: \"37ec51fd-7e4c-4217-9a70-389efca8e855\") " pod="openshift-marketplace/redhat-operators-w2gmf" Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.889133 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37ec51fd-7e4c-4217-9a70-389efca8e855-utilities\") pod \"redhat-operators-w2gmf\" (UID: \"37ec51fd-7e4c-4217-9a70-389efca8e855\") " pod="openshift-marketplace/redhat-operators-w2gmf" Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.889189 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37ec51fd-7e4c-4217-9a70-389efca8e855-catalog-content\") pod \"redhat-operators-w2gmf\" (UID: \"37ec51fd-7e4c-4217-9a70-389efca8e855\") " pod="openshift-marketplace/redhat-operators-w2gmf" Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.889232 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9n2j\" (UniqueName: \"kubernetes.io/projected/37ec51fd-7e4c-4217-9a70-389efca8e855-kube-api-access-b9n2j\") pod \"redhat-operators-w2gmf\" (UID: \"37ec51fd-7e4c-4217-9a70-389efca8e855\") " pod="openshift-marketplace/redhat-operators-w2gmf" Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.890161 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37ec51fd-7e4c-4217-9a70-389efca8e855-utilities\") pod \"redhat-operators-w2gmf\" (UID: \"37ec51fd-7e4c-4217-9a70-389efca8e855\") " pod="openshift-marketplace/redhat-operators-w2gmf" Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.890239 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37ec51fd-7e4c-4217-9a70-389efca8e855-catalog-content\") pod \"redhat-operators-w2gmf\" (UID: \"37ec51fd-7e4c-4217-9a70-389efca8e855\") " pod="openshift-marketplace/redhat-operators-w2gmf" Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.899400 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.899435 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.908367 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:15:02 crc kubenswrapper[4790]: I1124 13:15:02.913303 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9n2j\" (UniqueName: \"kubernetes.io/projected/37ec51fd-7e4c-4217-9a70-389efca8e855-kube-api-access-b9n2j\") pod \"redhat-operators-w2gmf\" (UID: \"37ec51fd-7e4c-4217-9a70-389efca8e855\") " pod="openshift-marketplace/redhat-operators-w2gmf" Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.020190 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w2gmf" Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.099667 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mv8f6"] Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.101563 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mv8f6" Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.162934 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mv8f6"] Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.194359 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mszvk\" (UniqueName: \"kubernetes.io/projected/d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1-kube-api-access-mszvk\") pod \"redhat-operators-mv8f6\" (UID: \"d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1\") " pod="openshift-marketplace/redhat-operators-mv8f6" Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.194498 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1-utilities\") pod \"redhat-operators-mv8f6\" (UID: \"d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1\") " pod="openshift-marketplace/redhat-operators-mv8f6" Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.194727 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1-catalog-content\") pod \"redhat-operators-mv8f6\" (UID: \"d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1\") " pod="openshift-marketplace/redhat-operators-mv8f6" Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.217129 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-t5fjs" Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.217208 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-t5fjs" Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.220535 4790 patch_prober.go:28] interesting pod/console-f9d7485db-t5fjs container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.220603 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-t5fjs" podUID="aef429a6-aa76-4776-a6e2-30096a5e7ccb" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.296531 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mszvk\" (UniqueName: \"kubernetes.io/projected/d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1-kube-api-access-mszvk\") pod \"redhat-operators-mv8f6\" (UID: \"d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1\") " pod="openshift-marketplace/redhat-operators-mv8f6" Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.296604 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1-utilities\") pod \"redhat-operators-mv8f6\" (UID: \"d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1\") " pod="openshift-marketplace/redhat-operators-mv8f6" Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.296717 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1-catalog-content\") pod \"redhat-operators-mv8f6\" (UID: \"d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1\") " pod="openshift-marketplace/redhat-operators-mv8f6" Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.298336 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1-catalog-content\") pod \"redhat-operators-mv8f6\" (UID: \"d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1\") " pod="openshift-marketplace/redhat-operators-mv8f6" Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.299078 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1-utilities\") pod \"redhat-operators-mv8f6\" (UID: \"d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1\") " pod="openshift-marketplace/redhat-operators-mv8f6" Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.299772 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w2gmf"] Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.321920 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mszvk\" (UniqueName: \"kubernetes.io/projected/d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1-kube-api-access-mszvk\") pod \"redhat-operators-mv8f6\" (UID: \"d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1\") " pod="openshift-marketplace/redhat-operators-mv8f6" Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.366374 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-thd68" event={"ID":"6381724d-d16b-4d7b-9d9f-613451b93c77","Type":"ContainerStarted","Data":"53d539899e076f66813e25210633c7d3b3031b89c9e9cc5d92881c68595075f9"} Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.366432 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-thd68" event={"ID":"6381724d-d16b-4d7b-9d9f-613451b93c77","Type":"ContainerStarted","Data":"b3f06e4ad38a76ed234e9f3fb96db72680ff8656cb1b928307a804e0fc9a8fe3"} Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.367481 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.374480 4790 generic.go:334] "Generic (PLEG): container finished" podID="be665b83-d54a-46e0-9a52-6f74c116aa61" containerID="8e3357ee792f0b56f63aa7cb6effecef2103ed3e8c06f0c42cab62e72301dbb9" exitCode=0 Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.374714 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"be665b83-d54a-46e0-9a52-6f74c116aa61","Type":"ContainerDied","Data":"8e3357ee792f0b56f63aa7cb6effecef2103ed3e8c06f0c42cab62e72301dbb9"} Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.378033 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w2gmf" event={"ID":"37ec51fd-7e4c-4217-9a70-389efca8e855","Type":"ContainerStarted","Data":"294d557adae8a5e5a92d82bf90d00e69291a09f27976c073fddd23fdd74c19b6"} Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.378128 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-4ldk4" Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.386816 4790 generic.go:334] "Generic (PLEG): container finished" podID="d53d9ecd-c244-4a4a-9b0d-f21530ed721e" containerID="dda0d898f910127a5a2a4d11663736ae85ad4df023b7494669f0aa204c5cfd4c" exitCode=0 Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.386935 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4r724" event={"ID":"d53d9ecd-c244-4a4a-9b0d-f21530ed721e","Type":"ContainerDied","Data":"dda0d898f910127a5a2a4d11663736ae85ad4df023b7494669f0aa204c5cfd4c"} Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.409066 4790 patch_prober.go:28] interesting pod/router-default-5444994796-4ldk4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:15:03 crc kubenswrapper[4790]: [-]has-synced failed: reason withheld Nov 24 13:15:03 crc kubenswrapper[4790]: [+]process-running ok Nov 24 13:15:03 crc kubenswrapper[4790]: healthz check failed Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.409659 4790 generic.go:334] "Generic (PLEG): container finished" podID="620bb83f-d131-4cce-b059-63ac16d078db" containerID="35193b2e078bbc0405057f2ab163fa82c6b85b36d9d0b00ae061fc71843346c1" exitCode=0 Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.409767 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ldk4" podUID="c36b34a9-5050-4b55-886b-83c0dd80a3df" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.411655 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-8lrlt" event={"ID":"620bb83f-d131-4cce-b059-63ac16d078db","Type":"ContainerDied","Data":"35193b2e078bbc0405057f2ab163fa82c6b85b36d9d0b00ae061fc71843346c1"} Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.420186 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-8jkhv" Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.425743 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mv8f6" Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.425917 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-thd68" podStartSLOduration=127.425892579 podStartE2EDuration="2m7.425892579s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:15:03.405511626 +0000 UTC m=+151.785405288" watchObservedRunningTime="2025-11-24 13:15:03.425892579 +0000 UTC m=+151.805786241" Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.512536 4790 patch_prober.go:28] interesting pod/downloads-7954f5f757-hkbxm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.512618 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-hkbxm" podUID="207655ea-f838-43a7-ad87-2bee51686a16" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.513099 4790 patch_prober.go:28] interesting pod/downloads-7954f5f757-hkbxm container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.513133 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-hkbxm" podUID="207655ea-f838-43a7-ad87-2bee51686a16" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" Nov 24 13:15:03 crc kubenswrapper[4790]: I1124 13:15:03.738690 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-s7x26" Nov 24 13:15:04 crc kubenswrapper[4790]: I1124 13:15:04.031088 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mv8f6"] Nov 24 13:15:04 crc kubenswrapper[4790]: I1124 13:15:04.383342 4790 patch_prober.go:28] interesting pod/router-default-5444994796-4ldk4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:15:04 crc kubenswrapper[4790]: [-]has-synced failed: reason withheld Nov 24 13:15:04 crc kubenswrapper[4790]: [+]process-running ok Nov 24 13:15:04 crc kubenswrapper[4790]: healthz check failed Nov 24 13:15:04 crc kubenswrapper[4790]: I1124 13:15:04.383680 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ldk4" podUID="c36b34a9-5050-4b55-886b-83c0dd80a3df" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:15:04 crc kubenswrapper[4790]: I1124 13:15:04.430291 4790 generic.go:334] "Generic (PLEG): container finished" podID="37ec51fd-7e4c-4217-9a70-389efca8e855" containerID="4ba17f52f39ed2f3f39e53f8da98fef69c050d9079a9123eb0da0849c194c5f2" exitCode=0 Nov 24 13:15:04 crc kubenswrapper[4790]: I1124 13:15:04.430394 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w2gmf" event={"ID":"37ec51fd-7e4c-4217-9a70-389efca8e855","Type":"ContainerDied","Data":"4ba17f52f39ed2f3f39e53f8da98fef69c050d9079a9123eb0da0849c194c5f2"} Nov 24 13:15:04 crc kubenswrapper[4790]: I1124 13:15:04.445859 4790 generic.go:334] "Generic (PLEG): container finished" podID="d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1" containerID="4eee42fbff43685e52cb3f5de3cd8caa32ef06d3c8ae0e40292bd0c268f92cdd" exitCode=0 Nov 24 13:15:04 crc kubenswrapper[4790]: I1124 13:15:04.446774 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mv8f6" event={"ID":"d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1","Type":"ContainerDied","Data":"4eee42fbff43685e52cb3f5de3cd8caa32ef06d3c8ae0e40292bd0c268f92cdd"} Nov 24 13:15:04 crc kubenswrapper[4790]: I1124 13:15:04.446803 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mv8f6" event={"ID":"d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1","Type":"ContainerStarted","Data":"0e6b15d1a276e0d320d54bd0e7740f859de75116d051cd91be1083e34f4031d9"} Nov 24 13:15:04 crc kubenswrapper[4790]: I1124 13:15:04.815720 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-8lrlt" Nov 24 13:15:04 crc kubenswrapper[4790]: I1124 13:15:04.828784 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 13:15:04 crc kubenswrapper[4790]: I1124 13:15:04.854541 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kn6v\" (UniqueName: \"kubernetes.io/projected/620bb83f-d131-4cce-b059-63ac16d078db-kube-api-access-2kn6v\") pod \"620bb83f-d131-4cce-b059-63ac16d078db\" (UID: \"620bb83f-d131-4cce-b059-63ac16d078db\") " Nov 24 13:15:04 crc kubenswrapper[4790]: I1124 13:15:04.854601 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/be665b83-d54a-46e0-9a52-6f74c116aa61-kubelet-dir\") pod \"be665b83-d54a-46e0-9a52-6f74c116aa61\" (UID: \"be665b83-d54a-46e0-9a52-6f74c116aa61\") " Nov 24 13:15:04 crc kubenswrapper[4790]: I1124 13:15:04.854652 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/be665b83-d54a-46e0-9a52-6f74c116aa61-kube-api-access\") pod \"be665b83-d54a-46e0-9a52-6f74c116aa61\" (UID: \"be665b83-d54a-46e0-9a52-6f74c116aa61\") " Nov 24 13:15:04 crc kubenswrapper[4790]: I1124 13:15:04.854682 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/620bb83f-d131-4cce-b059-63ac16d078db-secret-volume\") pod \"620bb83f-d131-4cce-b059-63ac16d078db\" (UID: \"620bb83f-d131-4cce-b059-63ac16d078db\") " Nov 24 13:15:04 crc kubenswrapper[4790]: I1124 13:15:04.854711 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/620bb83f-d131-4cce-b059-63ac16d078db-config-volume\") pod \"620bb83f-d131-4cce-b059-63ac16d078db\" (UID: \"620bb83f-d131-4cce-b059-63ac16d078db\") " Nov 24 13:15:04 crc kubenswrapper[4790]: I1124 13:15:04.854769 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/be665b83-d54a-46e0-9a52-6f74c116aa61-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "be665b83-d54a-46e0-9a52-6f74c116aa61" (UID: "be665b83-d54a-46e0-9a52-6f74c116aa61"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:15:04 crc kubenswrapper[4790]: I1124 13:15:04.854899 4790 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/be665b83-d54a-46e0-9a52-6f74c116aa61-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 24 13:15:04 crc kubenswrapper[4790]: I1124 13:15:04.855747 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/620bb83f-d131-4cce-b059-63ac16d078db-config-volume" (OuterVolumeSpecName: "config-volume") pod "620bb83f-d131-4cce-b059-63ac16d078db" (UID: "620bb83f-d131-4cce-b059-63ac16d078db"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:15:04 crc kubenswrapper[4790]: I1124 13:15:04.864997 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/620bb83f-d131-4cce-b059-63ac16d078db-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "620bb83f-d131-4cce-b059-63ac16d078db" (UID: "620bb83f-d131-4cce-b059-63ac16d078db"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:15:04 crc kubenswrapper[4790]: I1124 13:15:04.865002 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be665b83-d54a-46e0-9a52-6f74c116aa61-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "be665b83-d54a-46e0-9a52-6f74c116aa61" (UID: "be665b83-d54a-46e0-9a52-6f74c116aa61"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:15:04 crc kubenswrapper[4790]: I1124 13:15:04.902253 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/620bb83f-d131-4cce-b059-63ac16d078db-kube-api-access-2kn6v" (OuterVolumeSpecName: "kube-api-access-2kn6v") pod "620bb83f-d131-4cce-b059-63ac16d078db" (UID: "620bb83f-d131-4cce-b059-63ac16d078db"). InnerVolumeSpecName "kube-api-access-2kn6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:15:04 crc kubenswrapper[4790]: I1124 13:15:04.957816 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/be665b83-d54a-46e0-9a52-6f74c116aa61-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 13:15:04 crc kubenswrapper[4790]: I1124 13:15:04.957908 4790 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/620bb83f-d131-4cce-b059-63ac16d078db-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 13:15:04 crc kubenswrapper[4790]: I1124 13:15:04.957951 4790 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/620bb83f-d131-4cce-b059-63ac16d078db-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 13:15:04 crc kubenswrapper[4790]: I1124 13:15:04.957969 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kn6v\" (UniqueName: \"kubernetes.io/projected/620bb83f-d131-4cce-b059-63ac16d078db-kube-api-access-2kn6v\") on node \"crc\" DevicePath \"\"" Nov 24 13:15:05 crc kubenswrapper[4790]: I1124 13:15:05.382712 4790 patch_prober.go:28] interesting pod/router-default-5444994796-4ldk4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:15:05 crc kubenswrapper[4790]: [-]has-synced failed: reason withheld Nov 24 13:15:05 crc kubenswrapper[4790]: [+]process-running ok Nov 24 13:15:05 crc kubenswrapper[4790]: healthz check failed Nov 24 13:15:05 crc kubenswrapper[4790]: I1124 13:15:05.383283 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ldk4" podUID="c36b34a9-5050-4b55-886b-83c0dd80a3df" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:15:05 crc kubenswrapper[4790]: I1124 13:15:05.468521 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"be665b83-d54a-46e0-9a52-6f74c116aa61","Type":"ContainerDied","Data":"ba909656c441255ec02e6f55857101af5c593ae88ed3ce10bbaf96e5e9e9688c"} Nov 24 13:15:05 crc kubenswrapper[4790]: I1124 13:15:05.468561 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba909656c441255ec02e6f55857101af5c593ae88ed3ce10bbaf96e5e9e9688c" Nov 24 13:15:05 crc kubenswrapper[4790]: I1124 13:15:05.468539 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 13:15:05 crc kubenswrapper[4790]: I1124 13:15:05.476210 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-8lrlt" event={"ID":"620bb83f-d131-4cce-b059-63ac16d078db","Type":"ContainerDied","Data":"e4dcc9d6f7c37ca4866c3463a7f23ed766c02c704e704e53a5c8925d0aff97eb"} Nov 24 13:15:05 crc kubenswrapper[4790]: I1124 13:15:05.476263 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-8lrlt" Nov 24 13:15:05 crc kubenswrapper[4790]: I1124 13:15:05.476300 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4dcc9d6f7c37ca4866c3463a7f23ed766c02c704e704e53a5c8925d0aff97eb" Nov 24 13:15:05 crc kubenswrapper[4790]: I1124 13:15:05.593287 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-csctv" Nov 24 13:15:05 crc kubenswrapper[4790]: I1124 13:15:05.978550 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 13:15:05 crc kubenswrapper[4790]: E1124 13:15:05.978787 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be665b83-d54a-46e0-9a52-6f74c116aa61" containerName="pruner" Nov 24 13:15:05 crc kubenswrapper[4790]: I1124 13:15:05.978799 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="be665b83-d54a-46e0-9a52-6f74c116aa61" containerName="pruner" Nov 24 13:15:05 crc kubenswrapper[4790]: E1124 13:15:05.978810 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="620bb83f-d131-4cce-b059-63ac16d078db" containerName="collect-profiles" Nov 24 13:15:05 crc kubenswrapper[4790]: I1124 13:15:05.978816 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="620bb83f-d131-4cce-b059-63ac16d078db" containerName="collect-profiles" Nov 24 13:15:05 crc kubenswrapper[4790]: I1124 13:15:05.978941 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="620bb83f-d131-4cce-b059-63ac16d078db" containerName="collect-profiles" Nov 24 13:15:05 crc kubenswrapper[4790]: I1124 13:15:05.978955 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="be665b83-d54a-46e0-9a52-6f74c116aa61" containerName="pruner" Nov 24 13:15:05 crc kubenswrapper[4790]: I1124 13:15:05.979305 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 13:15:05 crc kubenswrapper[4790]: I1124 13:15:05.982148 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 24 13:15:05 crc kubenswrapper[4790]: I1124 13:15:05.982429 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 24 13:15:06 crc kubenswrapper[4790]: I1124 13:15:06.025111 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 13:15:06 crc kubenswrapper[4790]: I1124 13:15:06.081757 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2420a87e-7aee-42d7-a61e-720f3a014682-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"2420a87e-7aee-42d7-a61e-720f3a014682\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 13:15:06 crc kubenswrapper[4790]: I1124 13:15:06.081842 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2420a87e-7aee-42d7-a61e-720f3a014682-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"2420a87e-7aee-42d7-a61e-720f3a014682\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 13:15:06 crc kubenswrapper[4790]: I1124 13:15:06.182692 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2420a87e-7aee-42d7-a61e-720f3a014682-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"2420a87e-7aee-42d7-a61e-720f3a014682\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 13:15:06 crc kubenswrapper[4790]: I1124 13:15:06.182772 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2420a87e-7aee-42d7-a61e-720f3a014682-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"2420a87e-7aee-42d7-a61e-720f3a014682\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 13:15:06 crc kubenswrapper[4790]: I1124 13:15:06.182956 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2420a87e-7aee-42d7-a61e-720f3a014682-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"2420a87e-7aee-42d7-a61e-720f3a014682\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 13:15:06 crc kubenswrapper[4790]: I1124 13:15:06.209240 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2420a87e-7aee-42d7-a61e-720f3a014682-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"2420a87e-7aee-42d7-a61e-720f3a014682\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 13:15:06 crc kubenswrapper[4790]: I1124 13:15:06.308780 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 13:15:06 crc kubenswrapper[4790]: I1124 13:15:06.381810 4790 patch_prober.go:28] interesting pod/router-default-5444994796-4ldk4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:15:06 crc kubenswrapper[4790]: [-]has-synced failed: reason withheld Nov 24 13:15:06 crc kubenswrapper[4790]: [+]process-running ok Nov 24 13:15:06 crc kubenswrapper[4790]: healthz check failed Nov 24 13:15:06 crc kubenswrapper[4790]: I1124 13:15:06.381930 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ldk4" podUID="c36b34a9-5050-4b55-886b-83c0dd80a3df" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:15:06 crc kubenswrapper[4790]: I1124 13:15:06.873442 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 13:15:07 crc kubenswrapper[4790]: I1124 13:15:07.387256 4790 patch_prober.go:28] interesting pod/router-default-5444994796-4ldk4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:15:07 crc kubenswrapper[4790]: [-]has-synced failed: reason withheld Nov 24 13:15:07 crc kubenswrapper[4790]: [+]process-running ok Nov 24 13:15:07 crc kubenswrapper[4790]: healthz check failed Nov 24 13:15:07 crc kubenswrapper[4790]: I1124 13:15:07.387510 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ldk4" podUID="c36b34a9-5050-4b55-886b-83c0dd80a3df" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:15:07 crc kubenswrapper[4790]: I1124 13:15:07.541092 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"2420a87e-7aee-42d7-a61e-720f3a014682","Type":"ContainerStarted","Data":"9e5f1da513025699ef5b3fe4cd868b480752fc92e461a26e38a7404619bb8a2e"} Nov 24 13:15:08 crc kubenswrapper[4790]: I1124 13:15:08.382664 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-4ldk4" Nov 24 13:15:08 crc kubenswrapper[4790]: I1124 13:15:08.386055 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-4ldk4" Nov 24 13:15:08 crc kubenswrapper[4790]: I1124 13:15:08.585443 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"2420a87e-7aee-42d7-a61e-720f3a014682","Type":"ContainerStarted","Data":"a79db9d72a9df6368db0caaa68b90e3e3411afcaf8133b2c874bb6801d72d07d"} Nov 24 13:15:09 crc kubenswrapper[4790]: I1124 13:15:09.614553 4790 generic.go:334] "Generic (PLEG): container finished" podID="2420a87e-7aee-42d7-a61e-720f3a014682" containerID="a79db9d72a9df6368db0caaa68b90e3e3411afcaf8133b2c874bb6801d72d07d" exitCode=0 Nov 24 13:15:09 crc kubenswrapper[4790]: I1124 13:15:09.614628 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"2420a87e-7aee-42d7-a61e-720f3a014682","Type":"ContainerDied","Data":"a79db9d72a9df6368db0caaa68b90e3e3411afcaf8133b2c874bb6801d72d07d"} Nov 24 13:15:13 crc kubenswrapper[4790]: I1124 13:15:13.228102 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-t5fjs" Nov 24 13:15:13 crc kubenswrapper[4790]: I1124 13:15:13.233909 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-t5fjs" Nov 24 13:15:13 crc kubenswrapper[4790]: I1124 13:15:13.505707 4790 patch_prober.go:28] interesting pod/downloads-7954f5f757-hkbxm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Nov 24 13:15:13 crc kubenswrapper[4790]: I1124 13:15:13.505755 4790 patch_prober.go:28] interesting pod/downloads-7954f5f757-hkbxm container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Nov 24 13:15:13 crc kubenswrapper[4790]: I1124 13:15:13.506031 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-hkbxm" podUID="207655ea-f838-43a7-ad87-2bee51686a16" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" Nov 24 13:15:13 crc kubenswrapper[4790]: I1124 13:15:13.506091 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-hkbxm" podUID="207655ea-f838-43a7-ad87-2bee51686a16" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" Nov 24 13:15:13 crc kubenswrapper[4790]: I1124 13:15:13.939176 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:15:13 crc kubenswrapper[4790]: I1124 13:15:13.939268 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:15:18 crc kubenswrapper[4790]: I1124 13:15:18.603594 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bef2f789-12a8-45b6-b875-7307992176ee-metrics-certs\") pod \"network-metrics-daemon-vdtxl\" (UID: \"bef2f789-12a8-45b6-b875-7307992176ee\") " pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:15:18 crc kubenswrapper[4790]: I1124 13:15:18.615139 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bef2f789-12a8-45b6-b875-7307992176ee-metrics-certs\") pod \"network-metrics-daemon-vdtxl\" (UID: \"bef2f789-12a8-45b6-b875-7307992176ee\") " pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:15:18 crc kubenswrapper[4790]: I1124 13:15:18.761478 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdtxl" Nov 24 13:15:20 crc kubenswrapper[4790]: I1124 13:15:20.778788 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 13:15:20 crc kubenswrapper[4790]: I1124 13:15:20.834686 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2420a87e-7aee-42d7-a61e-720f3a014682-kubelet-dir\") pod \"2420a87e-7aee-42d7-a61e-720f3a014682\" (UID: \"2420a87e-7aee-42d7-a61e-720f3a014682\") " Nov 24 13:15:20 crc kubenswrapper[4790]: I1124 13:15:20.834792 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2420a87e-7aee-42d7-a61e-720f3a014682-kube-api-access\") pod \"2420a87e-7aee-42d7-a61e-720f3a014682\" (UID: \"2420a87e-7aee-42d7-a61e-720f3a014682\") " Nov 24 13:15:20 crc kubenswrapper[4790]: I1124 13:15:20.835012 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2420a87e-7aee-42d7-a61e-720f3a014682-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "2420a87e-7aee-42d7-a61e-720f3a014682" (UID: "2420a87e-7aee-42d7-a61e-720f3a014682"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:15:20 crc kubenswrapper[4790]: I1124 13:15:20.835350 4790 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2420a87e-7aee-42d7-a61e-720f3a014682-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 24 13:15:20 crc kubenswrapper[4790]: I1124 13:15:20.839222 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2420a87e-7aee-42d7-a61e-720f3a014682-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "2420a87e-7aee-42d7-a61e-720f3a014682" (UID: "2420a87e-7aee-42d7-a61e-720f3a014682"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:15:20 crc kubenswrapper[4790]: I1124 13:15:20.936569 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2420a87e-7aee-42d7-a61e-720f3a014682-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 13:15:21 crc kubenswrapper[4790]: I1124 13:15:21.718611 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"2420a87e-7aee-42d7-a61e-720f3a014682","Type":"ContainerDied","Data":"9e5f1da513025699ef5b3fe4cd868b480752fc92e461a26e38a7404619bb8a2e"} Nov 24 13:15:21 crc kubenswrapper[4790]: I1124 13:15:21.718942 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e5f1da513025699ef5b3fe4cd868b480752fc92e461a26e38a7404619bb8a2e" Nov 24 13:15:21 crc kubenswrapper[4790]: I1124 13:15:21.719011 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 13:15:21 crc kubenswrapper[4790]: I1124 13:15:21.791552 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:15:23 crc kubenswrapper[4790]: I1124 13:15:23.510202 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-hkbxm" Nov 24 13:15:32 crc kubenswrapper[4790]: E1124 13:15:32.274495 4790 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 24 13:15:32 crc kubenswrapper[4790]: E1124 13:15:32.275177 4790 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qrnmj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-7sjpz_openshift-marketplace(9420c18f-997b-4d89-b537-3cb99fe22682): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 13:15:32 crc kubenswrapper[4790]: E1124 13:15:32.276673 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-7sjpz" podUID="9420c18f-997b-4d89-b537-3cb99fe22682" Nov 24 13:15:33 crc kubenswrapper[4790]: E1124 13:15:33.568462 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7sjpz" podUID="9420c18f-997b-4d89-b537-3cb99fe22682" Nov 24 13:15:33 crc kubenswrapper[4790]: I1124 13:15:33.818173 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bfmvc" Nov 24 13:15:36 crc kubenswrapper[4790]: E1124 13:15:36.224896 4790 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 24 13:15:36 crc kubenswrapper[4790]: E1124 13:15:36.225849 4790 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b9n2j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-w2gmf_openshift-marketplace(37ec51fd-7e4c-4217-9a70-389efca8e855): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 13:15:36 crc kubenswrapper[4790]: E1124 13:15:36.232877 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-w2gmf" podUID="37ec51fd-7e4c-4217-9a70-389efca8e855" Nov 24 13:15:36 crc kubenswrapper[4790]: E1124 13:15:36.246741 4790 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 24 13:15:36 crc kubenswrapper[4790]: E1124 13:15:36.246907 4790 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vb77f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-4r724_openshift-marketplace(d53d9ecd-c244-4a4a-9b0d-f21530ed721e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 13:15:36 crc kubenswrapper[4790]: E1124 13:15:36.248223 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-4r724" podUID="d53d9ecd-c244-4a4a-9b0d-f21530ed721e" Nov 24 13:15:36 crc kubenswrapper[4790]: E1124 13:15:36.290641 4790 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 24 13:15:36 crc kubenswrapper[4790]: E1124 13:15:36.290809 4790 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mszvk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-mv8f6_openshift-marketplace(d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 13:15:36 crc kubenswrapper[4790]: E1124 13:15:36.291999 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-mv8f6" podUID="d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1" Nov 24 13:15:37 crc kubenswrapper[4790]: E1124 13:15:37.384799 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-w2gmf" podUID="37ec51fd-7e4c-4217-9a70-389efca8e855" Nov 24 13:15:37 crc kubenswrapper[4790]: E1124 13:15:37.384867 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-mv8f6" podUID="d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1" Nov 24 13:15:37 crc kubenswrapper[4790]: E1124 13:15:37.384931 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-4r724" podUID="d53d9ecd-c244-4a4a-9b0d-f21530ed721e" Nov 24 13:15:37 crc kubenswrapper[4790]: E1124 13:15:37.475259 4790 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 24 13:15:37 crc kubenswrapper[4790]: E1124 13:15:37.475500 4790 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mgvhh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-g5nsf_openshift-marketplace(3c3ca964-4656-4248-bbcd-cd29957699cf): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 13:15:37 crc kubenswrapper[4790]: E1124 13:15:37.480045 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-g5nsf" podUID="3c3ca964-4656-4248-bbcd-cd29957699cf" Nov 24 13:15:37 crc kubenswrapper[4790]: E1124 13:15:37.480242 4790 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 24 13:15:37 crc kubenswrapper[4790]: E1124 13:15:37.480378 4790 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vvdlx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-qxxm7_openshift-marketplace(594c7e16-73fa-4653-aa63-9f93214efd12): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 13:15:37 crc kubenswrapper[4790]: E1124 13:15:37.481556 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-qxxm7" podUID="594c7e16-73fa-4653-aa63-9f93214efd12" Nov 24 13:15:37 crc kubenswrapper[4790]: E1124 13:15:37.497380 4790 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 24 13:15:37 crc kubenswrapper[4790]: E1124 13:15:37.499733 4790 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p5wmq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-7mwgd_openshift-marketplace(036855a4-3747-4525-bcda-03c03daf917c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 13:15:37 crc kubenswrapper[4790]: E1124 13:15:37.501594 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-7mwgd" podUID="036855a4-3747-4525-bcda-03c03daf917c" Nov 24 13:15:37 crc kubenswrapper[4790]: I1124 13:15:37.806992 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4gq7f" event={"ID":"6551fecc-df79-44f6-8600-fd016469a1e9","Type":"ContainerStarted","Data":"665468361d8cd7a472d4411803d78ac2e935b6b1d74605e0771e7aa03b4a442d"} Nov 24 13:15:37 crc kubenswrapper[4790]: E1124 13:15:37.810589 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-7mwgd" podUID="036855a4-3747-4525-bcda-03c03daf917c" Nov 24 13:15:37 crc kubenswrapper[4790]: E1124 13:15:37.810590 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-qxxm7" podUID="594c7e16-73fa-4653-aa63-9f93214efd12" Nov 24 13:15:37 crc kubenswrapper[4790]: E1124 13:15:37.813246 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-g5nsf" podUID="3c3ca964-4656-4248-bbcd-cd29957699cf" Nov 24 13:15:37 crc kubenswrapper[4790]: I1124 13:15:37.821533 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-vdtxl"] Nov 24 13:15:37 crc kubenswrapper[4790]: W1124 13:15:37.834261 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbef2f789_12a8_45b6_b875_7307992176ee.slice/crio-d3d3de9b917e9b72581ee1f761abfa62b773c38e9d8d427e011845744bc0e2f0 WatchSource:0}: Error finding container d3d3de9b917e9b72581ee1f761abfa62b773c38e9d8d427e011845744bc0e2f0: Status 404 returned error can't find the container with id d3d3de9b917e9b72581ee1f761abfa62b773c38e9d8d427e011845744bc0e2f0 Nov 24 13:15:38 crc kubenswrapper[4790]: I1124 13:15:38.821982 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-vdtxl" event={"ID":"bef2f789-12a8-45b6-b875-7307992176ee","Type":"ContainerStarted","Data":"88f5dfb7410cba47c53b90706da64acfee1af4acc9f57d78c85a3ca5fe5a01d3"} Nov 24 13:15:38 crc kubenswrapper[4790]: I1124 13:15:38.823589 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-vdtxl" event={"ID":"bef2f789-12a8-45b6-b875-7307992176ee","Type":"ContainerStarted","Data":"f00c1e6675b002a7606ffe9c0f63a792c84ce86573ba3775cfdc95006c8f871f"} Nov 24 13:15:38 crc kubenswrapper[4790]: I1124 13:15:38.823688 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-vdtxl" event={"ID":"bef2f789-12a8-45b6-b875-7307992176ee","Type":"ContainerStarted","Data":"d3d3de9b917e9b72581ee1f761abfa62b773c38e9d8d427e011845744bc0e2f0"} Nov 24 13:15:38 crc kubenswrapper[4790]: I1124 13:15:38.828642 4790 generic.go:334] "Generic (PLEG): container finished" podID="6551fecc-df79-44f6-8600-fd016469a1e9" containerID="665468361d8cd7a472d4411803d78ac2e935b6b1d74605e0771e7aa03b4a442d" exitCode=0 Nov 24 13:15:38 crc kubenswrapper[4790]: I1124 13:15:38.828721 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4gq7f" event={"ID":"6551fecc-df79-44f6-8600-fd016469a1e9","Type":"ContainerDied","Data":"665468361d8cd7a472d4411803d78ac2e935b6b1d74605e0771e7aa03b4a442d"} Nov 24 13:15:38 crc kubenswrapper[4790]: I1124 13:15:38.844214 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-vdtxl" podStartSLOduration=162.844181952 podStartE2EDuration="2m42.844181952s" podCreationTimestamp="2025-11-24 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:15:38.840391179 +0000 UTC m=+187.220284851" watchObservedRunningTime="2025-11-24 13:15:38.844181952 +0000 UTC m=+187.224075614" Nov 24 13:15:39 crc kubenswrapper[4790]: I1124 13:15:39.551354 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:39 crc kubenswrapper[4790]: I1124 13:15:39.836689 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4gq7f" event={"ID":"6551fecc-df79-44f6-8600-fd016469a1e9","Type":"ContainerStarted","Data":"508973ce581e9b75831512a919b7fc9f5dc487ac35cedec3025e43725ac6c764"} Nov 24 13:15:39 crc kubenswrapper[4790]: I1124 13:15:39.856286 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4gq7f" podStartSLOduration=3.982471602 podStartE2EDuration="40.856261798s" podCreationTimestamp="2025-11-24 13:14:59 +0000 UTC" firstStartedPulling="2025-11-24 13:15:02.36601418 +0000 UTC m=+150.745907852" lastFinishedPulling="2025-11-24 13:15:39.239804386 +0000 UTC m=+187.619698048" observedRunningTime="2025-11-24 13:15:39.855833404 +0000 UTC m=+188.235727056" watchObservedRunningTime="2025-11-24 13:15:39.856261798 +0000 UTC m=+188.236155460" Nov 24 13:15:40 crc kubenswrapper[4790]: I1124 13:15:40.366933 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4gq7f" Nov 24 13:15:40 crc kubenswrapper[4790]: I1124 13:15:40.368955 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4gq7f" Nov 24 13:15:41 crc kubenswrapper[4790]: I1124 13:15:41.490149 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-4gq7f" podUID="6551fecc-df79-44f6-8600-fd016469a1e9" containerName="registry-server" probeResult="failure" output=< Nov 24 13:15:41 crc kubenswrapper[4790]: timeout: failed to connect service ":50051" within 1s Nov 24 13:15:41 crc kubenswrapper[4790]: > Nov 24 13:15:43 crc kubenswrapper[4790]: I1124 13:15:43.938463 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:15:43 crc kubenswrapper[4790]: I1124 13:15:43.938821 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:15:47 crc kubenswrapper[4790]: I1124 13:15:47.883217 4790 generic.go:334] "Generic (PLEG): container finished" podID="9420c18f-997b-4d89-b537-3cb99fe22682" containerID="09ed628c5000247436c66eb024238eacbf57acd0e1ff823774254afda5265851" exitCode=0 Nov 24 13:15:47 crc kubenswrapper[4790]: I1124 13:15:47.883270 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sjpz" event={"ID":"9420c18f-997b-4d89-b537-3cb99fe22682","Type":"ContainerDied","Data":"09ed628c5000247436c66eb024238eacbf57acd0e1ff823774254afda5265851"} Nov 24 13:15:48 crc kubenswrapper[4790]: I1124 13:15:48.890934 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sjpz" event={"ID":"9420c18f-997b-4d89-b537-3cb99fe22682","Type":"ContainerStarted","Data":"f4c51cbb081f5f037e19488611a405c5eeadc45596e77273893334fb27307810"} Nov 24 13:15:48 crc kubenswrapper[4790]: I1124 13:15:48.909265 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7sjpz" podStartSLOduration=1.981897013 podStartE2EDuration="47.909236369s" podCreationTimestamp="2025-11-24 13:15:01 +0000 UTC" firstStartedPulling="2025-11-24 13:15:02.366652231 +0000 UTC m=+150.746545903" lastFinishedPulling="2025-11-24 13:15:48.293991597 +0000 UTC m=+196.673885259" observedRunningTime="2025-11-24 13:15:48.908477046 +0000 UTC m=+197.288370708" watchObservedRunningTime="2025-11-24 13:15:48.909236369 +0000 UTC m=+197.289130051" Nov 24 13:15:49 crc kubenswrapper[4790]: I1124 13:15:49.900687 4790 generic.go:334] "Generic (PLEG): container finished" podID="036855a4-3747-4525-bcda-03c03daf917c" containerID="848e1a0f31932ec85245be378c302690cb605b6f5d4e21c22a41163b25965dd7" exitCode=0 Nov 24 13:15:49 crc kubenswrapper[4790]: I1124 13:15:49.900774 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mwgd" event={"ID":"036855a4-3747-4525-bcda-03c03daf917c","Type":"ContainerDied","Data":"848e1a0f31932ec85245be378c302690cb605b6f5d4e21c22a41163b25965dd7"} Nov 24 13:15:49 crc kubenswrapper[4790]: I1124 13:15:49.904205 4790 generic.go:334] "Generic (PLEG): container finished" podID="d53d9ecd-c244-4a4a-9b0d-f21530ed721e" containerID="7b459b3f4a0b2f6b85c445fa2c512306a172190624da8b84bdd74ed3f76dce11" exitCode=0 Nov 24 13:15:49 crc kubenswrapper[4790]: I1124 13:15:49.904261 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4r724" event={"ID":"d53d9ecd-c244-4a4a-9b0d-f21530ed721e","Type":"ContainerDied","Data":"7b459b3f4a0b2f6b85c445fa2c512306a172190624da8b84bdd74ed3f76dce11"} Nov 24 13:15:50 crc kubenswrapper[4790]: I1124 13:15:50.423868 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4gq7f" Nov 24 13:15:50 crc kubenswrapper[4790]: I1124 13:15:50.465777 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4gq7f" Nov 24 13:15:50 crc kubenswrapper[4790]: I1124 13:15:50.911686 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4r724" event={"ID":"d53d9ecd-c244-4a4a-9b0d-f21530ed721e","Type":"ContainerStarted","Data":"9de303b5dc3411abcf96ae237bbea60e8b1e746d2ab1f26f1116ae2d7471ad78"} Nov 24 13:15:50 crc kubenswrapper[4790]: I1124 13:15:50.914364 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mwgd" event={"ID":"036855a4-3747-4525-bcda-03c03daf917c","Type":"ContainerStarted","Data":"516a5ce6bbcd9d9dd940e5d5335f6de4769093e15170825cb62c42e08ab53227"} Nov 24 13:15:50 crc kubenswrapper[4790]: I1124 13:15:50.916147 4790 generic.go:334] "Generic (PLEG): container finished" podID="3c3ca964-4656-4248-bbcd-cd29957699cf" containerID="79ad0251c9cc2fb0b4256b38f4688d937967244a73dcd1ed43737c96fa6773fa" exitCode=0 Nov 24 13:15:50 crc kubenswrapper[4790]: I1124 13:15:50.916215 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g5nsf" event={"ID":"3c3ca964-4656-4248-bbcd-cd29957699cf","Type":"ContainerDied","Data":"79ad0251c9cc2fb0b4256b38f4688d937967244a73dcd1ed43737c96fa6773fa"} Nov 24 13:15:50 crc kubenswrapper[4790]: I1124 13:15:50.935422 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4r724" podStartSLOduration=2.906930514 podStartE2EDuration="49.935405924s" podCreationTimestamp="2025-11-24 13:15:01 +0000 UTC" firstStartedPulling="2025-11-24 13:15:03.389666251 +0000 UTC m=+151.769559923" lastFinishedPulling="2025-11-24 13:15:50.418141661 +0000 UTC m=+198.798035333" observedRunningTime="2025-11-24 13:15:50.931626548 +0000 UTC m=+199.311520210" watchObservedRunningTime="2025-11-24 13:15:50.935405924 +0000 UTC m=+199.315299586" Nov 24 13:15:50 crc kubenswrapper[4790]: I1124 13:15:50.962081 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7mwgd" podStartSLOduration=3.934175024 podStartE2EDuration="51.962061872s" podCreationTimestamp="2025-11-24 13:14:59 +0000 UTC" firstStartedPulling="2025-11-24 13:15:02.342520607 +0000 UTC m=+150.722414259" lastFinishedPulling="2025-11-24 13:15:50.370407425 +0000 UTC m=+198.750301107" observedRunningTime="2025-11-24 13:15:50.958255675 +0000 UTC m=+199.338149337" watchObservedRunningTime="2025-11-24 13:15:50.962061872 +0000 UTC m=+199.341955534" Nov 24 13:15:51 crc kubenswrapper[4790]: I1124 13:15:51.670791 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7sjpz" Nov 24 13:15:51 crc kubenswrapper[4790]: I1124 13:15:51.671077 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7sjpz" Nov 24 13:15:51 crc kubenswrapper[4790]: I1124 13:15:51.708911 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4gq7f"] Nov 24 13:15:51 crc kubenswrapper[4790]: I1124 13:15:51.716531 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7sjpz" Nov 24 13:15:51 crc kubenswrapper[4790]: I1124 13:15:51.899032 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4r724" Nov 24 13:15:51 crc kubenswrapper[4790]: I1124 13:15:51.899097 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4r724" Nov 24 13:15:51 crc kubenswrapper[4790]: I1124 13:15:51.922559 4790 generic.go:334] "Generic (PLEG): container finished" podID="594c7e16-73fa-4653-aa63-9f93214efd12" containerID="b7369082161a056e6fd32844db72c40163c461bde65d6a74f23396550ff2430a" exitCode=0 Nov 24 13:15:51 crc kubenswrapper[4790]: I1124 13:15:51.922630 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qxxm7" event={"ID":"594c7e16-73fa-4653-aa63-9f93214efd12","Type":"ContainerDied","Data":"b7369082161a056e6fd32844db72c40163c461bde65d6a74f23396550ff2430a"} Nov 24 13:15:51 crc kubenswrapper[4790]: I1124 13:15:51.927876 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g5nsf" event={"ID":"3c3ca964-4656-4248-bbcd-cd29957699cf","Type":"ContainerStarted","Data":"b07f5e8a194a7916b770b8c37c36831bead0572f4ffe24e42040bb4b3f4b69ac"} Nov 24 13:15:51 crc kubenswrapper[4790]: I1124 13:15:51.928232 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4gq7f" podUID="6551fecc-df79-44f6-8600-fd016469a1e9" containerName="registry-server" containerID="cri-o://508973ce581e9b75831512a919b7fc9f5dc487ac35cedec3025e43725ac6c764" gracePeriod=2 Nov 24 13:15:51 crc kubenswrapper[4790]: I1124 13:15:51.938439 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4r724" Nov 24 13:15:51 crc kubenswrapper[4790]: I1124 13:15:51.960120 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-g5nsf" podStartSLOduration=3.964189906 podStartE2EDuration="52.960101228s" podCreationTimestamp="2025-11-24 13:14:59 +0000 UTC" firstStartedPulling="2025-11-24 13:15:02.342746624 +0000 UTC m=+150.722640276" lastFinishedPulling="2025-11-24 13:15:51.338657936 +0000 UTC m=+199.718551598" observedRunningTime="2025-11-24 13:15:51.957536579 +0000 UTC m=+200.337430241" watchObservedRunningTime="2025-11-24 13:15:51.960101228 +0000 UTC m=+200.339994890" Nov 24 13:15:52 crc kubenswrapper[4790]: I1124 13:15:52.297065 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4gq7f" Nov 24 13:15:52 crc kubenswrapper[4790]: I1124 13:15:52.381004 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zdnf\" (UniqueName: \"kubernetes.io/projected/6551fecc-df79-44f6-8600-fd016469a1e9-kube-api-access-9zdnf\") pod \"6551fecc-df79-44f6-8600-fd016469a1e9\" (UID: \"6551fecc-df79-44f6-8600-fd016469a1e9\") " Nov 24 13:15:52 crc kubenswrapper[4790]: I1124 13:15:52.381341 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6551fecc-df79-44f6-8600-fd016469a1e9-utilities\") pod \"6551fecc-df79-44f6-8600-fd016469a1e9\" (UID: \"6551fecc-df79-44f6-8600-fd016469a1e9\") " Nov 24 13:15:52 crc kubenswrapper[4790]: I1124 13:15:52.381382 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6551fecc-df79-44f6-8600-fd016469a1e9-catalog-content\") pod \"6551fecc-df79-44f6-8600-fd016469a1e9\" (UID: \"6551fecc-df79-44f6-8600-fd016469a1e9\") " Nov 24 13:15:52 crc kubenswrapper[4790]: I1124 13:15:52.382210 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6551fecc-df79-44f6-8600-fd016469a1e9-utilities" (OuterVolumeSpecName: "utilities") pod "6551fecc-df79-44f6-8600-fd016469a1e9" (UID: "6551fecc-df79-44f6-8600-fd016469a1e9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:15:52 crc kubenswrapper[4790]: I1124 13:15:52.387027 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6551fecc-df79-44f6-8600-fd016469a1e9-kube-api-access-9zdnf" (OuterVolumeSpecName: "kube-api-access-9zdnf") pod "6551fecc-df79-44f6-8600-fd016469a1e9" (UID: "6551fecc-df79-44f6-8600-fd016469a1e9"). InnerVolumeSpecName "kube-api-access-9zdnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:15:52 crc kubenswrapper[4790]: I1124 13:15:52.442083 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6551fecc-df79-44f6-8600-fd016469a1e9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6551fecc-df79-44f6-8600-fd016469a1e9" (UID: "6551fecc-df79-44f6-8600-fd016469a1e9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:15:52 crc kubenswrapper[4790]: I1124 13:15:52.482963 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zdnf\" (UniqueName: \"kubernetes.io/projected/6551fecc-df79-44f6-8600-fd016469a1e9-kube-api-access-9zdnf\") on node \"crc\" DevicePath \"\"" Nov 24 13:15:52 crc kubenswrapper[4790]: I1124 13:15:52.483024 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6551fecc-df79-44f6-8600-fd016469a1e9-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:15:52 crc kubenswrapper[4790]: I1124 13:15:52.483035 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6551fecc-df79-44f6-8600-fd016469a1e9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:15:52 crc kubenswrapper[4790]: I1124 13:15:52.941284 4790 generic.go:334] "Generic (PLEG): container finished" podID="d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1" containerID="ed7f92a568fae36d75494745a78a6070e611e35929300357d74edc8c4f3feac3" exitCode=0 Nov 24 13:15:52 crc kubenswrapper[4790]: I1124 13:15:52.941346 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mv8f6" event={"ID":"d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1","Type":"ContainerDied","Data":"ed7f92a568fae36d75494745a78a6070e611e35929300357d74edc8c4f3feac3"} Nov 24 13:15:52 crc kubenswrapper[4790]: I1124 13:15:52.944287 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w2gmf" event={"ID":"37ec51fd-7e4c-4217-9a70-389efca8e855","Type":"ContainerStarted","Data":"122f3825d2a17abfbe00cc7de53c51aeb11b73ec6ae8cbef2dfdb8244b10e316"} Nov 24 13:15:52 crc kubenswrapper[4790]: I1124 13:15:52.948133 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qxxm7" event={"ID":"594c7e16-73fa-4653-aa63-9f93214efd12","Type":"ContainerStarted","Data":"96435c3c2d12f19bca126da859c7d48765a73390f48fcd58ff1a8674413f2525"} Nov 24 13:15:52 crc kubenswrapper[4790]: I1124 13:15:52.950380 4790 generic.go:334] "Generic (PLEG): container finished" podID="6551fecc-df79-44f6-8600-fd016469a1e9" containerID="508973ce581e9b75831512a919b7fc9f5dc487ac35cedec3025e43725ac6c764" exitCode=0 Nov 24 13:15:52 crc kubenswrapper[4790]: I1124 13:15:52.950433 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4gq7f" Nov 24 13:15:52 crc kubenswrapper[4790]: I1124 13:15:52.950468 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4gq7f" event={"ID":"6551fecc-df79-44f6-8600-fd016469a1e9","Type":"ContainerDied","Data":"508973ce581e9b75831512a919b7fc9f5dc487ac35cedec3025e43725ac6c764"} Nov 24 13:15:52 crc kubenswrapper[4790]: I1124 13:15:52.950489 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4gq7f" event={"ID":"6551fecc-df79-44f6-8600-fd016469a1e9","Type":"ContainerDied","Data":"76bcc228cfd641d37790e8e140007306401c42e33ea3545b4181f5f22205cfe0"} Nov 24 13:15:52 crc kubenswrapper[4790]: I1124 13:15:52.950505 4790 scope.go:117] "RemoveContainer" containerID="508973ce581e9b75831512a919b7fc9f5dc487ac35cedec3025e43725ac6c764" Nov 24 13:15:52 crc kubenswrapper[4790]: I1124 13:15:52.966233 4790 scope.go:117] "RemoveContainer" containerID="665468361d8cd7a472d4411803d78ac2e935b6b1d74605e0771e7aa03b4a442d" Nov 24 13:15:52 crc kubenswrapper[4790]: I1124 13:15:52.982317 4790 scope.go:117] "RemoveContainer" containerID="bbdb78fe3c3610827c9f7ae7a89b7acaca4a30ab557ba170224ed4500313107d" Nov 24 13:15:53 crc kubenswrapper[4790]: I1124 13:15:53.000634 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qxxm7" podStartSLOduration=3.963235538 podStartE2EDuration="54.000536136s" podCreationTimestamp="2025-11-24 13:14:59 +0000 UTC" firstStartedPulling="2025-11-24 13:15:02.343643923 +0000 UTC m=+150.723537575" lastFinishedPulling="2025-11-24 13:15:52.380944511 +0000 UTC m=+200.760838173" observedRunningTime="2025-11-24 13:15:52.998841314 +0000 UTC m=+201.378734986" watchObservedRunningTime="2025-11-24 13:15:53.000536136 +0000 UTC m=+201.380429798" Nov 24 13:15:53 crc kubenswrapper[4790]: I1124 13:15:53.004827 4790 scope.go:117] "RemoveContainer" containerID="508973ce581e9b75831512a919b7fc9f5dc487ac35cedec3025e43725ac6c764" Nov 24 13:15:53 crc kubenswrapper[4790]: E1124 13:15:53.005508 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"508973ce581e9b75831512a919b7fc9f5dc487ac35cedec3025e43725ac6c764\": container with ID starting with 508973ce581e9b75831512a919b7fc9f5dc487ac35cedec3025e43725ac6c764 not found: ID does not exist" containerID="508973ce581e9b75831512a919b7fc9f5dc487ac35cedec3025e43725ac6c764" Nov 24 13:15:53 crc kubenswrapper[4790]: I1124 13:15:53.005545 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"508973ce581e9b75831512a919b7fc9f5dc487ac35cedec3025e43725ac6c764"} err="failed to get container status \"508973ce581e9b75831512a919b7fc9f5dc487ac35cedec3025e43725ac6c764\": rpc error: code = NotFound desc = could not find container \"508973ce581e9b75831512a919b7fc9f5dc487ac35cedec3025e43725ac6c764\": container with ID starting with 508973ce581e9b75831512a919b7fc9f5dc487ac35cedec3025e43725ac6c764 not found: ID does not exist" Nov 24 13:15:53 crc kubenswrapper[4790]: I1124 13:15:53.005588 4790 scope.go:117] "RemoveContainer" containerID="665468361d8cd7a472d4411803d78ac2e935b6b1d74605e0771e7aa03b4a442d" Nov 24 13:15:53 crc kubenswrapper[4790]: E1124 13:15:53.006145 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"665468361d8cd7a472d4411803d78ac2e935b6b1d74605e0771e7aa03b4a442d\": container with ID starting with 665468361d8cd7a472d4411803d78ac2e935b6b1d74605e0771e7aa03b4a442d not found: ID does not exist" containerID="665468361d8cd7a472d4411803d78ac2e935b6b1d74605e0771e7aa03b4a442d" Nov 24 13:15:53 crc kubenswrapper[4790]: I1124 13:15:53.006179 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"665468361d8cd7a472d4411803d78ac2e935b6b1d74605e0771e7aa03b4a442d"} err="failed to get container status \"665468361d8cd7a472d4411803d78ac2e935b6b1d74605e0771e7aa03b4a442d\": rpc error: code = NotFound desc = could not find container \"665468361d8cd7a472d4411803d78ac2e935b6b1d74605e0771e7aa03b4a442d\": container with ID starting with 665468361d8cd7a472d4411803d78ac2e935b6b1d74605e0771e7aa03b4a442d not found: ID does not exist" Nov 24 13:15:53 crc kubenswrapper[4790]: I1124 13:15:53.006198 4790 scope.go:117] "RemoveContainer" containerID="bbdb78fe3c3610827c9f7ae7a89b7acaca4a30ab557ba170224ed4500313107d" Nov 24 13:15:53 crc kubenswrapper[4790]: E1124 13:15:53.006546 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbdb78fe3c3610827c9f7ae7a89b7acaca4a30ab557ba170224ed4500313107d\": container with ID starting with bbdb78fe3c3610827c9f7ae7a89b7acaca4a30ab557ba170224ed4500313107d not found: ID does not exist" containerID="bbdb78fe3c3610827c9f7ae7a89b7acaca4a30ab557ba170224ed4500313107d" Nov 24 13:15:53 crc kubenswrapper[4790]: I1124 13:15:53.006574 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbdb78fe3c3610827c9f7ae7a89b7acaca4a30ab557ba170224ed4500313107d"} err="failed to get container status \"bbdb78fe3c3610827c9f7ae7a89b7acaca4a30ab557ba170224ed4500313107d\": rpc error: code = NotFound desc = could not find container \"bbdb78fe3c3610827c9f7ae7a89b7acaca4a30ab557ba170224ed4500313107d\": container with ID starting with bbdb78fe3c3610827c9f7ae7a89b7acaca4a30ab557ba170224ed4500313107d not found: ID does not exist" Nov 24 13:15:53 crc kubenswrapper[4790]: I1124 13:15:53.012278 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4gq7f"] Nov 24 13:15:53 crc kubenswrapper[4790]: I1124 13:15:53.017290 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4gq7f"] Nov 24 13:15:53 crc kubenswrapper[4790]: I1124 13:15:53.963667 4790 generic.go:334] "Generic (PLEG): container finished" podID="37ec51fd-7e4c-4217-9a70-389efca8e855" containerID="122f3825d2a17abfbe00cc7de53c51aeb11b73ec6ae8cbef2dfdb8244b10e316" exitCode=0 Nov 24 13:15:53 crc kubenswrapper[4790]: I1124 13:15:53.963746 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w2gmf" event={"ID":"37ec51fd-7e4c-4217-9a70-389efca8e855","Type":"ContainerDied","Data":"122f3825d2a17abfbe00cc7de53c51aeb11b73ec6ae8cbef2dfdb8244b10e316"} Nov 24 13:15:54 crc kubenswrapper[4790]: I1124 13:15:54.321291 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6551fecc-df79-44f6-8600-fd016469a1e9" path="/var/lib/kubelet/pods/6551fecc-df79-44f6-8600-fd016469a1e9/volumes" Nov 24 13:15:55 crc kubenswrapper[4790]: I1124 13:15:55.982412 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mv8f6" event={"ID":"d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1","Type":"ContainerStarted","Data":"04370e176b596d415c2ce1d243372d9eaf1b45160b0fdd834e0dbd47eb2e65f3"} Nov 24 13:15:57 crc kubenswrapper[4790]: I1124 13:15:57.007103 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w2gmf" event={"ID":"37ec51fd-7e4c-4217-9a70-389efca8e855","Type":"ContainerStarted","Data":"2d6e78ba26ee897ac86ff20e363a1134a5b1e4751da5763bf285a97fd7eecdf3"} Nov 24 13:15:57 crc kubenswrapper[4790]: I1124 13:15:57.035002 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mv8f6" podStartSLOduration=3.257201965 podStartE2EDuration="54.034983836s" podCreationTimestamp="2025-11-24 13:15:03 +0000 UTC" firstStartedPulling="2025-11-24 13:15:04.449238731 +0000 UTC m=+152.829132383" lastFinishedPulling="2025-11-24 13:15:55.227020562 +0000 UTC m=+203.606914254" observedRunningTime="2025-11-24 13:15:57.030445427 +0000 UTC m=+205.410339089" watchObservedRunningTime="2025-11-24 13:15:57.034983836 +0000 UTC m=+205.414877498" Nov 24 13:15:58 crc kubenswrapper[4790]: I1124 13:15:58.041292 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-w2gmf" podStartSLOduration=3.986156567 podStartE2EDuration="56.041259165s" podCreationTimestamp="2025-11-24 13:15:02 +0000 UTC" firstStartedPulling="2025-11-24 13:15:04.44059054 +0000 UTC m=+152.820484192" lastFinishedPulling="2025-11-24 13:15:56.495693128 +0000 UTC m=+204.875586790" observedRunningTime="2025-11-24 13:15:58.039102709 +0000 UTC m=+206.418996371" watchObservedRunningTime="2025-11-24 13:15:58.041259165 +0000 UTC m=+206.421152867" Nov 24 13:15:59 crc kubenswrapper[4790]: I1124 13:15:59.717337 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qxxm7" Nov 24 13:15:59 crc kubenswrapper[4790]: I1124 13:15:59.717397 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qxxm7" Nov 24 13:15:59 crc kubenswrapper[4790]: I1124 13:15:59.763130 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qxxm7" Nov 24 13:15:59 crc kubenswrapper[4790]: I1124 13:15:59.956550 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-g5nsf" Nov 24 13:15:59 crc kubenswrapper[4790]: I1124 13:15:59.956897 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-g5nsf" Nov 24 13:16:00 crc kubenswrapper[4790]: I1124 13:15:59.999935 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-g5nsf" Nov 24 13:16:00 crc kubenswrapper[4790]: I1124 13:16:00.059195 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7mwgd" Nov 24 13:16:00 crc kubenswrapper[4790]: I1124 13:16:00.059246 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7mwgd" Nov 24 13:16:00 crc kubenswrapper[4790]: I1124 13:16:00.060509 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qxxm7" Nov 24 13:16:00 crc kubenswrapper[4790]: I1124 13:16:00.071508 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-g5nsf" Nov 24 13:16:00 crc kubenswrapper[4790]: I1124 13:16:00.113593 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7mwgd" Nov 24 13:16:01 crc kubenswrapper[4790]: I1124 13:16:01.074501 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7mwgd" Nov 24 13:16:01 crc kubenswrapper[4790]: I1124 13:16:01.727156 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7sjpz" Nov 24 13:16:01 crc kubenswrapper[4790]: I1124 13:16:01.934657 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4r724" Nov 24 13:16:02 crc kubenswrapper[4790]: I1124 13:16:02.510005 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7mwgd"] Nov 24 13:16:03 crc kubenswrapper[4790]: I1124 13:16:03.020569 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-w2gmf" Nov 24 13:16:03 crc kubenswrapper[4790]: I1124 13:16:03.020625 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-w2gmf" Nov 24 13:16:03 crc kubenswrapper[4790]: I1124 13:16:03.038063 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7mwgd" podUID="036855a4-3747-4525-bcda-03c03daf917c" containerName="registry-server" containerID="cri-o://516a5ce6bbcd9d9dd940e5d5335f6de4769093e15170825cb62c42e08ab53227" gracePeriod=2 Nov 24 13:16:03 crc kubenswrapper[4790]: I1124 13:16:03.085560 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-w2gmf" Nov 24 13:16:03 crc kubenswrapper[4790]: I1124 13:16:03.128847 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-w2gmf" Nov 24 13:16:03 crc kubenswrapper[4790]: I1124 13:16:03.397440 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7mwgd" Nov 24 13:16:03 crc kubenswrapper[4790]: I1124 13:16:03.427968 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mv8f6" Nov 24 13:16:03 crc kubenswrapper[4790]: I1124 13:16:03.428015 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mv8f6" Nov 24 13:16:03 crc kubenswrapper[4790]: I1124 13:16:03.474349 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mv8f6" Nov 24 13:16:03 crc kubenswrapper[4790]: I1124 13:16:03.564149 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/036855a4-3747-4525-bcda-03c03daf917c-utilities\") pod \"036855a4-3747-4525-bcda-03c03daf917c\" (UID: \"036855a4-3747-4525-bcda-03c03daf917c\") " Nov 24 13:16:03 crc kubenswrapper[4790]: I1124 13:16:03.564398 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/036855a4-3747-4525-bcda-03c03daf917c-catalog-content\") pod \"036855a4-3747-4525-bcda-03c03daf917c\" (UID: \"036855a4-3747-4525-bcda-03c03daf917c\") " Nov 24 13:16:03 crc kubenswrapper[4790]: I1124 13:16:03.564515 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p5wmq\" (UniqueName: \"kubernetes.io/projected/036855a4-3747-4525-bcda-03c03daf917c-kube-api-access-p5wmq\") pod \"036855a4-3747-4525-bcda-03c03daf917c\" (UID: \"036855a4-3747-4525-bcda-03c03daf917c\") " Nov 24 13:16:03 crc kubenswrapper[4790]: I1124 13:16:03.565380 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/036855a4-3747-4525-bcda-03c03daf917c-utilities" (OuterVolumeSpecName: "utilities") pod "036855a4-3747-4525-bcda-03c03daf917c" (UID: "036855a4-3747-4525-bcda-03c03daf917c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:16:03 crc kubenswrapper[4790]: I1124 13:16:03.569298 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/036855a4-3747-4525-bcda-03c03daf917c-kube-api-access-p5wmq" (OuterVolumeSpecName: "kube-api-access-p5wmq") pod "036855a4-3747-4525-bcda-03c03daf917c" (UID: "036855a4-3747-4525-bcda-03c03daf917c"). InnerVolumeSpecName "kube-api-access-p5wmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:16:03 crc kubenswrapper[4790]: I1124 13:16:03.611996 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/036855a4-3747-4525-bcda-03c03daf917c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "036855a4-3747-4525-bcda-03c03daf917c" (UID: "036855a4-3747-4525-bcda-03c03daf917c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:16:03 crc kubenswrapper[4790]: I1124 13:16:03.665720 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/036855a4-3747-4525-bcda-03c03daf917c-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:03 crc kubenswrapper[4790]: I1124 13:16:03.665763 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/036855a4-3747-4525-bcda-03c03daf917c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:03 crc kubenswrapper[4790]: I1124 13:16:03.665778 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p5wmq\" (UniqueName: \"kubernetes.io/projected/036855a4-3747-4525-bcda-03c03daf917c-kube-api-access-p5wmq\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:04 crc kubenswrapper[4790]: I1124 13:16:04.046393 4790 generic.go:334] "Generic (PLEG): container finished" podID="036855a4-3747-4525-bcda-03c03daf917c" containerID="516a5ce6bbcd9d9dd940e5d5335f6de4769093e15170825cb62c42e08ab53227" exitCode=0 Nov 24 13:16:04 crc kubenswrapper[4790]: I1124 13:16:04.046454 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7mwgd" Nov 24 13:16:04 crc kubenswrapper[4790]: I1124 13:16:04.046460 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mwgd" event={"ID":"036855a4-3747-4525-bcda-03c03daf917c","Type":"ContainerDied","Data":"516a5ce6bbcd9d9dd940e5d5335f6de4769093e15170825cb62c42e08ab53227"} Nov 24 13:16:04 crc kubenswrapper[4790]: I1124 13:16:04.046511 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mwgd" event={"ID":"036855a4-3747-4525-bcda-03c03daf917c","Type":"ContainerDied","Data":"4e97071d7f696f2e9a0d01f6f27d8a5c57fa969055c10d4accb3a5b7b6838462"} Nov 24 13:16:04 crc kubenswrapper[4790]: I1124 13:16:04.046533 4790 scope.go:117] "RemoveContainer" containerID="516a5ce6bbcd9d9dd940e5d5335f6de4769093e15170825cb62c42e08ab53227" Nov 24 13:16:04 crc kubenswrapper[4790]: I1124 13:16:04.068590 4790 scope.go:117] "RemoveContainer" containerID="848e1a0f31932ec85245be378c302690cb605b6f5d4e21c22a41163b25965dd7" Nov 24 13:16:04 crc kubenswrapper[4790]: I1124 13:16:04.081001 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7mwgd"] Nov 24 13:16:04 crc kubenswrapper[4790]: I1124 13:16:04.085653 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7mwgd"] Nov 24 13:16:04 crc kubenswrapper[4790]: I1124 13:16:04.096687 4790 scope.go:117] "RemoveContainer" containerID="9d341d978351dc4ba171434b3fb3a4ef21bb8d0db3793155383bd465c04c080a" Nov 24 13:16:04 crc kubenswrapper[4790]: I1124 13:16:04.112576 4790 scope.go:117] "RemoveContainer" containerID="516a5ce6bbcd9d9dd940e5d5335f6de4769093e15170825cb62c42e08ab53227" Nov 24 13:16:04 crc kubenswrapper[4790]: I1124 13:16:04.112699 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mv8f6" Nov 24 13:16:04 crc kubenswrapper[4790]: E1124 13:16:04.113176 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"516a5ce6bbcd9d9dd940e5d5335f6de4769093e15170825cb62c42e08ab53227\": container with ID starting with 516a5ce6bbcd9d9dd940e5d5335f6de4769093e15170825cb62c42e08ab53227 not found: ID does not exist" containerID="516a5ce6bbcd9d9dd940e5d5335f6de4769093e15170825cb62c42e08ab53227" Nov 24 13:16:04 crc kubenswrapper[4790]: I1124 13:16:04.113232 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"516a5ce6bbcd9d9dd940e5d5335f6de4769093e15170825cb62c42e08ab53227"} err="failed to get container status \"516a5ce6bbcd9d9dd940e5d5335f6de4769093e15170825cb62c42e08ab53227\": rpc error: code = NotFound desc = could not find container \"516a5ce6bbcd9d9dd940e5d5335f6de4769093e15170825cb62c42e08ab53227\": container with ID starting with 516a5ce6bbcd9d9dd940e5d5335f6de4769093e15170825cb62c42e08ab53227 not found: ID does not exist" Nov 24 13:16:04 crc kubenswrapper[4790]: I1124 13:16:04.113267 4790 scope.go:117] "RemoveContainer" containerID="848e1a0f31932ec85245be378c302690cb605b6f5d4e21c22a41163b25965dd7" Nov 24 13:16:04 crc kubenswrapper[4790]: E1124 13:16:04.113637 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"848e1a0f31932ec85245be378c302690cb605b6f5d4e21c22a41163b25965dd7\": container with ID starting with 848e1a0f31932ec85245be378c302690cb605b6f5d4e21c22a41163b25965dd7 not found: ID does not exist" containerID="848e1a0f31932ec85245be378c302690cb605b6f5d4e21c22a41163b25965dd7" Nov 24 13:16:04 crc kubenswrapper[4790]: I1124 13:16:04.113675 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"848e1a0f31932ec85245be378c302690cb605b6f5d4e21c22a41163b25965dd7"} err="failed to get container status \"848e1a0f31932ec85245be378c302690cb605b6f5d4e21c22a41163b25965dd7\": rpc error: code = NotFound desc = could not find container \"848e1a0f31932ec85245be378c302690cb605b6f5d4e21c22a41163b25965dd7\": container with ID starting with 848e1a0f31932ec85245be378c302690cb605b6f5d4e21c22a41163b25965dd7 not found: ID does not exist" Nov 24 13:16:04 crc kubenswrapper[4790]: I1124 13:16:04.113704 4790 scope.go:117] "RemoveContainer" containerID="9d341d978351dc4ba171434b3fb3a4ef21bb8d0db3793155383bd465c04c080a" Nov 24 13:16:04 crc kubenswrapper[4790]: E1124 13:16:04.114607 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d341d978351dc4ba171434b3fb3a4ef21bb8d0db3793155383bd465c04c080a\": container with ID starting with 9d341d978351dc4ba171434b3fb3a4ef21bb8d0db3793155383bd465c04c080a not found: ID does not exist" containerID="9d341d978351dc4ba171434b3fb3a4ef21bb8d0db3793155383bd465c04c080a" Nov 24 13:16:04 crc kubenswrapper[4790]: I1124 13:16:04.114639 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d341d978351dc4ba171434b3fb3a4ef21bb8d0db3793155383bd465c04c080a"} err="failed to get container status \"9d341d978351dc4ba171434b3fb3a4ef21bb8d0db3793155383bd465c04c080a\": rpc error: code = NotFound desc = could not find container \"9d341d978351dc4ba171434b3fb3a4ef21bb8d0db3793155383bd465c04c080a\": container with ID starting with 9d341d978351dc4ba171434b3fb3a4ef21bb8d0db3793155383bd465c04c080a not found: ID does not exist" Nov 24 13:16:04 crc kubenswrapper[4790]: I1124 13:16:04.309459 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4r724"] Nov 24 13:16:04 crc kubenswrapper[4790]: I1124 13:16:04.310055 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4r724" podUID="d53d9ecd-c244-4a4a-9b0d-f21530ed721e" containerName="registry-server" containerID="cri-o://9de303b5dc3411abcf96ae237bbea60e8b1e746d2ab1f26f1116ae2d7471ad78" gracePeriod=2 Nov 24 13:16:04 crc kubenswrapper[4790]: I1124 13:16:04.323216 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="036855a4-3747-4525-bcda-03c03daf917c" path="/var/lib/kubelet/pods/036855a4-3747-4525-bcda-03c03daf917c/volumes" Nov 24 13:16:04 crc kubenswrapper[4790]: I1124 13:16:04.677575 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4r724" Nov 24 13:16:04 crc kubenswrapper[4790]: I1124 13:16:04.683449 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vb77f\" (UniqueName: \"kubernetes.io/projected/d53d9ecd-c244-4a4a-9b0d-f21530ed721e-kube-api-access-vb77f\") pod \"d53d9ecd-c244-4a4a-9b0d-f21530ed721e\" (UID: \"d53d9ecd-c244-4a4a-9b0d-f21530ed721e\") " Nov 24 13:16:04 crc kubenswrapper[4790]: I1124 13:16:04.683521 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d53d9ecd-c244-4a4a-9b0d-f21530ed721e-catalog-content\") pod \"d53d9ecd-c244-4a4a-9b0d-f21530ed721e\" (UID: \"d53d9ecd-c244-4a4a-9b0d-f21530ed721e\") " Nov 24 13:16:04 crc kubenswrapper[4790]: I1124 13:16:04.683570 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d53d9ecd-c244-4a4a-9b0d-f21530ed721e-utilities\") pod \"d53d9ecd-c244-4a4a-9b0d-f21530ed721e\" (UID: \"d53d9ecd-c244-4a4a-9b0d-f21530ed721e\") " Nov 24 13:16:04 crc kubenswrapper[4790]: I1124 13:16:04.689254 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d53d9ecd-c244-4a4a-9b0d-f21530ed721e-utilities" (OuterVolumeSpecName: "utilities") pod "d53d9ecd-c244-4a4a-9b0d-f21530ed721e" (UID: "d53d9ecd-c244-4a4a-9b0d-f21530ed721e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:16:04 crc kubenswrapper[4790]: I1124 13:16:04.697945 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d53d9ecd-c244-4a4a-9b0d-f21530ed721e-kube-api-access-vb77f" (OuterVolumeSpecName: "kube-api-access-vb77f") pod "d53d9ecd-c244-4a4a-9b0d-f21530ed721e" (UID: "d53d9ecd-c244-4a4a-9b0d-f21530ed721e"). InnerVolumeSpecName "kube-api-access-vb77f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:16:04 crc kubenswrapper[4790]: I1124 13:16:04.707485 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d53d9ecd-c244-4a4a-9b0d-f21530ed721e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d53d9ecd-c244-4a4a-9b0d-f21530ed721e" (UID: "d53d9ecd-c244-4a4a-9b0d-f21530ed721e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:16:04 crc kubenswrapper[4790]: I1124 13:16:04.789076 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d53d9ecd-c244-4a4a-9b0d-f21530ed721e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:04 crc kubenswrapper[4790]: I1124 13:16:04.789123 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d53d9ecd-c244-4a4a-9b0d-f21530ed721e-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:04 crc kubenswrapper[4790]: I1124 13:16:04.789136 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vb77f\" (UniqueName: \"kubernetes.io/projected/d53d9ecd-c244-4a4a-9b0d-f21530ed721e-kube-api-access-vb77f\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:05 crc kubenswrapper[4790]: I1124 13:16:05.055787 4790 generic.go:334] "Generic (PLEG): container finished" podID="d53d9ecd-c244-4a4a-9b0d-f21530ed721e" containerID="9de303b5dc3411abcf96ae237bbea60e8b1e746d2ab1f26f1116ae2d7471ad78" exitCode=0 Nov 24 13:16:05 crc kubenswrapper[4790]: I1124 13:16:05.055860 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4r724" event={"ID":"d53d9ecd-c244-4a4a-9b0d-f21530ed721e","Type":"ContainerDied","Data":"9de303b5dc3411abcf96ae237bbea60e8b1e746d2ab1f26f1116ae2d7471ad78"} Nov 24 13:16:05 crc kubenswrapper[4790]: I1124 13:16:05.055919 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4r724" event={"ID":"d53d9ecd-c244-4a4a-9b0d-f21530ed721e","Type":"ContainerDied","Data":"b43ae547dc9e75f6eace74a9633b475e05ba96aa4fb5286159e996401dadfbc3"} Nov 24 13:16:05 crc kubenswrapper[4790]: I1124 13:16:05.055947 4790 scope.go:117] "RemoveContainer" containerID="9de303b5dc3411abcf96ae237bbea60e8b1e746d2ab1f26f1116ae2d7471ad78" Nov 24 13:16:05 crc kubenswrapper[4790]: I1124 13:16:05.055938 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4r724" Nov 24 13:16:05 crc kubenswrapper[4790]: I1124 13:16:05.073976 4790 scope.go:117] "RemoveContainer" containerID="7b459b3f4a0b2f6b85c445fa2c512306a172190624da8b84bdd74ed3f76dce11" Nov 24 13:16:05 crc kubenswrapper[4790]: I1124 13:16:05.087501 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4r724"] Nov 24 13:16:05 crc kubenswrapper[4790]: I1124 13:16:05.094807 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4r724"] Nov 24 13:16:05 crc kubenswrapper[4790]: I1124 13:16:05.105317 4790 scope.go:117] "RemoveContainer" containerID="dda0d898f910127a5a2a4d11663736ae85ad4df023b7494669f0aa204c5cfd4c" Nov 24 13:16:05 crc kubenswrapper[4790]: I1124 13:16:05.120855 4790 scope.go:117] "RemoveContainer" containerID="9de303b5dc3411abcf96ae237bbea60e8b1e746d2ab1f26f1116ae2d7471ad78" Nov 24 13:16:05 crc kubenswrapper[4790]: E1124 13:16:05.121245 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9de303b5dc3411abcf96ae237bbea60e8b1e746d2ab1f26f1116ae2d7471ad78\": container with ID starting with 9de303b5dc3411abcf96ae237bbea60e8b1e746d2ab1f26f1116ae2d7471ad78 not found: ID does not exist" containerID="9de303b5dc3411abcf96ae237bbea60e8b1e746d2ab1f26f1116ae2d7471ad78" Nov 24 13:16:05 crc kubenswrapper[4790]: I1124 13:16:05.121320 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9de303b5dc3411abcf96ae237bbea60e8b1e746d2ab1f26f1116ae2d7471ad78"} err="failed to get container status \"9de303b5dc3411abcf96ae237bbea60e8b1e746d2ab1f26f1116ae2d7471ad78\": rpc error: code = NotFound desc = could not find container \"9de303b5dc3411abcf96ae237bbea60e8b1e746d2ab1f26f1116ae2d7471ad78\": container with ID starting with 9de303b5dc3411abcf96ae237bbea60e8b1e746d2ab1f26f1116ae2d7471ad78 not found: ID does not exist" Nov 24 13:16:05 crc kubenswrapper[4790]: I1124 13:16:05.121356 4790 scope.go:117] "RemoveContainer" containerID="7b459b3f4a0b2f6b85c445fa2c512306a172190624da8b84bdd74ed3f76dce11" Nov 24 13:16:05 crc kubenswrapper[4790]: E1124 13:16:05.121743 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b459b3f4a0b2f6b85c445fa2c512306a172190624da8b84bdd74ed3f76dce11\": container with ID starting with 7b459b3f4a0b2f6b85c445fa2c512306a172190624da8b84bdd74ed3f76dce11 not found: ID does not exist" containerID="7b459b3f4a0b2f6b85c445fa2c512306a172190624da8b84bdd74ed3f76dce11" Nov 24 13:16:05 crc kubenswrapper[4790]: I1124 13:16:05.121775 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b459b3f4a0b2f6b85c445fa2c512306a172190624da8b84bdd74ed3f76dce11"} err="failed to get container status \"7b459b3f4a0b2f6b85c445fa2c512306a172190624da8b84bdd74ed3f76dce11\": rpc error: code = NotFound desc = could not find container \"7b459b3f4a0b2f6b85c445fa2c512306a172190624da8b84bdd74ed3f76dce11\": container with ID starting with 7b459b3f4a0b2f6b85c445fa2c512306a172190624da8b84bdd74ed3f76dce11 not found: ID does not exist" Nov 24 13:16:05 crc kubenswrapper[4790]: I1124 13:16:05.121796 4790 scope.go:117] "RemoveContainer" containerID="dda0d898f910127a5a2a4d11663736ae85ad4df023b7494669f0aa204c5cfd4c" Nov 24 13:16:05 crc kubenswrapper[4790]: E1124 13:16:05.122126 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dda0d898f910127a5a2a4d11663736ae85ad4df023b7494669f0aa204c5cfd4c\": container with ID starting with dda0d898f910127a5a2a4d11663736ae85ad4df023b7494669f0aa204c5cfd4c not found: ID does not exist" containerID="dda0d898f910127a5a2a4d11663736ae85ad4df023b7494669f0aa204c5cfd4c" Nov 24 13:16:05 crc kubenswrapper[4790]: I1124 13:16:05.122159 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dda0d898f910127a5a2a4d11663736ae85ad4df023b7494669f0aa204c5cfd4c"} err="failed to get container status \"dda0d898f910127a5a2a4d11663736ae85ad4df023b7494669f0aa204c5cfd4c\": rpc error: code = NotFound desc = could not find container \"dda0d898f910127a5a2a4d11663736ae85ad4df023b7494669f0aa204c5cfd4c\": container with ID starting with dda0d898f910127a5a2a4d11663736ae85ad4df023b7494669f0aa204c5cfd4c not found: ID does not exist" Nov 24 13:16:06 crc kubenswrapper[4790]: I1124 13:16:06.325327 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d53d9ecd-c244-4a4a-9b0d-f21530ed721e" path="/var/lib/kubelet/pods/d53d9ecd-c244-4a4a-9b0d-f21530ed721e/volumes" Nov 24 13:16:06 crc kubenswrapper[4790]: I1124 13:16:06.918240 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mv8f6"] Nov 24 13:16:06 crc kubenswrapper[4790]: I1124 13:16:06.918668 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mv8f6" podUID="d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1" containerName="registry-server" containerID="cri-o://04370e176b596d415c2ce1d243372d9eaf1b45160b0fdd834e0dbd47eb2e65f3" gracePeriod=2 Nov 24 13:16:07 crc kubenswrapper[4790]: I1124 13:16:07.075557 4790 generic.go:334] "Generic (PLEG): container finished" podID="d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1" containerID="04370e176b596d415c2ce1d243372d9eaf1b45160b0fdd834e0dbd47eb2e65f3" exitCode=0 Nov 24 13:16:07 crc kubenswrapper[4790]: I1124 13:16:07.075628 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mv8f6" event={"ID":"d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1","Type":"ContainerDied","Data":"04370e176b596d415c2ce1d243372d9eaf1b45160b0fdd834e0dbd47eb2e65f3"} Nov 24 13:16:07 crc kubenswrapper[4790]: I1124 13:16:07.310569 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mv8f6" Nov 24 13:16:07 crc kubenswrapper[4790]: I1124 13:16:07.321567 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1-utilities\") pod \"d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1\" (UID: \"d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1\") " Nov 24 13:16:07 crc kubenswrapper[4790]: I1124 13:16:07.321637 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1-catalog-content\") pod \"d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1\" (UID: \"d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1\") " Nov 24 13:16:07 crc kubenswrapper[4790]: I1124 13:16:07.321682 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mszvk\" (UniqueName: \"kubernetes.io/projected/d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1-kube-api-access-mszvk\") pod \"d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1\" (UID: \"d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1\") " Nov 24 13:16:07 crc kubenswrapper[4790]: I1124 13:16:07.324455 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1-utilities" (OuterVolumeSpecName: "utilities") pod "d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1" (UID: "d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:16:07 crc kubenswrapper[4790]: I1124 13:16:07.330939 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:07 crc kubenswrapper[4790]: I1124 13:16:07.347147 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1-kube-api-access-mszvk" (OuterVolumeSpecName: "kube-api-access-mszvk") pod "d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1" (UID: "d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1"). InnerVolumeSpecName "kube-api-access-mszvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:16:07 crc kubenswrapper[4790]: I1124 13:16:07.427278 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1" (UID: "d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:16:07 crc kubenswrapper[4790]: I1124 13:16:07.432734 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mszvk\" (UniqueName: \"kubernetes.io/projected/d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1-kube-api-access-mszvk\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:07 crc kubenswrapper[4790]: I1124 13:16:07.432781 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:08 crc kubenswrapper[4790]: I1124 13:16:08.082143 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mv8f6" event={"ID":"d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1","Type":"ContainerDied","Data":"0e6b15d1a276e0d320d54bd0e7740f859de75116d051cd91be1083e34f4031d9"} Nov 24 13:16:08 crc kubenswrapper[4790]: I1124 13:16:08.082193 4790 scope.go:117] "RemoveContainer" containerID="04370e176b596d415c2ce1d243372d9eaf1b45160b0fdd834e0dbd47eb2e65f3" Nov 24 13:16:08 crc kubenswrapper[4790]: I1124 13:16:08.082325 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mv8f6" Nov 24 13:16:08 crc kubenswrapper[4790]: I1124 13:16:08.100946 4790 scope.go:117] "RemoveContainer" containerID="ed7f92a568fae36d75494745a78a6070e611e35929300357d74edc8c4f3feac3" Nov 24 13:16:08 crc kubenswrapper[4790]: I1124 13:16:08.117379 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mv8f6"] Nov 24 13:16:08 crc kubenswrapper[4790]: I1124 13:16:08.121506 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mv8f6"] Nov 24 13:16:08 crc kubenswrapper[4790]: I1124 13:16:08.128679 4790 scope.go:117] "RemoveContainer" containerID="4eee42fbff43685e52cb3f5de3cd8caa32ef06d3c8ae0e40292bd0c268f92cdd" Nov 24 13:16:08 crc kubenswrapper[4790]: I1124 13:16:08.325009 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1" path="/var/lib/kubelet/pods/d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1/volumes" Nov 24 13:16:13 crc kubenswrapper[4790]: I1124 13:16:13.129322 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-zxkb7"] Nov 24 13:16:13 crc kubenswrapper[4790]: I1124 13:16:13.938427 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:16:13 crc kubenswrapper[4790]: I1124 13:16:13.938832 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:16:13 crc kubenswrapper[4790]: I1124 13:16:13.938954 4790 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 13:16:13 crc kubenswrapper[4790]: I1124 13:16:13.939814 4790 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989"} pod="openshift-machine-config-operator/machine-config-daemon-xz49t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 13:16:13 crc kubenswrapper[4790]: I1124 13:16:13.939960 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" containerID="cri-o://b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989" gracePeriod=600 Nov 24 13:16:14 crc kubenswrapper[4790]: I1124 13:16:14.121029 4790 generic.go:334] "Generic (PLEG): container finished" podID="0d73b133-48f0-455f-8f6a-742e633f631a" containerID="b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989" exitCode=0 Nov 24 13:16:14 crc kubenswrapper[4790]: I1124 13:16:14.121079 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerDied","Data":"b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989"} Nov 24 13:16:15 crc kubenswrapper[4790]: I1124 13:16:15.129117 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"abd5645bbc014c541440d33742c811f564b9e51e2066d093c2f1fef3b356b89b"} Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.166696 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" podUID="a86ba968-c6dd-42c8-91fb-207bc839b8ec" containerName="oauth-openshift" containerID="cri-o://0c06dbf4c49aaeeb09163e81c971e4b532cc334825adcec02f48a8f9eafa867e" gracePeriod=15 Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.530112 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.553514 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a86ba968-c6dd-42c8-91fb-207bc839b8ec-audit-policies\") pod \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.553565 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-user-idp-0-file-data\") pod \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.553688 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-user-template-login\") pod \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.553714 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a86ba968-c6dd-42c8-91fb-207bc839b8ec-audit-dir\") pod \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.553743 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5zww\" (UniqueName: \"kubernetes.io/projected/a86ba968-c6dd-42c8-91fb-207bc839b8ec-kube-api-access-t5zww\") pod \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.553767 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-router-certs\") pod \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.553791 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-user-template-provider-selection\") pod \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.554725 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a86ba968-c6dd-42c8-91fb-207bc839b8ec-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "a86ba968-c6dd-42c8-91fb-207bc839b8ec" (UID: "a86ba968-c6dd-42c8-91fb-207bc839b8ec"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.554843 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a86ba968-c6dd-42c8-91fb-207bc839b8ec-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "a86ba968-c6dd-42c8-91fb-207bc839b8ec" (UID: "a86ba968-c6dd-42c8-91fb-207bc839b8ec"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.566197 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "a86ba968-c6dd-42c8-91fb-207bc839b8ec" (UID: "a86ba968-c6dd-42c8-91fb-207bc839b8ec"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.569173 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "a86ba968-c6dd-42c8-91fb-207bc839b8ec" (UID: "a86ba968-c6dd-42c8-91fb-207bc839b8ec"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.572455 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "a86ba968-c6dd-42c8-91fb-207bc839b8ec" (UID: "a86ba968-c6dd-42c8-91fb-207bc839b8ec"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.573784 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-56c748df47-nnzs5"] Nov 24 13:16:38 crc kubenswrapper[4790]: E1124 13:16:38.574187 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="036855a4-3747-4525-bcda-03c03daf917c" containerName="extract-utilities" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.574211 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="036855a4-3747-4525-bcda-03c03daf917c" containerName="extract-utilities" Nov 24 13:16:38 crc kubenswrapper[4790]: E1124 13:16:38.574223 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d53d9ecd-c244-4a4a-9b0d-f21530ed721e" containerName="extract-utilities" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.574234 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="d53d9ecd-c244-4a4a-9b0d-f21530ed721e" containerName="extract-utilities" Nov 24 13:16:38 crc kubenswrapper[4790]: E1124 13:16:38.574248 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1" containerName="extract-utilities" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.574257 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1" containerName="extract-utilities" Nov 24 13:16:38 crc kubenswrapper[4790]: E1124 13:16:38.574276 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a86ba968-c6dd-42c8-91fb-207bc839b8ec" containerName="oauth-openshift" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.574287 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="a86ba968-c6dd-42c8-91fb-207bc839b8ec" containerName="oauth-openshift" Nov 24 13:16:38 crc kubenswrapper[4790]: E1124 13:16:38.574302 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d53d9ecd-c244-4a4a-9b0d-f21530ed721e" containerName="extract-content" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.574312 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="d53d9ecd-c244-4a4a-9b0d-f21530ed721e" containerName="extract-content" Nov 24 13:16:38 crc kubenswrapper[4790]: E1124 13:16:38.574325 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="036855a4-3747-4525-bcda-03c03daf917c" containerName="registry-server" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.574336 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="036855a4-3747-4525-bcda-03c03daf917c" containerName="registry-server" Nov 24 13:16:38 crc kubenswrapper[4790]: E1124 13:16:38.574345 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1" containerName="extract-content" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.574353 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1" containerName="extract-content" Nov 24 13:16:38 crc kubenswrapper[4790]: E1124 13:16:38.574365 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d53d9ecd-c244-4a4a-9b0d-f21530ed721e" containerName="registry-server" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.574373 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="d53d9ecd-c244-4a4a-9b0d-f21530ed721e" containerName="registry-server" Nov 24 13:16:38 crc kubenswrapper[4790]: E1124 13:16:38.574381 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2420a87e-7aee-42d7-a61e-720f3a014682" containerName="pruner" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.574389 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="2420a87e-7aee-42d7-a61e-720f3a014682" containerName="pruner" Nov 24 13:16:38 crc kubenswrapper[4790]: E1124 13:16:38.574399 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6551fecc-df79-44f6-8600-fd016469a1e9" containerName="extract-content" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.574406 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="6551fecc-df79-44f6-8600-fd016469a1e9" containerName="extract-content" Nov 24 13:16:38 crc kubenswrapper[4790]: E1124 13:16:38.574418 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="036855a4-3747-4525-bcda-03c03daf917c" containerName="extract-content" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.574426 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="036855a4-3747-4525-bcda-03c03daf917c" containerName="extract-content" Nov 24 13:16:38 crc kubenswrapper[4790]: E1124 13:16:38.574434 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1" containerName="registry-server" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.574442 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1" containerName="registry-server" Nov 24 13:16:38 crc kubenswrapper[4790]: E1124 13:16:38.574455 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6551fecc-df79-44f6-8600-fd016469a1e9" containerName="registry-server" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.574463 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="6551fecc-df79-44f6-8600-fd016469a1e9" containerName="registry-server" Nov 24 13:16:38 crc kubenswrapper[4790]: E1124 13:16:38.574473 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6551fecc-df79-44f6-8600-fd016469a1e9" containerName="extract-utilities" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.574481 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="6551fecc-df79-44f6-8600-fd016469a1e9" containerName="extract-utilities" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.574607 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="036855a4-3747-4525-bcda-03c03daf917c" containerName="registry-server" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.574617 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="d53d9ecd-c244-4a4a-9b0d-f21530ed721e" containerName="registry-server" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.574633 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8cad7f0-dedb-4532-86e4-4a9fd8aaebd1" containerName="registry-server" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.574643 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="a86ba968-c6dd-42c8-91fb-207bc839b8ec" containerName="oauth-openshift" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.574655 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="6551fecc-df79-44f6-8600-fd016469a1e9" containerName="registry-server" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.574668 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="2420a87e-7aee-42d7-a61e-720f3a014682" containerName="pruner" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.575118 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.576301 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "a86ba968-c6dd-42c8-91fb-207bc839b8ec" (UID: "a86ba968-c6dd-42c8-91fb-207bc839b8ec"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.577226 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a86ba968-c6dd-42c8-91fb-207bc839b8ec-kube-api-access-t5zww" (OuterVolumeSpecName: "kube-api-access-t5zww") pod "a86ba968-c6dd-42c8-91fb-207bc839b8ec" (UID: "a86ba968-c6dd-42c8-91fb-207bc839b8ec"). InnerVolumeSpecName "kube-api-access-t5zww". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.588539 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-56c748df47-nnzs5"] Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.655189 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-session\") pod \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.655276 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-trusted-ca-bundle\") pod \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.655353 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-ocp-branding-template\") pod \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.655385 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-serving-cert\") pod \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.655449 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-cliconfig\") pod \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.655474 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-service-ca\") pod \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.655528 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-user-template-error\") pod \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\" (UID: \"a86ba968-c6dd-42c8-91fb-207bc839b8ec\") " Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.655917 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "a86ba968-c6dd-42c8-91fb-207bc839b8ec" (UID: "a86ba968-c6dd-42c8-91fb-207bc839b8ec"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.656298 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "a86ba968-c6dd-42c8-91fb-207bc839b8ec" (UID: "a86ba968-c6dd-42c8-91fb-207bc839b8ec"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.656451 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.656483 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lw4kj\" (UniqueName: \"kubernetes.io/projected/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-kube-api-access-lw4kj\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.656604 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.656623 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-audit-dir\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.656640 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-v4-0-config-system-service-ca\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.656662 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-v4-0-config-user-template-error\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.661495 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "a86ba968-c6dd-42c8-91fb-207bc839b8ec" (UID: "a86ba968-c6dd-42c8-91fb-207bc839b8ec"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.661920 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "a86ba968-c6dd-42c8-91fb-207bc839b8ec" (UID: "a86ba968-c6dd-42c8-91fb-207bc839b8ec"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.662106 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "a86ba968-c6dd-42c8-91fb-207bc839b8ec" (UID: "a86ba968-c6dd-42c8-91fb-207bc839b8ec"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.662260 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "a86ba968-c6dd-42c8-91fb-207bc839b8ec" (UID: "a86ba968-c6dd-42c8-91fb-207bc839b8ec"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.662648 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-v4-0-config-system-session\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.662733 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.662783 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-v4-0-config-user-template-login\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.662812 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-audit-policies\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.662845 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.662873 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-v4-0-config-system-router-certs\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.662910 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.662978 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.663086 4790 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a86ba968-c6dd-42c8-91fb-207bc839b8ec-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.663101 4790 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.663114 4790 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.663130 4790 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a86ba968-c6dd-42c8-91fb-207bc839b8ec-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.663141 4790 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.663154 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5zww\" (UniqueName: \"kubernetes.io/projected/a86ba968-c6dd-42c8-91fb-207bc839b8ec-kube-api-access-t5zww\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.663164 4790 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.663178 4790 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.663190 4790 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.663200 4790 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.663212 4790 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.663222 4790 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.663232 4790 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.662412 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "a86ba968-c6dd-42c8-91fb-207bc839b8ec" (UID: "a86ba968-c6dd-42c8-91fb-207bc839b8ec"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.764824 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.765659 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-v4-0-config-user-template-login\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.765697 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-audit-policies\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.765738 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.765771 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-v4-0-config-system-router-certs\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.765793 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.765835 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.765903 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.765952 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lw4kj\" (UniqueName: \"kubernetes.io/projected/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-kube-api-access-lw4kj\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.766016 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.766051 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-v4-0-config-system-service-ca\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.766085 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-audit-dir\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.766112 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-v4-0-config-user-template-error\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.766186 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-v4-0-config-system-session\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.766251 4790 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a86ba968-c6dd-42c8-91fb-207bc839b8ec-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.766521 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-audit-dir\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.766872 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-audit-policies\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.768158 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.768611 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.768990 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.769328 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-v4-0-config-system-service-ca\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.769669 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.770079 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.771205 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.772429 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-v4-0-config-user-template-error\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.772437 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-v4-0-config-system-session\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.772496 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-v4-0-config-system-router-certs\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.773320 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-v4-0-config-user-template-login\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.783465 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lw4kj\" (UniqueName: \"kubernetes.io/projected/542e1e31-28eb-4741-b9fe-8db4ebd2fda0-kube-api-access-lw4kj\") pod \"oauth-openshift-56c748df47-nnzs5\" (UID: \"542e1e31-28eb-4741-b9fe-8db4ebd2fda0\") " pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:38 crc kubenswrapper[4790]: I1124 13:16:38.934154 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:39 crc kubenswrapper[4790]: I1124 13:16:39.239525 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-56c748df47-nnzs5"] Nov 24 13:16:39 crc kubenswrapper[4790]: W1124 13:16:39.251272 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod542e1e31_28eb_4741_b9fe_8db4ebd2fda0.slice/crio-712ae3e9f10efb831209779b6cf3938884251adfd975c349017bd54845b201dc WatchSource:0}: Error finding container 712ae3e9f10efb831209779b6cf3938884251adfd975c349017bd54845b201dc: Status 404 returned error can't find the container with id 712ae3e9f10efb831209779b6cf3938884251adfd975c349017bd54845b201dc Nov 24 13:16:39 crc kubenswrapper[4790]: I1124 13:16:39.263214 4790 generic.go:334] "Generic (PLEG): container finished" podID="a86ba968-c6dd-42c8-91fb-207bc839b8ec" containerID="0c06dbf4c49aaeeb09163e81c971e4b532cc334825adcec02f48a8f9eafa867e" exitCode=0 Nov 24 13:16:39 crc kubenswrapper[4790]: I1124 13:16:39.263266 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" event={"ID":"a86ba968-c6dd-42c8-91fb-207bc839b8ec","Type":"ContainerDied","Data":"0c06dbf4c49aaeeb09163e81c971e4b532cc334825adcec02f48a8f9eafa867e"} Nov 24 13:16:39 crc kubenswrapper[4790]: I1124 13:16:39.263302 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" event={"ID":"a86ba968-c6dd-42c8-91fb-207bc839b8ec","Type":"ContainerDied","Data":"52aa339943e7758dc3adc470725aa069f384cd6d35be2b78083c3b060213239f"} Nov 24 13:16:39 crc kubenswrapper[4790]: I1124 13:16:39.263322 4790 scope.go:117] "RemoveContainer" containerID="0c06dbf4c49aaeeb09163e81c971e4b532cc334825adcec02f48a8f9eafa867e" Nov 24 13:16:39 crc kubenswrapper[4790]: I1124 13:16:39.263353 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-zxkb7" Nov 24 13:16:39 crc kubenswrapper[4790]: I1124 13:16:39.264922 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" event={"ID":"542e1e31-28eb-4741-b9fe-8db4ebd2fda0","Type":"ContainerStarted","Data":"712ae3e9f10efb831209779b6cf3938884251adfd975c349017bd54845b201dc"} Nov 24 13:16:39 crc kubenswrapper[4790]: I1124 13:16:39.285779 4790 scope.go:117] "RemoveContainer" containerID="0c06dbf4c49aaeeb09163e81c971e4b532cc334825adcec02f48a8f9eafa867e" Nov 24 13:16:39 crc kubenswrapper[4790]: E1124 13:16:39.286489 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c06dbf4c49aaeeb09163e81c971e4b532cc334825adcec02f48a8f9eafa867e\": container with ID starting with 0c06dbf4c49aaeeb09163e81c971e4b532cc334825adcec02f48a8f9eafa867e not found: ID does not exist" containerID="0c06dbf4c49aaeeb09163e81c971e4b532cc334825adcec02f48a8f9eafa867e" Nov 24 13:16:39 crc kubenswrapper[4790]: I1124 13:16:39.286632 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c06dbf4c49aaeeb09163e81c971e4b532cc334825adcec02f48a8f9eafa867e"} err="failed to get container status \"0c06dbf4c49aaeeb09163e81c971e4b532cc334825adcec02f48a8f9eafa867e\": rpc error: code = NotFound desc = could not find container \"0c06dbf4c49aaeeb09163e81c971e4b532cc334825adcec02f48a8f9eafa867e\": container with ID starting with 0c06dbf4c49aaeeb09163e81c971e4b532cc334825adcec02f48a8f9eafa867e not found: ID does not exist" Nov 24 13:16:39 crc kubenswrapper[4790]: I1124 13:16:39.363684 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-zxkb7"] Nov 24 13:16:39 crc kubenswrapper[4790]: I1124 13:16:39.367041 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-zxkb7"] Nov 24 13:16:40 crc kubenswrapper[4790]: I1124 13:16:40.273856 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" event={"ID":"542e1e31-28eb-4741-b9fe-8db4ebd2fda0","Type":"ContainerStarted","Data":"43885fe52f6658e735e6b21aa52404f29961b163c634e813e7e788b161c8190a"} Nov 24 13:16:40 crc kubenswrapper[4790]: I1124 13:16:40.275420 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:40 crc kubenswrapper[4790]: I1124 13:16:40.283674 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" Nov 24 13:16:40 crc kubenswrapper[4790]: I1124 13:16:40.313648 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-56c748df47-nnzs5" podStartSLOduration=27.313625451 podStartE2EDuration="27.313625451s" podCreationTimestamp="2025-11-24 13:16:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:40.311066703 +0000 UTC m=+248.690960405" watchObservedRunningTime="2025-11-24 13:16:40.313625451 +0000 UTC m=+248.693519123" Nov 24 13:16:40 crc kubenswrapper[4790]: I1124 13:16:40.323603 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a86ba968-c6dd-42c8-91fb-207bc839b8ec" path="/var/lib/kubelet/pods/a86ba968-c6dd-42c8-91fb-207bc839b8ec/volumes" Nov 24 13:16:57 crc kubenswrapper[4790]: I1124 13:16:57.934065 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qxxm7"] Nov 24 13:16:57 crc kubenswrapper[4790]: I1124 13:16:57.937767 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qxxm7" podUID="594c7e16-73fa-4653-aa63-9f93214efd12" containerName="registry-server" containerID="cri-o://96435c3c2d12f19bca126da859c7d48765a73390f48fcd58ff1a8674413f2525" gracePeriod=30 Nov 24 13:16:57 crc kubenswrapper[4790]: I1124 13:16:57.940311 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g5nsf"] Nov 24 13:16:57 crc kubenswrapper[4790]: I1124 13:16:57.940552 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-g5nsf" podUID="3c3ca964-4656-4248-bbcd-cd29957699cf" containerName="registry-server" containerID="cri-o://b07f5e8a194a7916b770b8c37c36831bead0572f4ffe24e42040bb4b3f4b69ac" gracePeriod=30 Nov 24 13:16:57 crc kubenswrapper[4790]: I1124 13:16:57.946387 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-s7x26"] Nov 24 13:16:57 crc kubenswrapper[4790]: I1124 13:16:57.946597 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-s7x26" podUID="951c1cb8-6613-404d-b0ec-ca5c8b66ec30" containerName="marketplace-operator" containerID="cri-o://2c08f3df7115b36a938fa7e6de536a3abb3482f5ed97fb396fd826fbdd252083" gracePeriod=30 Nov 24 13:16:57 crc kubenswrapper[4790]: I1124 13:16:57.964013 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7sjpz"] Nov 24 13:16:57 crc kubenswrapper[4790]: I1124 13:16:57.964332 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7sjpz" podUID="9420c18f-997b-4d89-b537-3cb99fe22682" containerName="registry-server" containerID="cri-o://f4c51cbb081f5f037e19488611a405c5eeadc45596e77273893334fb27307810" gracePeriod=30 Nov 24 13:16:57 crc kubenswrapper[4790]: I1124 13:16:57.965667 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w2gmf"] Nov 24 13:16:57 crc kubenswrapper[4790]: I1124 13:16:57.965960 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-w2gmf" podUID="37ec51fd-7e4c-4217-9a70-389efca8e855" containerName="registry-server" containerID="cri-o://2d6e78ba26ee897ac86ff20e363a1134a5b1e4751da5763bf285a97fd7eecdf3" gracePeriod=30 Nov 24 13:16:57 crc kubenswrapper[4790]: I1124 13:16:57.977356 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vj9c9"] Nov 24 13:16:57 crc kubenswrapper[4790]: I1124 13:16:57.978244 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vj9c9" Nov 24 13:16:57 crc kubenswrapper[4790]: I1124 13:16:57.989395 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vj9c9"] Nov 24 13:16:57 crc kubenswrapper[4790]: I1124 13:16:57.999354 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c843c1ad-11d6-4577-b337-40cdd992ae55-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vj9c9\" (UID: \"c843c1ad-11d6-4577-b337-40cdd992ae55\") " pod="openshift-marketplace/marketplace-operator-79b997595-vj9c9" Nov 24 13:16:57 crc kubenswrapper[4790]: I1124 13:16:57.999422 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c843c1ad-11d6-4577-b337-40cdd992ae55-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vj9c9\" (UID: \"c843c1ad-11d6-4577-b337-40cdd992ae55\") " pod="openshift-marketplace/marketplace-operator-79b997595-vj9c9" Nov 24 13:16:57 crc kubenswrapper[4790]: I1124 13:16:57.999490 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xt5mk\" (UniqueName: \"kubernetes.io/projected/c843c1ad-11d6-4577-b337-40cdd992ae55-kube-api-access-xt5mk\") pod \"marketplace-operator-79b997595-vj9c9\" (UID: \"c843c1ad-11d6-4577-b337-40cdd992ae55\") " pod="openshift-marketplace/marketplace-operator-79b997595-vj9c9" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.100760 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xt5mk\" (UniqueName: \"kubernetes.io/projected/c843c1ad-11d6-4577-b337-40cdd992ae55-kube-api-access-xt5mk\") pod \"marketplace-operator-79b997595-vj9c9\" (UID: \"c843c1ad-11d6-4577-b337-40cdd992ae55\") " pod="openshift-marketplace/marketplace-operator-79b997595-vj9c9" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.101102 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c843c1ad-11d6-4577-b337-40cdd992ae55-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vj9c9\" (UID: \"c843c1ad-11d6-4577-b337-40cdd992ae55\") " pod="openshift-marketplace/marketplace-operator-79b997595-vj9c9" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.101134 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c843c1ad-11d6-4577-b337-40cdd992ae55-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vj9c9\" (UID: \"c843c1ad-11d6-4577-b337-40cdd992ae55\") " pod="openshift-marketplace/marketplace-operator-79b997595-vj9c9" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.102962 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c843c1ad-11d6-4577-b337-40cdd992ae55-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vj9c9\" (UID: \"c843c1ad-11d6-4577-b337-40cdd992ae55\") " pod="openshift-marketplace/marketplace-operator-79b997595-vj9c9" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.107338 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c843c1ad-11d6-4577-b337-40cdd992ae55-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vj9c9\" (UID: \"c843c1ad-11d6-4577-b337-40cdd992ae55\") " pod="openshift-marketplace/marketplace-operator-79b997595-vj9c9" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.116719 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xt5mk\" (UniqueName: \"kubernetes.io/projected/c843c1ad-11d6-4577-b337-40cdd992ae55-kube-api-access-xt5mk\") pod \"marketplace-operator-79b997595-vj9c9\" (UID: \"c843c1ad-11d6-4577-b337-40cdd992ae55\") " pod="openshift-marketplace/marketplace-operator-79b997595-vj9c9" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.334759 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vj9c9" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.362454 4790 generic.go:334] "Generic (PLEG): container finished" podID="3c3ca964-4656-4248-bbcd-cd29957699cf" containerID="b07f5e8a194a7916b770b8c37c36831bead0572f4ffe24e42040bb4b3f4b69ac" exitCode=0 Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.362487 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g5nsf" event={"ID":"3c3ca964-4656-4248-bbcd-cd29957699cf","Type":"ContainerDied","Data":"b07f5e8a194a7916b770b8c37c36831bead0572f4ffe24e42040bb4b3f4b69ac"} Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.362537 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g5nsf" event={"ID":"3c3ca964-4656-4248-bbcd-cd29957699cf","Type":"ContainerDied","Data":"c6da93a593b6a7c000003a0430d78bbc1166d5f9239729cda1cf6e7e5e08ddb3"} Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.362551 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6da93a593b6a7c000003a0430d78bbc1166d5f9239729cda1cf6e7e5e08ddb3" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.363636 4790 generic.go:334] "Generic (PLEG): container finished" podID="951c1cb8-6613-404d-b0ec-ca5c8b66ec30" containerID="2c08f3df7115b36a938fa7e6de536a3abb3482f5ed97fb396fd826fbdd252083" exitCode=0 Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.363687 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-s7x26" event={"ID":"951c1cb8-6613-404d-b0ec-ca5c8b66ec30","Type":"ContainerDied","Data":"2c08f3df7115b36a938fa7e6de536a3abb3482f5ed97fb396fd826fbdd252083"} Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.363705 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-s7x26" event={"ID":"951c1cb8-6613-404d-b0ec-ca5c8b66ec30","Type":"ContainerDied","Data":"1cc2d2a338e96c9db5db751721c7e862551d2a108944c87a05f96e2990853495"} Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.363715 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1cc2d2a338e96c9db5db751721c7e862551d2a108944c87a05f96e2990853495" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.365801 4790 generic.go:334] "Generic (PLEG): container finished" podID="37ec51fd-7e4c-4217-9a70-389efca8e855" containerID="2d6e78ba26ee897ac86ff20e363a1134a5b1e4751da5763bf285a97fd7eecdf3" exitCode=0 Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.365868 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w2gmf" event={"ID":"37ec51fd-7e4c-4217-9a70-389efca8e855","Type":"ContainerDied","Data":"2d6e78ba26ee897ac86ff20e363a1134a5b1e4751da5763bf285a97fd7eecdf3"} Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.365917 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w2gmf" event={"ID":"37ec51fd-7e4c-4217-9a70-389efca8e855","Type":"ContainerDied","Data":"294d557adae8a5e5a92d82bf90d00e69291a09f27976c073fddd23fdd74c19b6"} Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.365931 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="294d557adae8a5e5a92d82bf90d00e69291a09f27976c073fddd23fdd74c19b6" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.369152 4790 generic.go:334] "Generic (PLEG): container finished" podID="594c7e16-73fa-4653-aa63-9f93214efd12" containerID="96435c3c2d12f19bca126da859c7d48765a73390f48fcd58ff1a8674413f2525" exitCode=0 Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.369201 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qxxm7" event={"ID":"594c7e16-73fa-4653-aa63-9f93214efd12","Type":"ContainerDied","Data":"96435c3c2d12f19bca126da859c7d48765a73390f48fcd58ff1a8674413f2525"} Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.369218 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qxxm7" event={"ID":"594c7e16-73fa-4653-aa63-9f93214efd12","Type":"ContainerDied","Data":"87897b4e662cb478bcf88adf746a1f56cc38ebb43e296b983da5ae30160b21f7"} Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.369229 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87897b4e662cb478bcf88adf746a1f56cc38ebb43e296b983da5ae30160b21f7" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.371995 4790 generic.go:334] "Generic (PLEG): container finished" podID="9420c18f-997b-4d89-b537-3cb99fe22682" containerID="f4c51cbb081f5f037e19488611a405c5eeadc45596e77273893334fb27307810" exitCode=0 Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.372035 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sjpz" event={"ID":"9420c18f-997b-4d89-b537-3cb99fe22682","Type":"ContainerDied","Data":"f4c51cbb081f5f037e19488611a405c5eeadc45596e77273893334fb27307810"} Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.380233 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g5nsf" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.386402 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qxxm7" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.390627 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-s7x26" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.411372 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w2gmf" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.454914 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7sjpz" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.507272 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dq57\" (UniqueName: \"kubernetes.io/projected/951c1cb8-6613-404d-b0ec-ca5c8b66ec30-kube-api-access-2dq57\") pod \"951c1cb8-6613-404d-b0ec-ca5c8b66ec30\" (UID: \"951c1cb8-6613-404d-b0ec-ca5c8b66ec30\") " Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.507361 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgvhh\" (UniqueName: \"kubernetes.io/projected/3c3ca964-4656-4248-bbcd-cd29957699cf-kube-api-access-mgvhh\") pod \"3c3ca964-4656-4248-bbcd-cd29957699cf\" (UID: \"3c3ca964-4656-4248-bbcd-cd29957699cf\") " Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.507440 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c3ca964-4656-4248-bbcd-cd29957699cf-utilities\") pod \"3c3ca964-4656-4248-bbcd-cd29957699cf\" (UID: \"3c3ca964-4656-4248-bbcd-cd29957699cf\") " Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.507506 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/951c1cb8-6613-404d-b0ec-ca5c8b66ec30-marketplace-trusted-ca\") pod \"951c1cb8-6613-404d-b0ec-ca5c8b66ec30\" (UID: \"951c1cb8-6613-404d-b0ec-ca5c8b66ec30\") " Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.507539 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/594c7e16-73fa-4653-aa63-9f93214efd12-utilities\") pod \"594c7e16-73fa-4653-aa63-9f93214efd12\" (UID: \"594c7e16-73fa-4653-aa63-9f93214efd12\") " Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.507606 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/951c1cb8-6613-404d-b0ec-ca5c8b66ec30-marketplace-operator-metrics\") pod \"951c1cb8-6613-404d-b0ec-ca5c8b66ec30\" (UID: \"951c1cb8-6613-404d-b0ec-ca5c8b66ec30\") " Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.507677 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/594c7e16-73fa-4653-aa63-9f93214efd12-catalog-content\") pod \"594c7e16-73fa-4653-aa63-9f93214efd12\" (UID: \"594c7e16-73fa-4653-aa63-9f93214efd12\") " Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.507759 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvdlx\" (UniqueName: \"kubernetes.io/projected/594c7e16-73fa-4653-aa63-9f93214efd12-kube-api-access-vvdlx\") pod \"594c7e16-73fa-4653-aa63-9f93214efd12\" (UID: \"594c7e16-73fa-4653-aa63-9f93214efd12\") " Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.507801 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c3ca964-4656-4248-bbcd-cd29957699cf-catalog-content\") pod \"3c3ca964-4656-4248-bbcd-cd29957699cf\" (UID: \"3c3ca964-4656-4248-bbcd-cd29957699cf\") " Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.509154 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/594c7e16-73fa-4653-aa63-9f93214efd12-utilities" (OuterVolumeSpecName: "utilities") pod "594c7e16-73fa-4653-aa63-9f93214efd12" (UID: "594c7e16-73fa-4653-aa63-9f93214efd12"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.510664 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/594c7e16-73fa-4653-aa63-9f93214efd12-kube-api-access-vvdlx" (OuterVolumeSpecName: "kube-api-access-vvdlx") pod "594c7e16-73fa-4653-aa63-9f93214efd12" (UID: "594c7e16-73fa-4653-aa63-9f93214efd12"). InnerVolumeSpecName "kube-api-access-vvdlx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.510811 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/951c1cb8-6613-404d-b0ec-ca5c8b66ec30-kube-api-access-2dq57" (OuterVolumeSpecName: "kube-api-access-2dq57") pod "951c1cb8-6613-404d-b0ec-ca5c8b66ec30" (UID: "951c1cb8-6613-404d-b0ec-ca5c8b66ec30"). InnerVolumeSpecName "kube-api-access-2dq57". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.510834 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/951c1cb8-6613-404d-b0ec-ca5c8b66ec30-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "951c1cb8-6613-404d-b0ec-ca5c8b66ec30" (UID: "951c1cb8-6613-404d-b0ec-ca5c8b66ec30"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.511626 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/951c1cb8-6613-404d-b0ec-ca5c8b66ec30-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "951c1cb8-6613-404d-b0ec-ca5c8b66ec30" (UID: "951c1cb8-6613-404d-b0ec-ca5c8b66ec30"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.512075 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c3ca964-4656-4248-bbcd-cd29957699cf-utilities" (OuterVolumeSpecName: "utilities") pod "3c3ca964-4656-4248-bbcd-cd29957699cf" (UID: "3c3ca964-4656-4248-bbcd-cd29957699cf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.512854 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c3ca964-4656-4248-bbcd-cd29957699cf-kube-api-access-mgvhh" (OuterVolumeSpecName: "kube-api-access-mgvhh") pod "3c3ca964-4656-4248-bbcd-cd29957699cf" (UID: "3c3ca964-4656-4248-bbcd-cd29957699cf"). InnerVolumeSpecName "kube-api-access-mgvhh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.569425 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/594c7e16-73fa-4653-aa63-9f93214efd12-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "594c7e16-73fa-4653-aa63-9f93214efd12" (UID: "594c7e16-73fa-4653-aa63-9f93214efd12"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.580932 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c3ca964-4656-4248-bbcd-cd29957699cf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3c3ca964-4656-4248-bbcd-cd29957699cf" (UID: "3c3ca964-4656-4248-bbcd-cd29957699cf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.608918 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9420c18f-997b-4d89-b537-3cb99fe22682-utilities\") pod \"9420c18f-997b-4d89-b537-3cb99fe22682\" (UID: \"9420c18f-997b-4d89-b537-3cb99fe22682\") " Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.609078 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9n2j\" (UniqueName: \"kubernetes.io/projected/37ec51fd-7e4c-4217-9a70-389efca8e855-kube-api-access-b9n2j\") pod \"37ec51fd-7e4c-4217-9a70-389efca8e855\" (UID: \"37ec51fd-7e4c-4217-9a70-389efca8e855\") " Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.609121 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9420c18f-997b-4d89-b537-3cb99fe22682-catalog-content\") pod \"9420c18f-997b-4d89-b537-3cb99fe22682\" (UID: \"9420c18f-997b-4d89-b537-3cb99fe22682\") " Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.609185 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37ec51fd-7e4c-4217-9a70-389efca8e855-catalog-content\") pod \"37ec51fd-7e4c-4217-9a70-389efca8e855\" (UID: \"37ec51fd-7e4c-4217-9a70-389efca8e855\") " Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.609212 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrnmj\" (UniqueName: \"kubernetes.io/projected/9420c18f-997b-4d89-b537-3cb99fe22682-kube-api-access-qrnmj\") pod \"9420c18f-997b-4d89-b537-3cb99fe22682\" (UID: \"9420c18f-997b-4d89-b537-3cb99fe22682\") " Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.609242 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37ec51fd-7e4c-4217-9a70-389efca8e855-utilities\") pod \"37ec51fd-7e4c-4217-9a70-389efca8e855\" (UID: \"37ec51fd-7e4c-4217-9a70-389efca8e855\") " Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.609476 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvdlx\" (UniqueName: \"kubernetes.io/projected/594c7e16-73fa-4653-aa63-9f93214efd12-kube-api-access-vvdlx\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.609494 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c3ca964-4656-4248-bbcd-cd29957699cf-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.609503 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dq57\" (UniqueName: \"kubernetes.io/projected/951c1cb8-6613-404d-b0ec-ca5c8b66ec30-kube-api-access-2dq57\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.609512 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgvhh\" (UniqueName: \"kubernetes.io/projected/3c3ca964-4656-4248-bbcd-cd29957699cf-kube-api-access-mgvhh\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.609520 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c3ca964-4656-4248-bbcd-cd29957699cf-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.609529 4790 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/951c1cb8-6613-404d-b0ec-ca5c8b66ec30-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.609537 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/594c7e16-73fa-4653-aa63-9f93214efd12-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.609545 4790 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/951c1cb8-6613-404d-b0ec-ca5c8b66ec30-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.609554 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/594c7e16-73fa-4653-aa63-9f93214efd12-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.609821 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9420c18f-997b-4d89-b537-3cb99fe22682-utilities" (OuterVolumeSpecName: "utilities") pod "9420c18f-997b-4d89-b537-3cb99fe22682" (UID: "9420c18f-997b-4d89-b537-3cb99fe22682"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.610402 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37ec51fd-7e4c-4217-9a70-389efca8e855-utilities" (OuterVolumeSpecName: "utilities") pod "37ec51fd-7e4c-4217-9a70-389efca8e855" (UID: "37ec51fd-7e4c-4217-9a70-389efca8e855"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.613377 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9420c18f-997b-4d89-b537-3cb99fe22682-kube-api-access-qrnmj" (OuterVolumeSpecName: "kube-api-access-qrnmj") pod "9420c18f-997b-4d89-b537-3cb99fe22682" (UID: "9420c18f-997b-4d89-b537-3cb99fe22682"). InnerVolumeSpecName "kube-api-access-qrnmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.613468 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37ec51fd-7e4c-4217-9a70-389efca8e855-kube-api-access-b9n2j" (OuterVolumeSpecName: "kube-api-access-b9n2j") pod "37ec51fd-7e4c-4217-9a70-389efca8e855" (UID: "37ec51fd-7e4c-4217-9a70-389efca8e855"). InnerVolumeSpecName "kube-api-access-b9n2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.631663 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9420c18f-997b-4d89-b537-3cb99fe22682-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9420c18f-997b-4d89-b537-3cb99fe22682" (UID: "9420c18f-997b-4d89-b537-3cb99fe22682"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.700601 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37ec51fd-7e4c-4217-9a70-389efca8e855-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "37ec51fd-7e4c-4217-9a70-389efca8e855" (UID: "37ec51fd-7e4c-4217-9a70-389efca8e855"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.710520 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9420c18f-997b-4d89-b537-3cb99fe22682-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.710567 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9n2j\" (UniqueName: \"kubernetes.io/projected/37ec51fd-7e4c-4217-9a70-389efca8e855-kube-api-access-b9n2j\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.710581 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9420c18f-997b-4d89-b537-3cb99fe22682-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.710593 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37ec51fd-7e4c-4217-9a70-389efca8e855-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.710603 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrnmj\" (UniqueName: \"kubernetes.io/projected/9420c18f-997b-4d89-b537-3cb99fe22682-kube-api-access-qrnmj\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.710614 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37ec51fd-7e4c-4217-9a70-389efca8e855-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:16:58 crc kubenswrapper[4790]: I1124 13:16:58.803236 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vj9c9"] Nov 24 13:16:58 crc kubenswrapper[4790]: W1124 13:16:58.808698 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc843c1ad_11d6_4577_b337_40cdd992ae55.slice/crio-051a64bd3828fd6e9794518cf60f4643ad31aa5896d9af9e822fa49989624dad WatchSource:0}: Error finding container 051a64bd3828fd6e9794518cf60f4643ad31aa5896d9af9e822fa49989624dad: Status 404 returned error can't find the container with id 051a64bd3828fd6e9794518cf60f4643ad31aa5896d9af9e822fa49989624dad Nov 24 13:16:59 crc kubenswrapper[4790]: I1124 13:16:59.378480 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7sjpz" Nov 24 13:16:59 crc kubenswrapper[4790]: I1124 13:16:59.378447 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sjpz" event={"ID":"9420c18f-997b-4d89-b537-3cb99fe22682","Type":"ContainerDied","Data":"de788473233b1593d640a745ebd3401285b49f955f81e4a48d3e1b6b7c8cfe56"} Nov 24 13:16:59 crc kubenswrapper[4790]: I1124 13:16:59.378826 4790 scope.go:117] "RemoveContainer" containerID="f4c51cbb081f5f037e19488611a405c5eeadc45596e77273893334fb27307810" Nov 24 13:16:59 crc kubenswrapper[4790]: I1124 13:16:59.379802 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vj9c9" event={"ID":"c843c1ad-11d6-4577-b337-40cdd992ae55","Type":"ContainerStarted","Data":"f1af23190b3d120d8dbf459710d4b27993399cfd087aa522871978bcf3d6b406"} Nov 24 13:16:59 crc kubenswrapper[4790]: I1124 13:16:59.379843 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-s7x26" Nov 24 13:16:59 crc kubenswrapper[4790]: I1124 13:16:59.379850 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vj9c9" event={"ID":"c843c1ad-11d6-4577-b337-40cdd992ae55","Type":"ContainerStarted","Data":"051a64bd3828fd6e9794518cf60f4643ad31aa5896d9af9e822fa49989624dad"} Nov 24 13:16:59 crc kubenswrapper[4790]: I1124 13:16:59.379857 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w2gmf" Nov 24 13:16:59 crc kubenswrapper[4790]: I1124 13:16:59.379842 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qxxm7" Nov 24 13:16:59 crc kubenswrapper[4790]: I1124 13:16:59.380067 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g5nsf" Nov 24 13:16:59 crc kubenswrapper[4790]: I1124 13:16:59.380088 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-vj9c9" Nov 24 13:16:59 crc kubenswrapper[4790]: I1124 13:16:59.398600 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-vj9c9" Nov 24 13:16:59 crc kubenswrapper[4790]: I1124 13:16:59.410175 4790 scope.go:117] "RemoveContainer" containerID="09ed628c5000247436c66eb024238eacbf57acd0e1ff823774254afda5265851" Nov 24 13:16:59 crc kubenswrapper[4790]: I1124 13:16:59.446821 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-vj9c9" podStartSLOduration=2.446795349 podStartE2EDuration="2.446795349s" podCreationTimestamp="2025-11-24 13:16:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:59.421388606 +0000 UTC m=+267.801282268" watchObservedRunningTime="2025-11-24 13:16:59.446795349 +0000 UTC m=+267.826689011" Nov 24 13:16:59 crc kubenswrapper[4790]: I1124 13:16:59.455834 4790 scope.go:117] "RemoveContainer" containerID="ee0a266eda880b7d7eb3ee729adeb053d6151082a6b64edfd8f7690905f84bbf" Nov 24 13:16:59 crc kubenswrapper[4790]: I1124 13:16:59.456309 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7sjpz"] Nov 24 13:16:59 crc kubenswrapper[4790]: I1124 13:16:59.461567 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7sjpz"] Nov 24 13:16:59 crc kubenswrapper[4790]: I1124 13:16:59.472929 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w2gmf"] Nov 24 13:16:59 crc kubenswrapper[4790]: I1124 13:16:59.477952 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-w2gmf"] Nov 24 13:16:59 crc kubenswrapper[4790]: I1124 13:16:59.484858 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-s7x26"] Nov 24 13:16:59 crc kubenswrapper[4790]: I1124 13:16:59.487443 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-s7x26"] Nov 24 13:16:59 crc kubenswrapper[4790]: I1124 13:16:59.500868 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g5nsf"] Nov 24 13:16:59 crc kubenswrapper[4790]: I1124 13:16:59.504188 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-g5nsf"] Nov 24 13:16:59 crc kubenswrapper[4790]: I1124 13:16:59.510871 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qxxm7"] Nov 24 13:16:59 crc kubenswrapper[4790]: I1124 13:16:59.517731 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qxxm7"] Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.150555 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-r266z"] Nov 24 13:17:00 crc kubenswrapper[4790]: E1124 13:17:00.150808 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="951c1cb8-6613-404d-b0ec-ca5c8b66ec30" containerName="marketplace-operator" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.150824 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="951c1cb8-6613-404d-b0ec-ca5c8b66ec30" containerName="marketplace-operator" Nov 24 13:17:00 crc kubenswrapper[4790]: E1124 13:17:00.150837 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c3ca964-4656-4248-bbcd-cd29957699cf" containerName="extract-utilities" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.150845 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c3ca964-4656-4248-bbcd-cd29957699cf" containerName="extract-utilities" Nov 24 13:17:00 crc kubenswrapper[4790]: E1124 13:17:00.150856 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c3ca964-4656-4248-bbcd-cd29957699cf" containerName="extract-content" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.150864 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c3ca964-4656-4248-bbcd-cd29957699cf" containerName="extract-content" Nov 24 13:17:00 crc kubenswrapper[4790]: E1124 13:17:00.150875 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37ec51fd-7e4c-4217-9a70-389efca8e855" containerName="registry-server" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.150898 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="37ec51fd-7e4c-4217-9a70-389efca8e855" containerName="registry-server" Nov 24 13:17:00 crc kubenswrapper[4790]: E1124 13:17:00.150916 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9420c18f-997b-4d89-b537-3cb99fe22682" containerName="extract-utilities" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.150926 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="9420c18f-997b-4d89-b537-3cb99fe22682" containerName="extract-utilities" Nov 24 13:17:00 crc kubenswrapper[4790]: E1124 13:17:00.150939 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="594c7e16-73fa-4653-aa63-9f93214efd12" containerName="registry-server" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.150947 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="594c7e16-73fa-4653-aa63-9f93214efd12" containerName="registry-server" Nov 24 13:17:00 crc kubenswrapper[4790]: E1124 13:17:00.150959 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="594c7e16-73fa-4653-aa63-9f93214efd12" containerName="extract-content" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.150967 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="594c7e16-73fa-4653-aa63-9f93214efd12" containerName="extract-content" Nov 24 13:17:00 crc kubenswrapper[4790]: E1124 13:17:00.150979 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="594c7e16-73fa-4653-aa63-9f93214efd12" containerName="extract-utilities" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.150987 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="594c7e16-73fa-4653-aa63-9f93214efd12" containerName="extract-utilities" Nov 24 13:17:00 crc kubenswrapper[4790]: E1124 13:17:00.150999 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9420c18f-997b-4d89-b537-3cb99fe22682" containerName="extract-content" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.151007 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="9420c18f-997b-4d89-b537-3cb99fe22682" containerName="extract-content" Nov 24 13:17:00 crc kubenswrapper[4790]: E1124 13:17:00.151018 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9420c18f-997b-4d89-b537-3cb99fe22682" containerName="registry-server" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.151026 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="9420c18f-997b-4d89-b537-3cb99fe22682" containerName="registry-server" Nov 24 13:17:00 crc kubenswrapper[4790]: E1124 13:17:00.151038 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c3ca964-4656-4248-bbcd-cd29957699cf" containerName="registry-server" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.151045 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c3ca964-4656-4248-bbcd-cd29957699cf" containerName="registry-server" Nov 24 13:17:00 crc kubenswrapper[4790]: E1124 13:17:00.151056 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37ec51fd-7e4c-4217-9a70-389efca8e855" containerName="extract-content" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.151061 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="37ec51fd-7e4c-4217-9a70-389efca8e855" containerName="extract-content" Nov 24 13:17:00 crc kubenswrapper[4790]: E1124 13:17:00.151072 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37ec51fd-7e4c-4217-9a70-389efca8e855" containerName="extract-utilities" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.151078 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="37ec51fd-7e4c-4217-9a70-389efca8e855" containerName="extract-utilities" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.151159 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="37ec51fd-7e4c-4217-9a70-389efca8e855" containerName="registry-server" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.151175 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="594c7e16-73fa-4653-aa63-9f93214efd12" containerName="registry-server" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.151182 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="951c1cb8-6613-404d-b0ec-ca5c8b66ec30" containerName="marketplace-operator" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.151190 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c3ca964-4656-4248-bbcd-cd29957699cf" containerName="registry-server" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.151197 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="9420c18f-997b-4d89-b537-3cb99fe22682" containerName="registry-server" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.151829 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r266z" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.153455 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.159516 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r266z"] Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.321583 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37ec51fd-7e4c-4217-9a70-389efca8e855" path="/var/lib/kubelet/pods/37ec51fd-7e4c-4217-9a70-389efca8e855/volumes" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.322235 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c3ca964-4656-4248-bbcd-cd29957699cf" path="/var/lib/kubelet/pods/3c3ca964-4656-4248-bbcd-cd29957699cf/volumes" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.322813 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="594c7e16-73fa-4653-aa63-9f93214efd12" path="/var/lib/kubelet/pods/594c7e16-73fa-4653-aa63-9f93214efd12/volumes" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.323872 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9420c18f-997b-4d89-b537-3cb99fe22682" path="/var/lib/kubelet/pods/9420c18f-997b-4d89-b537-3cb99fe22682/volumes" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.324545 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="951c1cb8-6613-404d-b0ec-ca5c8b66ec30" path="/var/lib/kubelet/pods/951c1cb8-6613-404d-b0ec-ca5c8b66ec30/volumes" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.335402 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97jnw\" (UniqueName: \"kubernetes.io/projected/92ba1be5-af31-4941-94ef-3977f17b297e-kube-api-access-97jnw\") pod \"redhat-marketplace-r266z\" (UID: \"92ba1be5-af31-4941-94ef-3977f17b297e\") " pod="openshift-marketplace/redhat-marketplace-r266z" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.335481 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92ba1be5-af31-4941-94ef-3977f17b297e-utilities\") pod \"redhat-marketplace-r266z\" (UID: \"92ba1be5-af31-4941-94ef-3977f17b297e\") " pod="openshift-marketplace/redhat-marketplace-r266z" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.335525 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92ba1be5-af31-4941-94ef-3977f17b297e-catalog-content\") pod \"redhat-marketplace-r266z\" (UID: \"92ba1be5-af31-4941-94ef-3977f17b297e\") " pod="openshift-marketplace/redhat-marketplace-r266z" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.349707 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-j69qv"] Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.350760 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j69qv" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.352440 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.364052 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j69qv"] Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.436496 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92ba1be5-af31-4941-94ef-3977f17b297e-utilities\") pod \"redhat-marketplace-r266z\" (UID: \"92ba1be5-af31-4941-94ef-3977f17b297e\") " pod="openshift-marketplace/redhat-marketplace-r266z" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.436548 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/324e1511-0e35-441f-a819-6faf957fe189-catalog-content\") pod \"redhat-operators-j69qv\" (UID: \"324e1511-0e35-441f-a819-6faf957fe189\") " pod="openshift-marketplace/redhat-operators-j69qv" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.436611 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92ba1be5-af31-4941-94ef-3977f17b297e-catalog-content\") pod \"redhat-marketplace-r266z\" (UID: \"92ba1be5-af31-4941-94ef-3977f17b297e\") " pod="openshift-marketplace/redhat-marketplace-r266z" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.436642 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crb5n\" (UniqueName: \"kubernetes.io/projected/324e1511-0e35-441f-a819-6faf957fe189-kube-api-access-crb5n\") pod \"redhat-operators-j69qv\" (UID: \"324e1511-0e35-441f-a819-6faf957fe189\") " pod="openshift-marketplace/redhat-operators-j69qv" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.436673 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97jnw\" (UniqueName: \"kubernetes.io/projected/92ba1be5-af31-4941-94ef-3977f17b297e-kube-api-access-97jnw\") pod \"redhat-marketplace-r266z\" (UID: \"92ba1be5-af31-4941-94ef-3977f17b297e\") " pod="openshift-marketplace/redhat-marketplace-r266z" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.437138 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/324e1511-0e35-441f-a819-6faf957fe189-utilities\") pod \"redhat-operators-j69qv\" (UID: \"324e1511-0e35-441f-a819-6faf957fe189\") " pod="openshift-marketplace/redhat-operators-j69qv" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.437213 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92ba1be5-af31-4941-94ef-3977f17b297e-catalog-content\") pod \"redhat-marketplace-r266z\" (UID: \"92ba1be5-af31-4941-94ef-3977f17b297e\") " pod="openshift-marketplace/redhat-marketplace-r266z" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.437455 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92ba1be5-af31-4941-94ef-3977f17b297e-utilities\") pod \"redhat-marketplace-r266z\" (UID: \"92ba1be5-af31-4941-94ef-3977f17b297e\") " pod="openshift-marketplace/redhat-marketplace-r266z" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.454514 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97jnw\" (UniqueName: \"kubernetes.io/projected/92ba1be5-af31-4941-94ef-3977f17b297e-kube-api-access-97jnw\") pod \"redhat-marketplace-r266z\" (UID: \"92ba1be5-af31-4941-94ef-3977f17b297e\") " pod="openshift-marketplace/redhat-marketplace-r266z" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.468379 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r266z" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.540000 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/324e1511-0e35-441f-a819-6faf957fe189-catalog-content\") pod \"redhat-operators-j69qv\" (UID: \"324e1511-0e35-441f-a819-6faf957fe189\") " pod="openshift-marketplace/redhat-operators-j69qv" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.540109 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crb5n\" (UniqueName: \"kubernetes.io/projected/324e1511-0e35-441f-a819-6faf957fe189-kube-api-access-crb5n\") pod \"redhat-operators-j69qv\" (UID: \"324e1511-0e35-441f-a819-6faf957fe189\") " pod="openshift-marketplace/redhat-operators-j69qv" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.540173 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/324e1511-0e35-441f-a819-6faf957fe189-utilities\") pod \"redhat-operators-j69qv\" (UID: \"324e1511-0e35-441f-a819-6faf957fe189\") " pod="openshift-marketplace/redhat-operators-j69qv" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.540488 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/324e1511-0e35-441f-a819-6faf957fe189-catalog-content\") pod \"redhat-operators-j69qv\" (UID: \"324e1511-0e35-441f-a819-6faf957fe189\") " pod="openshift-marketplace/redhat-operators-j69qv" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.540563 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/324e1511-0e35-441f-a819-6faf957fe189-utilities\") pod \"redhat-operators-j69qv\" (UID: \"324e1511-0e35-441f-a819-6faf957fe189\") " pod="openshift-marketplace/redhat-operators-j69qv" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.560907 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crb5n\" (UniqueName: \"kubernetes.io/projected/324e1511-0e35-441f-a819-6faf957fe189-kube-api-access-crb5n\") pod \"redhat-operators-j69qv\" (UID: \"324e1511-0e35-441f-a819-6faf957fe189\") " pod="openshift-marketplace/redhat-operators-j69qv" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.637453 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r266z"] Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.668248 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j69qv" Nov 24 13:17:00 crc kubenswrapper[4790]: I1124 13:17:00.865815 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j69qv"] Nov 24 13:17:00 crc kubenswrapper[4790]: W1124 13:17:00.872176 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod324e1511_0e35_441f_a819_6faf957fe189.slice/crio-44e31064f8c48b15d799d6a7f3f5b1d4e94f1702466c412bbcdb4e7de8676b54 WatchSource:0}: Error finding container 44e31064f8c48b15d799d6a7f3f5b1d4e94f1702466c412bbcdb4e7de8676b54: Status 404 returned error can't find the container with id 44e31064f8c48b15d799d6a7f3f5b1d4e94f1702466c412bbcdb4e7de8676b54 Nov 24 13:17:01 crc kubenswrapper[4790]: I1124 13:17:01.390826 4790 generic.go:334] "Generic (PLEG): container finished" podID="324e1511-0e35-441f-a819-6faf957fe189" containerID="bccf83f5efa4bb596c5a1856f780440b6fa155b46a95ebaf00a5d582f20392e1" exitCode=0 Nov 24 13:17:01 crc kubenswrapper[4790]: I1124 13:17:01.390911 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j69qv" event={"ID":"324e1511-0e35-441f-a819-6faf957fe189","Type":"ContainerDied","Data":"bccf83f5efa4bb596c5a1856f780440b6fa155b46a95ebaf00a5d582f20392e1"} Nov 24 13:17:01 crc kubenswrapper[4790]: I1124 13:17:01.390972 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j69qv" event={"ID":"324e1511-0e35-441f-a819-6faf957fe189","Type":"ContainerStarted","Data":"44e31064f8c48b15d799d6a7f3f5b1d4e94f1702466c412bbcdb4e7de8676b54"} Nov 24 13:17:01 crc kubenswrapper[4790]: I1124 13:17:01.392917 4790 generic.go:334] "Generic (PLEG): container finished" podID="92ba1be5-af31-4941-94ef-3977f17b297e" containerID="b99a3457bee20678589edc37bd77cc07a4adbbdd2edbc6e1a4ad9fb5ff7a045d" exitCode=0 Nov 24 13:17:01 crc kubenswrapper[4790]: I1124 13:17:01.393016 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r266z" event={"ID":"92ba1be5-af31-4941-94ef-3977f17b297e","Type":"ContainerDied","Data":"b99a3457bee20678589edc37bd77cc07a4adbbdd2edbc6e1a4ad9fb5ff7a045d"} Nov 24 13:17:01 crc kubenswrapper[4790]: I1124 13:17:01.393064 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r266z" event={"ID":"92ba1be5-af31-4941-94ef-3977f17b297e","Type":"ContainerStarted","Data":"e48225bcc3d199f1791b0f4089249783a97deba2eb6ab1273f85df45993c10ec"} Nov 24 13:17:02 crc kubenswrapper[4790]: I1124 13:17:02.398448 4790 generic.go:334] "Generic (PLEG): container finished" podID="92ba1be5-af31-4941-94ef-3977f17b297e" containerID="db95cc3702fd1dc5f8e91139b22ea2e8e05c352c6e81f254e8504ab89cee7322" exitCode=0 Nov 24 13:17:02 crc kubenswrapper[4790]: I1124 13:17:02.398546 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r266z" event={"ID":"92ba1be5-af31-4941-94ef-3977f17b297e","Type":"ContainerDied","Data":"db95cc3702fd1dc5f8e91139b22ea2e8e05c352c6e81f254e8504ab89cee7322"} Nov 24 13:17:02 crc kubenswrapper[4790]: I1124 13:17:02.400538 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j69qv" event={"ID":"324e1511-0e35-441f-a819-6faf957fe189","Type":"ContainerStarted","Data":"c5ed873b1002b1b75dadfe4f6037daa841206c87c98fb9df269ab7a2d7c4f465"} Nov 24 13:17:02 crc kubenswrapper[4790]: I1124 13:17:02.549027 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-r2j97"] Nov 24 13:17:02 crc kubenswrapper[4790]: I1124 13:17:02.550022 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r2j97" Nov 24 13:17:02 crc kubenswrapper[4790]: I1124 13:17:02.557023 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 24 13:17:02 crc kubenswrapper[4790]: I1124 13:17:02.567117 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r2j97"] Nov 24 13:17:02 crc kubenswrapper[4790]: I1124 13:17:02.575555 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmn7x\" (UniqueName: \"kubernetes.io/projected/b790d0d0-0038-4153-8508-b5f974dda04c-kube-api-access-cmn7x\") pod \"community-operators-r2j97\" (UID: \"b790d0d0-0038-4153-8508-b5f974dda04c\") " pod="openshift-marketplace/community-operators-r2j97" Nov 24 13:17:02 crc kubenswrapper[4790]: I1124 13:17:02.575604 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b790d0d0-0038-4153-8508-b5f974dda04c-utilities\") pod \"community-operators-r2j97\" (UID: \"b790d0d0-0038-4153-8508-b5f974dda04c\") " pod="openshift-marketplace/community-operators-r2j97" Nov 24 13:17:02 crc kubenswrapper[4790]: I1124 13:17:02.575628 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b790d0d0-0038-4153-8508-b5f974dda04c-catalog-content\") pod \"community-operators-r2j97\" (UID: \"b790d0d0-0038-4153-8508-b5f974dda04c\") " pod="openshift-marketplace/community-operators-r2j97" Nov 24 13:17:02 crc kubenswrapper[4790]: I1124 13:17:02.676633 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmn7x\" (UniqueName: \"kubernetes.io/projected/b790d0d0-0038-4153-8508-b5f974dda04c-kube-api-access-cmn7x\") pod \"community-operators-r2j97\" (UID: \"b790d0d0-0038-4153-8508-b5f974dda04c\") " pod="openshift-marketplace/community-operators-r2j97" Nov 24 13:17:02 crc kubenswrapper[4790]: I1124 13:17:02.677221 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b790d0d0-0038-4153-8508-b5f974dda04c-utilities\") pod \"community-operators-r2j97\" (UID: \"b790d0d0-0038-4153-8508-b5f974dda04c\") " pod="openshift-marketplace/community-operators-r2j97" Nov 24 13:17:02 crc kubenswrapper[4790]: I1124 13:17:02.677276 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b790d0d0-0038-4153-8508-b5f974dda04c-catalog-content\") pod \"community-operators-r2j97\" (UID: \"b790d0d0-0038-4153-8508-b5f974dda04c\") " pod="openshift-marketplace/community-operators-r2j97" Nov 24 13:17:02 crc kubenswrapper[4790]: I1124 13:17:02.678854 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b790d0d0-0038-4153-8508-b5f974dda04c-catalog-content\") pod \"community-operators-r2j97\" (UID: \"b790d0d0-0038-4153-8508-b5f974dda04c\") " pod="openshift-marketplace/community-operators-r2j97" Nov 24 13:17:02 crc kubenswrapper[4790]: I1124 13:17:02.679015 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b790d0d0-0038-4153-8508-b5f974dda04c-utilities\") pod \"community-operators-r2j97\" (UID: \"b790d0d0-0038-4153-8508-b5f974dda04c\") " pod="openshift-marketplace/community-operators-r2j97" Nov 24 13:17:02 crc kubenswrapper[4790]: I1124 13:17:02.698765 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmn7x\" (UniqueName: \"kubernetes.io/projected/b790d0d0-0038-4153-8508-b5f974dda04c-kube-api-access-cmn7x\") pod \"community-operators-r2j97\" (UID: \"b790d0d0-0038-4153-8508-b5f974dda04c\") " pod="openshift-marketplace/community-operators-r2j97" Nov 24 13:17:02 crc kubenswrapper[4790]: I1124 13:17:02.753236 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pn2zj"] Nov 24 13:17:02 crc kubenswrapper[4790]: I1124 13:17:02.754353 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pn2zj" Nov 24 13:17:02 crc kubenswrapper[4790]: I1124 13:17:02.758685 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 24 13:17:02 crc kubenswrapper[4790]: I1124 13:17:02.769368 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pn2zj"] Nov 24 13:17:02 crc kubenswrapper[4790]: I1124 13:17:02.779260 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/730d2a27-6ef4-4a97-8e6b-5272e26aa247-catalog-content\") pod \"certified-operators-pn2zj\" (UID: \"730d2a27-6ef4-4a97-8e6b-5272e26aa247\") " pod="openshift-marketplace/certified-operators-pn2zj" Nov 24 13:17:02 crc kubenswrapper[4790]: I1124 13:17:02.779408 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/730d2a27-6ef4-4a97-8e6b-5272e26aa247-utilities\") pod \"certified-operators-pn2zj\" (UID: \"730d2a27-6ef4-4a97-8e6b-5272e26aa247\") " pod="openshift-marketplace/certified-operators-pn2zj" Nov 24 13:17:02 crc kubenswrapper[4790]: I1124 13:17:02.779438 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnnxf\" (UniqueName: \"kubernetes.io/projected/730d2a27-6ef4-4a97-8e6b-5272e26aa247-kube-api-access-mnnxf\") pod \"certified-operators-pn2zj\" (UID: \"730d2a27-6ef4-4a97-8e6b-5272e26aa247\") " pod="openshift-marketplace/certified-operators-pn2zj" Nov 24 13:17:02 crc kubenswrapper[4790]: I1124 13:17:02.867785 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r2j97" Nov 24 13:17:02 crc kubenswrapper[4790]: I1124 13:17:02.880035 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/730d2a27-6ef4-4a97-8e6b-5272e26aa247-catalog-content\") pod \"certified-operators-pn2zj\" (UID: \"730d2a27-6ef4-4a97-8e6b-5272e26aa247\") " pod="openshift-marketplace/certified-operators-pn2zj" Nov 24 13:17:02 crc kubenswrapper[4790]: I1124 13:17:02.880097 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/730d2a27-6ef4-4a97-8e6b-5272e26aa247-utilities\") pod \"certified-operators-pn2zj\" (UID: \"730d2a27-6ef4-4a97-8e6b-5272e26aa247\") " pod="openshift-marketplace/certified-operators-pn2zj" Nov 24 13:17:02 crc kubenswrapper[4790]: I1124 13:17:02.880121 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnnxf\" (UniqueName: \"kubernetes.io/projected/730d2a27-6ef4-4a97-8e6b-5272e26aa247-kube-api-access-mnnxf\") pod \"certified-operators-pn2zj\" (UID: \"730d2a27-6ef4-4a97-8e6b-5272e26aa247\") " pod="openshift-marketplace/certified-operators-pn2zj" Nov 24 13:17:02 crc kubenswrapper[4790]: I1124 13:17:02.880492 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/730d2a27-6ef4-4a97-8e6b-5272e26aa247-catalog-content\") pod \"certified-operators-pn2zj\" (UID: \"730d2a27-6ef4-4a97-8e6b-5272e26aa247\") " pod="openshift-marketplace/certified-operators-pn2zj" Nov 24 13:17:02 crc kubenswrapper[4790]: I1124 13:17:02.880523 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/730d2a27-6ef4-4a97-8e6b-5272e26aa247-utilities\") pod \"certified-operators-pn2zj\" (UID: \"730d2a27-6ef4-4a97-8e6b-5272e26aa247\") " pod="openshift-marketplace/certified-operators-pn2zj" Nov 24 13:17:02 crc kubenswrapper[4790]: I1124 13:17:02.899160 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnnxf\" (UniqueName: \"kubernetes.io/projected/730d2a27-6ef4-4a97-8e6b-5272e26aa247-kube-api-access-mnnxf\") pod \"certified-operators-pn2zj\" (UID: \"730d2a27-6ef4-4a97-8e6b-5272e26aa247\") " pod="openshift-marketplace/certified-operators-pn2zj" Nov 24 13:17:03 crc kubenswrapper[4790]: I1124 13:17:03.109616 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pn2zj" Nov 24 13:17:03 crc kubenswrapper[4790]: I1124 13:17:03.285422 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r2j97"] Nov 24 13:17:03 crc kubenswrapper[4790]: W1124 13:17:03.291568 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb790d0d0_0038_4153_8508_b5f974dda04c.slice/crio-2b998ed567a370390fc27eb766a29f0317efd7587263e94e797e501fca823d98 WatchSource:0}: Error finding container 2b998ed567a370390fc27eb766a29f0317efd7587263e94e797e501fca823d98: Status 404 returned error can't find the container with id 2b998ed567a370390fc27eb766a29f0317efd7587263e94e797e501fca823d98 Nov 24 13:17:03 crc kubenswrapper[4790]: I1124 13:17:03.408726 4790 generic.go:334] "Generic (PLEG): container finished" podID="324e1511-0e35-441f-a819-6faf957fe189" containerID="c5ed873b1002b1b75dadfe4f6037daa841206c87c98fb9df269ab7a2d7c4f465" exitCode=0 Nov 24 13:17:03 crc kubenswrapper[4790]: I1124 13:17:03.408796 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j69qv" event={"ID":"324e1511-0e35-441f-a819-6faf957fe189","Type":"ContainerDied","Data":"c5ed873b1002b1b75dadfe4f6037daa841206c87c98fb9df269ab7a2d7c4f465"} Nov 24 13:17:03 crc kubenswrapper[4790]: I1124 13:17:03.412139 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r2j97" event={"ID":"b790d0d0-0038-4153-8508-b5f974dda04c","Type":"ContainerStarted","Data":"8c635a22d6853ab2b43811d154d38b26922332371ba5f476e3809cdba17fd607"} Nov 24 13:17:03 crc kubenswrapper[4790]: I1124 13:17:03.412186 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r2j97" event={"ID":"b790d0d0-0038-4153-8508-b5f974dda04c","Type":"ContainerStarted","Data":"2b998ed567a370390fc27eb766a29f0317efd7587263e94e797e501fca823d98"} Nov 24 13:17:03 crc kubenswrapper[4790]: I1124 13:17:03.422335 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r266z" event={"ID":"92ba1be5-af31-4941-94ef-3977f17b297e","Type":"ContainerStarted","Data":"206f5f81626f132fc0c5124d6275aff9f21848b1d48b5874746f012ad6c8ad91"} Nov 24 13:17:03 crc kubenswrapper[4790]: I1124 13:17:03.443901 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-r266z" podStartSLOduration=2.041649816 podStartE2EDuration="3.443859632s" podCreationTimestamp="2025-11-24 13:17:00 +0000 UTC" firstStartedPulling="2025-11-24 13:17:01.394598646 +0000 UTC m=+269.774492308" lastFinishedPulling="2025-11-24 13:17:02.796808472 +0000 UTC m=+271.176702124" observedRunningTime="2025-11-24 13:17:03.443717807 +0000 UTC m=+271.823611469" watchObservedRunningTime="2025-11-24 13:17:03.443859632 +0000 UTC m=+271.823753304" Nov 24 13:17:03 crc kubenswrapper[4790]: I1124 13:17:03.486297 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pn2zj"] Nov 24 13:17:03 crc kubenswrapper[4790]: W1124 13:17:03.492220 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod730d2a27_6ef4_4a97_8e6b_5272e26aa247.slice/crio-93faa936195d38d35036222f7f9c1107654d55dc421d50c7af36ff41d35fea26 WatchSource:0}: Error finding container 93faa936195d38d35036222f7f9c1107654d55dc421d50c7af36ff41d35fea26: Status 404 returned error can't find the container with id 93faa936195d38d35036222f7f9c1107654d55dc421d50c7af36ff41d35fea26 Nov 24 13:17:04 crc kubenswrapper[4790]: I1124 13:17:04.429046 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j69qv" event={"ID":"324e1511-0e35-441f-a819-6faf957fe189","Type":"ContainerStarted","Data":"e208ec35513fd0ab410faec2c705e25f136f5d5626eddaf8636794dd56ddef2b"} Nov 24 13:17:04 crc kubenswrapper[4790]: I1124 13:17:04.430987 4790 generic.go:334] "Generic (PLEG): container finished" podID="730d2a27-6ef4-4a97-8e6b-5272e26aa247" containerID="4acc542cd92b0a895489568ad066e7ad98b6159a6423d2738df93d81adce0f80" exitCode=0 Nov 24 13:17:04 crc kubenswrapper[4790]: I1124 13:17:04.431018 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pn2zj" event={"ID":"730d2a27-6ef4-4a97-8e6b-5272e26aa247","Type":"ContainerDied","Data":"4acc542cd92b0a895489568ad066e7ad98b6159a6423d2738df93d81adce0f80"} Nov 24 13:17:04 crc kubenswrapper[4790]: I1124 13:17:04.431052 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pn2zj" event={"ID":"730d2a27-6ef4-4a97-8e6b-5272e26aa247","Type":"ContainerStarted","Data":"93faa936195d38d35036222f7f9c1107654d55dc421d50c7af36ff41d35fea26"} Nov 24 13:17:04 crc kubenswrapper[4790]: I1124 13:17:04.437086 4790 generic.go:334] "Generic (PLEG): container finished" podID="b790d0d0-0038-4153-8508-b5f974dda04c" containerID="8c635a22d6853ab2b43811d154d38b26922332371ba5f476e3809cdba17fd607" exitCode=0 Nov 24 13:17:04 crc kubenswrapper[4790]: I1124 13:17:04.437150 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r2j97" event={"ID":"b790d0d0-0038-4153-8508-b5f974dda04c","Type":"ContainerDied","Data":"8c635a22d6853ab2b43811d154d38b26922332371ba5f476e3809cdba17fd607"} Nov 24 13:17:04 crc kubenswrapper[4790]: I1124 13:17:04.437184 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r2j97" event={"ID":"b790d0d0-0038-4153-8508-b5f974dda04c","Type":"ContainerStarted","Data":"671625f2690507a78224d63267e12dd69a2e222abb0770bd6a718f7a7ff7d773"} Nov 24 13:17:04 crc kubenswrapper[4790]: I1124 13:17:04.448765 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-j69qv" podStartSLOduration=2.009091101 podStartE2EDuration="4.448748613s" podCreationTimestamp="2025-11-24 13:17:00 +0000 UTC" firstStartedPulling="2025-11-24 13:17:01.3924353 +0000 UTC m=+269.772328962" lastFinishedPulling="2025-11-24 13:17:03.832092812 +0000 UTC m=+272.211986474" observedRunningTime="2025-11-24 13:17:04.446317638 +0000 UTC m=+272.826211300" watchObservedRunningTime="2025-11-24 13:17:04.448748613 +0000 UTC m=+272.828642275" Nov 24 13:17:05 crc kubenswrapper[4790]: I1124 13:17:05.446237 4790 generic.go:334] "Generic (PLEG): container finished" podID="b790d0d0-0038-4153-8508-b5f974dda04c" containerID="671625f2690507a78224d63267e12dd69a2e222abb0770bd6a718f7a7ff7d773" exitCode=0 Nov 24 13:17:05 crc kubenswrapper[4790]: I1124 13:17:05.447699 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r2j97" event={"ID":"b790d0d0-0038-4153-8508-b5f974dda04c","Type":"ContainerDied","Data":"671625f2690507a78224d63267e12dd69a2e222abb0770bd6a718f7a7ff7d773"} Nov 24 13:17:06 crc kubenswrapper[4790]: I1124 13:17:06.453150 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r2j97" event={"ID":"b790d0d0-0038-4153-8508-b5f974dda04c","Type":"ContainerStarted","Data":"1009438344b82ad72e3d2bc97fd6bf27af2d1b40ec336f7a8a508c0e939c478e"} Nov 24 13:17:06 crc kubenswrapper[4790]: I1124 13:17:06.484515 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-r2j97" podStartSLOduration=2.031777165 podStartE2EDuration="4.484499086s" podCreationTimestamp="2025-11-24 13:17:02 +0000 UTC" firstStartedPulling="2025-11-24 13:17:03.420552734 +0000 UTC m=+271.800446396" lastFinishedPulling="2025-11-24 13:17:05.873274655 +0000 UTC m=+274.253168317" observedRunningTime="2025-11-24 13:17:06.481240695 +0000 UTC m=+274.861134357" watchObservedRunningTime="2025-11-24 13:17:06.484499086 +0000 UTC m=+274.864392748" Nov 24 13:17:08 crc kubenswrapper[4790]: I1124 13:17:08.465050 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pn2zj" event={"ID":"730d2a27-6ef4-4a97-8e6b-5272e26aa247","Type":"ContainerStarted","Data":"67b32023432d3f56c81185e30864fa6f05fe6e07d006872288ce88ee7913fb20"} Nov 24 13:17:09 crc kubenswrapper[4790]: I1124 13:17:09.471045 4790 generic.go:334] "Generic (PLEG): container finished" podID="730d2a27-6ef4-4a97-8e6b-5272e26aa247" containerID="67b32023432d3f56c81185e30864fa6f05fe6e07d006872288ce88ee7913fb20" exitCode=0 Nov 24 13:17:09 crc kubenswrapper[4790]: I1124 13:17:09.471106 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pn2zj" event={"ID":"730d2a27-6ef4-4a97-8e6b-5272e26aa247","Type":"ContainerDied","Data":"67b32023432d3f56c81185e30864fa6f05fe6e07d006872288ce88ee7913fb20"} Nov 24 13:17:10 crc kubenswrapper[4790]: I1124 13:17:10.469187 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-r266z" Nov 24 13:17:10 crc kubenswrapper[4790]: I1124 13:17:10.469554 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-r266z" Nov 24 13:17:10 crc kubenswrapper[4790]: I1124 13:17:10.478959 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pn2zj" event={"ID":"730d2a27-6ef4-4a97-8e6b-5272e26aa247","Type":"ContainerStarted","Data":"d39e79890018b3cf73d68d6bed504a60b38b6539c4d31478378a16e563642b19"} Nov 24 13:17:10 crc kubenswrapper[4790]: I1124 13:17:10.499575 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pn2zj" podStartSLOduration=2.973801775 podStartE2EDuration="8.499548991s" podCreationTimestamp="2025-11-24 13:17:02 +0000 UTC" firstStartedPulling="2025-11-24 13:17:04.432927489 +0000 UTC m=+272.812821151" lastFinishedPulling="2025-11-24 13:17:09.958674705 +0000 UTC m=+278.338568367" observedRunningTime="2025-11-24 13:17:10.496289669 +0000 UTC m=+278.876183361" watchObservedRunningTime="2025-11-24 13:17:10.499548991 +0000 UTC m=+278.879442653" Nov 24 13:17:10 crc kubenswrapper[4790]: I1124 13:17:10.518069 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-r266z" Nov 24 13:17:10 crc kubenswrapper[4790]: I1124 13:17:10.607178 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-r266z" Nov 24 13:17:10 crc kubenswrapper[4790]: I1124 13:17:10.669732 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-j69qv" Nov 24 13:17:10 crc kubenswrapper[4790]: I1124 13:17:10.669797 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-j69qv" Nov 24 13:17:10 crc kubenswrapper[4790]: I1124 13:17:10.709058 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-j69qv" Nov 24 13:17:11 crc kubenswrapper[4790]: I1124 13:17:11.519831 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-j69qv" Nov 24 13:17:12 crc kubenswrapper[4790]: I1124 13:17:12.868777 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-r2j97" Nov 24 13:17:12 crc kubenswrapper[4790]: I1124 13:17:12.869185 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-r2j97" Nov 24 13:17:12 crc kubenswrapper[4790]: I1124 13:17:12.927329 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-r2j97" Nov 24 13:17:13 crc kubenswrapper[4790]: I1124 13:17:13.110359 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pn2zj" Nov 24 13:17:13 crc kubenswrapper[4790]: I1124 13:17:13.110798 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pn2zj" Nov 24 13:17:13 crc kubenswrapper[4790]: I1124 13:17:13.146672 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pn2zj" Nov 24 13:17:13 crc kubenswrapper[4790]: I1124 13:17:13.529654 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-r2j97" Nov 24 13:17:23 crc kubenswrapper[4790]: I1124 13:17:23.158542 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pn2zj" Nov 24 13:18:43 crc kubenswrapper[4790]: I1124 13:18:43.939258 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:18:43 crc kubenswrapper[4790]: I1124 13:18:43.939852 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:19:13 crc kubenswrapper[4790]: I1124 13:19:13.938708 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:19:13 crc kubenswrapper[4790]: I1124 13:19:13.939252 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:19:43 crc kubenswrapper[4790]: I1124 13:19:43.938794 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:19:43 crc kubenswrapper[4790]: I1124 13:19:43.939925 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:19:43 crc kubenswrapper[4790]: I1124 13:19:43.940011 4790 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 13:19:43 crc kubenswrapper[4790]: I1124 13:19:43.941290 4790 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"abd5645bbc014c541440d33742c811f564b9e51e2066d093c2f1fef3b356b89b"} pod="openshift-machine-config-operator/machine-config-daemon-xz49t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 13:19:43 crc kubenswrapper[4790]: I1124 13:19:43.941553 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" containerID="cri-o://abd5645bbc014c541440d33742c811f564b9e51e2066d093c2f1fef3b356b89b" gracePeriod=600 Nov 24 13:19:44 crc kubenswrapper[4790]: I1124 13:19:44.389880 4790 generic.go:334] "Generic (PLEG): container finished" podID="0d73b133-48f0-455f-8f6a-742e633f631a" containerID="abd5645bbc014c541440d33742c811f564b9e51e2066d093c2f1fef3b356b89b" exitCode=0 Nov 24 13:19:44 crc kubenswrapper[4790]: I1124 13:19:44.389998 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerDied","Data":"abd5645bbc014c541440d33742c811f564b9e51e2066d093c2f1fef3b356b89b"} Nov 24 13:19:44 crc kubenswrapper[4790]: I1124 13:19:44.390495 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"dcbd627924cf1dc4266a99e99dc330119809da6d3c320cfad690c481fc6af519"} Nov 24 13:19:44 crc kubenswrapper[4790]: I1124 13:19:44.390521 4790 scope.go:117] "RemoveContainer" containerID="b98ae4c01c92169e9f4b33633fc6508020f934e2f11530beb17b6fdcdcb2f989" Nov 24 13:20:36 crc kubenswrapper[4790]: I1124 13:20:36.326593 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-fxw8n"] Nov 24 13:20:36 crc kubenswrapper[4790]: I1124 13:20:36.328102 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-fxw8n" Nov 24 13:20:36 crc kubenswrapper[4790]: I1124 13:20:36.378917 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-fxw8n"] Nov 24 13:20:36 crc kubenswrapper[4790]: I1124 13:20:36.487782 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-fxw8n\" (UID: \"a577183b-a9a0-4d10-8c9f-5c481cd4caba\") " pod="openshift-image-registry/image-registry-66df7c8f76-fxw8n" Nov 24 13:20:36 crc kubenswrapper[4790]: I1124 13:20:36.488053 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a577183b-a9a0-4d10-8c9f-5c481cd4caba-ca-trust-extracted\") pod \"image-registry-66df7c8f76-fxw8n\" (UID: \"a577183b-a9a0-4d10-8c9f-5c481cd4caba\") " pod="openshift-image-registry/image-registry-66df7c8f76-fxw8n" Nov 24 13:20:36 crc kubenswrapper[4790]: I1124 13:20:36.488219 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a577183b-a9a0-4d10-8c9f-5c481cd4caba-registry-certificates\") pod \"image-registry-66df7c8f76-fxw8n\" (UID: \"a577183b-a9a0-4d10-8c9f-5c481cd4caba\") " pod="openshift-image-registry/image-registry-66df7c8f76-fxw8n" Nov 24 13:20:36 crc kubenswrapper[4790]: I1124 13:20:36.488308 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a577183b-a9a0-4d10-8c9f-5c481cd4caba-bound-sa-token\") pod \"image-registry-66df7c8f76-fxw8n\" (UID: \"a577183b-a9a0-4d10-8c9f-5c481cd4caba\") " pod="openshift-image-registry/image-registry-66df7c8f76-fxw8n" Nov 24 13:20:36 crc kubenswrapper[4790]: I1124 13:20:36.488351 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a577183b-a9a0-4d10-8c9f-5c481cd4caba-installation-pull-secrets\") pod \"image-registry-66df7c8f76-fxw8n\" (UID: \"a577183b-a9a0-4d10-8c9f-5c481cd4caba\") " pod="openshift-image-registry/image-registry-66df7c8f76-fxw8n" Nov 24 13:20:36 crc kubenswrapper[4790]: I1124 13:20:36.488486 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a577183b-a9a0-4d10-8c9f-5c481cd4caba-trusted-ca\") pod \"image-registry-66df7c8f76-fxw8n\" (UID: \"a577183b-a9a0-4d10-8c9f-5c481cd4caba\") " pod="openshift-image-registry/image-registry-66df7c8f76-fxw8n" Nov 24 13:20:36 crc kubenswrapper[4790]: I1124 13:20:36.488655 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a577183b-a9a0-4d10-8c9f-5c481cd4caba-registry-tls\") pod \"image-registry-66df7c8f76-fxw8n\" (UID: \"a577183b-a9a0-4d10-8c9f-5c481cd4caba\") " pod="openshift-image-registry/image-registry-66df7c8f76-fxw8n" Nov 24 13:20:36 crc kubenswrapper[4790]: I1124 13:20:36.488751 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqnkm\" (UniqueName: \"kubernetes.io/projected/a577183b-a9a0-4d10-8c9f-5c481cd4caba-kube-api-access-tqnkm\") pod \"image-registry-66df7c8f76-fxw8n\" (UID: \"a577183b-a9a0-4d10-8c9f-5c481cd4caba\") " pod="openshift-image-registry/image-registry-66df7c8f76-fxw8n" Nov 24 13:20:36 crc kubenswrapper[4790]: I1124 13:20:36.514834 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-fxw8n\" (UID: \"a577183b-a9a0-4d10-8c9f-5c481cd4caba\") " pod="openshift-image-registry/image-registry-66df7c8f76-fxw8n" Nov 24 13:20:36 crc kubenswrapper[4790]: I1124 13:20:36.591016 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a577183b-a9a0-4d10-8c9f-5c481cd4caba-registry-tls\") pod \"image-registry-66df7c8f76-fxw8n\" (UID: \"a577183b-a9a0-4d10-8c9f-5c481cd4caba\") " pod="openshift-image-registry/image-registry-66df7c8f76-fxw8n" Nov 24 13:20:36 crc kubenswrapper[4790]: I1124 13:20:36.591863 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqnkm\" (UniqueName: \"kubernetes.io/projected/a577183b-a9a0-4d10-8c9f-5c481cd4caba-kube-api-access-tqnkm\") pod \"image-registry-66df7c8f76-fxw8n\" (UID: \"a577183b-a9a0-4d10-8c9f-5c481cd4caba\") " pod="openshift-image-registry/image-registry-66df7c8f76-fxw8n" Nov 24 13:20:36 crc kubenswrapper[4790]: I1124 13:20:36.591933 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a577183b-a9a0-4d10-8c9f-5c481cd4caba-ca-trust-extracted\") pod \"image-registry-66df7c8f76-fxw8n\" (UID: \"a577183b-a9a0-4d10-8c9f-5c481cd4caba\") " pod="openshift-image-registry/image-registry-66df7c8f76-fxw8n" Nov 24 13:20:36 crc kubenswrapper[4790]: I1124 13:20:36.591991 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a577183b-a9a0-4d10-8c9f-5c481cd4caba-registry-certificates\") pod \"image-registry-66df7c8f76-fxw8n\" (UID: \"a577183b-a9a0-4d10-8c9f-5c481cd4caba\") " pod="openshift-image-registry/image-registry-66df7c8f76-fxw8n" Nov 24 13:20:36 crc kubenswrapper[4790]: I1124 13:20:36.592052 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a577183b-a9a0-4d10-8c9f-5c481cd4caba-bound-sa-token\") pod \"image-registry-66df7c8f76-fxw8n\" (UID: \"a577183b-a9a0-4d10-8c9f-5c481cd4caba\") " pod="openshift-image-registry/image-registry-66df7c8f76-fxw8n" Nov 24 13:20:36 crc kubenswrapper[4790]: I1124 13:20:36.592096 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a577183b-a9a0-4d10-8c9f-5c481cd4caba-installation-pull-secrets\") pod \"image-registry-66df7c8f76-fxw8n\" (UID: \"a577183b-a9a0-4d10-8c9f-5c481cd4caba\") " pod="openshift-image-registry/image-registry-66df7c8f76-fxw8n" Nov 24 13:20:36 crc kubenswrapper[4790]: I1124 13:20:36.592129 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a577183b-a9a0-4d10-8c9f-5c481cd4caba-trusted-ca\") pod \"image-registry-66df7c8f76-fxw8n\" (UID: \"a577183b-a9a0-4d10-8c9f-5c481cd4caba\") " pod="openshift-image-registry/image-registry-66df7c8f76-fxw8n" Nov 24 13:20:36 crc kubenswrapper[4790]: I1124 13:20:36.592456 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a577183b-a9a0-4d10-8c9f-5c481cd4caba-ca-trust-extracted\") pod \"image-registry-66df7c8f76-fxw8n\" (UID: \"a577183b-a9a0-4d10-8c9f-5c481cd4caba\") " pod="openshift-image-registry/image-registry-66df7c8f76-fxw8n" Nov 24 13:20:36 crc kubenswrapper[4790]: I1124 13:20:36.594388 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a577183b-a9a0-4d10-8c9f-5c481cd4caba-trusted-ca\") pod \"image-registry-66df7c8f76-fxw8n\" (UID: \"a577183b-a9a0-4d10-8c9f-5c481cd4caba\") " pod="openshift-image-registry/image-registry-66df7c8f76-fxw8n" Nov 24 13:20:36 crc kubenswrapper[4790]: I1124 13:20:36.594776 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a577183b-a9a0-4d10-8c9f-5c481cd4caba-registry-certificates\") pod \"image-registry-66df7c8f76-fxw8n\" (UID: \"a577183b-a9a0-4d10-8c9f-5c481cd4caba\") " pod="openshift-image-registry/image-registry-66df7c8f76-fxw8n" Nov 24 13:20:36 crc kubenswrapper[4790]: I1124 13:20:36.601349 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a577183b-a9a0-4d10-8c9f-5c481cd4caba-installation-pull-secrets\") pod \"image-registry-66df7c8f76-fxw8n\" (UID: \"a577183b-a9a0-4d10-8c9f-5c481cd4caba\") " pod="openshift-image-registry/image-registry-66df7c8f76-fxw8n" Nov 24 13:20:36 crc kubenswrapper[4790]: I1124 13:20:36.603010 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a577183b-a9a0-4d10-8c9f-5c481cd4caba-registry-tls\") pod \"image-registry-66df7c8f76-fxw8n\" (UID: \"a577183b-a9a0-4d10-8c9f-5c481cd4caba\") " pod="openshift-image-registry/image-registry-66df7c8f76-fxw8n" Nov 24 13:20:36 crc kubenswrapper[4790]: I1124 13:20:36.613831 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a577183b-a9a0-4d10-8c9f-5c481cd4caba-bound-sa-token\") pod \"image-registry-66df7c8f76-fxw8n\" (UID: \"a577183b-a9a0-4d10-8c9f-5c481cd4caba\") " pod="openshift-image-registry/image-registry-66df7c8f76-fxw8n" Nov 24 13:20:36 crc kubenswrapper[4790]: I1124 13:20:36.615182 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqnkm\" (UniqueName: \"kubernetes.io/projected/a577183b-a9a0-4d10-8c9f-5c481cd4caba-kube-api-access-tqnkm\") pod \"image-registry-66df7c8f76-fxw8n\" (UID: \"a577183b-a9a0-4d10-8c9f-5c481cd4caba\") " pod="openshift-image-registry/image-registry-66df7c8f76-fxw8n" Nov 24 13:20:36 crc kubenswrapper[4790]: I1124 13:20:36.654876 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-fxw8n" Nov 24 13:20:36 crc kubenswrapper[4790]: I1124 13:20:36.899520 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-fxw8n"] Nov 24 13:20:37 crc kubenswrapper[4790]: I1124 13:20:37.706864 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-fxw8n" event={"ID":"a577183b-a9a0-4d10-8c9f-5c481cd4caba","Type":"ContainerStarted","Data":"55fc78480509a0200809fe274476f0a4ebfe8ad0f8a310cc71d134dd3a46855e"} Nov 24 13:20:37 crc kubenswrapper[4790]: I1124 13:20:37.707392 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-fxw8n" Nov 24 13:20:37 crc kubenswrapper[4790]: I1124 13:20:37.707407 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-fxw8n" event={"ID":"a577183b-a9a0-4d10-8c9f-5c481cd4caba","Type":"ContainerStarted","Data":"492df05f7b1f517737a1d94b0f93a36ac37d0ac2dcaa3a0197ec867cbd3a6ae3"} Nov 24 13:20:37 crc kubenswrapper[4790]: I1124 13:20:37.736182 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-fxw8n" podStartSLOduration=1.736159883 podStartE2EDuration="1.736159883s" podCreationTimestamp="2025-11-24 13:20:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:20:37.734603158 +0000 UTC m=+486.114496830" watchObservedRunningTime="2025-11-24 13:20:37.736159883 +0000 UTC m=+486.116053545" Nov 24 13:20:56 crc kubenswrapper[4790]: I1124 13:20:56.660864 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-fxw8n" Nov 24 13:20:56 crc kubenswrapper[4790]: I1124 13:20:56.729533 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-thd68"] Nov 24 13:21:21 crc kubenswrapper[4790]: I1124 13:21:21.769185 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-thd68" podUID="6381724d-d16b-4d7b-9d9f-613451b93c77" containerName="registry" containerID="cri-o://53d539899e076f66813e25210633c7d3b3031b89c9e9cc5d92881c68595075f9" gracePeriod=30 Nov 24 13:21:21 crc kubenswrapper[4790]: I1124 13:21:21.787216 4790 patch_prober.go:28] interesting pod/image-registry-697d97f7c8-thd68 container/registry namespace/openshift-image-registry: Readiness probe status=failure output="Get \"https://10.217.0.19:5000/healthz\": dial tcp 10.217.0.19:5000: connect: connection refused" start-of-body= Nov 24 13:21:21 crc kubenswrapper[4790]: I1124 13:21:21.787356 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-image-registry/image-registry-697d97f7c8-thd68" podUID="6381724d-d16b-4d7b-9d9f-613451b93c77" containerName="registry" probeResult="failure" output="Get \"https://10.217.0.19:5000/healthz\": dial tcp 10.217.0.19:5000: connect: connection refused" Nov 24 13:21:21 crc kubenswrapper[4790]: I1124 13:21:21.978394 4790 generic.go:334] "Generic (PLEG): container finished" podID="6381724d-d16b-4d7b-9d9f-613451b93c77" containerID="53d539899e076f66813e25210633c7d3b3031b89c9e9cc5d92881c68595075f9" exitCode=0 Nov 24 13:21:21 crc kubenswrapper[4790]: I1124 13:21:21.978462 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-thd68" event={"ID":"6381724d-d16b-4d7b-9d9f-613451b93c77","Type":"ContainerDied","Data":"53d539899e076f66813e25210633c7d3b3031b89c9e9cc5d92881c68595075f9"} Nov 24 13:21:22 crc kubenswrapper[4790]: I1124 13:21:22.133769 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:21:22 crc kubenswrapper[4790]: I1124 13:21:22.257917 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6381724d-d16b-4d7b-9d9f-613451b93c77-installation-pull-secrets\") pod \"6381724d-d16b-4d7b-9d9f-613451b93c77\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " Nov 24 13:21:22 crc kubenswrapper[4790]: I1124 13:21:22.257982 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6381724d-d16b-4d7b-9d9f-613451b93c77-trusted-ca\") pod \"6381724d-d16b-4d7b-9d9f-613451b93c77\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " Nov 24 13:21:22 crc kubenswrapper[4790]: I1124 13:21:22.258043 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6381724d-d16b-4d7b-9d9f-613451b93c77-registry-tls\") pod \"6381724d-d16b-4d7b-9d9f-613451b93c77\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " Nov 24 13:21:22 crc kubenswrapper[4790]: I1124 13:21:22.258090 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6381724d-d16b-4d7b-9d9f-613451b93c77-bound-sa-token\") pod \"6381724d-d16b-4d7b-9d9f-613451b93c77\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " Nov 24 13:21:22 crc kubenswrapper[4790]: I1124 13:21:22.258143 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6381724d-d16b-4d7b-9d9f-613451b93c77-registry-certificates\") pod \"6381724d-d16b-4d7b-9d9f-613451b93c77\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " Nov 24 13:21:22 crc kubenswrapper[4790]: I1124 13:21:22.258208 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9zzk\" (UniqueName: \"kubernetes.io/projected/6381724d-d16b-4d7b-9d9f-613451b93c77-kube-api-access-t9zzk\") pod \"6381724d-d16b-4d7b-9d9f-613451b93c77\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " Nov 24 13:21:22 crc kubenswrapper[4790]: I1124 13:21:22.258230 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6381724d-d16b-4d7b-9d9f-613451b93c77-ca-trust-extracted\") pod \"6381724d-d16b-4d7b-9d9f-613451b93c77\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " Nov 24 13:21:22 crc kubenswrapper[4790]: I1124 13:21:22.258427 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"6381724d-d16b-4d7b-9d9f-613451b93c77\" (UID: \"6381724d-d16b-4d7b-9d9f-613451b93c77\") " Nov 24 13:21:22 crc kubenswrapper[4790]: I1124 13:21:22.259400 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6381724d-d16b-4d7b-9d9f-613451b93c77-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "6381724d-d16b-4d7b-9d9f-613451b93c77" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:21:22 crc kubenswrapper[4790]: I1124 13:21:22.259440 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6381724d-d16b-4d7b-9d9f-613451b93c77-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "6381724d-d16b-4d7b-9d9f-613451b93c77" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:21:22 crc kubenswrapper[4790]: I1124 13:21:22.265215 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6381724d-d16b-4d7b-9d9f-613451b93c77-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "6381724d-d16b-4d7b-9d9f-613451b93c77" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:21:22 crc kubenswrapper[4790]: I1124 13:21:22.265401 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6381724d-d16b-4d7b-9d9f-613451b93c77-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "6381724d-d16b-4d7b-9d9f-613451b93c77" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:21:22 crc kubenswrapper[4790]: I1124 13:21:22.266227 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6381724d-d16b-4d7b-9d9f-613451b93c77-kube-api-access-t9zzk" (OuterVolumeSpecName: "kube-api-access-t9zzk") pod "6381724d-d16b-4d7b-9d9f-613451b93c77" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77"). InnerVolumeSpecName "kube-api-access-t9zzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:21:22 crc kubenswrapper[4790]: I1124 13:21:22.268494 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6381724d-d16b-4d7b-9d9f-613451b93c77-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "6381724d-d16b-4d7b-9d9f-613451b93c77" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:21:22 crc kubenswrapper[4790]: I1124 13:21:22.277907 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "6381724d-d16b-4d7b-9d9f-613451b93c77" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 13:21:22 crc kubenswrapper[4790]: I1124 13:21:22.284358 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6381724d-d16b-4d7b-9d9f-613451b93c77-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "6381724d-d16b-4d7b-9d9f-613451b93c77" (UID: "6381724d-d16b-4d7b-9d9f-613451b93c77"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:21:22 crc kubenswrapper[4790]: I1124 13:21:22.361155 4790 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6381724d-d16b-4d7b-9d9f-613451b93c77-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:21:22 crc kubenswrapper[4790]: I1124 13:21:22.361204 4790 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6381724d-d16b-4d7b-9d9f-613451b93c77-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 24 13:21:22 crc kubenswrapper[4790]: I1124 13:21:22.361219 4790 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6381724d-d16b-4d7b-9d9f-613451b93c77-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:21:22 crc kubenswrapper[4790]: I1124 13:21:22.361230 4790 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6381724d-d16b-4d7b-9d9f-613451b93c77-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 13:21:22 crc kubenswrapper[4790]: I1124 13:21:22.361244 4790 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6381724d-d16b-4d7b-9d9f-613451b93c77-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 24 13:21:22 crc kubenswrapper[4790]: I1124 13:21:22.361252 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9zzk\" (UniqueName: \"kubernetes.io/projected/6381724d-d16b-4d7b-9d9f-613451b93c77-kube-api-access-t9zzk\") on node \"crc\" DevicePath \"\"" Nov 24 13:21:22 crc kubenswrapper[4790]: I1124 13:21:22.361263 4790 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6381724d-d16b-4d7b-9d9f-613451b93c77-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 24 13:21:22 crc kubenswrapper[4790]: I1124 13:21:22.986837 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-thd68" event={"ID":"6381724d-d16b-4d7b-9d9f-613451b93c77","Type":"ContainerDied","Data":"b3f06e4ad38a76ed234e9f3fb96db72680ff8656cb1b928307a804e0fc9a8fe3"} Nov 24 13:21:22 crc kubenswrapper[4790]: I1124 13:21:22.986957 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-thd68" Nov 24 13:21:22 crc kubenswrapper[4790]: I1124 13:21:22.987773 4790 scope.go:117] "RemoveContainer" containerID="53d539899e076f66813e25210633c7d3b3031b89c9e9cc5d92881c68595075f9" Nov 24 13:21:23 crc kubenswrapper[4790]: I1124 13:21:23.016537 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-thd68"] Nov 24 13:21:23 crc kubenswrapper[4790]: I1124 13:21:23.021524 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-thd68"] Nov 24 13:21:24 crc kubenswrapper[4790]: I1124 13:21:24.325008 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6381724d-d16b-4d7b-9d9f-613451b93c77" path="/var/lib/kubelet/pods/6381724d-d16b-4d7b-9d9f-613451b93c77/volumes" Nov 24 13:21:37 crc kubenswrapper[4790]: I1124 13:21:37.577055 4790 scope.go:117] "RemoveContainer" containerID="2c08f3df7115b36a938fa7e6de536a3abb3482f5ed97fb396fd826fbdd252083" Nov 24 13:21:37 crc kubenswrapper[4790]: I1124 13:21:37.604214 4790 scope.go:117] "RemoveContainer" containerID="eabdb0f6a84865cf987b267782d98e2708385c7918d7c2b55bdc40888793a9b0" Nov 24 13:21:37 crc kubenswrapper[4790]: I1124 13:21:37.668169 4790 scope.go:117] "RemoveContainer" containerID="4ba17f52f39ed2f3f39e53f8da98fef69c050d9079a9123eb0da0849c194c5f2" Nov 24 13:21:37 crc kubenswrapper[4790]: I1124 13:21:37.692599 4790 scope.go:117] "RemoveContainer" containerID="382422e39267f8e3cffa06519a5c49d19a2912c3172f2013f441180e59b7dcac" Nov 24 13:22:13 crc kubenswrapper[4790]: I1124 13:22:13.938704 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:22:13 crc kubenswrapper[4790]: I1124 13:22:13.939711 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:22:37 crc kubenswrapper[4790]: I1124 13:22:37.757322 4790 scope.go:117] "RemoveContainer" containerID="122f3825d2a17abfbe00cc7de53c51aeb11b73ec6ae8cbef2dfdb8244b10e316" Nov 24 13:22:37 crc kubenswrapper[4790]: I1124 13:22:37.784818 4790 scope.go:117] "RemoveContainer" containerID="b7369082161a056e6fd32844db72c40163c461bde65d6a74f23396550ff2430a" Nov 24 13:22:37 crc kubenswrapper[4790]: I1124 13:22:37.809279 4790 scope.go:117] "RemoveContainer" containerID="2d6e78ba26ee897ac86ff20e363a1134a5b1e4751da5763bf285a97fd7eecdf3" Nov 24 13:22:37 crc kubenswrapper[4790]: I1124 13:22:37.825160 4790 scope.go:117] "RemoveContainer" containerID="96435c3c2d12f19bca126da859c7d48765a73390f48fcd58ff1a8674413f2525" Nov 24 13:22:37 crc kubenswrapper[4790]: I1124 13:22:37.839843 4790 scope.go:117] "RemoveContainer" containerID="b07f5e8a194a7916b770b8c37c36831bead0572f4ffe24e42040bb4b3f4b69ac" Nov 24 13:22:37 crc kubenswrapper[4790]: I1124 13:22:37.853477 4790 scope.go:117] "RemoveContainer" containerID="79ad0251c9cc2fb0b4256b38f4688d937967244a73dcd1ed43737c96fa6773fa" Nov 24 13:22:43 crc kubenswrapper[4790]: I1124 13:22:43.938945 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:22:43 crc kubenswrapper[4790]: I1124 13:22:43.939540 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:23:13 crc kubenswrapper[4790]: I1124 13:23:13.938814 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:23:13 crc kubenswrapper[4790]: I1124 13:23:13.939788 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:23:13 crc kubenswrapper[4790]: I1124 13:23:13.939895 4790 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 13:23:13 crc kubenswrapper[4790]: I1124 13:23:13.941118 4790 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dcbd627924cf1dc4266a99e99dc330119809da6d3c320cfad690c481fc6af519"} pod="openshift-machine-config-operator/machine-config-daemon-xz49t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 13:23:13 crc kubenswrapper[4790]: I1124 13:23:13.941194 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" containerID="cri-o://dcbd627924cf1dc4266a99e99dc330119809da6d3c320cfad690c481fc6af519" gracePeriod=600 Nov 24 13:23:14 crc kubenswrapper[4790]: I1124 13:23:14.728771 4790 generic.go:334] "Generic (PLEG): container finished" podID="0d73b133-48f0-455f-8f6a-742e633f631a" containerID="dcbd627924cf1dc4266a99e99dc330119809da6d3c320cfad690c481fc6af519" exitCode=0 Nov 24 13:23:14 crc kubenswrapper[4790]: I1124 13:23:14.728854 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerDied","Data":"dcbd627924cf1dc4266a99e99dc330119809da6d3c320cfad690c481fc6af519"} Nov 24 13:23:14 crc kubenswrapper[4790]: I1124 13:23:14.729306 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"039103a6742b65a25f073d9f23e4fb704fe4fd769ff785303a1ff06bb7cfe2d2"} Nov 24 13:23:14 crc kubenswrapper[4790]: I1124 13:23:14.729339 4790 scope.go:117] "RemoveContainer" containerID="abd5645bbc014c541440d33742c811f564b9e51e2066d093c2f1fef3b356b89b" Nov 24 13:24:55 crc kubenswrapper[4790]: I1124 13:24:55.459395 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-cjdhk"] Nov 24 13:24:55 crc kubenswrapper[4790]: I1124 13:24:55.460773 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-cjdhk" podUID="69940b0d-4afc-4a7f-8ff7-cb128ce58913" containerName="controller-manager" containerID="cri-o://675c34e9a7963fc1e18def04b460ba4e931cbb81bbc230859d8cce0fae5bfe06" gracePeriod=30 Nov 24 13:24:55 crc kubenswrapper[4790]: I1124 13:24:55.557494 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wnlxq"] Nov 24 13:24:55 crc kubenswrapper[4790]: I1124 13:24:55.558107 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wnlxq" podUID="f50b116f-b25b-4985-8975-fbb523e1ddf6" containerName="route-controller-manager" containerID="cri-o://223274807b4aa4b9eb85d9dbb4d301afa8de029c7481e07a66699b8e1ef76c0f" gracePeriod=30 Nov 24 13:24:55 crc kubenswrapper[4790]: I1124 13:24:55.893495 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-cjdhk" Nov 24 13:24:55 crc kubenswrapper[4790]: I1124 13:24:55.989277 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wnlxq" Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.095475 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f50b116f-b25b-4985-8975-fbb523e1ddf6-client-ca\") pod \"f50b116f-b25b-4985-8975-fbb523e1ddf6\" (UID: \"f50b116f-b25b-4985-8975-fbb523e1ddf6\") " Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.095597 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f50b116f-b25b-4985-8975-fbb523e1ddf6-serving-cert\") pod \"f50b116f-b25b-4985-8975-fbb523e1ddf6\" (UID: \"f50b116f-b25b-4985-8975-fbb523e1ddf6\") " Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.095648 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/69940b0d-4afc-4a7f-8ff7-cb128ce58913-proxy-ca-bundles\") pod \"69940b0d-4afc-4a7f-8ff7-cb128ce58913\" (UID: \"69940b0d-4afc-4a7f-8ff7-cb128ce58913\") " Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.095686 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f50b116f-b25b-4985-8975-fbb523e1ddf6-config\") pod \"f50b116f-b25b-4985-8975-fbb523e1ddf6\" (UID: \"f50b116f-b25b-4985-8975-fbb523e1ddf6\") " Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.095723 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzgqr\" (UniqueName: \"kubernetes.io/projected/f50b116f-b25b-4985-8975-fbb523e1ddf6-kube-api-access-dzgqr\") pod \"f50b116f-b25b-4985-8975-fbb523e1ddf6\" (UID: \"f50b116f-b25b-4985-8975-fbb523e1ddf6\") " Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.095779 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69940b0d-4afc-4a7f-8ff7-cb128ce58913-config\") pod \"69940b0d-4afc-4a7f-8ff7-cb128ce58913\" (UID: \"69940b0d-4afc-4a7f-8ff7-cb128ce58913\") " Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.095808 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/69940b0d-4afc-4a7f-8ff7-cb128ce58913-serving-cert\") pod \"69940b0d-4afc-4a7f-8ff7-cb128ce58913\" (UID: \"69940b0d-4afc-4a7f-8ff7-cb128ce58913\") " Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.095870 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/69940b0d-4afc-4a7f-8ff7-cb128ce58913-client-ca\") pod \"69940b0d-4afc-4a7f-8ff7-cb128ce58913\" (UID: \"69940b0d-4afc-4a7f-8ff7-cb128ce58913\") " Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.095951 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2x49b\" (UniqueName: \"kubernetes.io/projected/69940b0d-4afc-4a7f-8ff7-cb128ce58913-kube-api-access-2x49b\") pod \"69940b0d-4afc-4a7f-8ff7-cb128ce58913\" (UID: \"69940b0d-4afc-4a7f-8ff7-cb128ce58913\") " Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.097099 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69940b0d-4afc-4a7f-8ff7-cb128ce58913-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "69940b0d-4afc-4a7f-8ff7-cb128ce58913" (UID: "69940b0d-4afc-4a7f-8ff7-cb128ce58913"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.097118 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f50b116f-b25b-4985-8975-fbb523e1ddf6-config" (OuterVolumeSpecName: "config") pod "f50b116f-b25b-4985-8975-fbb523e1ddf6" (UID: "f50b116f-b25b-4985-8975-fbb523e1ddf6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.097252 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69940b0d-4afc-4a7f-8ff7-cb128ce58913-config" (OuterVolumeSpecName: "config") pod "69940b0d-4afc-4a7f-8ff7-cb128ce58913" (UID: "69940b0d-4afc-4a7f-8ff7-cb128ce58913"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.097918 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69940b0d-4afc-4a7f-8ff7-cb128ce58913-client-ca" (OuterVolumeSpecName: "client-ca") pod "69940b0d-4afc-4a7f-8ff7-cb128ce58913" (UID: "69940b0d-4afc-4a7f-8ff7-cb128ce58913"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.098038 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f50b116f-b25b-4985-8975-fbb523e1ddf6-client-ca" (OuterVolumeSpecName: "client-ca") pod "f50b116f-b25b-4985-8975-fbb523e1ddf6" (UID: "f50b116f-b25b-4985-8975-fbb523e1ddf6"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.104795 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f50b116f-b25b-4985-8975-fbb523e1ddf6-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f50b116f-b25b-4985-8975-fbb523e1ddf6" (UID: "f50b116f-b25b-4985-8975-fbb523e1ddf6"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.105054 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f50b116f-b25b-4985-8975-fbb523e1ddf6-kube-api-access-dzgqr" (OuterVolumeSpecName: "kube-api-access-dzgqr") pod "f50b116f-b25b-4985-8975-fbb523e1ddf6" (UID: "f50b116f-b25b-4985-8975-fbb523e1ddf6"). InnerVolumeSpecName "kube-api-access-dzgqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.105039 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69940b0d-4afc-4a7f-8ff7-cb128ce58913-kube-api-access-2x49b" (OuterVolumeSpecName: "kube-api-access-2x49b") pod "69940b0d-4afc-4a7f-8ff7-cb128ce58913" (UID: "69940b0d-4afc-4a7f-8ff7-cb128ce58913"). InnerVolumeSpecName "kube-api-access-2x49b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.105224 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69940b0d-4afc-4a7f-8ff7-cb128ce58913-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "69940b0d-4afc-4a7f-8ff7-cb128ce58913" (UID: "69940b0d-4afc-4a7f-8ff7-cb128ce58913"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.197416 4790 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/69940b0d-4afc-4a7f-8ff7-cb128ce58913-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.197469 4790 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/69940b0d-4afc-4a7f-8ff7-cb128ce58913-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.197481 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2x49b\" (UniqueName: \"kubernetes.io/projected/69940b0d-4afc-4a7f-8ff7-cb128ce58913-kube-api-access-2x49b\") on node \"crc\" DevicePath \"\"" Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.197495 4790 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f50b116f-b25b-4985-8975-fbb523e1ddf6-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.197509 4790 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f50b116f-b25b-4985-8975-fbb523e1ddf6-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.197518 4790 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/69940b0d-4afc-4a7f-8ff7-cb128ce58913-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.197532 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f50b116f-b25b-4985-8975-fbb523e1ddf6-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.197545 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzgqr\" (UniqueName: \"kubernetes.io/projected/f50b116f-b25b-4985-8975-fbb523e1ddf6-kube-api-access-dzgqr\") on node \"crc\" DevicePath \"\"" Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.197560 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69940b0d-4afc-4a7f-8ff7-cb128ce58913-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.378161 4790 generic.go:334] "Generic (PLEG): container finished" podID="69940b0d-4afc-4a7f-8ff7-cb128ce58913" containerID="675c34e9a7963fc1e18def04b460ba4e931cbb81bbc230859d8cce0fae5bfe06" exitCode=0 Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.378255 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-cjdhk" Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.378254 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-cjdhk" event={"ID":"69940b0d-4afc-4a7f-8ff7-cb128ce58913","Type":"ContainerDied","Data":"675c34e9a7963fc1e18def04b460ba4e931cbb81bbc230859d8cce0fae5bfe06"} Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.378858 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-cjdhk" event={"ID":"69940b0d-4afc-4a7f-8ff7-cb128ce58913","Type":"ContainerDied","Data":"960cd279ac7a1f420dd9ab80bdc74bdffb985ef736d6e01607277f0782682069"} Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.378903 4790 scope.go:117] "RemoveContainer" containerID="675c34e9a7963fc1e18def04b460ba4e931cbb81bbc230859d8cce0fae5bfe06" Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.380911 4790 generic.go:334] "Generic (PLEG): container finished" podID="f50b116f-b25b-4985-8975-fbb523e1ddf6" containerID="223274807b4aa4b9eb85d9dbb4d301afa8de029c7481e07a66699b8e1ef76c0f" exitCode=0 Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.380973 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wnlxq" event={"ID":"f50b116f-b25b-4985-8975-fbb523e1ddf6","Type":"ContainerDied","Data":"223274807b4aa4b9eb85d9dbb4d301afa8de029c7481e07a66699b8e1ef76c0f"} Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.381025 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wnlxq" event={"ID":"f50b116f-b25b-4985-8975-fbb523e1ddf6","Type":"ContainerDied","Data":"7f9d98c4a9b7868ebb84c9e55fd33a58756669f8116a53c60e50931c32ab45b5"} Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.381069 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wnlxq" Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.399454 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-cjdhk"] Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.403396 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-cjdhk"] Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.414541 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wnlxq"] Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.415373 4790 scope.go:117] "RemoveContainer" containerID="675c34e9a7963fc1e18def04b460ba4e931cbb81bbc230859d8cce0fae5bfe06" Nov 24 13:24:56 crc kubenswrapper[4790]: E1124 13:24:56.416025 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"675c34e9a7963fc1e18def04b460ba4e931cbb81bbc230859d8cce0fae5bfe06\": container with ID starting with 675c34e9a7963fc1e18def04b460ba4e931cbb81bbc230859d8cce0fae5bfe06 not found: ID does not exist" containerID="675c34e9a7963fc1e18def04b460ba4e931cbb81bbc230859d8cce0fae5bfe06" Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.416150 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"675c34e9a7963fc1e18def04b460ba4e931cbb81bbc230859d8cce0fae5bfe06"} err="failed to get container status \"675c34e9a7963fc1e18def04b460ba4e931cbb81bbc230859d8cce0fae5bfe06\": rpc error: code = NotFound desc = could not find container \"675c34e9a7963fc1e18def04b460ba4e931cbb81bbc230859d8cce0fae5bfe06\": container with ID starting with 675c34e9a7963fc1e18def04b460ba4e931cbb81bbc230859d8cce0fae5bfe06 not found: ID does not exist" Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.416251 4790 scope.go:117] "RemoveContainer" containerID="223274807b4aa4b9eb85d9dbb4d301afa8de029c7481e07a66699b8e1ef76c0f" Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.417754 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wnlxq"] Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.430184 4790 scope.go:117] "RemoveContainer" containerID="223274807b4aa4b9eb85d9dbb4d301afa8de029c7481e07a66699b8e1ef76c0f" Nov 24 13:24:56 crc kubenswrapper[4790]: E1124 13:24:56.430640 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"223274807b4aa4b9eb85d9dbb4d301afa8de029c7481e07a66699b8e1ef76c0f\": container with ID starting with 223274807b4aa4b9eb85d9dbb4d301afa8de029c7481e07a66699b8e1ef76c0f not found: ID does not exist" containerID="223274807b4aa4b9eb85d9dbb4d301afa8de029c7481e07a66699b8e1ef76c0f" Nov 24 13:24:56 crc kubenswrapper[4790]: I1124 13:24:56.430704 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"223274807b4aa4b9eb85d9dbb4d301afa8de029c7481e07a66699b8e1ef76c0f"} err="failed to get container status \"223274807b4aa4b9eb85d9dbb4d301afa8de029c7481e07a66699b8e1ef76c0f\": rpc error: code = NotFound desc = could not find container \"223274807b4aa4b9eb85d9dbb4d301afa8de029c7481e07a66699b8e1ef76c0f\": container with ID starting with 223274807b4aa4b9eb85d9dbb4d301afa8de029c7481e07a66699b8e1ef76c0f not found: ID does not exist" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.102354 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f5f5f5484-mks6s"] Nov 24 13:24:57 crc kubenswrapper[4790]: E1124 13:24:57.102732 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f50b116f-b25b-4985-8975-fbb523e1ddf6" containerName="route-controller-manager" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.102748 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="f50b116f-b25b-4985-8975-fbb523e1ddf6" containerName="route-controller-manager" Nov 24 13:24:57 crc kubenswrapper[4790]: E1124 13:24:57.102768 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6381724d-d16b-4d7b-9d9f-613451b93c77" containerName="registry" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.102774 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="6381724d-d16b-4d7b-9d9f-613451b93c77" containerName="registry" Nov 24 13:24:57 crc kubenswrapper[4790]: E1124 13:24:57.102784 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69940b0d-4afc-4a7f-8ff7-cb128ce58913" containerName="controller-manager" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.102792 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="69940b0d-4afc-4a7f-8ff7-cb128ce58913" containerName="controller-manager" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.102913 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="f50b116f-b25b-4985-8975-fbb523e1ddf6" containerName="route-controller-manager" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.102927 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="6381724d-d16b-4d7b-9d9f-613451b93c77" containerName="registry" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.102936 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="69940b0d-4afc-4a7f-8ff7-cb128ce58913" containerName="controller-manager" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.103612 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f5f5f5484-mks6s" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.106224 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5f785cc7f5-kvk69"] Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.107436 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f785cc7f5-kvk69" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.107551 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.108011 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.108501 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.108609 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.109404 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.109438 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.110048 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.110195 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.110978 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e9c2b265-fabc-401f-95d1-37a822254f4e-proxy-ca-bundles\") pod \"controller-manager-5f785cc7f5-kvk69\" (UID: \"e9c2b265-fabc-401f-95d1-37a822254f4e\") " pod="openshift-controller-manager/controller-manager-5f785cc7f5-kvk69" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.111022 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjbc5\" (UniqueName: \"kubernetes.io/projected/e9c2b265-fabc-401f-95d1-37a822254f4e-kube-api-access-kjbc5\") pod \"controller-manager-5f785cc7f5-kvk69\" (UID: \"e9c2b265-fabc-401f-95d1-37a822254f4e\") " pod="openshift-controller-manager/controller-manager-5f785cc7f5-kvk69" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.111053 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9c2b265-fabc-401f-95d1-37a822254f4e-serving-cert\") pod \"controller-manager-5f785cc7f5-kvk69\" (UID: \"e9c2b265-fabc-401f-95d1-37a822254f4e\") " pod="openshift-controller-manager/controller-manager-5f785cc7f5-kvk69" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.111070 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wj6sb\" (UniqueName: \"kubernetes.io/projected/52c032e9-3a39-48bb-92f4-0fee5465e58d-kube-api-access-wj6sb\") pod \"route-controller-manager-6f5f5f5484-mks6s\" (UID: \"52c032e9-3a39-48bb-92f4-0fee5465e58d\") " pod="openshift-route-controller-manager/route-controller-manager-6f5f5f5484-mks6s" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.111090 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52c032e9-3a39-48bb-92f4-0fee5465e58d-config\") pod \"route-controller-manager-6f5f5f5484-mks6s\" (UID: \"52c032e9-3a39-48bb-92f4-0fee5465e58d\") " pod="openshift-route-controller-manager/route-controller-manager-6f5f5f5484-mks6s" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.111107 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/52c032e9-3a39-48bb-92f4-0fee5465e58d-client-ca\") pod \"route-controller-manager-6f5f5f5484-mks6s\" (UID: \"52c032e9-3a39-48bb-92f4-0fee5465e58d\") " pod="openshift-route-controller-manager/route-controller-manager-6f5f5f5484-mks6s" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.111125 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9c2b265-fabc-401f-95d1-37a822254f4e-config\") pod \"controller-manager-5f785cc7f5-kvk69\" (UID: \"e9c2b265-fabc-401f-95d1-37a822254f4e\") " pod="openshift-controller-manager/controller-manager-5f785cc7f5-kvk69" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.111163 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e9c2b265-fabc-401f-95d1-37a822254f4e-client-ca\") pod \"controller-manager-5f785cc7f5-kvk69\" (UID: \"e9c2b265-fabc-401f-95d1-37a822254f4e\") " pod="openshift-controller-manager/controller-manager-5f785cc7f5-kvk69" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.111191 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/52c032e9-3a39-48bb-92f4-0fee5465e58d-serving-cert\") pod \"route-controller-manager-6f5f5f5484-mks6s\" (UID: \"52c032e9-3a39-48bb-92f4-0fee5465e58d\") " pod="openshift-route-controller-manager/route-controller-manager-6f5f5f5484-mks6s" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.112746 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.116826 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.117212 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.117421 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.122009 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5f785cc7f5-kvk69"] Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.128626 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f5f5f5484-mks6s"] Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.167105 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.212013 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e9c2b265-fabc-401f-95d1-37a822254f4e-client-ca\") pod \"controller-manager-5f785cc7f5-kvk69\" (UID: \"e9c2b265-fabc-401f-95d1-37a822254f4e\") " pod="openshift-controller-manager/controller-manager-5f785cc7f5-kvk69" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.212081 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/52c032e9-3a39-48bb-92f4-0fee5465e58d-serving-cert\") pod \"route-controller-manager-6f5f5f5484-mks6s\" (UID: \"52c032e9-3a39-48bb-92f4-0fee5465e58d\") " pod="openshift-route-controller-manager/route-controller-manager-6f5f5f5484-mks6s" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.212119 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e9c2b265-fabc-401f-95d1-37a822254f4e-proxy-ca-bundles\") pod \"controller-manager-5f785cc7f5-kvk69\" (UID: \"e9c2b265-fabc-401f-95d1-37a822254f4e\") " pod="openshift-controller-manager/controller-manager-5f785cc7f5-kvk69" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.212144 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjbc5\" (UniqueName: \"kubernetes.io/projected/e9c2b265-fabc-401f-95d1-37a822254f4e-kube-api-access-kjbc5\") pod \"controller-manager-5f785cc7f5-kvk69\" (UID: \"e9c2b265-fabc-401f-95d1-37a822254f4e\") " pod="openshift-controller-manager/controller-manager-5f785cc7f5-kvk69" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.212170 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9c2b265-fabc-401f-95d1-37a822254f4e-serving-cert\") pod \"controller-manager-5f785cc7f5-kvk69\" (UID: \"e9c2b265-fabc-401f-95d1-37a822254f4e\") " pod="openshift-controller-manager/controller-manager-5f785cc7f5-kvk69" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.212186 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wj6sb\" (UniqueName: \"kubernetes.io/projected/52c032e9-3a39-48bb-92f4-0fee5465e58d-kube-api-access-wj6sb\") pod \"route-controller-manager-6f5f5f5484-mks6s\" (UID: \"52c032e9-3a39-48bb-92f4-0fee5465e58d\") " pod="openshift-route-controller-manager/route-controller-manager-6f5f5f5484-mks6s" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.212213 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52c032e9-3a39-48bb-92f4-0fee5465e58d-config\") pod \"route-controller-manager-6f5f5f5484-mks6s\" (UID: \"52c032e9-3a39-48bb-92f4-0fee5465e58d\") " pod="openshift-route-controller-manager/route-controller-manager-6f5f5f5484-mks6s" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.212233 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/52c032e9-3a39-48bb-92f4-0fee5465e58d-client-ca\") pod \"route-controller-manager-6f5f5f5484-mks6s\" (UID: \"52c032e9-3a39-48bb-92f4-0fee5465e58d\") " pod="openshift-route-controller-manager/route-controller-manager-6f5f5f5484-mks6s" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.212249 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9c2b265-fabc-401f-95d1-37a822254f4e-config\") pod \"controller-manager-5f785cc7f5-kvk69\" (UID: \"e9c2b265-fabc-401f-95d1-37a822254f4e\") " pod="openshift-controller-manager/controller-manager-5f785cc7f5-kvk69" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.213609 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e9c2b265-fabc-401f-95d1-37a822254f4e-client-ca\") pod \"controller-manager-5f785cc7f5-kvk69\" (UID: \"e9c2b265-fabc-401f-95d1-37a822254f4e\") " pod="openshift-controller-manager/controller-manager-5f785cc7f5-kvk69" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.213947 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/52c032e9-3a39-48bb-92f4-0fee5465e58d-client-ca\") pod \"route-controller-manager-6f5f5f5484-mks6s\" (UID: \"52c032e9-3a39-48bb-92f4-0fee5465e58d\") " pod="openshift-route-controller-manager/route-controller-manager-6f5f5f5484-mks6s" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.214071 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e9c2b265-fabc-401f-95d1-37a822254f4e-proxy-ca-bundles\") pod \"controller-manager-5f785cc7f5-kvk69\" (UID: \"e9c2b265-fabc-401f-95d1-37a822254f4e\") " pod="openshift-controller-manager/controller-manager-5f785cc7f5-kvk69" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.214177 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52c032e9-3a39-48bb-92f4-0fee5465e58d-config\") pod \"route-controller-manager-6f5f5f5484-mks6s\" (UID: \"52c032e9-3a39-48bb-92f4-0fee5465e58d\") " pod="openshift-route-controller-manager/route-controller-manager-6f5f5f5484-mks6s" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.214846 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9c2b265-fabc-401f-95d1-37a822254f4e-config\") pod \"controller-manager-5f785cc7f5-kvk69\" (UID: \"e9c2b265-fabc-401f-95d1-37a822254f4e\") " pod="openshift-controller-manager/controller-manager-5f785cc7f5-kvk69" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.218436 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9c2b265-fabc-401f-95d1-37a822254f4e-serving-cert\") pod \"controller-manager-5f785cc7f5-kvk69\" (UID: \"e9c2b265-fabc-401f-95d1-37a822254f4e\") " pod="openshift-controller-manager/controller-manager-5f785cc7f5-kvk69" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.225317 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/52c032e9-3a39-48bb-92f4-0fee5465e58d-serving-cert\") pod \"route-controller-manager-6f5f5f5484-mks6s\" (UID: \"52c032e9-3a39-48bb-92f4-0fee5465e58d\") " pod="openshift-route-controller-manager/route-controller-manager-6f5f5f5484-mks6s" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.229580 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjbc5\" (UniqueName: \"kubernetes.io/projected/e9c2b265-fabc-401f-95d1-37a822254f4e-kube-api-access-kjbc5\") pod \"controller-manager-5f785cc7f5-kvk69\" (UID: \"e9c2b265-fabc-401f-95d1-37a822254f4e\") " pod="openshift-controller-manager/controller-manager-5f785cc7f5-kvk69" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.230966 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wj6sb\" (UniqueName: \"kubernetes.io/projected/52c032e9-3a39-48bb-92f4-0fee5465e58d-kube-api-access-wj6sb\") pod \"route-controller-manager-6f5f5f5484-mks6s\" (UID: \"52c032e9-3a39-48bb-92f4-0fee5465e58d\") " pod="openshift-route-controller-manager/route-controller-manager-6f5f5f5484-mks6s" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.457746 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f5f5f5484-mks6s" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.471220 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f785cc7f5-kvk69" Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.687621 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f5f5f5484-mks6s"] Nov 24 13:24:57 crc kubenswrapper[4790]: I1124 13:24:57.743614 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5f785cc7f5-kvk69"] Nov 24 13:24:58 crc kubenswrapper[4790]: I1124 13:24:58.327173 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69940b0d-4afc-4a7f-8ff7-cb128ce58913" path="/var/lib/kubelet/pods/69940b0d-4afc-4a7f-8ff7-cb128ce58913/volumes" Nov 24 13:24:58 crc kubenswrapper[4790]: I1124 13:24:58.328155 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f50b116f-b25b-4985-8975-fbb523e1ddf6" path="/var/lib/kubelet/pods/f50b116f-b25b-4985-8975-fbb523e1ddf6/volumes" Nov 24 13:24:58 crc kubenswrapper[4790]: I1124 13:24:58.400175 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f785cc7f5-kvk69" event={"ID":"e9c2b265-fabc-401f-95d1-37a822254f4e","Type":"ContainerStarted","Data":"5d773dc28a99c9485753ecb41a17c581cbfee3d92e1ceb2711bda970d85df6ce"} Nov 24 13:24:58 crc kubenswrapper[4790]: I1124 13:24:58.400229 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f785cc7f5-kvk69" event={"ID":"e9c2b265-fabc-401f-95d1-37a822254f4e","Type":"ContainerStarted","Data":"647d29e78bfb9ee780f54cd17f917c5238f86fdfe7dae17c6e58b4a640d317d7"} Nov 24 13:24:58 crc kubenswrapper[4790]: I1124 13:24:58.400465 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5f785cc7f5-kvk69" Nov 24 13:24:58 crc kubenswrapper[4790]: I1124 13:24:58.402339 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f5f5f5484-mks6s" event={"ID":"52c032e9-3a39-48bb-92f4-0fee5465e58d","Type":"ContainerStarted","Data":"5d6df7dc12e0412b0c15c34507cb6178836801df9c635164539aaa9810aa8cd8"} Nov 24 13:24:58 crc kubenswrapper[4790]: I1124 13:24:58.402369 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f5f5f5484-mks6s" event={"ID":"52c032e9-3a39-48bb-92f4-0fee5465e58d","Type":"ContainerStarted","Data":"32cc0ecee14a7e1a50fbee851b7f0e3178cc4d5f3c070a1fac9a08bba6c5fe98"} Nov 24 13:24:58 crc kubenswrapper[4790]: I1124 13:24:58.402675 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6f5f5f5484-mks6s" Nov 24 13:24:58 crc kubenswrapper[4790]: I1124 13:24:58.406283 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5f785cc7f5-kvk69" Nov 24 13:24:58 crc kubenswrapper[4790]: I1124 13:24:58.410554 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6f5f5f5484-mks6s" Nov 24 13:24:58 crc kubenswrapper[4790]: I1124 13:24:58.427737 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5f785cc7f5-kvk69" podStartSLOduration=3.427709287 podStartE2EDuration="3.427709287s" podCreationTimestamp="2025-11-24 13:24:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:24:58.425677007 +0000 UTC m=+746.805570689" watchObservedRunningTime="2025-11-24 13:24:58.427709287 +0000 UTC m=+746.807602949" Nov 24 13:24:58 crc kubenswrapper[4790]: I1124 13:24:58.450043 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6f5f5f5484-mks6s" podStartSLOduration=3.450017685 podStartE2EDuration="3.450017685s" podCreationTimestamp="2025-11-24 13:24:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:24:58.448731627 +0000 UTC m=+746.828625299" watchObservedRunningTime="2025-11-24 13:24:58.450017685 +0000 UTC m=+746.829911347" Nov 24 13:25:08 crc kubenswrapper[4790]: I1124 13:25:08.193877 4790 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 13:25:43 crc kubenswrapper[4790]: I1124 13:25:43.939300 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:25:43 crc kubenswrapper[4790]: I1124 13:25:43.939811 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:25:44 crc kubenswrapper[4790]: I1124 13:25:44.724533 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-f6b92"] Nov 24 13:25:44 crc kubenswrapper[4790]: I1124 13:25:44.725897 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f6b92" Nov 24 13:25:44 crc kubenswrapper[4790]: I1124 13:25:44.739498 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f6b92"] Nov 24 13:25:44 crc kubenswrapper[4790]: I1124 13:25:44.854407 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llzzj\" (UniqueName: \"kubernetes.io/projected/9292a43b-dfe0-45a7-8007-434efafabf33-kube-api-access-llzzj\") pod \"community-operators-f6b92\" (UID: \"9292a43b-dfe0-45a7-8007-434efafabf33\") " pod="openshift-marketplace/community-operators-f6b92" Nov 24 13:25:44 crc kubenswrapper[4790]: I1124 13:25:44.854527 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9292a43b-dfe0-45a7-8007-434efafabf33-catalog-content\") pod \"community-operators-f6b92\" (UID: \"9292a43b-dfe0-45a7-8007-434efafabf33\") " pod="openshift-marketplace/community-operators-f6b92" Nov 24 13:25:44 crc kubenswrapper[4790]: I1124 13:25:44.854598 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9292a43b-dfe0-45a7-8007-434efafabf33-utilities\") pod \"community-operators-f6b92\" (UID: \"9292a43b-dfe0-45a7-8007-434efafabf33\") " pod="openshift-marketplace/community-operators-f6b92" Nov 24 13:25:44 crc kubenswrapper[4790]: I1124 13:25:44.956189 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llzzj\" (UniqueName: \"kubernetes.io/projected/9292a43b-dfe0-45a7-8007-434efafabf33-kube-api-access-llzzj\") pod \"community-operators-f6b92\" (UID: \"9292a43b-dfe0-45a7-8007-434efafabf33\") " pod="openshift-marketplace/community-operators-f6b92" Nov 24 13:25:44 crc kubenswrapper[4790]: I1124 13:25:44.956292 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9292a43b-dfe0-45a7-8007-434efafabf33-catalog-content\") pod \"community-operators-f6b92\" (UID: \"9292a43b-dfe0-45a7-8007-434efafabf33\") " pod="openshift-marketplace/community-operators-f6b92" Nov 24 13:25:44 crc kubenswrapper[4790]: I1124 13:25:44.956327 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9292a43b-dfe0-45a7-8007-434efafabf33-utilities\") pod \"community-operators-f6b92\" (UID: \"9292a43b-dfe0-45a7-8007-434efafabf33\") " pod="openshift-marketplace/community-operators-f6b92" Nov 24 13:25:44 crc kubenswrapper[4790]: I1124 13:25:44.957106 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9292a43b-dfe0-45a7-8007-434efafabf33-catalog-content\") pod \"community-operators-f6b92\" (UID: \"9292a43b-dfe0-45a7-8007-434efafabf33\") " pod="openshift-marketplace/community-operators-f6b92" Nov 24 13:25:44 crc kubenswrapper[4790]: I1124 13:25:44.957098 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9292a43b-dfe0-45a7-8007-434efafabf33-utilities\") pod \"community-operators-f6b92\" (UID: \"9292a43b-dfe0-45a7-8007-434efafabf33\") " pod="openshift-marketplace/community-operators-f6b92" Nov 24 13:25:44 crc kubenswrapper[4790]: I1124 13:25:44.982295 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llzzj\" (UniqueName: \"kubernetes.io/projected/9292a43b-dfe0-45a7-8007-434efafabf33-kube-api-access-llzzj\") pod \"community-operators-f6b92\" (UID: \"9292a43b-dfe0-45a7-8007-434efafabf33\") " pod="openshift-marketplace/community-operators-f6b92" Nov 24 13:25:45 crc kubenswrapper[4790]: I1124 13:25:45.168298 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f6b92" Nov 24 13:25:45 crc kubenswrapper[4790]: I1124 13:25:45.496526 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f6b92"] Nov 24 13:25:45 crc kubenswrapper[4790]: I1124 13:25:45.745062 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f6b92" event={"ID":"9292a43b-dfe0-45a7-8007-434efafabf33","Type":"ContainerStarted","Data":"24fe77c1d7c3c0bfbbb7d3049ab5265e42405ef86fd2f97d639cd8ebfb36f69b"} Nov 24 13:25:45 crc kubenswrapper[4790]: I1124 13:25:45.839689 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5x94h"] Nov 24 13:25:45 crc kubenswrapper[4790]: I1124 13:25:45.840233 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="ovn-controller" containerID="cri-o://908f7697965f4e6d820bfe32f7f59960e917dedea04373ff0e2b8126bb835163" gracePeriod=30 Nov 24 13:25:45 crc kubenswrapper[4790]: I1124 13:25:45.840286 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://1a2bf135bf819ef8ce489337a9f6b7b0bb626ff335b37ec7e6a74e0e13a1ba40" gracePeriod=30 Nov 24 13:25:45 crc kubenswrapper[4790]: I1124 13:25:45.840316 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="ovn-acl-logging" containerID="cri-o://23e4305b0dc57da3ad6d92e8f04f1f18780c1ebaf63ee19140a8443b99197bdd" gracePeriod=30 Nov 24 13:25:45 crc kubenswrapper[4790]: I1124 13:25:45.840356 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="sbdb" containerID="cri-o://b16be7938ff339aa0f92c4fb0d444114ac253439e80bf12ceb55c7e9a1944603" gracePeriod=30 Nov 24 13:25:45 crc kubenswrapper[4790]: I1124 13:25:45.840398 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="nbdb" containerID="cri-o://9f3d5e18b5f456b7627a9154ff8fb00945ec026d4fb8797d4c4f3e2a2fbb1db7" gracePeriod=30 Nov 24 13:25:45 crc kubenswrapper[4790]: I1124 13:25:45.840405 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="northd" containerID="cri-o://ad72e45458f88569f2bb77a55dc98e2bad29a7e451e5258d371722616042bebd" gracePeriod=30 Nov 24 13:25:45 crc kubenswrapper[4790]: I1124 13:25:45.840450 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="kube-rbac-proxy-node" containerID="cri-o://ec06ff65e703be63a81e99901326ec1eaa20c2176ddbd0764cc7cdc0b5761bd5" gracePeriod=30 Nov 24 13:25:45 crc kubenswrapper[4790]: I1124 13:25:45.968955 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="ovnkube-controller" containerID="cri-o://fbeaf8db5bfb90ff3667f7182b5506b392454f0f0fa6a32223ed0a832e1a86d1" gracePeriod=30 Nov 24 13:25:46 crc kubenswrapper[4790]: I1124 13:25:46.755751 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5x94h_528a4b22-34ba-41b8-8c7d-07d98eebe02b/ovnkube-controller/3.log" Nov 24 13:25:46 crc kubenswrapper[4790]: I1124 13:25:46.758224 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5x94h_528a4b22-34ba-41b8-8c7d-07d98eebe02b/ovn-acl-logging/0.log" Nov 24 13:25:46 crc kubenswrapper[4790]: I1124 13:25:46.758834 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5x94h_528a4b22-34ba-41b8-8c7d-07d98eebe02b/ovn-controller/0.log" Nov 24 13:25:46 crc kubenswrapper[4790]: I1124 13:25:46.759336 4790 generic.go:334] "Generic (PLEG): container finished" podID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerID="fbeaf8db5bfb90ff3667f7182b5506b392454f0f0fa6a32223ed0a832e1a86d1" exitCode=0 Nov 24 13:25:46 crc kubenswrapper[4790]: I1124 13:25:46.759439 4790 generic.go:334] "Generic (PLEG): container finished" podID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerID="b16be7938ff339aa0f92c4fb0d444114ac253439e80bf12ceb55c7e9a1944603" exitCode=0 Nov 24 13:25:46 crc kubenswrapper[4790]: I1124 13:25:46.759512 4790 generic.go:334] "Generic (PLEG): container finished" podID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerID="9f3d5e18b5f456b7627a9154ff8fb00945ec026d4fb8797d4c4f3e2a2fbb1db7" exitCode=0 Nov 24 13:25:46 crc kubenswrapper[4790]: I1124 13:25:46.759582 4790 generic.go:334] "Generic (PLEG): container finished" podID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerID="ad72e45458f88569f2bb77a55dc98e2bad29a7e451e5258d371722616042bebd" exitCode=0 Nov 24 13:25:46 crc kubenswrapper[4790]: I1124 13:25:46.759643 4790 generic.go:334] "Generic (PLEG): container finished" podID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerID="1a2bf135bf819ef8ce489337a9f6b7b0bb626ff335b37ec7e6a74e0e13a1ba40" exitCode=0 Nov 24 13:25:46 crc kubenswrapper[4790]: I1124 13:25:46.759702 4790 generic.go:334] "Generic (PLEG): container finished" podID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerID="ec06ff65e703be63a81e99901326ec1eaa20c2176ddbd0764cc7cdc0b5761bd5" exitCode=0 Nov 24 13:25:46 crc kubenswrapper[4790]: I1124 13:25:46.759756 4790 generic.go:334] "Generic (PLEG): container finished" podID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerID="23e4305b0dc57da3ad6d92e8f04f1f18780c1ebaf63ee19140a8443b99197bdd" exitCode=143 Nov 24 13:25:46 crc kubenswrapper[4790]: I1124 13:25:46.759809 4790 generic.go:334] "Generic (PLEG): container finished" podID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerID="908f7697965f4e6d820bfe32f7f59960e917dedea04373ff0e2b8126bb835163" exitCode=143 Nov 24 13:25:46 crc kubenswrapper[4790]: I1124 13:25:46.759950 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" event={"ID":"528a4b22-34ba-41b8-8c7d-07d98eebe02b","Type":"ContainerDied","Data":"fbeaf8db5bfb90ff3667f7182b5506b392454f0f0fa6a32223ed0a832e1a86d1"} Nov 24 13:25:46 crc kubenswrapper[4790]: I1124 13:25:46.760101 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" event={"ID":"528a4b22-34ba-41b8-8c7d-07d98eebe02b","Type":"ContainerDied","Data":"b16be7938ff339aa0f92c4fb0d444114ac253439e80bf12ceb55c7e9a1944603"} Nov 24 13:25:46 crc kubenswrapper[4790]: I1124 13:25:46.760196 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" event={"ID":"528a4b22-34ba-41b8-8c7d-07d98eebe02b","Type":"ContainerDied","Data":"9f3d5e18b5f456b7627a9154ff8fb00945ec026d4fb8797d4c4f3e2a2fbb1db7"} Nov 24 13:25:46 crc kubenswrapper[4790]: I1124 13:25:46.760275 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" event={"ID":"528a4b22-34ba-41b8-8c7d-07d98eebe02b","Type":"ContainerDied","Data":"ad72e45458f88569f2bb77a55dc98e2bad29a7e451e5258d371722616042bebd"} Nov 24 13:25:46 crc kubenswrapper[4790]: I1124 13:25:46.760351 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" event={"ID":"528a4b22-34ba-41b8-8c7d-07d98eebe02b","Type":"ContainerDied","Data":"1a2bf135bf819ef8ce489337a9f6b7b0bb626ff335b37ec7e6a74e0e13a1ba40"} Nov 24 13:25:46 crc kubenswrapper[4790]: I1124 13:25:46.760426 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" event={"ID":"528a4b22-34ba-41b8-8c7d-07d98eebe02b","Type":"ContainerDied","Data":"ec06ff65e703be63a81e99901326ec1eaa20c2176ddbd0764cc7cdc0b5761bd5"} Nov 24 13:25:46 crc kubenswrapper[4790]: I1124 13:25:46.760515 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" event={"ID":"528a4b22-34ba-41b8-8c7d-07d98eebe02b","Type":"ContainerDied","Data":"23e4305b0dc57da3ad6d92e8f04f1f18780c1ebaf63ee19140a8443b99197bdd"} Nov 24 13:25:46 crc kubenswrapper[4790]: I1124 13:25:46.760235 4790 scope.go:117] "RemoveContainer" containerID="92683d18aac71f52361dbb0877043c6ae0cf8452eff0df475873ab76bda9d6a9" Nov 24 13:25:46 crc kubenswrapper[4790]: I1124 13:25:46.760605 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" event={"ID":"528a4b22-34ba-41b8-8c7d-07d98eebe02b","Type":"ContainerDied","Data":"908f7697965f4e6d820bfe32f7f59960e917dedea04373ff0e2b8126bb835163"} Nov 24 13:25:46 crc kubenswrapper[4790]: I1124 13:25:46.762322 4790 generic.go:334] "Generic (PLEG): container finished" podID="9292a43b-dfe0-45a7-8007-434efafabf33" containerID="63e7ee560b607c9889d138792de432cf8cc0728034ffcf946f8331fa01d30f20" exitCode=0 Nov 24 13:25:46 crc kubenswrapper[4790]: I1124 13:25:46.762408 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f6b92" event={"ID":"9292a43b-dfe0-45a7-8007-434efafabf33","Type":"ContainerDied","Data":"63e7ee560b607c9889d138792de432cf8cc0728034ffcf946f8331fa01d30f20"} Nov 24 13:25:46 crc kubenswrapper[4790]: I1124 13:25:46.767841 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-76rkg_6c0e3499-87ac-481f-a010-708a3a4a730f/kube-multus/2.log" Nov 24 13:25:46 crc kubenswrapper[4790]: I1124 13:25:46.768272 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-76rkg_6c0e3499-87ac-481f-a010-708a3a4a730f/kube-multus/1.log" Nov 24 13:25:46 crc kubenswrapper[4790]: I1124 13:25:46.768310 4790 generic.go:334] "Generic (PLEG): container finished" podID="6c0e3499-87ac-481f-a010-708a3a4a730f" containerID="24025d94118c5c9d2f93c0792f7c151a99d9f5b9e3b328284faad63d2f3f2a23" exitCode=2 Nov 24 13:25:46 crc kubenswrapper[4790]: I1124 13:25:46.768347 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-76rkg" event={"ID":"6c0e3499-87ac-481f-a010-708a3a4a730f","Type":"ContainerDied","Data":"24025d94118c5c9d2f93c0792f7c151a99d9f5b9e3b328284faad63d2f3f2a23"} Nov 24 13:25:46 crc kubenswrapper[4790]: I1124 13:25:46.768950 4790 scope.go:117] "RemoveContainer" containerID="24025d94118c5c9d2f93c0792f7c151a99d9f5b9e3b328284faad63d2f3f2a23" Nov 24 13:25:46 crc kubenswrapper[4790]: I1124 13:25:46.777686 4790 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 13:25:46 crc kubenswrapper[4790]: I1124 13:25:46.917929 4790 scope.go:117] "RemoveContainer" containerID="6af0748fbdc734fe38245ac5426022cd063620ec033c5e4285073feb4a376019" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.092166 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5x94h_528a4b22-34ba-41b8-8c7d-07d98eebe02b/ovn-acl-logging/0.log" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.092636 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5x94h_528a4b22-34ba-41b8-8c7d-07d98eebe02b/ovn-controller/0.log" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.093051 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.156184 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-tnc7l"] Nov 24 13:25:47 crc kubenswrapper[4790]: E1124 13:25:47.156518 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="ovnkube-controller" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.156538 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="ovnkube-controller" Nov 24 13:25:47 crc kubenswrapper[4790]: E1124 13:25:47.156550 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="ovn-acl-logging" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.156557 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="ovn-acl-logging" Nov 24 13:25:47 crc kubenswrapper[4790]: E1124 13:25:47.156570 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="ovnkube-controller" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.156578 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="ovnkube-controller" Nov 24 13:25:47 crc kubenswrapper[4790]: E1124 13:25:47.156587 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="ovnkube-controller" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.156597 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="ovnkube-controller" Nov 24 13:25:47 crc kubenswrapper[4790]: E1124 13:25:47.156605 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="kube-rbac-proxy-node" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.156612 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="kube-rbac-proxy-node" Nov 24 13:25:47 crc kubenswrapper[4790]: E1124 13:25:47.156622 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="nbdb" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.156629 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="nbdb" Nov 24 13:25:47 crc kubenswrapper[4790]: E1124 13:25:47.156637 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="ovnkube-controller" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.156645 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="ovnkube-controller" Nov 24 13:25:47 crc kubenswrapper[4790]: E1124 13:25:47.156666 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="kubecfg-setup" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.156676 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="kubecfg-setup" Nov 24 13:25:47 crc kubenswrapper[4790]: E1124 13:25:47.156688 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="sbdb" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.156696 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="sbdb" Nov 24 13:25:47 crc kubenswrapper[4790]: E1124 13:25:47.156709 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="ovn-controller" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.156716 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="ovn-controller" Nov 24 13:25:47 crc kubenswrapper[4790]: E1124 13:25:47.156730 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.156738 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 13:25:47 crc kubenswrapper[4790]: E1124 13:25:47.156749 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="northd" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.156757 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="northd" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.156913 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="ovnkube-controller" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.156928 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.156941 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="nbdb" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.156950 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="ovnkube-controller" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.156958 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="northd" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.156966 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="ovn-acl-logging" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.156975 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="ovn-controller" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.156985 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="kube-rbac-proxy-node" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.156997 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="sbdb" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.157005 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="ovnkube-controller" Nov 24 13:25:47 crc kubenswrapper[4790]: E1124 13:25:47.157155 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="ovnkube-controller" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.157163 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="ovnkube-controller" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.157258 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="ovnkube-controller" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.157275 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" containerName="ovnkube-controller" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.159682 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.210687 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-var-lib-openvswitch\") pod \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.211228 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/528a4b22-34ba-41b8-8c7d-07d98eebe02b-ovnkube-script-lib\") pod \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.211257 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-log-socket\") pod \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.211298 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-cni-bin\") pod \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.211382 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-run-openvswitch\") pod \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.211426 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-kubelet\") pod \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.211441 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-run-systemd\") pod \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.211477 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-node-log\") pod \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.211503 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-run-ovn-kubernetes\") pod \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.211534 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-etc-openvswitch\") pod \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.211553 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-slash\") pod \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.211585 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-run-netns\") pod \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.211620 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bcnzl\" (UniqueName: \"kubernetes.io/projected/528a4b22-34ba-41b8-8c7d-07d98eebe02b-kube-api-access-bcnzl\") pod \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.211644 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-cni-netd\") pod \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.211678 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.211706 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/528a4b22-34ba-41b8-8c7d-07d98eebe02b-ovn-node-metrics-cert\") pod \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.211734 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/528a4b22-34ba-41b8-8c7d-07d98eebe02b-env-overrides\") pod \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.211756 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/528a4b22-34ba-41b8-8c7d-07d98eebe02b-ovnkube-config\") pod \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.211777 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-run-ovn\") pod \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.211792 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-systemd-units\") pod \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\" (UID: \"528a4b22-34ba-41b8-8c7d-07d98eebe02b\") " Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.212025 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fea3675c-4038-418b-87f8-d326c01f13b1-ovnkube-script-lib\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.212068 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-run-systemd\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.212094 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-log-socket\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.210862 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "528a4b22-34ba-41b8-8c7d-07d98eebe02b" (UID: "528a4b22-34ba-41b8-8c7d-07d98eebe02b"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.212141 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-node-log" (OuterVolumeSpecName: "node-log") pod "528a4b22-34ba-41b8-8c7d-07d98eebe02b" (UID: "528a4b22-34ba-41b8-8c7d-07d98eebe02b"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.212123 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fea3675c-4038-418b-87f8-d326c01f13b1-ovnkube-config\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.212071 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "528a4b22-34ba-41b8-8c7d-07d98eebe02b" (UID: "528a4b22-34ba-41b8-8c7d-07d98eebe02b"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.212069 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "528a4b22-34ba-41b8-8c7d-07d98eebe02b" (UID: "528a4b22-34ba-41b8-8c7d-07d98eebe02b"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.212318 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-host-run-netns\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.212330 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "528a4b22-34ba-41b8-8c7d-07d98eebe02b" (UID: "528a4b22-34ba-41b8-8c7d-07d98eebe02b"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.212096 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-log-socket" (OuterVolumeSpecName: "log-socket") pod "528a4b22-34ba-41b8-8c7d-07d98eebe02b" (UID: "528a4b22-34ba-41b8-8c7d-07d98eebe02b"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.212101 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "528a4b22-34ba-41b8-8c7d-07d98eebe02b" (UID: "528a4b22-34ba-41b8-8c7d-07d98eebe02b"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.212335 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-slash" (OuterVolumeSpecName: "host-slash") pod "528a4b22-34ba-41b8-8c7d-07d98eebe02b" (UID: "528a4b22-34ba-41b8-8c7d-07d98eebe02b"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.212118 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "528a4b22-34ba-41b8-8c7d-07d98eebe02b" (UID: "528a4b22-34ba-41b8-8c7d-07d98eebe02b"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.212209 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "528a4b22-34ba-41b8-8c7d-07d98eebe02b" (UID: "528a4b22-34ba-41b8-8c7d-07d98eebe02b"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.212240 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "528a4b22-34ba-41b8-8c7d-07d98eebe02b" (UID: "528a4b22-34ba-41b8-8c7d-07d98eebe02b"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.212279 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "528a4b22-34ba-41b8-8c7d-07d98eebe02b" (UID: "528a4b22-34ba-41b8-8c7d-07d98eebe02b"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.212301 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "528a4b22-34ba-41b8-8c7d-07d98eebe02b" (UID: "528a4b22-34ba-41b8-8c7d-07d98eebe02b"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.212303 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "528a4b22-34ba-41b8-8c7d-07d98eebe02b" (UID: "528a4b22-34ba-41b8-8c7d-07d98eebe02b"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.212450 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.212619 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-var-lib-openvswitch\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.212680 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-host-kubelet\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.212721 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-host-cni-netd\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.212741 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-etc-openvswitch\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.212762 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-host-slash\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.212786 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-run-ovn\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.212831 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-systemd-units\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.212847 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fea3675c-4038-418b-87f8-d326c01f13b1-ovn-node-metrics-cert\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.212961 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-host-run-ovn-kubernetes\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.212982 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-host-cni-bin\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.212990 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/528a4b22-34ba-41b8-8c7d-07d98eebe02b-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "528a4b22-34ba-41b8-8c7d-07d98eebe02b" (UID: "528a4b22-34ba-41b8-8c7d-07d98eebe02b"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.213020 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/528a4b22-34ba-41b8-8c7d-07d98eebe02b-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "528a4b22-34ba-41b8-8c7d-07d98eebe02b" (UID: "528a4b22-34ba-41b8-8c7d-07d98eebe02b"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.213135 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-node-log\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.213143 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/528a4b22-34ba-41b8-8c7d-07d98eebe02b-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "528a4b22-34ba-41b8-8c7d-07d98eebe02b" (UID: "528a4b22-34ba-41b8-8c7d-07d98eebe02b"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.213166 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fea3675c-4038-418b-87f8-d326c01f13b1-env-overrides\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.213245 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-run-openvswitch\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.213305 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mb9d\" (UniqueName: \"kubernetes.io/projected/fea3675c-4038-418b-87f8-d326c01f13b1-kube-api-access-7mb9d\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.213408 4790 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-node-log\") on node \"crc\" DevicePath \"\"" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.213430 4790 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.213446 4790 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.213458 4790 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-slash\") on node \"crc\" DevicePath \"\"" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.213470 4790 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.213481 4790 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.213495 4790 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.213508 4790 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/528a4b22-34ba-41b8-8c7d-07d98eebe02b-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.213522 4790 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/528a4b22-34ba-41b8-8c7d-07d98eebe02b-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.213535 4790 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.213547 4790 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.213557 4790 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.213568 4790 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/528a4b22-34ba-41b8-8c7d-07d98eebe02b-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.213579 4790 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-log-socket\") on node \"crc\" DevicePath \"\"" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.213592 4790 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.213605 4790 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.213618 4790 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.224894 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/528a4b22-34ba-41b8-8c7d-07d98eebe02b-kube-api-access-bcnzl" (OuterVolumeSpecName: "kube-api-access-bcnzl") pod "528a4b22-34ba-41b8-8c7d-07d98eebe02b" (UID: "528a4b22-34ba-41b8-8c7d-07d98eebe02b"). InnerVolumeSpecName "kube-api-access-bcnzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.231562 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/528a4b22-34ba-41b8-8c7d-07d98eebe02b-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "528a4b22-34ba-41b8-8c7d-07d98eebe02b" (UID: "528a4b22-34ba-41b8-8c7d-07d98eebe02b"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.233483 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "528a4b22-34ba-41b8-8c7d-07d98eebe02b" (UID: "528a4b22-34ba-41b8-8c7d-07d98eebe02b"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.314892 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-run-systemd\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.314939 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-log-socket\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.314960 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-run-systemd\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.314961 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fea3675c-4038-418b-87f8-d326c01f13b1-ovnkube-config\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.315044 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-host-run-netns\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.315077 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.315109 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-var-lib-openvswitch\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.315137 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-host-kubelet\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.315163 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-host-cni-netd\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.315186 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-etc-openvswitch\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.315208 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-host-slash\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.315232 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-run-ovn\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.315262 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-systemd-units\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.315288 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fea3675c-4038-418b-87f8-d326c01f13b1-ovn-node-metrics-cert\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.315323 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-host-run-ovn-kubernetes\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.315349 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-host-cni-bin\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.315385 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-node-log\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.315412 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fea3675c-4038-418b-87f8-d326c01f13b1-env-overrides\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.315447 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-run-openvswitch\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.315482 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mb9d\" (UniqueName: \"kubernetes.io/projected/fea3675c-4038-418b-87f8-d326c01f13b1-kube-api-access-7mb9d\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.315518 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fea3675c-4038-418b-87f8-d326c01f13b1-ovnkube-script-lib\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.315576 4790 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/528a4b22-34ba-41b8-8c7d-07d98eebe02b-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.315594 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bcnzl\" (UniqueName: \"kubernetes.io/projected/528a4b22-34ba-41b8-8c7d-07d98eebe02b-kube-api-access-bcnzl\") on node \"crc\" DevicePath \"\"" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.315609 4790 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/528a4b22-34ba-41b8-8c7d-07d98eebe02b-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.315710 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fea3675c-4038-418b-87f8-d326c01f13b1-ovnkube-config\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.315755 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-log-socket\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.315784 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-run-ovn\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.315809 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-host-run-netns\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.315841 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.315904 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-var-lib-openvswitch\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.315941 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-host-kubelet\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.315976 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-host-cni-netd\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.316013 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-etc-openvswitch\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.316048 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-host-slash\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.316078 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-host-cni-bin\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.316111 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-systemd-units\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.316317 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fea3675c-4038-418b-87f8-d326c01f13b1-ovnkube-script-lib\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.316373 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-node-log\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.316725 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-run-openvswitch\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.316801 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fea3675c-4038-418b-87f8-d326c01f13b1-env-overrides\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.316861 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fea3675c-4038-418b-87f8-d326c01f13b1-host-run-ovn-kubernetes\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.320521 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fea3675c-4038-418b-87f8-d326c01f13b1-ovn-node-metrics-cert\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.338404 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mb9d\" (UniqueName: \"kubernetes.io/projected/fea3675c-4038-418b-87f8-d326c01f13b1-kube-api-access-7mb9d\") pod \"ovnkube-node-tnc7l\" (UID: \"fea3675c-4038-418b-87f8-d326c01f13b1\") " pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.493197 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:47 crc kubenswrapper[4790]: W1124 13:25:47.523562 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfea3675c_4038_418b_87f8_d326c01f13b1.slice/crio-64573cb82949e045742e5d59b513aa3ad347ab822e99c2885ff7c0153d4fbc40 WatchSource:0}: Error finding container 64573cb82949e045742e5d59b513aa3ad347ab822e99c2885ff7c0153d4fbc40: Status 404 returned error can't find the container with id 64573cb82949e045742e5d59b513aa3ad347ab822e99c2885ff7c0153d4fbc40 Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.777533 4790 generic.go:334] "Generic (PLEG): container finished" podID="fea3675c-4038-418b-87f8-d326c01f13b1" containerID="dd7fbd1322b81418092bf23f8171001307a093b4f7e8cf5e61f771405f84b57c" exitCode=0 Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.777964 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" event={"ID":"fea3675c-4038-418b-87f8-d326c01f13b1","Type":"ContainerDied","Data":"dd7fbd1322b81418092bf23f8171001307a093b4f7e8cf5e61f771405f84b57c"} Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.777999 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" event={"ID":"fea3675c-4038-418b-87f8-d326c01f13b1","Type":"ContainerStarted","Data":"64573cb82949e045742e5d59b513aa3ad347ab822e99c2885ff7c0153d4fbc40"} Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.783042 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5x94h_528a4b22-34ba-41b8-8c7d-07d98eebe02b/ovn-acl-logging/0.log" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.783630 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5x94h_528a4b22-34ba-41b8-8c7d-07d98eebe02b/ovn-controller/0.log" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.784376 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" event={"ID":"528a4b22-34ba-41b8-8c7d-07d98eebe02b","Type":"ContainerDied","Data":"af11f648dbb79e4be6d1491f965cd72809a6b9117b650da1e064e2270043adaf"} Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.784448 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5x94h" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.784497 4790 scope.go:117] "RemoveContainer" containerID="fbeaf8db5bfb90ff3667f7182b5506b392454f0f0fa6a32223ed0a832e1a86d1" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.787122 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-76rkg_6c0e3499-87ac-481f-a010-708a3a4a730f/kube-multus/2.log" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.787228 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-76rkg" event={"ID":"6c0e3499-87ac-481f-a010-708a3a4a730f","Type":"ContainerStarted","Data":"4f4b05f3717851ca674f051e1e7c17953497096a2c118ce401365ca6bc4a47f8"} Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.866215 4790 scope.go:117] "RemoveContainer" containerID="b16be7938ff339aa0f92c4fb0d444114ac253439e80bf12ceb55c7e9a1944603" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.884251 4790 scope.go:117] "RemoveContainer" containerID="9f3d5e18b5f456b7627a9154ff8fb00945ec026d4fb8797d4c4f3e2a2fbb1db7" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.905632 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5x94h"] Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.910010 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5x94h"] Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.914859 4790 scope.go:117] "RemoveContainer" containerID="ad72e45458f88569f2bb77a55dc98e2bad29a7e451e5258d371722616042bebd" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.928845 4790 scope.go:117] "RemoveContainer" containerID="1a2bf135bf819ef8ce489337a9f6b7b0bb626ff335b37ec7e6a74e0e13a1ba40" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.945750 4790 scope.go:117] "RemoveContainer" containerID="ec06ff65e703be63a81e99901326ec1eaa20c2176ddbd0764cc7cdc0b5761bd5" Nov 24 13:25:47 crc kubenswrapper[4790]: I1124 13:25:47.963103 4790 scope.go:117] "RemoveContainer" containerID="23e4305b0dc57da3ad6d92e8f04f1f18780c1ebaf63ee19140a8443b99197bdd" Nov 24 13:25:48 crc kubenswrapper[4790]: I1124 13:25:48.059114 4790 scope.go:117] "RemoveContainer" containerID="908f7697965f4e6d820bfe32f7f59960e917dedea04373ff0e2b8126bb835163" Nov 24 13:25:48 crc kubenswrapper[4790]: I1124 13:25:48.271530 4790 scope.go:117] "RemoveContainer" containerID="9e15a9058c2c7cce52ce390e23d4473062a1deee51e5f9b95f10f5b0a3cccf33" Nov 24 13:25:48 crc kubenswrapper[4790]: I1124 13:25:48.322488 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="528a4b22-34ba-41b8-8c7d-07d98eebe02b" path="/var/lib/kubelet/pods/528a4b22-34ba-41b8-8c7d-07d98eebe02b/volumes" Nov 24 13:25:49 crc kubenswrapper[4790]: I1124 13:25:49.109780 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nt2g7"] Nov 24 13:25:49 crc kubenswrapper[4790]: I1124 13:25:49.113719 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nt2g7" Nov 24 13:25:49 crc kubenswrapper[4790]: I1124 13:25:49.141013 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwm5p\" (UniqueName: \"kubernetes.io/projected/b319457e-1e71-442a-908f-2d003bcddbda-kube-api-access-jwm5p\") pod \"redhat-operators-nt2g7\" (UID: \"b319457e-1e71-442a-908f-2d003bcddbda\") " pod="openshift-marketplace/redhat-operators-nt2g7" Nov 24 13:25:49 crc kubenswrapper[4790]: I1124 13:25:49.141332 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b319457e-1e71-442a-908f-2d003bcddbda-utilities\") pod \"redhat-operators-nt2g7\" (UID: \"b319457e-1e71-442a-908f-2d003bcddbda\") " pod="openshift-marketplace/redhat-operators-nt2g7" Nov 24 13:25:49 crc kubenswrapper[4790]: I1124 13:25:49.141501 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b319457e-1e71-442a-908f-2d003bcddbda-catalog-content\") pod \"redhat-operators-nt2g7\" (UID: \"b319457e-1e71-442a-908f-2d003bcddbda\") " pod="openshift-marketplace/redhat-operators-nt2g7" Nov 24 13:25:49 crc kubenswrapper[4790]: I1124 13:25:49.243917 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b319457e-1e71-442a-908f-2d003bcddbda-utilities\") pod \"redhat-operators-nt2g7\" (UID: \"b319457e-1e71-442a-908f-2d003bcddbda\") " pod="openshift-marketplace/redhat-operators-nt2g7" Nov 24 13:25:49 crc kubenswrapper[4790]: I1124 13:25:49.244012 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b319457e-1e71-442a-908f-2d003bcddbda-catalog-content\") pod \"redhat-operators-nt2g7\" (UID: \"b319457e-1e71-442a-908f-2d003bcddbda\") " pod="openshift-marketplace/redhat-operators-nt2g7" Nov 24 13:25:49 crc kubenswrapper[4790]: I1124 13:25:49.244062 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwm5p\" (UniqueName: \"kubernetes.io/projected/b319457e-1e71-442a-908f-2d003bcddbda-kube-api-access-jwm5p\") pod \"redhat-operators-nt2g7\" (UID: \"b319457e-1e71-442a-908f-2d003bcddbda\") " pod="openshift-marketplace/redhat-operators-nt2g7" Nov 24 13:25:49 crc kubenswrapper[4790]: I1124 13:25:49.244929 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b319457e-1e71-442a-908f-2d003bcddbda-utilities\") pod \"redhat-operators-nt2g7\" (UID: \"b319457e-1e71-442a-908f-2d003bcddbda\") " pod="openshift-marketplace/redhat-operators-nt2g7" Nov 24 13:25:49 crc kubenswrapper[4790]: I1124 13:25:49.246660 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b319457e-1e71-442a-908f-2d003bcddbda-catalog-content\") pod \"redhat-operators-nt2g7\" (UID: \"b319457e-1e71-442a-908f-2d003bcddbda\") " pod="openshift-marketplace/redhat-operators-nt2g7" Nov 24 13:25:49 crc kubenswrapper[4790]: I1124 13:25:49.269679 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwm5p\" (UniqueName: \"kubernetes.io/projected/b319457e-1e71-442a-908f-2d003bcddbda-kube-api-access-jwm5p\") pod \"redhat-operators-nt2g7\" (UID: \"b319457e-1e71-442a-908f-2d003bcddbda\") " pod="openshift-marketplace/redhat-operators-nt2g7" Nov 24 13:25:49 crc kubenswrapper[4790]: I1124 13:25:49.504748 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nt2g7" Nov 24 13:25:49 crc kubenswrapper[4790]: E1124 13:25:49.536127 4790 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-operators-nt2g7_openshift-marketplace_b319457e-1e71-442a-908f-2d003bcddbda_0(22013f93450105b70e35d562e455b70a61cf4021bb08eefa2ce237a9336b5b44): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 13:25:49 crc kubenswrapper[4790]: E1124 13:25:49.536762 4790 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-operators-nt2g7_openshift-marketplace_b319457e-1e71-442a-908f-2d003bcddbda_0(22013f93450105b70e35d562e455b70a61cf4021bb08eefa2ce237a9336b5b44): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/redhat-operators-nt2g7" Nov 24 13:25:49 crc kubenswrapper[4790]: E1124 13:25:49.536801 4790 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-operators-nt2g7_openshift-marketplace_b319457e-1e71-442a-908f-2d003bcddbda_0(22013f93450105b70e35d562e455b70a61cf4021bb08eefa2ce237a9336b5b44): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/redhat-operators-nt2g7" Nov 24 13:25:49 crc kubenswrapper[4790]: E1124 13:25:49.536907 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"redhat-operators-nt2g7_openshift-marketplace(b319457e-1e71-442a-908f-2d003bcddbda)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"redhat-operators-nt2g7_openshift-marketplace(b319457e-1e71-442a-908f-2d003bcddbda)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-operators-nt2g7_openshift-marketplace_b319457e-1e71-442a-908f-2d003bcddbda_0(22013f93450105b70e35d562e455b70a61cf4021bb08eefa2ce237a9336b5b44): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/redhat-operators-nt2g7" podUID="b319457e-1e71-442a-908f-2d003bcddbda" Nov 24 13:25:49 crc kubenswrapper[4790]: I1124 13:25:49.806417 4790 generic.go:334] "Generic (PLEG): container finished" podID="9292a43b-dfe0-45a7-8007-434efafabf33" containerID="b5dff595bb3bccbbc3210d034799cbf8edd7cdf13279da1392eff7d1b72ab43c" exitCode=0 Nov 24 13:25:49 crc kubenswrapper[4790]: I1124 13:25:49.806475 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f6b92" event={"ID":"9292a43b-dfe0-45a7-8007-434efafabf33","Type":"ContainerDied","Data":"b5dff595bb3bccbbc3210d034799cbf8edd7cdf13279da1392eff7d1b72ab43c"} Nov 24 13:25:49 crc kubenswrapper[4790]: I1124 13:25:49.809999 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" event={"ID":"fea3675c-4038-418b-87f8-d326c01f13b1","Type":"ContainerStarted","Data":"64cb8c3eb43e7501747b3f46101842a45a64ba485e4e07eac68becf741566055"} Nov 24 13:25:49 crc kubenswrapper[4790]: I1124 13:25:49.810038 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" event={"ID":"fea3675c-4038-418b-87f8-d326c01f13b1","Type":"ContainerStarted","Data":"57b32dba412dd56195f8456b2d96858e3bf1339c6a8ff6c0afa2acf0d625697c"} Nov 24 13:25:49 crc kubenswrapper[4790]: I1124 13:25:49.810053 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" event={"ID":"fea3675c-4038-418b-87f8-d326c01f13b1","Type":"ContainerStarted","Data":"bb1baa5880f920ea6e3e14f3d226843b32d7e41b6b5d19fc5e022c3cf90f41f2"} Nov 24 13:25:50 crc kubenswrapper[4790]: I1124 13:25:50.818179 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f6b92" event={"ID":"9292a43b-dfe0-45a7-8007-434efafabf33","Type":"ContainerStarted","Data":"aae411f46cd512aed9968ae9a92d469277626f55ee29bc3b090ccc51144a36c9"} Nov 24 13:25:50 crc kubenswrapper[4790]: I1124 13:25:50.822299 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" event={"ID":"fea3675c-4038-418b-87f8-d326c01f13b1","Type":"ContainerStarted","Data":"bae5560e90d802d1751428faa5c97f2b1326654064f393ebf46d2c60e89d72d8"} Nov 24 13:25:50 crc kubenswrapper[4790]: I1124 13:25:50.822489 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" event={"ID":"fea3675c-4038-418b-87f8-d326c01f13b1","Type":"ContainerStarted","Data":"bfbac8aeef9ce723e4a89a9663d815732e1401828fdadd4fb1b69ba8a8af52e3"} Nov 24 13:25:50 crc kubenswrapper[4790]: I1124 13:25:50.822554 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" event={"ID":"fea3675c-4038-418b-87f8-d326c01f13b1","Type":"ContainerStarted","Data":"8a65cecb8d89fa8d90a35aed61fca947b91e454970c07d2fa10909c4c447341d"} Nov 24 13:25:50 crc kubenswrapper[4790]: I1124 13:25:50.846172 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-f6b92" podStartSLOduration=3.266480594 podStartE2EDuration="6.846134164s" podCreationTimestamp="2025-11-24 13:25:44 +0000 UTC" firstStartedPulling="2025-11-24 13:25:46.777277398 +0000 UTC m=+795.157171060" lastFinishedPulling="2025-11-24 13:25:50.356930968 +0000 UTC m=+798.736824630" observedRunningTime="2025-11-24 13:25:50.843559658 +0000 UTC m=+799.223453320" watchObservedRunningTime="2025-11-24 13:25:50.846134164 +0000 UTC m=+799.226027826" Nov 24 13:25:53 crc kubenswrapper[4790]: I1124 13:25:53.853611 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" event={"ID":"fea3675c-4038-418b-87f8-d326c01f13b1","Type":"ContainerStarted","Data":"77f62088cf80960672f18aae70a3ca8f52692638aadb6fb00ff06e82519adf41"} Nov 24 13:25:54 crc kubenswrapper[4790]: I1124 13:25:54.096621 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zqqc9"] Nov 24 13:25:54 crc kubenswrapper[4790]: I1124 13:25:54.098215 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zqqc9" Nov 24 13:25:54 crc kubenswrapper[4790]: I1124 13:25:54.230510 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38c94de9-58c8-4ca8-862a-0cedd910f16d-utilities\") pod \"redhat-marketplace-zqqc9\" (UID: \"38c94de9-58c8-4ca8-862a-0cedd910f16d\") " pod="openshift-marketplace/redhat-marketplace-zqqc9" Nov 24 13:25:54 crc kubenswrapper[4790]: I1124 13:25:54.231092 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwnnt\" (UniqueName: \"kubernetes.io/projected/38c94de9-58c8-4ca8-862a-0cedd910f16d-kube-api-access-lwnnt\") pod \"redhat-marketplace-zqqc9\" (UID: \"38c94de9-58c8-4ca8-862a-0cedd910f16d\") " pod="openshift-marketplace/redhat-marketplace-zqqc9" Nov 24 13:25:54 crc kubenswrapper[4790]: I1124 13:25:54.231128 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38c94de9-58c8-4ca8-862a-0cedd910f16d-catalog-content\") pod \"redhat-marketplace-zqqc9\" (UID: \"38c94de9-58c8-4ca8-862a-0cedd910f16d\") " pod="openshift-marketplace/redhat-marketplace-zqqc9" Nov 24 13:25:54 crc kubenswrapper[4790]: I1124 13:25:54.332331 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38c94de9-58c8-4ca8-862a-0cedd910f16d-utilities\") pod \"redhat-marketplace-zqqc9\" (UID: \"38c94de9-58c8-4ca8-862a-0cedd910f16d\") " pod="openshift-marketplace/redhat-marketplace-zqqc9" Nov 24 13:25:54 crc kubenswrapper[4790]: I1124 13:25:54.332801 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwnnt\" (UniqueName: \"kubernetes.io/projected/38c94de9-58c8-4ca8-862a-0cedd910f16d-kube-api-access-lwnnt\") pod \"redhat-marketplace-zqqc9\" (UID: \"38c94de9-58c8-4ca8-862a-0cedd910f16d\") " pod="openshift-marketplace/redhat-marketplace-zqqc9" Nov 24 13:25:54 crc kubenswrapper[4790]: I1124 13:25:54.332940 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38c94de9-58c8-4ca8-862a-0cedd910f16d-catalog-content\") pod \"redhat-marketplace-zqqc9\" (UID: \"38c94de9-58c8-4ca8-862a-0cedd910f16d\") " pod="openshift-marketplace/redhat-marketplace-zqqc9" Nov 24 13:25:54 crc kubenswrapper[4790]: I1124 13:25:54.332986 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38c94de9-58c8-4ca8-862a-0cedd910f16d-utilities\") pod \"redhat-marketplace-zqqc9\" (UID: \"38c94de9-58c8-4ca8-862a-0cedd910f16d\") " pod="openshift-marketplace/redhat-marketplace-zqqc9" Nov 24 13:25:54 crc kubenswrapper[4790]: I1124 13:25:54.333568 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38c94de9-58c8-4ca8-862a-0cedd910f16d-catalog-content\") pod \"redhat-marketplace-zqqc9\" (UID: \"38c94de9-58c8-4ca8-862a-0cedd910f16d\") " pod="openshift-marketplace/redhat-marketplace-zqqc9" Nov 24 13:25:54 crc kubenswrapper[4790]: I1124 13:25:54.359186 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwnnt\" (UniqueName: \"kubernetes.io/projected/38c94de9-58c8-4ca8-862a-0cedd910f16d-kube-api-access-lwnnt\") pod \"redhat-marketplace-zqqc9\" (UID: \"38c94de9-58c8-4ca8-862a-0cedd910f16d\") " pod="openshift-marketplace/redhat-marketplace-zqqc9" Nov 24 13:25:54 crc kubenswrapper[4790]: I1124 13:25:54.414300 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zqqc9" Nov 24 13:25:54 crc kubenswrapper[4790]: E1124 13:25:54.446808 4790 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-marketplace-zqqc9_openshift-marketplace_38c94de9-58c8-4ca8-862a-0cedd910f16d_0(87e813b2c86dca9103f57920e686edaeb51c1f509f7daf9e872e84db29655ca6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 13:25:54 crc kubenswrapper[4790]: E1124 13:25:54.447006 4790 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-marketplace-zqqc9_openshift-marketplace_38c94de9-58c8-4ca8-862a-0cedd910f16d_0(87e813b2c86dca9103f57920e686edaeb51c1f509f7daf9e872e84db29655ca6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/redhat-marketplace-zqqc9" Nov 24 13:25:54 crc kubenswrapper[4790]: E1124 13:25:54.447093 4790 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-marketplace-zqqc9_openshift-marketplace_38c94de9-58c8-4ca8-862a-0cedd910f16d_0(87e813b2c86dca9103f57920e686edaeb51c1f509f7daf9e872e84db29655ca6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/redhat-marketplace-zqqc9" Nov 24 13:25:54 crc kubenswrapper[4790]: E1124 13:25:54.447202 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"redhat-marketplace-zqqc9_openshift-marketplace(38c94de9-58c8-4ca8-862a-0cedd910f16d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"redhat-marketplace-zqqc9_openshift-marketplace(38c94de9-58c8-4ca8-862a-0cedd910f16d)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-marketplace-zqqc9_openshift-marketplace_38c94de9-58c8-4ca8-862a-0cedd910f16d_0(87e813b2c86dca9103f57920e686edaeb51c1f509f7daf9e872e84db29655ca6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/redhat-marketplace-zqqc9" podUID="38c94de9-58c8-4ca8-862a-0cedd910f16d" Nov 24 13:25:54 crc kubenswrapper[4790]: I1124 13:25:54.993340 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-kpg6g"] Nov 24 13:25:54 crc kubenswrapper[4790]: I1124 13:25:54.994335 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-kpg6g" Nov 24 13:25:54 crc kubenswrapper[4790]: I1124 13:25:54.997130 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Nov 24 13:25:54 crc kubenswrapper[4790]: I1124 13:25:54.998264 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Nov 24 13:25:54 crc kubenswrapper[4790]: I1124 13:25:54.998421 4790 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-xfpzp" Nov 24 13:25:55 crc kubenswrapper[4790]: I1124 13:25:55.007330 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Nov 24 13:25:55 crc kubenswrapper[4790]: I1124 13:25:55.045220 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5-crc-storage\") pod \"crc-storage-crc-kpg6g\" (UID: \"412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5\") " pod="crc-storage/crc-storage-crc-kpg6g" Nov 24 13:25:55 crc kubenswrapper[4790]: I1124 13:25:55.045327 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqx8l\" (UniqueName: \"kubernetes.io/projected/412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5-kube-api-access-rqx8l\") pod \"crc-storage-crc-kpg6g\" (UID: \"412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5\") " pod="crc-storage/crc-storage-crc-kpg6g" Nov 24 13:25:55 crc kubenswrapper[4790]: I1124 13:25:55.045397 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5-node-mnt\") pod \"crc-storage-crc-kpg6g\" (UID: \"412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5\") " pod="crc-storage/crc-storage-crc-kpg6g" Nov 24 13:25:55 crc kubenswrapper[4790]: I1124 13:25:55.147058 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5-node-mnt\") pod \"crc-storage-crc-kpg6g\" (UID: \"412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5\") " pod="crc-storage/crc-storage-crc-kpg6g" Nov 24 13:25:55 crc kubenswrapper[4790]: I1124 13:25:55.147218 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5-crc-storage\") pod \"crc-storage-crc-kpg6g\" (UID: \"412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5\") " pod="crc-storage/crc-storage-crc-kpg6g" Nov 24 13:25:55 crc kubenswrapper[4790]: I1124 13:25:55.147318 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqx8l\" (UniqueName: \"kubernetes.io/projected/412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5-kube-api-access-rqx8l\") pod \"crc-storage-crc-kpg6g\" (UID: \"412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5\") " pod="crc-storage/crc-storage-crc-kpg6g" Nov 24 13:25:55 crc kubenswrapper[4790]: I1124 13:25:55.147655 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5-node-mnt\") pod \"crc-storage-crc-kpg6g\" (UID: \"412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5\") " pod="crc-storage/crc-storage-crc-kpg6g" Nov 24 13:25:55 crc kubenswrapper[4790]: I1124 13:25:55.151432 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5-crc-storage\") pod \"crc-storage-crc-kpg6g\" (UID: \"412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5\") " pod="crc-storage/crc-storage-crc-kpg6g" Nov 24 13:25:55 crc kubenswrapper[4790]: I1124 13:25:55.167137 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqx8l\" (UniqueName: \"kubernetes.io/projected/412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5-kube-api-access-rqx8l\") pod \"crc-storage-crc-kpg6g\" (UID: \"412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5\") " pod="crc-storage/crc-storage-crc-kpg6g" Nov 24 13:25:55 crc kubenswrapper[4790]: I1124 13:25:55.168627 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-f6b92" Nov 24 13:25:55 crc kubenswrapper[4790]: I1124 13:25:55.168699 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-f6b92" Nov 24 13:25:55 crc kubenswrapper[4790]: I1124 13:25:55.216538 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-f6b92" Nov 24 13:25:55 crc kubenswrapper[4790]: I1124 13:25:55.311807 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-kpg6g" Nov 24 13:25:55 crc kubenswrapper[4790]: E1124 13:25:55.335422 4790 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-kpg6g_crc-storage_412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5_0(9ca82c19875f03dc6374f6b047f317f066877eff52f5dbc05d15514d365425cb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 13:25:55 crc kubenswrapper[4790]: E1124 13:25:55.335500 4790 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-kpg6g_crc-storage_412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5_0(9ca82c19875f03dc6374f6b047f317f066877eff52f5dbc05d15514d365425cb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-kpg6g" Nov 24 13:25:55 crc kubenswrapper[4790]: E1124 13:25:55.335535 4790 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-kpg6g_crc-storage_412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5_0(9ca82c19875f03dc6374f6b047f317f066877eff52f5dbc05d15514d365425cb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-kpg6g" Nov 24 13:25:55 crc kubenswrapper[4790]: E1124 13:25:55.335609 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-kpg6g_crc-storage(412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-kpg6g_crc-storage(412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-kpg6g_crc-storage_412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5_0(9ca82c19875f03dc6374f6b047f317f066877eff52f5dbc05d15514d365425cb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-kpg6g" podUID="412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5" Nov 24 13:25:55 crc kubenswrapper[4790]: I1124 13:25:55.872385 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" event={"ID":"fea3675c-4038-418b-87f8-d326c01f13b1","Type":"ContainerStarted","Data":"977462eb1d0cf21956004b8d04cb1b4a472c1f122a43194ef13db6782b21e0c9"} Nov 24 13:25:55 crc kubenswrapper[4790]: I1124 13:25:55.915550 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" podStartSLOduration=8.915517155 podStartE2EDuration="8.915517155s" podCreationTimestamp="2025-11-24 13:25:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:25:55.908341164 +0000 UTC m=+804.288234866" watchObservedRunningTime="2025-11-24 13:25:55.915517155 +0000 UTC m=+804.295410817" Nov 24 13:25:55 crc kubenswrapper[4790]: I1124 13:25:55.942385 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-f6b92" Nov 24 13:25:56 crc kubenswrapper[4790]: I1124 13:25:56.284139 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zqqc9"] Nov 24 13:25:56 crc kubenswrapper[4790]: I1124 13:25:56.284814 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zqqc9" Nov 24 13:25:56 crc kubenswrapper[4790]: I1124 13:25:56.285623 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zqqc9" Nov 24 13:25:56 crc kubenswrapper[4790]: I1124 13:25:56.290570 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-kpg6g"] Nov 24 13:25:56 crc kubenswrapper[4790]: I1124 13:25:56.290764 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-kpg6g" Nov 24 13:25:56 crc kubenswrapper[4790]: I1124 13:25:56.291461 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-kpg6g" Nov 24 13:25:56 crc kubenswrapper[4790]: I1124 13:25:56.328296 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nt2g7"] Nov 24 13:25:56 crc kubenswrapper[4790]: I1124 13:25:56.328638 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nt2g7" Nov 24 13:25:56 crc kubenswrapper[4790]: I1124 13:25:56.329444 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nt2g7" Nov 24 13:25:56 crc kubenswrapper[4790]: E1124 13:25:56.335090 4790 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-kpg6g_crc-storage_412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5_0(047a48d6b3761766c130586da72105457355c026c58b28e05d17e799973b7f39): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 13:25:56 crc kubenswrapper[4790]: E1124 13:25:56.335175 4790 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-kpg6g_crc-storage_412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5_0(047a48d6b3761766c130586da72105457355c026c58b28e05d17e799973b7f39): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-kpg6g" Nov 24 13:25:56 crc kubenswrapper[4790]: E1124 13:25:56.335210 4790 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-kpg6g_crc-storage_412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5_0(047a48d6b3761766c130586da72105457355c026c58b28e05d17e799973b7f39): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-kpg6g" Nov 24 13:25:56 crc kubenswrapper[4790]: E1124 13:25:56.335287 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-kpg6g_crc-storage(412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-kpg6g_crc-storage(412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-kpg6g_crc-storage_412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5_0(047a48d6b3761766c130586da72105457355c026c58b28e05d17e799973b7f39): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-kpg6g" podUID="412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5" Nov 24 13:25:56 crc kubenswrapper[4790]: E1124 13:25:56.351294 4790 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-marketplace-zqqc9_openshift-marketplace_38c94de9-58c8-4ca8-862a-0cedd910f16d_0(541e0e4642ec8245ea6cee9e0a74fc5c70e84ab1a1bf6b4d31733ba6c3e8a371): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 13:25:56 crc kubenswrapper[4790]: E1124 13:25:56.351385 4790 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-marketplace-zqqc9_openshift-marketplace_38c94de9-58c8-4ca8-862a-0cedd910f16d_0(541e0e4642ec8245ea6cee9e0a74fc5c70e84ab1a1bf6b4d31733ba6c3e8a371): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/redhat-marketplace-zqqc9" Nov 24 13:25:56 crc kubenswrapper[4790]: E1124 13:25:56.351414 4790 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-marketplace-zqqc9_openshift-marketplace_38c94de9-58c8-4ca8-862a-0cedd910f16d_0(541e0e4642ec8245ea6cee9e0a74fc5c70e84ab1a1bf6b4d31733ba6c3e8a371): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/redhat-marketplace-zqqc9" Nov 24 13:25:56 crc kubenswrapper[4790]: E1124 13:25:56.351480 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"redhat-marketplace-zqqc9_openshift-marketplace(38c94de9-58c8-4ca8-862a-0cedd910f16d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"redhat-marketplace-zqqc9_openshift-marketplace(38c94de9-58c8-4ca8-862a-0cedd910f16d)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-marketplace-zqqc9_openshift-marketplace_38c94de9-58c8-4ca8-862a-0cedd910f16d_0(541e0e4642ec8245ea6cee9e0a74fc5c70e84ab1a1bf6b4d31733ba6c3e8a371): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/redhat-marketplace-zqqc9" podUID="38c94de9-58c8-4ca8-862a-0cedd910f16d" Nov 24 13:25:56 crc kubenswrapper[4790]: E1124 13:25:56.366366 4790 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-operators-nt2g7_openshift-marketplace_b319457e-1e71-442a-908f-2d003bcddbda_0(bb399ba230d76c08a655f464e03b019113d75afc2a1f7f0b56a6e9a29ad66355): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 13:25:56 crc kubenswrapper[4790]: E1124 13:25:56.366526 4790 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-operators-nt2g7_openshift-marketplace_b319457e-1e71-442a-908f-2d003bcddbda_0(bb399ba230d76c08a655f464e03b019113d75afc2a1f7f0b56a6e9a29ad66355): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/redhat-operators-nt2g7" Nov 24 13:25:56 crc kubenswrapper[4790]: E1124 13:25:56.366567 4790 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-operators-nt2g7_openshift-marketplace_b319457e-1e71-442a-908f-2d003bcddbda_0(bb399ba230d76c08a655f464e03b019113d75afc2a1f7f0b56a6e9a29ad66355): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/redhat-operators-nt2g7" Nov 24 13:25:56 crc kubenswrapper[4790]: E1124 13:25:56.366637 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"redhat-operators-nt2g7_openshift-marketplace(b319457e-1e71-442a-908f-2d003bcddbda)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"redhat-operators-nt2g7_openshift-marketplace(b319457e-1e71-442a-908f-2d003bcddbda)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-operators-nt2g7_openshift-marketplace_b319457e-1e71-442a-908f-2d003bcddbda_0(bb399ba230d76c08a655f464e03b019113d75afc2a1f7f0b56a6e9a29ad66355): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/redhat-operators-nt2g7" podUID="b319457e-1e71-442a-908f-2d003bcddbda" Nov 24 13:25:56 crc kubenswrapper[4790]: I1124 13:25:56.910567 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:56 crc kubenswrapper[4790]: I1124 13:25:56.910627 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:56 crc kubenswrapper[4790]: I1124 13:25:56.910644 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:56 crc kubenswrapper[4790]: I1124 13:25:56.968103 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:56 crc kubenswrapper[4790]: I1124 13:25:56.972956 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:25:57 crc kubenswrapper[4790]: I1124 13:25:57.493042 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f6b92"] Nov 24 13:25:57 crc kubenswrapper[4790]: I1124 13:25:57.913795 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-f6b92" podUID="9292a43b-dfe0-45a7-8007-434efafabf33" containerName="registry-server" containerID="cri-o://aae411f46cd512aed9968ae9a92d469277626f55ee29bc3b090ccc51144a36c9" gracePeriod=2 Nov 24 13:25:58 crc kubenswrapper[4790]: I1124 13:25:58.134059 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f6b92" Nov 24 13:25:58 crc kubenswrapper[4790]: I1124 13:25:58.295261 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llzzj\" (UniqueName: \"kubernetes.io/projected/9292a43b-dfe0-45a7-8007-434efafabf33-kube-api-access-llzzj\") pod \"9292a43b-dfe0-45a7-8007-434efafabf33\" (UID: \"9292a43b-dfe0-45a7-8007-434efafabf33\") " Nov 24 13:25:58 crc kubenswrapper[4790]: I1124 13:25:58.295806 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9292a43b-dfe0-45a7-8007-434efafabf33-utilities\") pod \"9292a43b-dfe0-45a7-8007-434efafabf33\" (UID: \"9292a43b-dfe0-45a7-8007-434efafabf33\") " Nov 24 13:25:58 crc kubenswrapper[4790]: I1124 13:25:58.295838 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9292a43b-dfe0-45a7-8007-434efafabf33-catalog-content\") pod \"9292a43b-dfe0-45a7-8007-434efafabf33\" (UID: \"9292a43b-dfe0-45a7-8007-434efafabf33\") " Nov 24 13:25:58 crc kubenswrapper[4790]: I1124 13:25:58.296931 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9292a43b-dfe0-45a7-8007-434efafabf33-utilities" (OuterVolumeSpecName: "utilities") pod "9292a43b-dfe0-45a7-8007-434efafabf33" (UID: "9292a43b-dfe0-45a7-8007-434efafabf33"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:25:58 crc kubenswrapper[4790]: I1124 13:25:58.302303 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9292a43b-dfe0-45a7-8007-434efafabf33-kube-api-access-llzzj" (OuterVolumeSpecName: "kube-api-access-llzzj") pod "9292a43b-dfe0-45a7-8007-434efafabf33" (UID: "9292a43b-dfe0-45a7-8007-434efafabf33"). InnerVolumeSpecName "kube-api-access-llzzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:25:58 crc kubenswrapper[4790]: I1124 13:25:58.357613 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9292a43b-dfe0-45a7-8007-434efafabf33-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9292a43b-dfe0-45a7-8007-434efafabf33" (UID: "9292a43b-dfe0-45a7-8007-434efafabf33"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:25:58 crc kubenswrapper[4790]: I1124 13:25:58.398151 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llzzj\" (UniqueName: \"kubernetes.io/projected/9292a43b-dfe0-45a7-8007-434efafabf33-kube-api-access-llzzj\") on node \"crc\" DevicePath \"\"" Nov 24 13:25:58 crc kubenswrapper[4790]: I1124 13:25:58.398218 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9292a43b-dfe0-45a7-8007-434efafabf33-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:25:58 crc kubenswrapper[4790]: I1124 13:25:58.398233 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9292a43b-dfe0-45a7-8007-434efafabf33-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:25:58 crc kubenswrapper[4790]: I1124 13:25:58.921216 4790 generic.go:334] "Generic (PLEG): container finished" podID="9292a43b-dfe0-45a7-8007-434efafabf33" containerID="aae411f46cd512aed9968ae9a92d469277626f55ee29bc3b090ccc51144a36c9" exitCode=0 Nov 24 13:25:58 crc kubenswrapper[4790]: I1124 13:25:58.921292 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f6b92" event={"ID":"9292a43b-dfe0-45a7-8007-434efafabf33","Type":"ContainerDied","Data":"aae411f46cd512aed9968ae9a92d469277626f55ee29bc3b090ccc51144a36c9"} Nov 24 13:25:58 crc kubenswrapper[4790]: I1124 13:25:58.921358 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f6b92" event={"ID":"9292a43b-dfe0-45a7-8007-434efafabf33","Type":"ContainerDied","Data":"24fe77c1d7c3c0bfbbb7d3049ab5265e42405ef86fd2f97d639cd8ebfb36f69b"} Nov 24 13:25:58 crc kubenswrapper[4790]: I1124 13:25:58.921382 4790 scope.go:117] "RemoveContainer" containerID="aae411f46cd512aed9968ae9a92d469277626f55ee29bc3b090ccc51144a36c9" Nov 24 13:25:58 crc kubenswrapper[4790]: I1124 13:25:58.921384 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f6b92" Nov 24 13:25:58 crc kubenswrapper[4790]: I1124 13:25:58.948446 4790 scope.go:117] "RemoveContainer" containerID="b5dff595bb3bccbbc3210d034799cbf8edd7cdf13279da1392eff7d1b72ab43c" Nov 24 13:25:58 crc kubenswrapper[4790]: I1124 13:25:58.965061 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f6b92"] Nov 24 13:25:58 crc kubenswrapper[4790]: I1124 13:25:58.969682 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-f6b92"] Nov 24 13:25:58 crc kubenswrapper[4790]: I1124 13:25:58.987629 4790 scope.go:117] "RemoveContainer" containerID="63e7ee560b607c9889d138792de432cf8cc0728034ffcf946f8331fa01d30f20" Nov 24 13:25:59 crc kubenswrapper[4790]: I1124 13:25:59.003898 4790 scope.go:117] "RemoveContainer" containerID="aae411f46cd512aed9968ae9a92d469277626f55ee29bc3b090ccc51144a36c9" Nov 24 13:25:59 crc kubenswrapper[4790]: E1124 13:25:59.004698 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aae411f46cd512aed9968ae9a92d469277626f55ee29bc3b090ccc51144a36c9\": container with ID starting with aae411f46cd512aed9968ae9a92d469277626f55ee29bc3b090ccc51144a36c9 not found: ID does not exist" containerID="aae411f46cd512aed9968ae9a92d469277626f55ee29bc3b090ccc51144a36c9" Nov 24 13:25:59 crc kubenswrapper[4790]: I1124 13:25:59.004744 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aae411f46cd512aed9968ae9a92d469277626f55ee29bc3b090ccc51144a36c9"} err="failed to get container status \"aae411f46cd512aed9968ae9a92d469277626f55ee29bc3b090ccc51144a36c9\": rpc error: code = NotFound desc = could not find container \"aae411f46cd512aed9968ae9a92d469277626f55ee29bc3b090ccc51144a36c9\": container with ID starting with aae411f46cd512aed9968ae9a92d469277626f55ee29bc3b090ccc51144a36c9 not found: ID does not exist" Nov 24 13:25:59 crc kubenswrapper[4790]: I1124 13:25:59.004776 4790 scope.go:117] "RemoveContainer" containerID="b5dff595bb3bccbbc3210d034799cbf8edd7cdf13279da1392eff7d1b72ab43c" Nov 24 13:25:59 crc kubenswrapper[4790]: E1124 13:25:59.005298 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5dff595bb3bccbbc3210d034799cbf8edd7cdf13279da1392eff7d1b72ab43c\": container with ID starting with b5dff595bb3bccbbc3210d034799cbf8edd7cdf13279da1392eff7d1b72ab43c not found: ID does not exist" containerID="b5dff595bb3bccbbc3210d034799cbf8edd7cdf13279da1392eff7d1b72ab43c" Nov 24 13:25:59 crc kubenswrapper[4790]: I1124 13:25:59.005332 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5dff595bb3bccbbc3210d034799cbf8edd7cdf13279da1392eff7d1b72ab43c"} err="failed to get container status \"b5dff595bb3bccbbc3210d034799cbf8edd7cdf13279da1392eff7d1b72ab43c\": rpc error: code = NotFound desc = could not find container \"b5dff595bb3bccbbc3210d034799cbf8edd7cdf13279da1392eff7d1b72ab43c\": container with ID starting with b5dff595bb3bccbbc3210d034799cbf8edd7cdf13279da1392eff7d1b72ab43c not found: ID does not exist" Nov 24 13:25:59 crc kubenswrapper[4790]: I1124 13:25:59.005353 4790 scope.go:117] "RemoveContainer" containerID="63e7ee560b607c9889d138792de432cf8cc0728034ffcf946f8331fa01d30f20" Nov 24 13:25:59 crc kubenswrapper[4790]: E1124 13:25:59.005728 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63e7ee560b607c9889d138792de432cf8cc0728034ffcf946f8331fa01d30f20\": container with ID starting with 63e7ee560b607c9889d138792de432cf8cc0728034ffcf946f8331fa01d30f20 not found: ID does not exist" containerID="63e7ee560b607c9889d138792de432cf8cc0728034ffcf946f8331fa01d30f20" Nov 24 13:25:59 crc kubenswrapper[4790]: I1124 13:25:59.005797 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63e7ee560b607c9889d138792de432cf8cc0728034ffcf946f8331fa01d30f20"} err="failed to get container status \"63e7ee560b607c9889d138792de432cf8cc0728034ffcf946f8331fa01d30f20\": rpc error: code = NotFound desc = could not find container \"63e7ee560b607c9889d138792de432cf8cc0728034ffcf946f8331fa01d30f20\": container with ID starting with 63e7ee560b607c9889d138792de432cf8cc0728034ffcf946f8331fa01d30f20 not found: ID does not exist" Nov 24 13:26:00 crc kubenswrapper[4790]: I1124 13:26:00.321982 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9292a43b-dfe0-45a7-8007-434efafabf33" path="/var/lib/kubelet/pods/9292a43b-dfe0-45a7-8007-434efafabf33/volumes" Nov 24 13:26:07 crc kubenswrapper[4790]: I1124 13:26:07.314419 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zqqc9" Nov 24 13:26:07 crc kubenswrapper[4790]: I1124 13:26:07.316755 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zqqc9" Nov 24 13:26:07 crc kubenswrapper[4790]: I1124 13:26:07.567796 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zqqc9"] Nov 24 13:26:07 crc kubenswrapper[4790]: W1124 13:26:07.578553 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod38c94de9_58c8_4ca8_862a_0cedd910f16d.slice/crio-c8397cb2c5b57310db1c3c5e6107fdc308f25c419d3c148eb62f95159bf215bb WatchSource:0}: Error finding container c8397cb2c5b57310db1c3c5e6107fdc308f25c419d3c148eb62f95159bf215bb: Status 404 returned error can't find the container with id c8397cb2c5b57310db1c3c5e6107fdc308f25c419d3c148eb62f95159bf215bb Nov 24 13:26:08 crc kubenswrapper[4790]: I1124 13:26:08.010228 4790 generic.go:334] "Generic (PLEG): container finished" podID="38c94de9-58c8-4ca8-862a-0cedd910f16d" containerID="9f33ac36d47bc224619239da891ab758ba85ec15f5819e9c29ccfcd7b9ddd2ae" exitCode=0 Nov 24 13:26:08 crc kubenswrapper[4790]: I1124 13:26:08.010404 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zqqc9" event={"ID":"38c94de9-58c8-4ca8-862a-0cedd910f16d","Type":"ContainerDied","Data":"9f33ac36d47bc224619239da891ab758ba85ec15f5819e9c29ccfcd7b9ddd2ae"} Nov 24 13:26:08 crc kubenswrapper[4790]: I1124 13:26:08.010820 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zqqc9" event={"ID":"38c94de9-58c8-4ca8-862a-0cedd910f16d","Type":"ContainerStarted","Data":"c8397cb2c5b57310db1c3c5e6107fdc308f25c419d3c148eb62f95159bf215bb"} Nov 24 13:26:09 crc kubenswrapper[4790]: I1124 13:26:09.021465 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zqqc9" event={"ID":"38c94de9-58c8-4ca8-862a-0cedd910f16d","Type":"ContainerStarted","Data":"ed24d252e4c1274a9aa8c95f95506e9689b5b6fb0d330ca92c182a2382086532"} Nov 24 13:26:10 crc kubenswrapper[4790]: I1124 13:26:10.029382 4790 generic.go:334] "Generic (PLEG): container finished" podID="38c94de9-58c8-4ca8-862a-0cedd910f16d" containerID="ed24d252e4c1274a9aa8c95f95506e9689b5b6fb0d330ca92c182a2382086532" exitCode=0 Nov 24 13:26:10 crc kubenswrapper[4790]: I1124 13:26:10.029515 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zqqc9" event={"ID":"38c94de9-58c8-4ca8-862a-0cedd910f16d","Type":"ContainerDied","Data":"ed24d252e4c1274a9aa8c95f95506e9689b5b6fb0d330ca92c182a2382086532"} Nov 24 13:26:10 crc kubenswrapper[4790]: I1124 13:26:10.314073 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nt2g7" Nov 24 13:26:10 crc kubenswrapper[4790]: I1124 13:26:10.314167 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-kpg6g" Nov 24 13:26:10 crc kubenswrapper[4790]: I1124 13:26:10.314620 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nt2g7" Nov 24 13:26:10 crc kubenswrapper[4790]: I1124 13:26:10.315014 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-kpg6g" Nov 24 13:26:10 crc kubenswrapper[4790]: I1124 13:26:10.557336 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nt2g7"] Nov 24 13:26:10 crc kubenswrapper[4790]: W1124 13:26:10.557686 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb319457e_1e71_442a_908f_2d003bcddbda.slice/crio-d3e0ef1fc3aa52fb99bcb335b66af583832d821633bea1e0419e6249c44e20fd WatchSource:0}: Error finding container d3e0ef1fc3aa52fb99bcb335b66af583832d821633bea1e0419e6249c44e20fd: Status 404 returned error can't find the container with id d3e0ef1fc3aa52fb99bcb335b66af583832d821633bea1e0419e6249c44e20fd Nov 24 13:26:10 crc kubenswrapper[4790]: I1124 13:26:10.608293 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-kpg6g"] Nov 24 13:26:10 crc kubenswrapper[4790]: W1124 13:26:10.612162 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod412bae7a_3f92_47ce_bd17_5f0e5fd5c1e5.slice/crio-8ac0678bd10d8c41c134b2f09d9b6e523f65cce5436570ce8e0fc58b5017e411 WatchSource:0}: Error finding container 8ac0678bd10d8c41c134b2f09d9b6e523f65cce5436570ce8e0fc58b5017e411: Status 404 returned error can't find the container with id 8ac0678bd10d8c41c134b2f09d9b6e523f65cce5436570ce8e0fc58b5017e411 Nov 24 13:26:11 crc kubenswrapper[4790]: I1124 13:26:11.038355 4790 generic.go:334] "Generic (PLEG): container finished" podID="b319457e-1e71-442a-908f-2d003bcddbda" containerID="31df30c9ebeacabdbd8e67dab4c8818b026dce5669a2cca35355b0fa4804f0b6" exitCode=0 Nov 24 13:26:11 crc kubenswrapper[4790]: I1124 13:26:11.038494 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nt2g7" event={"ID":"b319457e-1e71-442a-908f-2d003bcddbda","Type":"ContainerDied","Data":"31df30c9ebeacabdbd8e67dab4c8818b026dce5669a2cca35355b0fa4804f0b6"} Nov 24 13:26:11 crc kubenswrapper[4790]: I1124 13:26:11.038559 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nt2g7" event={"ID":"b319457e-1e71-442a-908f-2d003bcddbda","Type":"ContainerStarted","Data":"d3e0ef1fc3aa52fb99bcb335b66af583832d821633bea1e0419e6249c44e20fd"} Nov 24 13:26:11 crc kubenswrapper[4790]: I1124 13:26:11.041806 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-kpg6g" event={"ID":"412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5","Type":"ContainerStarted","Data":"8ac0678bd10d8c41c134b2f09d9b6e523f65cce5436570ce8e0fc58b5017e411"} Nov 24 13:26:11 crc kubenswrapper[4790]: I1124 13:26:11.045536 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zqqc9" event={"ID":"38c94de9-58c8-4ca8-862a-0cedd910f16d","Type":"ContainerStarted","Data":"827fb7a0a6cfdaa2abe8be475d0de6a24bf2ee38019ca025e223e3eaa5c573f7"} Nov 24 13:26:11 crc kubenswrapper[4790]: I1124 13:26:11.089503 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zqqc9" podStartSLOduration=14.371811887 podStartE2EDuration="17.089469313s" podCreationTimestamp="2025-11-24 13:25:54 +0000 UTC" firstStartedPulling="2025-11-24 13:26:08.013630841 +0000 UTC m=+816.393524523" lastFinishedPulling="2025-11-24 13:26:10.731288287 +0000 UTC m=+819.111181949" observedRunningTime="2025-11-24 13:26:11.086257758 +0000 UTC m=+819.466151430" watchObservedRunningTime="2025-11-24 13:26:11.089469313 +0000 UTC m=+819.469362985" Nov 24 13:26:13 crc kubenswrapper[4790]: I1124 13:26:13.938723 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:26:13 crc kubenswrapper[4790]: I1124 13:26:13.939653 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:26:14 crc kubenswrapper[4790]: I1124 13:26:14.414993 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zqqc9" Nov 24 13:26:14 crc kubenswrapper[4790]: I1124 13:26:14.415101 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zqqc9" Nov 24 13:26:14 crc kubenswrapper[4790]: I1124 13:26:14.458780 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zqqc9" Nov 24 13:26:15 crc kubenswrapper[4790]: I1124 13:26:15.123522 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zqqc9" Nov 24 13:26:15 crc kubenswrapper[4790]: I1124 13:26:15.778754 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zqqc9"] Nov 24 13:26:17 crc kubenswrapper[4790]: I1124 13:26:17.097465 4790 generic.go:334] "Generic (PLEG): container finished" podID="b319457e-1e71-442a-908f-2d003bcddbda" containerID="0766bf5a3835fe06fc09a331b01edb2f0fe2030adc42d8a72cf8cd207ab73cc0" exitCode=0 Nov 24 13:26:17 crc kubenswrapper[4790]: I1124 13:26:17.097575 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nt2g7" event={"ID":"b319457e-1e71-442a-908f-2d003bcddbda","Type":"ContainerDied","Data":"0766bf5a3835fe06fc09a331b01edb2f0fe2030adc42d8a72cf8cd207ab73cc0"} Nov 24 13:26:17 crc kubenswrapper[4790]: I1124 13:26:17.101761 4790 generic.go:334] "Generic (PLEG): container finished" podID="412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5" containerID="64bc3c80920485efa9acd6d9ada588bc80cbf57817a1a94698a53a2910fbf1d8" exitCode=0 Nov 24 13:26:17 crc kubenswrapper[4790]: I1124 13:26:17.101854 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-kpg6g" event={"ID":"412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5","Type":"ContainerDied","Data":"64bc3c80920485efa9acd6d9ada588bc80cbf57817a1a94698a53a2910fbf1d8"} Nov 24 13:26:17 crc kubenswrapper[4790]: I1124 13:26:17.102121 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zqqc9" podUID="38c94de9-58c8-4ca8-862a-0cedd910f16d" containerName="registry-server" containerID="cri-o://827fb7a0a6cfdaa2abe8be475d0de6a24bf2ee38019ca025e223e3eaa5c573f7" gracePeriod=2 Nov 24 13:26:17 crc kubenswrapper[4790]: I1124 13:26:17.522693 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tnc7l" Nov 24 13:26:17 crc kubenswrapper[4790]: I1124 13:26:17.576824 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zqqc9" Nov 24 13:26:17 crc kubenswrapper[4790]: I1124 13:26:17.727986 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwnnt\" (UniqueName: \"kubernetes.io/projected/38c94de9-58c8-4ca8-862a-0cedd910f16d-kube-api-access-lwnnt\") pod \"38c94de9-58c8-4ca8-862a-0cedd910f16d\" (UID: \"38c94de9-58c8-4ca8-862a-0cedd910f16d\") " Nov 24 13:26:17 crc kubenswrapper[4790]: I1124 13:26:17.728112 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38c94de9-58c8-4ca8-862a-0cedd910f16d-catalog-content\") pod \"38c94de9-58c8-4ca8-862a-0cedd910f16d\" (UID: \"38c94de9-58c8-4ca8-862a-0cedd910f16d\") " Nov 24 13:26:17 crc kubenswrapper[4790]: I1124 13:26:17.728134 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38c94de9-58c8-4ca8-862a-0cedd910f16d-utilities\") pod \"38c94de9-58c8-4ca8-862a-0cedd910f16d\" (UID: \"38c94de9-58c8-4ca8-862a-0cedd910f16d\") " Nov 24 13:26:17 crc kubenswrapper[4790]: I1124 13:26:17.729094 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38c94de9-58c8-4ca8-862a-0cedd910f16d-utilities" (OuterVolumeSpecName: "utilities") pod "38c94de9-58c8-4ca8-862a-0cedd910f16d" (UID: "38c94de9-58c8-4ca8-862a-0cedd910f16d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:26:17 crc kubenswrapper[4790]: I1124 13:26:17.735554 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38c94de9-58c8-4ca8-862a-0cedd910f16d-kube-api-access-lwnnt" (OuterVolumeSpecName: "kube-api-access-lwnnt") pod "38c94de9-58c8-4ca8-862a-0cedd910f16d" (UID: "38c94de9-58c8-4ca8-862a-0cedd910f16d"). InnerVolumeSpecName "kube-api-access-lwnnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:26:17 crc kubenswrapper[4790]: I1124 13:26:17.748450 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38c94de9-58c8-4ca8-862a-0cedd910f16d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "38c94de9-58c8-4ca8-862a-0cedd910f16d" (UID: "38c94de9-58c8-4ca8-862a-0cedd910f16d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:26:17 crc kubenswrapper[4790]: I1124 13:26:17.829936 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38c94de9-58c8-4ca8-862a-0cedd910f16d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:17 crc kubenswrapper[4790]: I1124 13:26:17.830435 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38c94de9-58c8-4ca8-862a-0cedd910f16d-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:17 crc kubenswrapper[4790]: I1124 13:26:17.830453 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwnnt\" (UniqueName: \"kubernetes.io/projected/38c94de9-58c8-4ca8-862a-0cedd910f16d-kube-api-access-lwnnt\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:18 crc kubenswrapper[4790]: I1124 13:26:18.114394 4790 generic.go:334] "Generic (PLEG): container finished" podID="38c94de9-58c8-4ca8-862a-0cedd910f16d" containerID="827fb7a0a6cfdaa2abe8be475d0de6a24bf2ee38019ca025e223e3eaa5c573f7" exitCode=0 Nov 24 13:26:18 crc kubenswrapper[4790]: I1124 13:26:18.114509 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zqqc9" event={"ID":"38c94de9-58c8-4ca8-862a-0cedd910f16d","Type":"ContainerDied","Data":"827fb7a0a6cfdaa2abe8be475d0de6a24bf2ee38019ca025e223e3eaa5c573f7"} Nov 24 13:26:18 crc kubenswrapper[4790]: I1124 13:26:18.114570 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zqqc9" Nov 24 13:26:18 crc kubenswrapper[4790]: I1124 13:26:18.114607 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zqqc9" event={"ID":"38c94de9-58c8-4ca8-862a-0cedd910f16d","Type":"ContainerDied","Data":"c8397cb2c5b57310db1c3c5e6107fdc308f25c419d3c148eb62f95159bf215bb"} Nov 24 13:26:18 crc kubenswrapper[4790]: I1124 13:26:18.114641 4790 scope.go:117] "RemoveContainer" containerID="827fb7a0a6cfdaa2abe8be475d0de6a24bf2ee38019ca025e223e3eaa5c573f7" Nov 24 13:26:18 crc kubenswrapper[4790]: I1124 13:26:18.134673 4790 scope.go:117] "RemoveContainer" containerID="ed24d252e4c1274a9aa8c95f95506e9689b5b6fb0d330ca92c182a2382086532" Nov 24 13:26:18 crc kubenswrapper[4790]: I1124 13:26:18.152106 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zqqc9"] Nov 24 13:26:18 crc kubenswrapper[4790]: I1124 13:26:18.155282 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zqqc9"] Nov 24 13:26:18 crc kubenswrapper[4790]: I1124 13:26:18.159262 4790 scope.go:117] "RemoveContainer" containerID="9f33ac36d47bc224619239da891ab758ba85ec15f5819e9c29ccfcd7b9ddd2ae" Nov 24 13:26:18 crc kubenswrapper[4790]: I1124 13:26:18.176448 4790 scope.go:117] "RemoveContainer" containerID="827fb7a0a6cfdaa2abe8be475d0de6a24bf2ee38019ca025e223e3eaa5c573f7" Nov 24 13:26:18 crc kubenswrapper[4790]: E1124 13:26:18.179385 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"827fb7a0a6cfdaa2abe8be475d0de6a24bf2ee38019ca025e223e3eaa5c573f7\": container with ID starting with 827fb7a0a6cfdaa2abe8be475d0de6a24bf2ee38019ca025e223e3eaa5c573f7 not found: ID does not exist" containerID="827fb7a0a6cfdaa2abe8be475d0de6a24bf2ee38019ca025e223e3eaa5c573f7" Nov 24 13:26:18 crc kubenswrapper[4790]: I1124 13:26:18.179455 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"827fb7a0a6cfdaa2abe8be475d0de6a24bf2ee38019ca025e223e3eaa5c573f7"} err="failed to get container status \"827fb7a0a6cfdaa2abe8be475d0de6a24bf2ee38019ca025e223e3eaa5c573f7\": rpc error: code = NotFound desc = could not find container \"827fb7a0a6cfdaa2abe8be475d0de6a24bf2ee38019ca025e223e3eaa5c573f7\": container with ID starting with 827fb7a0a6cfdaa2abe8be475d0de6a24bf2ee38019ca025e223e3eaa5c573f7 not found: ID does not exist" Nov 24 13:26:18 crc kubenswrapper[4790]: I1124 13:26:18.179495 4790 scope.go:117] "RemoveContainer" containerID="ed24d252e4c1274a9aa8c95f95506e9689b5b6fb0d330ca92c182a2382086532" Nov 24 13:26:18 crc kubenswrapper[4790]: E1124 13:26:18.180020 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed24d252e4c1274a9aa8c95f95506e9689b5b6fb0d330ca92c182a2382086532\": container with ID starting with ed24d252e4c1274a9aa8c95f95506e9689b5b6fb0d330ca92c182a2382086532 not found: ID does not exist" containerID="ed24d252e4c1274a9aa8c95f95506e9689b5b6fb0d330ca92c182a2382086532" Nov 24 13:26:18 crc kubenswrapper[4790]: I1124 13:26:18.180073 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed24d252e4c1274a9aa8c95f95506e9689b5b6fb0d330ca92c182a2382086532"} err="failed to get container status \"ed24d252e4c1274a9aa8c95f95506e9689b5b6fb0d330ca92c182a2382086532\": rpc error: code = NotFound desc = could not find container \"ed24d252e4c1274a9aa8c95f95506e9689b5b6fb0d330ca92c182a2382086532\": container with ID starting with ed24d252e4c1274a9aa8c95f95506e9689b5b6fb0d330ca92c182a2382086532 not found: ID does not exist" Nov 24 13:26:18 crc kubenswrapper[4790]: I1124 13:26:18.180115 4790 scope.go:117] "RemoveContainer" containerID="9f33ac36d47bc224619239da891ab758ba85ec15f5819e9c29ccfcd7b9ddd2ae" Nov 24 13:26:18 crc kubenswrapper[4790]: E1124 13:26:18.180449 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f33ac36d47bc224619239da891ab758ba85ec15f5819e9c29ccfcd7b9ddd2ae\": container with ID starting with 9f33ac36d47bc224619239da891ab758ba85ec15f5819e9c29ccfcd7b9ddd2ae not found: ID does not exist" containerID="9f33ac36d47bc224619239da891ab758ba85ec15f5819e9c29ccfcd7b9ddd2ae" Nov 24 13:26:18 crc kubenswrapper[4790]: I1124 13:26:18.180518 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f33ac36d47bc224619239da891ab758ba85ec15f5819e9c29ccfcd7b9ddd2ae"} err="failed to get container status \"9f33ac36d47bc224619239da891ab758ba85ec15f5819e9c29ccfcd7b9ddd2ae\": rpc error: code = NotFound desc = could not find container \"9f33ac36d47bc224619239da891ab758ba85ec15f5819e9c29ccfcd7b9ddd2ae\": container with ID starting with 9f33ac36d47bc224619239da891ab758ba85ec15f5819e9c29ccfcd7b9ddd2ae not found: ID does not exist" Nov 24 13:26:18 crc kubenswrapper[4790]: I1124 13:26:18.322798 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38c94de9-58c8-4ca8-862a-0cedd910f16d" path="/var/lib/kubelet/pods/38c94de9-58c8-4ca8-862a-0cedd910f16d/volumes" Nov 24 13:26:18 crc kubenswrapper[4790]: I1124 13:26:18.411391 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-kpg6g" Nov 24 13:26:18 crc kubenswrapper[4790]: I1124 13:26:18.542117 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqx8l\" (UniqueName: \"kubernetes.io/projected/412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5-kube-api-access-rqx8l\") pod \"412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5\" (UID: \"412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5\") " Nov 24 13:26:18 crc kubenswrapper[4790]: I1124 13:26:18.542318 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5-crc-storage\") pod \"412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5\" (UID: \"412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5\") " Nov 24 13:26:18 crc kubenswrapper[4790]: I1124 13:26:18.542399 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5-node-mnt\") pod \"412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5\" (UID: \"412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5\") " Nov 24 13:26:18 crc kubenswrapper[4790]: I1124 13:26:18.542955 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5" (UID: "412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:26:18 crc kubenswrapper[4790]: I1124 13:26:18.543627 4790 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5-node-mnt\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:18 crc kubenswrapper[4790]: I1124 13:26:18.552157 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5-kube-api-access-rqx8l" (OuterVolumeSpecName: "kube-api-access-rqx8l") pod "412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5" (UID: "412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5"). InnerVolumeSpecName "kube-api-access-rqx8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:26:18 crc kubenswrapper[4790]: I1124 13:26:18.562211 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5" (UID: "412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:26:18 crc kubenswrapper[4790]: I1124 13:26:18.644632 4790 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5-crc-storage\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:18 crc kubenswrapper[4790]: I1124 13:26:18.645012 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqx8l\" (UniqueName: \"kubernetes.io/projected/412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5-kube-api-access-rqx8l\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:19 crc kubenswrapper[4790]: I1124 13:26:19.125928 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nt2g7" event={"ID":"b319457e-1e71-442a-908f-2d003bcddbda","Type":"ContainerStarted","Data":"ae0d7382da3d0739b6dab6a614719dcc67098762cb0ced60edd0de98ae0989f6"} Nov 24 13:26:19 crc kubenswrapper[4790]: I1124 13:26:19.128715 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-kpg6g" Nov 24 13:26:19 crc kubenswrapper[4790]: I1124 13:26:19.128746 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-kpg6g" event={"ID":"412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5","Type":"ContainerDied","Data":"8ac0678bd10d8c41c134b2f09d9b6e523f65cce5436570ce8e0fc58b5017e411"} Nov 24 13:26:19 crc kubenswrapper[4790]: I1124 13:26:19.128841 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ac0678bd10d8c41c134b2f09d9b6e523f65cce5436570ce8e0fc58b5017e411" Nov 24 13:26:19 crc kubenswrapper[4790]: I1124 13:26:19.153039 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nt2g7" podStartSLOduration=23.201080941 podStartE2EDuration="30.15301103s" podCreationTimestamp="2025-11-24 13:25:49 +0000 UTC" firstStartedPulling="2025-11-24 13:26:11.041696775 +0000 UTC m=+819.421590437" lastFinishedPulling="2025-11-24 13:26:17.993626854 +0000 UTC m=+826.373520526" observedRunningTime="2025-11-24 13:26:19.149185087 +0000 UTC m=+827.529078769" watchObservedRunningTime="2025-11-24 13:26:19.15301103 +0000 UTC m=+827.532904692" Nov 24 13:26:19 crc kubenswrapper[4790]: I1124 13:26:19.505489 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nt2g7" Nov 24 13:26:19 crc kubenswrapper[4790]: I1124 13:26:19.505716 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nt2g7" Nov 24 13:26:20 crc kubenswrapper[4790]: I1124 13:26:20.546258 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nt2g7" podUID="b319457e-1e71-442a-908f-2d003bcddbda" containerName="registry-server" probeResult="failure" output=< Nov 24 13:26:20 crc kubenswrapper[4790]: timeout: failed to connect service ":50051" within 1s Nov 24 13:26:20 crc kubenswrapper[4790]: > Nov 24 13:26:25 crc kubenswrapper[4790]: I1124 13:26:25.913357 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xpccg"] Nov 24 13:26:25 crc kubenswrapper[4790]: E1124 13:26:25.914534 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38c94de9-58c8-4ca8-862a-0cedd910f16d" containerName="extract-utilities" Nov 24 13:26:25 crc kubenswrapper[4790]: I1124 13:26:25.914551 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="38c94de9-58c8-4ca8-862a-0cedd910f16d" containerName="extract-utilities" Nov 24 13:26:25 crc kubenswrapper[4790]: E1124 13:26:25.914566 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9292a43b-dfe0-45a7-8007-434efafabf33" containerName="registry-server" Nov 24 13:26:25 crc kubenswrapper[4790]: I1124 13:26:25.914573 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="9292a43b-dfe0-45a7-8007-434efafabf33" containerName="registry-server" Nov 24 13:26:25 crc kubenswrapper[4790]: E1124 13:26:25.914583 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5" containerName="storage" Nov 24 13:26:25 crc kubenswrapper[4790]: I1124 13:26:25.914590 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5" containerName="storage" Nov 24 13:26:25 crc kubenswrapper[4790]: E1124 13:26:25.914603 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38c94de9-58c8-4ca8-862a-0cedd910f16d" containerName="registry-server" Nov 24 13:26:25 crc kubenswrapper[4790]: I1124 13:26:25.914608 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="38c94de9-58c8-4ca8-862a-0cedd910f16d" containerName="registry-server" Nov 24 13:26:25 crc kubenswrapper[4790]: E1124 13:26:25.914620 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38c94de9-58c8-4ca8-862a-0cedd910f16d" containerName="extract-content" Nov 24 13:26:25 crc kubenswrapper[4790]: I1124 13:26:25.914626 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="38c94de9-58c8-4ca8-862a-0cedd910f16d" containerName="extract-content" Nov 24 13:26:25 crc kubenswrapper[4790]: E1124 13:26:25.914635 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9292a43b-dfe0-45a7-8007-434efafabf33" containerName="extract-content" Nov 24 13:26:25 crc kubenswrapper[4790]: I1124 13:26:25.914640 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="9292a43b-dfe0-45a7-8007-434efafabf33" containerName="extract-content" Nov 24 13:26:25 crc kubenswrapper[4790]: E1124 13:26:25.914649 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9292a43b-dfe0-45a7-8007-434efafabf33" containerName="extract-utilities" Nov 24 13:26:25 crc kubenswrapper[4790]: I1124 13:26:25.914655 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="9292a43b-dfe0-45a7-8007-434efafabf33" containerName="extract-utilities" Nov 24 13:26:25 crc kubenswrapper[4790]: I1124 13:26:25.914757 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="9292a43b-dfe0-45a7-8007-434efafabf33" containerName="registry-server" Nov 24 13:26:25 crc kubenswrapper[4790]: I1124 13:26:25.914768 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5" containerName="storage" Nov 24 13:26:25 crc kubenswrapper[4790]: I1124 13:26:25.914781 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="38c94de9-58c8-4ca8-862a-0cedd910f16d" containerName="registry-server" Nov 24 13:26:25 crc kubenswrapper[4790]: I1124 13:26:25.915698 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xpccg" Nov 24 13:26:25 crc kubenswrapper[4790]: I1124 13:26:25.924903 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xpccg"] Nov 24 13:26:25 crc kubenswrapper[4790]: I1124 13:26:25.957496 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0dcf597-fdbd-418c-b739-4e61f2e9aa29-utilities\") pod \"certified-operators-xpccg\" (UID: \"f0dcf597-fdbd-418c-b739-4e61f2e9aa29\") " pod="openshift-marketplace/certified-operators-xpccg" Nov 24 13:26:25 crc kubenswrapper[4790]: I1124 13:26:25.957568 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0dcf597-fdbd-418c-b739-4e61f2e9aa29-catalog-content\") pod \"certified-operators-xpccg\" (UID: \"f0dcf597-fdbd-418c-b739-4e61f2e9aa29\") " pod="openshift-marketplace/certified-operators-xpccg" Nov 24 13:26:25 crc kubenswrapper[4790]: I1124 13:26:25.957604 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dhh4\" (UniqueName: \"kubernetes.io/projected/f0dcf597-fdbd-418c-b739-4e61f2e9aa29-kube-api-access-2dhh4\") pod \"certified-operators-xpccg\" (UID: \"f0dcf597-fdbd-418c-b739-4e61f2e9aa29\") " pod="openshift-marketplace/certified-operators-xpccg" Nov 24 13:26:26 crc kubenswrapper[4790]: I1124 13:26:26.059683 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0dcf597-fdbd-418c-b739-4e61f2e9aa29-utilities\") pod \"certified-operators-xpccg\" (UID: \"f0dcf597-fdbd-418c-b739-4e61f2e9aa29\") " pod="openshift-marketplace/certified-operators-xpccg" Nov 24 13:26:26 crc kubenswrapper[4790]: I1124 13:26:26.059772 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0dcf597-fdbd-418c-b739-4e61f2e9aa29-catalog-content\") pod \"certified-operators-xpccg\" (UID: \"f0dcf597-fdbd-418c-b739-4e61f2e9aa29\") " pod="openshift-marketplace/certified-operators-xpccg" Nov 24 13:26:26 crc kubenswrapper[4790]: I1124 13:26:26.059810 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dhh4\" (UniqueName: \"kubernetes.io/projected/f0dcf597-fdbd-418c-b739-4e61f2e9aa29-kube-api-access-2dhh4\") pod \"certified-operators-xpccg\" (UID: \"f0dcf597-fdbd-418c-b739-4e61f2e9aa29\") " pod="openshift-marketplace/certified-operators-xpccg" Nov 24 13:26:26 crc kubenswrapper[4790]: I1124 13:26:26.060439 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0dcf597-fdbd-418c-b739-4e61f2e9aa29-utilities\") pod \"certified-operators-xpccg\" (UID: \"f0dcf597-fdbd-418c-b739-4e61f2e9aa29\") " pod="openshift-marketplace/certified-operators-xpccg" Nov 24 13:26:26 crc kubenswrapper[4790]: I1124 13:26:26.060575 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0dcf597-fdbd-418c-b739-4e61f2e9aa29-catalog-content\") pod \"certified-operators-xpccg\" (UID: \"f0dcf597-fdbd-418c-b739-4e61f2e9aa29\") " pod="openshift-marketplace/certified-operators-xpccg" Nov 24 13:26:26 crc kubenswrapper[4790]: I1124 13:26:26.086628 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dhh4\" (UniqueName: \"kubernetes.io/projected/f0dcf597-fdbd-418c-b739-4e61f2e9aa29-kube-api-access-2dhh4\") pod \"certified-operators-xpccg\" (UID: \"f0dcf597-fdbd-418c-b739-4e61f2e9aa29\") " pod="openshift-marketplace/certified-operators-xpccg" Nov 24 13:26:26 crc kubenswrapper[4790]: I1124 13:26:26.239002 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xpccg" Nov 24 13:26:26 crc kubenswrapper[4790]: I1124 13:26:26.570644 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl"] Nov 24 13:26:26 crc kubenswrapper[4790]: I1124 13:26:26.572543 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl" Nov 24 13:26:26 crc kubenswrapper[4790]: I1124 13:26:26.575293 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 13:26:26 crc kubenswrapper[4790]: I1124 13:26:26.585764 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl"] Nov 24 13:26:26 crc kubenswrapper[4790]: I1124 13:26:26.721865 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a55d437e-ba5a-4e62-8bea-a9345a428e73-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl\" (UID: \"a55d437e-ba5a-4e62-8bea-a9345a428e73\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl" Nov 24 13:26:26 crc kubenswrapper[4790]: I1124 13:26:26.722140 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a55d437e-ba5a-4e62-8bea-a9345a428e73-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl\" (UID: \"a55d437e-ba5a-4e62-8bea-a9345a428e73\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl" Nov 24 13:26:26 crc kubenswrapper[4790]: I1124 13:26:26.722258 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zq2w7\" (UniqueName: \"kubernetes.io/projected/a55d437e-ba5a-4e62-8bea-a9345a428e73-kube-api-access-zq2w7\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl\" (UID: \"a55d437e-ba5a-4e62-8bea-a9345a428e73\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl" Nov 24 13:26:26 crc kubenswrapper[4790]: I1124 13:26:26.769791 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xpccg"] Nov 24 13:26:26 crc kubenswrapper[4790]: I1124 13:26:26.823551 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a55d437e-ba5a-4e62-8bea-a9345a428e73-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl\" (UID: \"a55d437e-ba5a-4e62-8bea-a9345a428e73\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl" Nov 24 13:26:26 crc kubenswrapper[4790]: I1124 13:26:26.823639 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zq2w7\" (UniqueName: \"kubernetes.io/projected/a55d437e-ba5a-4e62-8bea-a9345a428e73-kube-api-access-zq2w7\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl\" (UID: \"a55d437e-ba5a-4e62-8bea-a9345a428e73\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl" Nov 24 13:26:26 crc kubenswrapper[4790]: I1124 13:26:26.824169 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a55d437e-ba5a-4e62-8bea-a9345a428e73-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl\" (UID: \"a55d437e-ba5a-4e62-8bea-a9345a428e73\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl" Nov 24 13:26:26 crc kubenswrapper[4790]: I1124 13:26:26.824192 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a55d437e-ba5a-4e62-8bea-a9345a428e73-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl\" (UID: \"a55d437e-ba5a-4e62-8bea-a9345a428e73\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl" Nov 24 13:26:26 crc kubenswrapper[4790]: I1124 13:26:26.824659 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a55d437e-ba5a-4e62-8bea-a9345a428e73-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl\" (UID: \"a55d437e-ba5a-4e62-8bea-a9345a428e73\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl" Nov 24 13:26:26 crc kubenswrapper[4790]: I1124 13:26:26.844129 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zq2w7\" (UniqueName: \"kubernetes.io/projected/a55d437e-ba5a-4e62-8bea-a9345a428e73-kube-api-access-zq2w7\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl\" (UID: \"a55d437e-ba5a-4e62-8bea-a9345a428e73\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl" Nov 24 13:26:26 crc kubenswrapper[4790]: I1124 13:26:26.891673 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl" Nov 24 13:26:27 crc kubenswrapper[4790]: I1124 13:26:27.089167 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl"] Nov 24 13:26:27 crc kubenswrapper[4790]: W1124 13:26:27.103680 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda55d437e_ba5a_4e62_8bea_a9345a428e73.slice/crio-c85c9e9b44d5c8d90588eae777be2db9c434431255e272a7f45d3604eac5aeac WatchSource:0}: Error finding container c85c9e9b44d5c8d90588eae777be2db9c434431255e272a7f45d3604eac5aeac: Status 404 returned error can't find the container with id c85c9e9b44d5c8d90588eae777be2db9c434431255e272a7f45d3604eac5aeac Nov 24 13:26:27 crc kubenswrapper[4790]: I1124 13:26:27.183163 4790 generic.go:334] "Generic (PLEG): container finished" podID="f0dcf597-fdbd-418c-b739-4e61f2e9aa29" containerID="1e91f68484302d499351f61140294e0fce12da51c309294fbb09cf32dcfe6aff" exitCode=0 Nov 24 13:26:27 crc kubenswrapper[4790]: I1124 13:26:27.183254 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xpccg" event={"ID":"f0dcf597-fdbd-418c-b739-4e61f2e9aa29","Type":"ContainerDied","Data":"1e91f68484302d499351f61140294e0fce12da51c309294fbb09cf32dcfe6aff"} Nov 24 13:26:27 crc kubenswrapper[4790]: I1124 13:26:27.183390 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xpccg" event={"ID":"f0dcf597-fdbd-418c-b739-4e61f2e9aa29","Type":"ContainerStarted","Data":"0b89613db5cc26052a72535d88331938dcdd5c9772f80c19ad239a4f3c6e3e03"} Nov 24 13:26:27 crc kubenswrapper[4790]: I1124 13:26:27.185614 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl" event={"ID":"a55d437e-ba5a-4e62-8bea-a9345a428e73","Type":"ContainerStarted","Data":"c85c9e9b44d5c8d90588eae777be2db9c434431255e272a7f45d3604eac5aeac"} Nov 24 13:26:28 crc kubenswrapper[4790]: I1124 13:26:28.194788 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl" event={"ID":"a55d437e-ba5a-4e62-8bea-a9345a428e73","Type":"ContainerStarted","Data":"e11d5f2fdb7c499425773f2c809ff94a84e4caba9b82b37c4ff0633bcfda87cb"} Nov 24 13:26:29 crc kubenswrapper[4790]: I1124 13:26:29.560546 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nt2g7" Nov 24 13:26:29 crc kubenswrapper[4790]: I1124 13:26:29.607475 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nt2g7" Nov 24 13:26:31 crc kubenswrapper[4790]: I1124 13:26:31.222684 4790 generic.go:334] "Generic (PLEG): container finished" podID="a55d437e-ba5a-4e62-8bea-a9345a428e73" containerID="e11d5f2fdb7c499425773f2c809ff94a84e4caba9b82b37c4ff0633bcfda87cb" exitCode=0 Nov 24 13:26:31 crc kubenswrapper[4790]: I1124 13:26:31.222846 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl" event={"ID":"a55d437e-ba5a-4e62-8bea-a9345a428e73","Type":"ContainerDied","Data":"e11d5f2fdb7c499425773f2c809ff94a84e4caba9b82b37c4ff0633bcfda87cb"} Nov 24 13:26:32 crc kubenswrapper[4790]: I1124 13:26:32.232153 4790 generic.go:334] "Generic (PLEG): container finished" podID="f0dcf597-fdbd-418c-b739-4e61f2e9aa29" containerID="7578a1d4ef01aa582c4d05cfb85e0fc6d8e242e9f072f7eb02df1b18b2582492" exitCode=0 Nov 24 13:26:32 crc kubenswrapper[4790]: I1124 13:26:32.232219 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xpccg" event={"ID":"f0dcf597-fdbd-418c-b739-4e61f2e9aa29","Type":"ContainerDied","Data":"7578a1d4ef01aa582c4d05cfb85e0fc6d8e242e9f072f7eb02df1b18b2582492"} Nov 24 13:26:32 crc kubenswrapper[4790]: I1124 13:26:32.912627 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nt2g7"] Nov 24 13:26:32 crc kubenswrapper[4790]: I1124 13:26:32.913857 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nt2g7" podUID="b319457e-1e71-442a-908f-2d003bcddbda" containerName="registry-server" containerID="cri-o://ae0d7382da3d0739b6dab6a614719dcc67098762cb0ced60edd0de98ae0989f6" gracePeriod=2 Nov 24 13:26:33 crc kubenswrapper[4790]: I1124 13:26:33.250997 4790 generic.go:334] "Generic (PLEG): container finished" podID="b319457e-1e71-442a-908f-2d003bcddbda" containerID="ae0d7382da3d0739b6dab6a614719dcc67098762cb0ced60edd0de98ae0989f6" exitCode=0 Nov 24 13:26:33 crc kubenswrapper[4790]: I1124 13:26:33.251492 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nt2g7" event={"ID":"b319457e-1e71-442a-908f-2d003bcddbda","Type":"ContainerDied","Data":"ae0d7382da3d0739b6dab6a614719dcc67098762cb0ced60edd0de98ae0989f6"} Nov 24 13:26:33 crc kubenswrapper[4790]: I1124 13:26:33.353023 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nt2g7" Nov 24 13:26:33 crc kubenswrapper[4790]: I1124 13:26:33.436789 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b319457e-1e71-442a-908f-2d003bcddbda-utilities\") pod \"b319457e-1e71-442a-908f-2d003bcddbda\" (UID: \"b319457e-1e71-442a-908f-2d003bcddbda\") " Nov 24 13:26:33 crc kubenswrapper[4790]: I1124 13:26:33.436950 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwm5p\" (UniqueName: \"kubernetes.io/projected/b319457e-1e71-442a-908f-2d003bcddbda-kube-api-access-jwm5p\") pod \"b319457e-1e71-442a-908f-2d003bcddbda\" (UID: \"b319457e-1e71-442a-908f-2d003bcddbda\") " Nov 24 13:26:33 crc kubenswrapper[4790]: I1124 13:26:33.437007 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b319457e-1e71-442a-908f-2d003bcddbda-catalog-content\") pod \"b319457e-1e71-442a-908f-2d003bcddbda\" (UID: \"b319457e-1e71-442a-908f-2d003bcddbda\") " Nov 24 13:26:33 crc kubenswrapper[4790]: I1124 13:26:33.438822 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b319457e-1e71-442a-908f-2d003bcddbda-utilities" (OuterVolumeSpecName: "utilities") pod "b319457e-1e71-442a-908f-2d003bcddbda" (UID: "b319457e-1e71-442a-908f-2d003bcddbda"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:26:33 crc kubenswrapper[4790]: I1124 13:26:33.446207 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b319457e-1e71-442a-908f-2d003bcddbda-kube-api-access-jwm5p" (OuterVolumeSpecName: "kube-api-access-jwm5p") pod "b319457e-1e71-442a-908f-2d003bcddbda" (UID: "b319457e-1e71-442a-908f-2d003bcddbda"). InnerVolumeSpecName "kube-api-access-jwm5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:26:33 crc kubenswrapper[4790]: I1124 13:26:33.532664 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b319457e-1e71-442a-908f-2d003bcddbda-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b319457e-1e71-442a-908f-2d003bcddbda" (UID: "b319457e-1e71-442a-908f-2d003bcddbda"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:26:33 crc kubenswrapper[4790]: I1124 13:26:33.540467 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b319457e-1e71-442a-908f-2d003bcddbda-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:33 crc kubenswrapper[4790]: I1124 13:26:33.540556 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwm5p\" (UniqueName: \"kubernetes.io/projected/b319457e-1e71-442a-908f-2d003bcddbda-kube-api-access-jwm5p\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:33 crc kubenswrapper[4790]: I1124 13:26:33.540574 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b319457e-1e71-442a-908f-2d003bcddbda-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:34 crc kubenswrapper[4790]: I1124 13:26:34.261426 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xpccg" event={"ID":"f0dcf597-fdbd-418c-b739-4e61f2e9aa29","Type":"ContainerStarted","Data":"641a200a06f07ef7d422dfbc535f688ded397aa7214b1fa3af28849d02e4de2f"} Nov 24 13:26:34 crc kubenswrapper[4790]: I1124 13:26:34.264718 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nt2g7" event={"ID":"b319457e-1e71-442a-908f-2d003bcddbda","Type":"ContainerDied","Data":"d3e0ef1fc3aa52fb99bcb335b66af583832d821633bea1e0419e6249c44e20fd"} Nov 24 13:26:34 crc kubenswrapper[4790]: I1124 13:26:34.264776 4790 scope.go:117] "RemoveContainer" containerID="ae0d7382da3d0739b6dab6a614719dcc67098762cb0ced60edd0de98ae0989f6" Nov 24 13:26:34 crc kubenswrapper[4790]: I1124 13:26:34.264953 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nt2g7" Nov 24 13:26:34 crc kubenswrapper[4790]: I1124 13:26:34.276083 4790 generic.go:334] "Generic (PLEG): container finished" podID="a55d437e-ba5a-4e62-8bea-a9345a428e73" containerID="85a94615e43dc70c4fa5ebee6d5230b7ff7abed548c39a655b26facf27f2ef57" exitCode=0 Nov 24 13:26:34 crc kubenswrapper[4790]: I1124 13:26:34.276138 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl" event={"ID":"a55d437e-ba5a-4e62-8bea-a9345a428e73","Type":"ContainerDied","Data":"85a94615e43dc70c4fa5ebee6d5230b7ff7abed548c39a655b26facf27f2ef57"} Nov 24 13:26:34 crc kubenswrapper[4790]: I1124 13:26:34.294290 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xpccg" podStartSLOduration=3.334628965 podStartE2EDuration="9.294256483s" podCreationTimestamp="2025-11-24 13:26:25 +0000 UTC" firstStartedPulling="2025-11-24 13:26:27.185491749 +0000 UTC m=+835.565385411" lastFinishedPulling="2025-11-24 13:26:33.145119247 +0000 UTC m=+841.525012929" observedRunningTime="2025-11-24 13:26:34.289316646 +0000 UTC m=+842.669210328" watchObservedRunningTime="2025-11-24 13:26:34.294256483 +0000 UTC m=+842.674150145" Nov 24 13:26:34 crc kubenswrapper[4790]: I1124 13:26:34.305930 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nt2g7"] Nov 24 13:26:34 crc kubenswrapper[4790]: I1124 13:26:34.311189 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nt2g7"] Nov 24 13:26:34 crc kubenswrapper[4790]: I1124 13:26:34.312321 4790 scope.go:117] "RemoveContainer" containerID="0766bf5a3835fe06fc09a331b01edb2f0fe2030adc42d8a72cf8cd207ab73cc0" Nov 24 13:26:34 crc kubenswrapper[4790]: I1124 13:26:34.322315 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b319457e-1e71-442a-908f-2d003bcddbda" path="/var/lib/kubelet/pods/b319457e-1e71-442a-908f-2d003bcddbda/volumes" Nov 24 13:26:34 crc kubenswrapper[4790]: I1124 13:26:34.350129 4790 scope.go:117] "RemoveContainer" containerID="31df30c9ebeacabdbd8e67dab4c8818b026dce5669a2cca35355b0fa4804f0b6" Nov 24 13:26:35 crc kubenswrapper[4790]: I1124 13:26:35.286386 4790 generic.go:334] "Generic (PLEG): container finished" podID="a55d437e-ba5a-4e62-8bea-a9345a428e73" containerID="25ecba67b084a2b6fba6cddc74b00f641d712b03327f0e5e0634247d07dd5618" exitCode=0 Nov 24 13:26:35 crc kubenswrapper[4790]: I1124 13:26:35.286478 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl" event={"ID":"a55d437e-ba5a-4e62-8bea-a9345a428e73","Type":"ContainerDied","Data":"25ecba67b084a2b6fba6cddc74b00f641d712b03327f0e5e0634247d07dd5618"} Nov 24 13:26:36 crc kubenswrapper[4790]: I1124 13:26:36.239824 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xpccg" Nov 24 13:26:36 crc kubenswrapper[4790]: I1124 13:26:36.240004 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xpccg" Nov 24 13:26:36 crc kubenswrapper[4790]: I1124 13:26:36.281641 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xpccg" Nov 24 13:26:36 crc kubenswrapper[4790]: I1124 13:26:36.554706 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl" Nov 24 13:26:36 crc kubenswrapper[4790]: I1124 13:26:36.691431 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a55d437e-ba5a-4e62-8bea-a9345a428e73-util\") pod \"a55d437e-ba5a-4e62-8bea-a9345a428e73\" (UID: \"a55d437e-ba5a-4e62-8bea-a9345a428e73\") " Nov 24 13:26:36 crc kubenswrapper[4790]: I1124 13:26:36.691612 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a55d437e-ba5a-4e62-8bea-a9345a428e73-bundle\") pod \"a55d437e-ba5a-4e62-8bea-a9345a428e73\" (UID: \"a55d437e-ba5a-4e62-8bea-a9345a428e73\") " Nov 24 13:26:36 crc kubenswrapper[4790]: I1124 13:26:36.691703 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zq2w7\" (UniqueName: \"kubernetes.io/projected/a55d437e-ba5a-4e62-8bea-a9345a428e73-kube-api-access-zq2w7\") pod \"a55d437e-ba5a-4e62-8bea-a9345a428e73\" (UID: \"a55d437e-ba5a-4e62-8bea-a9345a428e73\") " Nov 24 13:26:36 crc kubenswrapper[4790]: I1124 13:26:36.692496 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a55d437e-ba5a-4e62-8bea-a9345a428e73-bundle" (OuterVolumeSpecName: "bundle") pod "a55d437e-ba5a-4e62-8bea-a9345a428e73" (UID: "a55d437e-ba5a-4e62-8bea-a9345a428e73"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:26:36 crc kubenswrapper[4790]: I1124 13:26:36.702128 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a55d437e-ba5a-4e62-8bea-a9345a428e73-kube-api-access-zq2w7" (OuterVolumeSpecName: "kube-api-access-zq2w7") pod "a55d437e-ba5a-4e62-8bea-a9345a428e73" (UID: "a55d437e-ba5a-4e62-8bea-a9345a428e73"). InnerVolumeSpecName "kube-api-access-zq2w7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:26:36 crc kubenswrapper[4790]: I1124 13:26:36.715216 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a55d437e-ba5a-4e62-8bea-a9345a428e73-util" (OuterVolumeSpecName: "util") pod "a55d437e-ba5a-4e62-8bea-a9345a428e73" (UID: "a55d437e-ba5a-4e62-8bea-a9345a428e73"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:26:36 crc kubenswrapper[4790]: I1124 13:26:36.793451 4790 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a55d437e-ba5a-4e62-8bea-a9345a428e73-util\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:36 crc kubenswrapper[4790]: I1124 13:26:36.793497 4790 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a55d437e-ba5a-4e62-8bea-a9345a428e73-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:36 crc kubenswrapper[4790]: I1124 13:26:36.793508 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zq2w7\" (UniqueName: \"kubernetes.io/projected/a55d437e-ba5a-4e62-8bea-a9345a428e73-kube-api-access-zq2w7\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:37 crc kubenswrapper[4790]: I1124 13:26:37.303470 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl" event={"ID":"a55d437e-ba5a-4e62-8bea-a9345a428e73","Type":"ContainerDied","Data":"c85c9e9b44d5c8d90588eae777be2db9c434431255e272a7f45d3604eac5aeac"} Nov 24 13:26:37 crc kubenswrapper[4790]: I1124 13:26:37.303558 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c85c9e9b44d5c8d90588eae777be2db9c434431255e272a7f45d3604eac5aeac" Nov 24 13:26:37 crc kubenswrapper[4790]: I1124 13:26:37.303562 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl" Nov 24 13:26:40 crc kubenswrapper[4790]: I1124 13:26:40.415707 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-kshj8"] Nov 24 13:26:40 crc kubenswrapper[4790]: E1124 13:26:40.416604 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a55d437e-ba5a-4e62-8bea-a9345a428e73" containerName="pull" Nov 24 13:26:40 crc kubenswrapper[4790]: I1124 13:26:40.416623 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="a55d437e-ba5a-4e62-8bea-a9345a428e73" containerName="pull" Nov 24 13:26:40 crc kubenswrapper[4790]: E1124 13:26:40.416638 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a55d437e-ba5a-4e62-8bea-a9345a428e73" containerName="util" Nov 24 13:26:40 crc kubenswrapper[4790]: I1124 13:26:40.416647 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="a55d437e-ba5a-4e62-8bea-a9345a428e73" containerName="util" Nov 24 13:26:40 crc kubenswrapper[4790]: E1124 13:26:40.416657 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b319457e-1e71-442a-908f-2d003bcddbda" containerName="extract-content" Nov 24 13:26:40 crc kubenswrapper[4790]: I1124 13:26:40.416666 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="b319457e-1e71-442a-908f-2d003bcddbda" containerName="extract-content" Nov 24 13:26:40 crc kubenswrapper[4790]: E1124 13:26:40.416685 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b319457e-1e71-442a-908f-2d003bcddbda" containerName="extract-utilities" Nov 24 13:26:40 crc kubenswrapper[4790]: I1124 13:26:40.416694 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="b319457e-1e71-442a-908f-2d003bcddbda" containerName="extract-utilities" Nov 24 13:26:40 crc kubenswrapper[4790]: E1124 13:26:40.416712 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b319457e-1e71-442a-908f-2d003bcddbda" containerName="registry-server" Nov 24 13:26:40 crc kubenswrapper[4790]: I1124 13:26:40.416720 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="b319457e-1e71-442a-908f-2d003bcddbda" containerName="registry-server" Nov 24 13:26:40 crc kubenswrapper[4790]: E1124 13:26:40.416734 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a55d437e-ba5a-4e62-8bea-a9345a428e73" containerName="extract" Nov 24 13:26:40 crc kubenswrapper[4790]: I1124 13:26:40.416742 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="a55d437e-ba5a-4e62-8bea-a9345a428e73" containerName="extract" Nov 24 13:26:40 crc kubenswrapper[4790]: I1124 13:26:40.416926 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="b319457e-1e71-442a-908f-2d003bcddbda" containerName="registry-server" Nov 24 13:26:40 crc kubenswrapper[4790]: I1124 13:26:40.416964 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="a55d437e-ba5a-4e62-8bea-a9345a428e73" containerName="extract" Nov 24 13:26:40 crc kubenswrapper[4790]: I1124 13:26:40.417704 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-kshj8" Nov 24 13:26:40 crc kubenswrapper[4790]: I1124 13:26:40.421205 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 24 13:26:40 crc kubenswrapper[4790]: I1124 13:26:40.421392 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-j5wmv" Nov 24 13:26:40 crc kubenswrapper[4790]: I1124 13:26:40.425007 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 24 13:26:40 crc kubenswrapper[4790]: I1124 13:26:40.432404 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-kshj8"] Nov 24 13:26:40 crc kubenswrapper[4790]: I1124 13:26:40.446388 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8l9h\" (UniqueName: \"kubernetes.io/projected/9f0af7f5-b65a-47dc-8c85-c452cb493912-kube-api-access-b8l9h\") pod \"nmstate-operator-557fdffb88-kshj8\" (UID: \"9f0af7f5-b65a-47dc-8c85-c452cb493912\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-kshj8" Nov 24 13:26:40 crc kubenswrapper[4790]: I1124 13:26:40.547668 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8l9h\" (UniqueName: \"kubernetes.io/projected/9f0af7f5-b65a-47dc-8c85-c452cb493912-kube-api-access-b8l9h\") pod \"nmstate-operator-557fdffb88-kshj8\" (UID: \"9f0af7f5-b65a-47dc-8c85-c452cb493912\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-kshj8" Nov 24 13:26:40 crc kubenswrapper[4790]: I1124 13:26:40.569299 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8l9h\" (UniqueName: \"kubernetes.io/projected/9f0af7f5-b65a-47dc-8c85-c452cb493912-kube-api-access-b8l9h\") pod \"nmstate-operator-557fdffb88-kshj8\" (UID: \"9f0af7f5-b65a-47dc-8c85-c452cb493912\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-kshj8" Nov 24 13:26:40 crc kubenswrapper[4790]: I1124 13:26:40.752962 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-kshj8" Nov 24 13:26:41 crc kubenswrapper[4790]: I1124 13:26:41.221121 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-kshj8"] Nov 24 13:26:41 crc kubenswrapper[4790]: W1124 13:26:41.247764 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f0af7f5_b65a_47dc_8c85_c452cb493912.slice/crio-8b7ac6be5d994066c9541b294f66471341572b0f978871130ee21ef99e88dd74 WatchSource:0}: Error finding container 8b7ac6be5d994066c9541b294f66471341572b0f978871130ee21ef99e88dd74: Status 404 returned error can't find the container with id 8b7ac6be5d994066c9541b294f66471341572b0f978871130ee21ef99e88dd74 Nov 24 13:26:41 crc kubenswrapper[4790]: I1124 13:26:41.329980 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-kshj8" event={"ID":"9f0af7f5-b65a-47dc-8c85-c452cb493912","Type":"ContainerStarted","Data":"8b7ac6be5d994066c9541b294f66471341572b0f978871130ee21ef99e88dd74"} Nov 24 13:26:43 crc kubenswrapper[4790]: I1124 13:26:43.939191 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:26:43 crc kubenswrapper[4790]: I1124 13:26:43.939705 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:26:43 crc kubenswrapper[4790]: I1124 13:26:43.939834 4790 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 13:26:43 crc kubenswrapper[4790]: I1124 13:26:43.940661 4790 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"039103a6742b65a25f073d9f23e4fb704fe4fd769ff785303a1ff06bb7cfe2d2"} pod="openshift-machine-config-operator/machine-config-daemon-xz49t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 13:26:43 crc kubenswrapper[4790]: I1124 13:26:43.940726 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" containerID="cri-o://039103a6742b65a25f073d9f23e4fb704fe4fd769ff785303a1ff06bb7cfe2d2" gracePeriod=600 Nov 24 13:26:44 crc kubenswrapper[4790]: I1124 13:26:44.368294 4790 generic.go:334] "Generic (PLEG): container finished" podID="0d73b133-48f0-455f-8f6a-742e633f631a" containerID="039103a6742b65a25f073d9f23e4fb704fe4fd769ff785303a1ff06bb7cfe2d2" exitCode=0 Nov 24 13:26:44 crc kubenswrapper[4790]: I1124 13:26:44.368380 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerDied","Data":"039103a6742b65a25f073d9f23e4fb704fe4fd769ff785303a1ff06bb7cfe2d2"} Nov 24 13:26:44 crc kubenswrapper[4790]: I1124 13:26:44.368473 4790 scope.go:117] "RemoveContainer" containerID="dcbd627924cf1dc4266a99e99dc330119809da6d3c320cfad690c481fc6af519" Nov 24 13:26:45 crc kubenswrapper[4790]: I1124 13:26:45.376573 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-kshj8" event={"ID":"9f0af7f5-b65a-47dc-8c85-c452cb493912","Type":"ContainerStarted","Data":"a106d818c35af436fb5470f3d7ab3f64014d50eef0554eaa3ee37722526329a6"} Nov 24 13:26:45 crc kubenswrapper[4790]: I1124 13:26:45.379461 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"e635b1d9a632257918b8c7ec939010f4c4491014f07d31a3ef3de514129b90d0"} Nov 24 13:26:45 crc kubenswrapper[4790]: I1124 13:26:45.397610 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-557fdffb88-kshj8" podStartSLOduration=2.024436636 podStartE2EDuration="5.39757924s" podCreationTimestamp="2025-11-24 13:26:40 +0000 UTC" firstStartedPulling="2025-11-24 13:26:41.2513489 +0000 UTC m=+849.631242572" lastFinishedPulling="2025-11-24 13:26:44.624491514 +0000 UTC m=+853.004385176" observedRunningTime="2025-11-24 13:26:45.395492232 +0000 UTC m=+853.775385904" watchObservedRunningTime="2025-11-24 13:26:45.39757924 +0000 UTC m=+853.777472902" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.292214 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xpccg" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.401765 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-xg8tt"] Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.402615 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-xg8tt" Nov 24 13:26:46 crc kubenswrapper[4790]: W1124 13:26:46.406840 4790 reflector.go:561] object-"openshift-nmstate"/"nmstate-handler-dockercfg-jtsc6": failed to list *v1.Secret: secrets "nmstate-handler-dockercfg-jtsc6" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-nmstate": no relationship found between node 'crc' and this object Nov 24 13:26:46 crc kubenswrapper[4790]: E1124 13:26:46.407251 4790 reflector.go:158] "Unhandled Error" err="object-\"openshift-nmstate\"/\"nmstate-handler-dockercfg-jtsc6\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"nmstate-handler-dockercfg-jtsc6\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-nmstate\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.430716 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-xg8tt"] Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.441981 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cq5q\" (UniqueName: \"kubernetes.io/projected/d9e69cb8-ffab-4a0d-be5d-e7d683a44c87-kube-api-access-9cq5q\") pod \"nmstate-metrics-5dcf9c57c5-xg8tt\" (UID: \"d9e69cb8-ffab-4a0d-be5d-e7d683a44c87\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-xg8tt" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.449181 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-kbn22"] Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.450461 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-kbn22" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.452343 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-ggp8t"] Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.456903 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-ggp8t" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.468674 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.477092 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-ggp8t"] Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.543173 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/da4996e7-f19c-47e9-a75c-000e4248dc2d-dbus-socket\") pod \"nmstate-handler-kbn22\" (UID: \"da4996e7-f19c-47e9-a75c-000e4248dc2d\") " pod="openshift-nmstate/nmstate-handler-kbn22" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.543240 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56flq\" (UniqueName: \"kubernetes.io/projected/c71e7468-3526-41e1-87bd-09809fe6ae07-kube-api-access-56flq\") pod \"nmstate-webhook-6b89b748d8-ggp8t\" (UID: \"c71e7468-3526-41e1-87bd-09809fe6ae07\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-ggp8t" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.543293 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cq5q\" (UniqueName: \"kubernetes.io/projected/d9e69cb8-ffab-4a0d-be5d-e7d683a44c87-kube-api-access-9cq5q\") pod \"nmstate-metrics-5dcf9c57c5-xg8tt\" (UID: \"d9e69cb8-ffab-4a0d-be5d-e7d683a44c87\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-xg8tt" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.543329 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmcv2\" (UniqueName: \"kubernetes.io/projected/da4996e7-f19c-47e9-a75c-000e4248dc2d-kube-api-access-pmcv2\") pod \"nmstate-handler-kbn22\" (UID: \"da4996e7-f19c-47e9-a75c-000e4248dc2d\") " pod="openshift-nmstate/nmstate-handler-kbn22" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.543377 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/da4996e7-f19c-47e9-a75c-000e4248dc2d-nmstate-lock\") pod \"nmstate-handler-kbn22\" (UID: \"da4996e7-f19c-47e9-a75c-000e4248dc2d\") " pod="openshift-nmstate/nmstate-handler-kbn22" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.543400 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/c71e7468-3526-41e1-87bd-09809fe6ae07-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-ggp8t\" (UID: \"c71e7468-3526-41e1-87bd-09809fe6ae07\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-ggp8t" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.543427 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/da4996e7-f19c-47e9-a75c-000e4248dc2d-ovs-socket\") pod \"nmstate-handler-kbn22\" (UID: \"da4996e7-f19c-47e9-a75c-000e4248dc2d\") " pod="openshift-nmstate/nmstate-handler-kbn22" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.587862 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cq5q\" (UniqueName: \"kubernetes.io/projected/d9e69cb8-ffab-4a0d-be5d-e7d683a44c87-kube-api-access-9cq5q\") pod \"nmstate-metrics-5dcf9c57c5-xg8tt\" (UID: \"d9e69cb8-ffab-4a0d-be5d-e7d683a44c87\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-xg8tt" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.618943 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-l2ltp"] Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.619664 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-l2ltp" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.624132 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.624340 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-d2gh2" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.633964 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.646220 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/da4996e7-f19c-47e9-a75c-000e4248dc2d-nmstate-lock\") pod \"nmstate-handler-kbn22\" (UID: \"da4996e7-f19c-47e9-a75c-000e4248dc2d\") " pod="openshift-nmstate/nmstate-handler-kbn22" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.646296 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/c71e7468-3526-41e1-87bd-09809fe6ae07-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-ggp8t\" (UID: \"c71e7468-3526-41e1-87bd-09809fe6ae07\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-ggp8t" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.646341 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/da4996e7-f19c-47e9-a75c-000e4248dc2d-ovs-socket\") pod \"nmstate-handler-kbn22\" (UID: \"da4996e7-f19c-47e9-a75c-000e4248dc2d\") " pod="openshift-nmstate/nmstate-handler-kbn22" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.646383 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/e6ee737d-a089-4ecb-a3a8-34ce580b84aa-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-l2ltp\" (UID: \"e6ee737d-a089-4ecb-a3a8-34ce580b84aa\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-l2ltp" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.646425 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/e6ee737d-a089-4ecb-a3a8-34ce580b84aa-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-l2ltp\" (UID: \"e6ee737d-a089-4ecb-a3a8-34ce580b84aa\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-l2ltp" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.646471 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/da4996e7-f19c-47e9-a75c-000e4248dc2d-dbus-socket\") pod \"nmstate-handler-kbn22\" (UID: \"da4996e7-f19c-47e9-a75c-000e4248dc2d\") " pod="openshift-nmstate/nmstate-handler-kbn22" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.646505 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56flq\" (UniqueName: \"kubernetes.io/projected/c71e7468-3526-41e1-87bd-09809fe6ae07-kube-api-access-56flq\") pod \"nmstate-webhook-6b89b748d8-ggp8t\" (UID: \"c71e7468-3526-41e1-87bd-09809fe6ae07\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-ggp8t" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.646549 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6kwp\" (UniqueName: \"kubernetes.io/projected/e6ee737d-a089-4ecb-a3a8-34ce580b84aa-kube-api-access-s6kwp\") pod \"nmstate-console-plugin-5874bd7bc5-l2ltp\" (UID: \"e6ee737d-a089-4ecb-a3a8-34ce580b84aa\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-l2ltp" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.646593 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmcv2\" (UniqueName: \"kubernetes.io/projected/da4996e7-f19c-47e9-a75c-000e4248dc2d-kube-api-access-pmcv2\") pod \"nmstate-handler-kbn22\" (UID: \"da4996e7-f19c-47e9-a75c-000e4248dc2d\") " pod="openshift-nmstate/nmstate-handler-kbn22" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.647246 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/da4996e7-f19c-47e9-a75c-000e4248dc2d-nmstate-lock\") pod \"nmstate-handler-kbn22\" (UID: \"da4996e7-f19c-47e9-a75c-000e4248dc2d\") " pod="openshift-nmstate/nmstate-handler-kbn22" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.648019 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/da4996e7-f19c-47e9-a75c-000e4248dc2d-ovs-socket\") pod \"nmstate-handler-kbn22\" (UID: \"da4996e7-f19c-47e9-a75c-000e4248dc2d\") " pod="openshift-nmstate/nmstate-handler-kbn22" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.648791 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/da4996e7-f19c-47e9-a75c-000e4248dc2d-dbus-socket\") pod \"nmstate-handler-kbn22\" (UID: \"da4996e7-f19c-47e9-a75c-000e4248dc2d\") " pod="openshift-nmstate/nmstate-handler-kbn22" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.659559 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-l2ltp"] Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.677228 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/c71e7468-3526-41e1-87bd-09809fe6ae07-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-ggp8t\" (UID: \"c71e7468-3526-41e1-87bd-09809fe6ae07\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-ggp8t" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.678108 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56flq\" (UniqueName: \"kubernetes.io/projected/c71e7468-3526-41e1-87bd-09809fe6ae07-kube-api-access-56flq\") pod \"nmstate-webhook-6b89b748d8-ggp8t\" (UID: \"c71e7468-3526-41e1-87bd-09809fe6ae07\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-ggp8t" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.689056 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmcv2\" (UniqueName: \"kubernetes.io/projected/da4996e7-f19c-47e9-a75c-000e4248dc2d-kube-api-access-pmcv2\") pod \"nmstate-handler-kbn22\" (UID: \"da4996e7-f19c-47e9-a75c-000e4248dc2d\") " pod="openshift-nmstate/nmstate-handler-kbn22" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.747911 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/e6ee737d-a089-4ecb-a3a8-34ce580b84aa-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-l2ltp\" (UID: \"e6ee737d-a089-4ecb-a3a8-34ce580b84aa\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-l2ltp" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.747979 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/e6ee737d-a089-4ecb-a3a8-34ce580b84aa-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-l2ltp\" (UID: \"e6ee737d-a089-4ecb-a3a8-34ce580b84aa\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-l2ltp" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.748028 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6kwp\" (UniqueName: \"kubernetes.io/projected/e6ee737d-a089-4ecb-a3a8-34ce580b84aa-kube-api-access-s6kwp\") pod \"nmstate-console-plugin-5874bd7bc5-l2ltp\" (UID: \"e6ee737d-a089-4ecb-a3a8-34ce580b84aa\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-l2ltp" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.749429 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/e6ee737d-a089-4ecb-a3a8-34ce580b84aa-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-l2ltp\" (UID: \"e6ee737d-a089-4ecb-a3a8-34ce580b84aa\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-l2ltp" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.760481 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/e6ee737d-a089-4ecb-a3a8-34ce580b84aa-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-l2ltp\" (UID: \"e6ee737d-a089-4ecb-a3a8-34ce580b84aa\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-l2ltp" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.772137 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6kwp\" (UniqueName: \"kubernetes.io/projected/e6ee737d-a089-4ecb-a3a8-34ce580b84aa-kube-api-access-s6kwp\") pod \"nmstate-console-plugin-5874bd7bc5-l2ltp\" (UID: \"e6ee737d-a089-4ecb-a3a8-34ce580b84aa\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-l2ltp" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.825361 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-59b6c6f859-nsktt"] Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.826591 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-59b6c6f859-nsktt" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.849450 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3941a5aa-ea4c-400a-8488-8339af3b122b-trusted-ca-bundle\") pod \"console-59b6c6f859-nsktt\" (UID: \"3941a5aa-ea4c-400a-8488-8339af3b122b\") " pod="openshift-console/console-59b6c6f859-nsktt" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.849515 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3941a5aa-ea4c-400a-8488-8339af3b122b-console-oauth-config\") pod \"console-59b6c6f859-nsktt\" (UID: \"3941a5aa-ea4c-400a-8488-8339af3b122b\") " pod="openshift-console/console-59b6c6f859-nsktt" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.849547 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3941a5aa-ea4c-400a-8488-8339af3b122b-console-config\") pod \"console-59b6c6f859-nsktt\" (UID: \"3941a5aa-ea4c-400a-8488-8339af3b122b\") " pod="openshift-console/console-59b6c6f859-nsktt" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.849577 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3941a5aa-ea4c-400a-8488-8339af3b122b-service-ca\") pod \"console-59b6c6f859-nsktt\" (UID: \"3941a5aa-ea4c-400a-8488-8339af3b122b\") " pod="openshift-console/console-59b6c6f859-nsktt" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.849602 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3941a5aa-ea4c-400a-8488-8339af3b122b-console-serving-cert\") pod \"console-59b6c6f859-nsktt\" (UID: \"3941a5aa-ea4c-400a-8488-8339af3b122b\") " pod="openshift-console/console-59b6c6f859-nsktt" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.849868 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3941a5aa-ea4c-400a-8488-8339af3b122b-oauth-serving-cert\") pod \"console-59b6c6f859-nsktt\" (UID: \"3941a5aa-ea4c-400a-8488-8339af3b122b\") " pod="openshift-console/console-59b6c6f859-nsktt" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.850090 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftx5z\" (UniqueName: \"kubernetes.io/projected/3941a5aa-ea4c-400a-8488-8339af3b122b-kube-api-access-ftx5z\") pod \"console-59b6c6f859-nsktt\" (UID: \"3941a5aa-ea4c-400a-8488-8339af3b122b\") " pod="openshift-console/console-59b6c6f859-nsktt" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.866355 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-59b6c6f859-nsktt"] Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.952207 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3941a5aa-ea4c-400a-8488-8339af3b122b-oauth-serving-cert\") pod \"console-59b6c6f859-nsktt\" (UID: \"3941a5aa-ea4c-400a-8488-8339af3b122b\") " pod="openshift-console/console-59b6c6f859-nsktt" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.952330 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftx5z\" (UniqueName: \"kubernetes.io/projected/3941a5aa-ea4c-400a-8488-8339af3b122b-kube-api-access-ftx5z\") pod \"console-59b6c6f859-nsktt\" (UID: \"3941a5aa-ea4c-400a-8488-8339af3b122b\") " pod="openshift-console/console-59b6c6f859-nsktt" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.952391 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3941a5aa-ea4c-400a-8488-8339af3b122b-trusted-ca-bundle\") pod \"console-59b6c6f859-nsktt\" (UID: \"3941a5aa-ea4c-400a-8488-8339af3b122b\") " pod="openshift-console/console-59b6c6f859-nsktt" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.952427 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3941a5aa-ea4c-400a-8488-8339af3b122b-console-oauth-config\") pod \"console-59b6c6f859-nsktt\" (UID: \"3941a5aa-ea4c-400a-8488-8339af3b122b\") " pod="openshift-console/console-59b6c6f859-nsktt" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.952481 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3941a5aa-ea4c-400a-8488-8339af3b122b-console-config\") pod \"console-59b6c6f859-nsktt\" (UID: \"3941a5aa-ea4c-400a-8488-8339af3b122b\") " pod="openshift-console/console-59b6c6f859-nsktt" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.952508 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3941a5aa-ea4c-400a-8488-8339af3b122b-service-ca\") pod \"console-59b6c6f859-nsktt\" (UID: \"3941a5aa-ea4c-400a-8488-8339af3b122b\") " pod="openshift-console/console-59b6c6f859-nsktt" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.952532 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3941a5aa-ea4c-400a-8488-8339af3b122b-console-serving-cert\") pod \"console-59b6c6f859-nsktt\" (UID: \"3941a5aa-ea4c-400a-8488-8339af3b122b\") " pod="openshift-console/console-59b6c6f859-nsktt" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.953783 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3941a5aa-ea4c-400a-8488-8339af3b122b-oauth-serving-cert\") pod \"console-59b6c6f859-nsktt\" (UID: \"3941a5aa-ea4c-400a-8488-8339af3b122b\") " pod="openshift-console/console-59b6c6f859-nsktt" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.954157 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3941a5aa-ea4c-400a-8488-8339af3b122b-console-config\") pod \"console-59b6c6f859-nsktt\" (UID: \"3941a5aa-ea4c-400a-8488-8339af3b122b\") " pod="openshift-console/console-59b6c6f859-nsktt" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.954463 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-l2ltp" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.954493 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3941a5aa-ea4c-400a-8488-8339af3b122b-trusted-ca-bundle\") pod \"console-59b6c6f859-nsktt\" (UID: \"3941a5aa-ea4c-400a-8488-8339af3b122b\") " pod="openshift-console/console-59b6c6f859-nsktt" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.955729 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3941a5aa-ea4c-400a-8488-8339af3b122b-service-ca\") pod \"console-59b6c6f859-nsktt\" (UID: \"3941a5aa-ea4c-400a-8488-8339af3b122b\") " pod="openshift-console/console-59b6c6f859-nsktt" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.957904 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3941a5aa-ea4c-400a-8488-8339af3b122b-console-serving-cert\") pod \"console-59b6c6f859-nsktt\" (UID: \"3941a5aa-ea4c-400a-8488-8339af3b122b\") " pod="openshift-console/console-59b6c6f859-nsktt" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.959686 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3941a5aa-ea4c-400a-8488-8339af3b122b-console-oauth-config\") pod \"console-59b6c6f859-nsktt\" (UID: \"3941a5aa-ea4c-400a-8488-8339af3b122b\") " pod="openshift-console/console-59b6c6f859-nsktt" Nov 24 13:26:46 crc kubenswrapper[4790]: I1124 13:26:46.984396 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftx5z\" (UniqueName: \"kubernetes.io/projected/3941a5aa-ea4c-400a-8488-8339af3b122b-kube-api-access-ftx5z\") pod \"console-59b6c6f859-nsktt\" (UID: \"3941a5aa-ea4c-400a-8488-8339af3b122b\") " pod="openshift-console/console-59b6c6f859-nsktt" Nov 24 13:26:47 crc kubenswrapper[4790]: I1124 13:26:47.152420 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-59b6c6f859-nsktt" Nov 24 13:26:47 crc kubenswrapper[4790]: I1124 13:26:47.394243 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-59b6c6f859-nsktt"] Nov 24 13:26:47 crc kubenswrapper[4790]: W1124 13:26:47.407399 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3941a5aa_ea4c_400a_8488_8339af3b122b.slice/crio-02e793a60a1be53a8f81de23ba1c80e51db240158a8973ed5abeddf44a336d7a WatchSource:0}: Error finding container 02e793a60a1be53a8f81de23ba1c80e51db240158a8973ed5abeddf44a336d7a: Status 404 returned error can't find the container with id 02e793a60a1be53a8f81de23ba1c80e51db240158a8973ed5abeddf44a336d7a Nov 24 13:26:47 crc kubenswrapper[4790]: I1124 13:26:47.473520 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-l2ltp"] Nov 24 13:26:47 crc kubenswrapper[4790]: W1124 13:26:47.478290 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6ee737d_a089_4ecb_a3a8_34ce580b84aa.slice/crio-530d00efe601c439a7cc3822c3704f3d696f56d7b800c1a41e73c47494584136 WatchSource:0}: Error finding container 530d00efe601c439a7cc3822c3704f3d696f56d7b800c1a41e73c47494584136: Status 404 returned error can't find the container with id 530d00efe601c439a7cc3822c3704f3d696f56d7b800c1a41e73c47494584136 Nov 24 13:26:47 crc kubenswrapper[4790]: I1124 13:26:47.509318 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-jtsc6" Nov 24 13:26:47 crc kubenswrapper[4790]: I1124 13:26:47.513629 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-xg8tt" Nov 24 13:26:47 crc kubenswrapper[4790]: I1124 13:26:47.514805 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-ggp8t" Nov 24 13:26:47 crc kubenswrapper[4790]: I1124 13:26:47.515881 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-kbn22" Nov 24 13:26:47 crc kubenswrapper[4790]: I1124 13:26:47.920960 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-xg8tt"] Nov 24 13:26:48 crc kubenswrapper[4790]: I1124 13:26:48.038203 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-ggp8t"] Nov 24 13:26:48 crc kubenswrapper[4790]: W1124 13:26:48.046016 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc71e7468_3526_41e1_87bd_09809fe6ae07.slice/crio-a5fb58c547869420872b6d195018eaf1fa88511f326af93f111a5902e1d95686 WatchSource:0}: Error finding container a5fb58c547869420872b6d195018eaf1fa88511f326af93f111a5902e1d95686: Status 404 returned error can't find the container with id a5fb58c547869420872b6d195018eaf1fa88511f326af93f111a5902e1d95686 Nov 24 13:26:48 crc kubenswrapper[4790]: I1124 13:26:48.106280 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xpccg"] Nov 24 13:26:48 crc kubenswrapper[4790]: I1124 13:26:48.106630 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xpccg" podUID="f0dcf597-fdbd-418c-b739-4e61f2e9aa29" containerName="registry-server" containerID="cri-o://641a200a06f07ef7d422dfbc535f688ded397aa7214b1fa3af28849d02e4de2f" gracePeriod=2 Nov 24 13:26:48 crc kubenswrapper[4790]: I1124 13:26:48.403456 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-l2ltp" event={"ID":"e6ee737d-a089-4ecb-a3a8-34ce580b84aa","Type":"ContainerStarted","Data":"530d00efe601c439a7cc3822c3704f3d696f56d7b800c1a41e73c47494584136"} Nov 24 13:26:48 crc kubenswrapper[4790]: I1124 13:26:48.404819 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-kbn22" event={"ID":"da4996e7-f19c-47e9-a75c-000e4248dc2d","Type":"ContainerStarted","Data":"531e738e0b574108123fe2d6337e820cc44d41c68b67ee5279e7607732ef00af"} Nov 24 13:26:48 crc kubenswrapper[4790]: I1124 13:26:48.406278 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-xg8tt" event={"ID":"d9e69cb8-ffab-4a0d-be5d-e7d683a44c87","Type":"ContainerStarted","Data":"9a783a5ff62d1907681923119c822f48e0e05ae144b5248f66dd6131ce562195"} Nov 24 13:26:48 crc kubenswrapper[4790]: I1124 13:26:48.407549 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-ggp8t" event={"ID":"c71e7468-3526-41e1-87bd-09809fe6ae07","Type":"ContainerStarted","Data":"a5fb58c547869420872b6d195018eaf1fa88511f326af93f111a5902e1d95686"} Nov 24 13:26:48 crc kubenswrapper[4790]: I1124 13:26:48.409328 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-59b6c6f859-nsktt" event={"ID":"3941a5aa-ea4c-400a-8488-8339af3b122b","Type":"ContainerStarted","Data":"aebed30e5f8999899e63b2067c49338cfd31d59177134952bc1d471018b217a1"} Nov 24 13:26:48 crc kubenswrapper[4790]: I1124 13:26:48.409388 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-59b6c6f859-nsktt" event={"ID":"3941a5aa-ea4c-400a-8488-8339af3b122b","Type":"ContainerStarted","Data":"02e793a60a1be53a8f81de23ba1c80e51db240158a8973ed5abeddf44a336d7a"} Nov 24 13:26:48 crc kubenswrapper[4790]: I1124 13:26:48.411818 4790 generic.go:334] "Generic (PLEG): container finished" podID="f0dcf597-fdbd-418c-b739-4e61f2e9aa29" containerID="641a200a06f07ef7d422dfbc535f688ded397aa7214b1fa3af28849d02e4de2f" exitCode=0 Nov 24 13:26:48 crc kubenswrapper[4790]: I1124 13:26:48.411902 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xpccg" event={"ID":"f0dcf597-fdbd-418c-b739-4e61f2e9aa29","Type":"ContainerDied","Data":"641a200a06f07ef7d422dfbc535f688ded397aa7214b1fa3af28849d02e4de2f"} Nov 24 13:26:48 crc kubenswrapper[4790]: I1124 13:26:48.432157 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-59b6c6f859-nsktt" podStartSLOduration=2.432124613 podStartE2EDuration="2.432124613s" podCreationTimestamp="2025-11-24 13:26:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:26:48.427527005 +0000 UTC m=+856.807420687" watchObservedRunningTime="2025-11-24 13:26:48.432124613 +0000 UTC m=+856.812018275" Nov 24 13:26:48 crc kubenswrapper[4790]: I1124 13:26:48.468431 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xpccg" Nov 24 13:26:48 crc kubenswrapper[4790]: I1124 13:26:48.481009 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0dcf597-fdbd-418c-b739-4e61f2e9aa29-catalog-content\") pod \"f0dcf597-fdbd-418c-b739-4e61f2e9aa29\" (UID: \"f0dcf597-fdbd-418c-b739-4e61f2e9aa29\") " Nov 24 13:26:48 crc kubenswrapper[4790]: I1124 13:26:48.481190 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0dcf597-fdbd-418c-b739-4e61f2e9aa29-utilities\") pod \"f0dcf597-fdbd-418c-b739-4e61f2e9aa29\" (UID: \"f0dcf597-fdbd-418c-b739-4e61f2e9aa29\") " Nov 24 13:26:48 crc kubenswrapper[4790]: I1124 13:26:48.481314 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dhh4\" (UniqueName: \"kubernetes.io/projected/f0dcf597-fdbd-418c-b739-4e61f2e9aa29-kube-api-access-2dhh4\") pod \"f0dcf597-fdbd-418c-b739-4e61f2e9aa29\" (UID: \"f0dcf597-fdbd-418c-b739-4e61f2e9aa29\") " Nov 24 13:26:48 crc kubenswrapper[4790]: I1124 13:26:48.483311 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0dcf597-fdbd-418c-b739-4e61f2e9aa29-utilities" (OuterVolumeSpecName: "utilities") pod "f0dcf597-fdbd-418c-b739-4e61f2e9aa29" (UID: "f0dcf597-fdbd-418c-b739-4e61f2e9aa29"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:26:48 crc kubenswrapper[4790]: I1124 13:26:48.497130 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0dcf597-fdbd-418c-b739-4e61f2e9aa29-kube-api-access-2dhh4" (OuterVolumeSpecName: "kube-api-access-2dhh4") pod "f0dcf597-fdbd-418c-b739-4e61f2e9aa29" (UID: "f0dcf597-fdbd-418c-b739-4e61f2e9aa29"). InnerVolumeSpecName "kube-api-access-2dhh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:26:48 crc kubenswrapper[4790]: I1124 13:26:48.542918 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0dcf597-fdbd-418c-b739-4e61f2e9aa29-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f0dcf597-fdbd-418c-b739-4e61f2e9aa29" (UID: "f0dcf597-fdbd-418c-b739-4e61f2e9aa29"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:26:48 crc kubenswrapper[4790]: I1124 13:26:48.582925 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0dcf597-fdbd-418c-b739-4e61f2e9aa29-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:48 crc kubenswrapper[4790]: I1124 13:26:48.583001 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dhh4\" (UniqueName: \"kubernetes.io/projected/f0dcf597-fdbd-418c-b739-4e61f2e9aa29-kube-api-access-2dhh4\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:48 crc kubenswrapper[4790]: I1124 13:26:48.583015 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0dcf597-fdbd-418c-b739-4e61f2e9aa29-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:49 crc kubenswrapper[4790]: I1124 13:26:49.424081 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xpccg" event={"ID":"f0dcf597-fdbd-418c-b739-4e61f2e9aa29","Type":"ContainerDied","Data":"0b89613db5cc26052a72535d88331938dcdd5c9772f80c19ad239a4f3c6e3e03"} Nov 24 13:26:49 crc kubenswrapper[4790]: I1124 13:26:49.424542 4790 scope.go:117] "RemoveContainer" containerID="641a200a06f07ef7d422dfbc535f688ded397aa7214b1fa3af28849d02e4de2f" Nov 24 13:26:49 crc kubenswrapper[4790]: I1124 13:26:49.424131 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xpccg" Nov 24 13:26:49 crc kubenswrapper[4790]: I1124 13:26:49.455082 4790 scope.go:117] "RemoveContainer" containerID="7578a1d4ef01aa582c4d05cfb85e0fc6d8e242e9f072f7eb02df1b18b2582492" Nov 24 13:26:49 crc kubenswrapper[4790]: I1124 13:26:49.467385 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xpccg"] Nov 24 13:26:49 crc kubenswrapper[4790]: I1124 13:26:49.473239 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xpccg"] Nov 24 13:26:49 crc kubenswrapper[4790]: I1124 13:26:49.507316 4790 scope.go:117] "RemoveContainer" containerID="1e91f68484302d499351f61140294e0fce12da51c309294fbb09cf32dcfe6aff" Nov 24 13:26:50 crc kubenswrapper[4790]: I1124 13:26:50.323546 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0dcf597-fdbd-418c-b739-4e61f2e9aa29" path="/var/lib/kubelet/pods/f0dcf597-fdbd-418c-b739-4e61f2e9aa29/volumes" Nov 24 13:26:52 crc kubenswrapper[4790]: I1124 13:26:52.446717 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-l2ltp" event={"ID":"e6ee737d-a089-4ecb-a3a8-34ce580b84aa","Type":"ContainerStarted","Data":"cc664079666d034b76d183bba5c6127d346b005a09b9bc95a4195336f5420934"} Nov 24 13:26:52 crc kubenswrapper[4790]: I1124 13:26:52.448980 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-kbn22" event={"ID":"da4996e7-f19c-47e9-a75c-000e4248dc2d","Type":"ContainerStarted","Data":"b8ff9c6f22b69ad8741b9006cd5be513ded8c81257583fcaed8d8188fc04fb76"} Nov 24 13:26:52 crc kubenswrapper[4790]: I1124 13:26:52.449164 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-kbn22" Nov 24 13:26:52 crc kubenswrapper[4790]: I1124 13:26:52.452812 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-xg8tt" event={"ID":"d9e69cb8-ffab-4a0d-be5d-e7d683a44c87","Type":"ContainerStarted","Data":"d6130ddaa8a683cf2a1b7daba3f4bf43e20ac18e476e659e97f5f9ff4b148941"} Nov 24 13:26:52 crc kubenswrapper[4790]: I1124 13:26:52.456161 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-ggp8t" event={"ID":"c71e7468-3526-41e1-87bd-09809fe6ae07","Type":"ContainerStarted","Data":"320792261f5fd430cd812178ed92a57040dc16025902861da63c2557fe77d875"} Nov 24 13:26:52 crc kubenswrapper[4790]: I1124 13:26:52.456459 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-ggp8t" Nov 24 13:26:52 crc kubenswrapper[4790]: I1124 13:26:52.466693 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-l2ltp" podStartSLOduration=2.740807723 podStartE2EDuration="6.466662599s" podCreationTimestamp="2025-11-24 13:26:46 +0000 UTC" firstStartedPulling="2025-11-24 13:26:47.481175563 +0000 UTC m=+855.861069225" lastFinishedPulling="2025-11-24 13:26:51.207030439 +0000 UTC m=+859.586924101" observedRunningTime="2025-11-24 13:26:52.462246516 +0000 UTC m=+860.842140188" watchObservedRunningTime="2025-11-24 13:26:52.466662599 +0000 UTC m=+860.846556261" Nov 24 13:26:52 crc kubenswrapper[4790]: I1124 13:26:52.492492 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-kbn22" podStartSLOduration=2.851748325 podStartE2EDuration="6.492455006s" podCreationTimestamp="2025-11-24 13:26:46 +0000 UTC" firstStartedPulling="2025-11-24 13:26:47.597592178 +0000 UTC m=+855.977485840" lastFinishedPulling="2025-11-24 13:26:51.238298858 +0000 UTC m=+859.618192521" observedRunningTime="2025-11-24 13:26:52.488438624 +0000 UTC m=+860.868332296" watchObservedRunningTime="2025-11-24 13:26:52.492455006 +0000 UTC m=+860.872348668" Nov 24 13:26:52 crc kubenswrapper[4790]: I1124 13:26:52.515802 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-ggp8t" podStartSLOduration=3.3591397179999998 podStartE2EDuration="6.515772534s" podCreationTimestamp="2025-11-24 13:26:46 +0000 UTC" firstStartedPulling="2025-11-24 13:26:48.050761274 +0000 UTC m=+856.430654936" lastFinishedPulling="2025-11-24 13:26:51.20739409 +0000 UTC m=+859.587287752" observedRunningTime="2025-11-24 13:26:52.51059236 +0000 UTC m=+860.890486022" watchObservedRunningTime="2025-11-24 13:26:52.515772534 +0000 UTC m=+860.895666196" Nov 24 13:26:56 crc kubenswrapper[4790]: I1124 13:26:56.490858 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-xg8tt" event={"ID":"d9e69cb8-ffab-4a0d-be5d-e7d683a44c87","Type":"ContainerStarted","Data":"93272e23b39ad8ed797310c2dbb6a93b2260d4032ec43df62b54a4d7e37859b0"} Nov 24 13:26:56 crc kubenswrapper[4790]: I1124 13:26:56.511734 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-xg8tt" podStartSLOduration=2.858092173 podStartE2EDuration="10.511699717s" podCreationTimestamp="2025-11-24 13:26:46 +0000 UTC" firstStartedPulling="2025-11-24 13:26:47.929057301 +0000 UTC m=+856.308950963" lastFinishedPulling="2025-11-24 13:26:55.582664845 +0000 UTC m=+863.962558507" observedRunningTime="2025-11-24 13:26:56.509132136 +0000 UTC m=+864.889025808" watchObservedRunningTime="2025-11-24 13:26:56.511699717 +0000 UTC m=+864.891593399" Nov 24 13:26:57 crc kubenswrapper[4790]: I1124 13:26:57.154295 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-59b6c6f859-nsktt" Nov 24 13:26:57 crc kubenswrapper[4790]: I1124 13:26:57.154368 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-59b6c6f859-nsktt" Nov 24 13:26:57 crc kubenswrapper[4790]: I1124 13:26:57.159943 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-59b6c6f859-nsktt" Nov 24 13:26:57 crc kubenswrapper[4790]: I1124 13:26:57.504167 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-59b6c6f859-nsktt" Nov 24 13:26:57 crc kubenswrapper[4790]: I1124 13:26:57.574893 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-t5fjs"] Nov 24 13:26:57 crc kubenswrapper[4790]: I1124 13:26:57.587852 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-kbn22" Nov 24 13:27:07 crc kubenswrapper[4790]: I1124 13:27:07.523574 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-ggp8t" Nov 24 13:27:20 crc kubenswrapper[4790]: I1124 13:27:20.616729 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t"] Nov 24 13:27:20 crc kubenswrapper[4790]: E1124 13:27:20.618083 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0dcf597-fdbd-418c-b739-4e61f2e9aa29" containerName="registry-server" Nov 24 13:27:20 crc kubenswrapper[4790]: I1124 13:27:20.618106 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0dcf597-fdbd-418c-b739-4e61f2e9aa29" containerName="registry-server" Nov 24 13:27:20 crc kubenswrapper[4790]: E1124 13:27:20.618127 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0dcf597-fdbd-418c-b739-4e61f2e9aa29" containerName="extract-content" Nov 24 13:27:20 crc kubenswrapper[4790]: I1124 13:27:20.618138 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0dcf597-fdbd-418c-b739-4e61f2e9aa29" containerName="extract-content" Nov 24 13:27:20 crc kubenswrapper[4790]: E1124 13:27:20.618155 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0dcf597-fdbd-418c-b739-4e61f2e9aa29" containerName="extract-utilities" Nov 24 13:27:20 crc kubenswrapper[4790]: I1124 13:27:20.618164 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0dcf597-fdbd-418c-b739-4e61f2e9aa29" containerName="extract-utilities" Nov 24 13:27:20 crc kubenswrapper[4790]: I1124 13:27:20.618316 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0dcf597-fdbd-418c-b739-4e61f2e9aa29" containerName="registry-server" Nov 24 13:27:20 crc kubenswrapper[4790]: I1124 13:27:20.619530 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t" Nov 24 13:27:20 crc kubenswrapper[4790]: I1124 13:27:20.624100 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 13:27:20 crc kubenswrapper[4790]: I1124 13:27:20.647560 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t"] Nov 24 13:27:20 crc kubenswrapper[4790]: I1124 13:27:20.731487 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvfqc\" (UniqueName: \"kubernetes.io/projected/7187360e-55ef-4709-a8b5-4aa49ebdf70e-kube-api-access-kvfqc\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t\" (UID: \"7187360e-55ef-4709-a8b5-4aa49ebdf70e\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t" Nov 24 13:27:20 crc kubenswrapper[4790]: I1124 13:27:20.731944 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7187360e-55ef-4709-a8b5-4aa49ebdf70e-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t\" (UID: \"7187360e-55ef-4709-a8b5-4aa49ebdf70e\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t" Nov 24 13:27:20 crc kubenswrapper[4790]: I1124 13:27:20.732065 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7187360e-55ef-4709-a8b5-4aa49ebdf70e-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t\" (UID: \"7187360e-55ef-4709-a8b5-4aa49ebdf70e\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t" Nov 24 13:27:20 crc kubenswrapper[4790]: I1124 13:27:20.833394 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7187360e-55ef-4709-a8b5-4aa49ebdf70e-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t\" (UID: \"7187360e-55ef-4709-a8b5-4aa49ebdf70e\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t" Nov 24 13:27:20 crc kubenswrapper[4790]: I1124 13:27:20.833528 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvfqc\" (UniqueName: \"kubernetes.io/projected/7187360e-55ef-4709-a8b5-4aa49ebdf70e-kube-api-access-kvfqc\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t\" (UID: \"7187360e-55ef-4709-a8b5-4aa49ebdf70e\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t" Nov 24 13:27:20 crc kubenswrapper[4790]: I1124 13:27:20.833574 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7187360e-55ef-4709-a8b5-4aa49ebdf70e-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t\" (UID: \"7187360e-55ef-4709-a8b5-4aa49ebdf70e\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t" Nov 24 13:27:20 crc kubenswrapper[4790]: I1124 13:27:20.834182 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7187360e-55ef-4709-a8b5-4aa49ebdf70e-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t\" (UID: \"7187360e-55ef-4709-a8b5-4aa49ebdf70e\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t" Nov 24 13:27:20 crc kubenswrapper[4790]: I1124 13:27:20.834302 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7187360e-55ef-4709-a8b5-4aa49ebdf70e-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t\" (UID: \"7187360e-55ef-4709-a8b5-4aa49ebdf70e\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t" Nov 24 13:27:20 crc kubenswrapper[4790]: I1124 13:27:20.857058 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvfqc\" (UniqueName: \"kubernetes.io/projected/7187360e-55ef-4709-a8b5-4aa49ebdf70e-kube-api-access-kvfqc\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t\" (UID: \"7187360e-55ef-4709-a8b5-4aa49ebdf70e\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t" Nov 24 13:27:20 crc kubenswrapper[4790]: I1124 13:27:20.943923 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t" Nov 24 13:27:21 crc kubenswrapper[4790]: I1124 13:27:21.168856 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t"] Nov 24 13:27:21 crc kubenswrapper[4790]: I1124 13:27:21.673170 4790 generic.go:334] "Generic (PLEG): container finished" podID="7187360e-55ef-4709-a8b5-4aa49ebdf70e" containerID="df27472e837d2c3d5235933ea4d663ece222f45e3070558b632ae16e74944f11" exitCode=0 Nov 24 13:27:21 crc kubenswrapper[4790]: I1124 13:27:21.673261 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t" event={"ID":"7187360e-55ef-4709-a8b5-4aa49ebdf70e","Type":"ContainerDied","Data":"df27472e837d2c3d5235933ea4d663ece222f45e3070558b632ae16e74944f11"} Nov 24 13:27:21 crc kubenswrapper[4790]: I1124 13:27:21.673360 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t" event={"ID":"7187360e-55ef-4709-a8b5-4aa49ebdf70e","Type":"ContainerStarted","Data":"07839163e2b5f35a60106d6bd561998893081787d3149b15944e513f48716c66"} Nov 24 13:27:22 crc kubenswrapper[4790]: I1124 13:27:22.642272 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-t5fjs" podUID="aef429a6-aa76-4776-a6e2-30096a5e7ccb" containerName="console" containerID="cri-o://8eaf51275e584017d75742d241cc6c04ee745e5aa9921263e22277a82e17a224" gracePeriod=15 Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.070175 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-t5fjs_aef429a6-aa76-4776-a6e2-30096a5e7ccb/console/0.log" Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.070674 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-t5fjs" Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.175941 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfndz\" (UniqueName: \"kubernetes.io/projected/aef429a6-aa76-4776-a6e2-30096a5e7ccb-kube-api-access-sfndz\") pod \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\" (UID: \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\") " Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.176017 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/aef429a6-aa76-4776-a6e2-30096a5e7ccb-console-config\") pod \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\" (UID: \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\") " Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.176049 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/aef429a6-aa76-4776-a6e2-30096a5e7ccb-service-ca\") pod \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\" (UID: \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\") " Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.176127 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aef429a6-aa76-4776-a6e2-30096a5e7ccb-trusted-ca-bundle\") pod \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\" (UID: \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\") " Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.176162 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/aef429a6-aa76-4776-a6e2-30096a5e7ccb-oauth-serving-cert\") pod \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\" (UID: \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\") " Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.176197 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/aef429a6-aa76-4776-a6e2-30096a5e7ccb-console-oauth-config\") pod \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\" (UID: \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\") " Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.176337 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/aef429a6-aa76-4776-a6e2-30096a5e7ccb-console-serving-cert\") pod \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\" (UID: \"aef429a6-aa76-4776-a6e2-30096a5e7ccb\") " Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.177771 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aef429a6-aa76-4776-a6e2-30096a5e7ccb-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "aef429a6-aa76-4776-a6e2-30096a5e7ccb" (UID: "aef429a6-aa76-4776-a6e2-30096a5e7ccb"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.177818 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aef429a6-aa76-4776-a6e2-30096a5e7ccb-service-ca" (OuterVolumeSpecName: "service-ca") pod "aef429a6-aa76-4776-a6e2-30096a5e7ccb" (UID: "aef429a6-aa76-4776-a6e2-30096a5e7ccb"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.177789 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aef429a6-aa76-4776-a6e2-30096a5e7ccb-console-config" (OuterVolumeSpecName: "console-config") pod "aef429a6-aa76-4776-a6e2-30096a5e7ccb" (UID: "aef429a6-aa76-4776-a6e2-30096a5e7ccb"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.178346 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aef429a6-aa76-4776-a6e2-30096a5e7ccb-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "aef429a6-aa76-4776-a6e2-30096a5e7ccb" (UID: "aef429a6-aa76-4776-a6e2-30096a5e7ccb"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.185674 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aef429a6-aa76-4776-a6e2-30096a5e7ccb-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "aef429a6-aa76-4776-a6e2-30096a5e7ccb" (UID: "aef429a6-aa76-4776-a6e2-30096a5e7ccb"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.186542 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aef429a6-aa76-4776-a6e2-30096a5e7ccb-kube-api-access-sfndz" (OuterVolumeSpecName: "kube-api-access-sfndz") pod "aef429a6-aa76-4776-a6e2-30096a5e7ccb" (UID: "aef429a6-aa76-4776-a6e2-30096a5e7ccb"). InnerVolumeSpecName "kube-api-access-sfndz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.186574 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aef429a6-aa76-4776-a6e2-30096a5e7ccb-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "aef429a6-aa76-4776-a6e2-30096a5e7ccb" (UID: "aef429a6-aa76-4776-a6e2-30096a5e7ccb"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.277783 4790 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/aef429a6-aa76-4776-a6e2-30096a5e7ccb-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.277833 4790 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aef429a6-aa76-4776-a6e2-30096a5e7ccb-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.277855 4790 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/aef429a6-aa76-4776-a6e2-30096a5e7ccb-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.277867 4790 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/aef429a6-aa76-4776-a6e2-30096a5e7ccb-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.277899 4790 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/aef429a6-aa76-4776-a6e2-30096a5e7ccb-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.277917 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfndz\" (UniqueName: \"kubernetes.io/projected/aef429a6-aa76-4776-a6e2-30096a5e7ccb-kube-api-access-sfndz\") on node \"crc\" DevicePath \"\"" Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.277927 4790 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/aef429a6-aa76-4776-a6e2-30096a5e7ccb-console-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:27:23 crc kubenswrapper[4790]: E1124 13:27:23.291046 4790 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7187360e_55ef_4709_a8b5_4aa49ebdf70e.slice/crio-conmon-87801887c1e1d60d881cb42782317936ec2d782d12be988dd567ad85b3bd5bdb.scope\": RecentStats: unable to find data in memory cache]" Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.695195 4790 generic.go:334] "Generic (PLEG): container finished" podID="7187360e-55ef-4709-a8b5-4aa49ebdf70e" containerID="87801887c1e1d60d881cb42782317936ec2d782d12be988dd567ad85b3bd5bdb" exitCode=0 Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.695297 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t" event={"ID":"7187360e-55ef-4709-a8b5-4aa49ebdf70e","Type":"ContainerDied","Data":"87801887c1e1d60d881cb42782317936ec2d782d12be988dd567ad85b3bd5bdb"} Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.708687 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-t5fjs_aef429a6-aa76-4776-a6e2-30096a5e7ccb/console/0.log" Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.708769 4790 generic.go:334] "Generic (PLEG): container finished" podID="aef429a6-aa76-4776-a6e2-30096a5e7ccb" containerID="8eaf51275e584017d75742d241cc6c04ee745e5aa9921263e22277a82e17a224" exitCode=2 Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.708852 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-t5fjs" event={"ID":"aef429a6-aa76-4776-a6e2-30096a5e7ccb","Type":"ContainerDied","Data":"8eaf51275e584017d75742d241cc6c04ee745e5aa9921263e22277a82e17a224"} Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.708961 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-t5fjs" event={"ID":"aef429a6-aa76-4776-a6e2-30096a5e7ccb","Type":"ContainerDied","Data":"0fa23a4ec0d51d080fca5722debd6883bf4e5a08d6775739c39c68f81d3d0f8b"} Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.708990 4790 scope.go:117] "RemoveContainer" containerID="8eaf51275e584017d75742d241cc6c04ee745e5aa9921263e22277a82e17a224" Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.708874 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-t5fjs" Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.749811 4790 scope.go:117] "RemoveContainer" containerID="8eaf51275e584017d75742d241cc6c04ee745e5aa9921263e22277a82e17a224" Nov 24 13:27:23 crc kubenswrapper[4790]: E1124 13:27:23.751733 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8eaf51275e584017d75742d241cc6c04ee745e5aa9921263e22277a82e17a224\": container with ID starting with 8eaf51275e584017d75742d241cc6c04ee745e5aa9921263e22277a82e17a224 not found: ID does not exist" containerID="8eaf51275e584017d75742d241cc6c04ee745e5aa9921263e22277a82e17a224" Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.751821 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8eaf51275e584017d75742d241cc6c04ee745e5aa9921263e22277a82e17a224"} err="failed to get container status \"8eaf51275e584017d75742d241cc6c04ee745e5aa9921263e22277a82e17a224\": rpc error: code = NotFound desc = could not find container \"8eaf51275e584017d75742d241cc6c04ee745e5aa9921263e22277a82e17a224\": container with ID starting with 8eaf51275e584017d75742d241cc6c04ee745e5aa9921263e22277a82e17a224 not found: ID does not exist" Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.761653 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-t5fjs"] Nov 24 13:27:23 crc kubenswrapper[4790]: I1124 13:27:23.761913 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-t5fjs"] Nov 24 13:27:24 crc kubenswrapper[4790]: I1124 13:27:24.327445 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aef429a6-aa76-4776-a6e2-30096a5e7ccb" path="/var/lib/kubelet/pods/aef429a6-aa76-4776-a6e2-30096a5e7ccb/volumes" Nov 24 13:27:24 crc kubenswrapper[4790]: I1124 13:27:24.720753 4790 generic.go:334] "Generic (PLEG): container finished" podID="7187360e-55ef-4709-a8b5-4aa49ebdf70e" containerID="46a57766b698244c64f833d7247cd533d79ae404a999954d4c86dc37b7b39249" exitCode=0 Nov 24 13:27:24 crc kubenswrapper[4790]: I1124 13:27:24.720841 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t" event={"ID":"7187360e-55ef-4709-a8b5-4aa49ebdf70e","Type":"ContainerDied","Data":"46a57766b698244c64f833d7247cd533d79ae404a999954d4c86dc37b7b39249"} Nov 24 13:27:26 crc kubenswrapper[4790]: I1124 13:27:26.012965 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t" Nov 24 13:27:26 crc kubenswrapper[4790]: I1124 13:27:26.125485 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7187360e-55ef-4709-a8b5-4aa49ebdf70e-bundle\") pod \"7187360e-55ef-4709-a8b5-4aa49ebdf70e\" (UID: \"7187360e-55ef-4709-a8b5-4aa49ebdf70e\") " Nov 24 13:27:26 crc kubenswrapper[4790]: I1124 13:27:26.126149 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7187360e-55ef-4709-a8b5-4aa49ebdf70e-util\") pod \"7187360e-55ef-4709-a8b5-4aa49ebdf70e\" (UID: \"7187360e-55ef-4709-a8b5-4aa49ebdf70e\") " Nov 24 13:27:26 crc kubenswrapper[4790]: I1124 13:27:26.126279 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvfqc\" (UniqueName: \"kubernetes.io/projected/7187360e-55ef-4709-a8b5-4aa49ebdf70e-kube-api-access-kvfqc\") pod \"7187360e-55ef-4709-a8b5-4aa49ebdf70e\" (UID: \"7187360e-55ef-4709-a8b5-4aa49ebdf70e\") " Nov 24 13:27:26 crc kubenswrapper[4790]: I1124 13:27:26.127093 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7187360e-55ef-4709-a8b5-4aa49ebdf70e-bundle" (OuterVolumeSpecName: "bundle") pod "7187360e-55ef-4709-a8b5-4aa49ebdf70e" (UID: "7187360e-55ef-4709-a8b5-4aa49ebdf70e"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:27:26 crc kubenswrapper[4790]: I1124 13:27:26.133902 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7187360e-55ef-4709-a8b5-4aa49ebdf70e-kube-api-access-kvfqc" (OuterVolumeSpecName: "kube-api-access-kvfqc") pod "7187360e-55ef-4709-a8b5-4aa49ebdf70e" (UID: "7187360e-55ef-4709-a8b5-4aa49ebdf70e"). InnerVolumeSpecName "kube-api-access-kvfqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:27:26 crc kubenswrapper[4790]: I1124 13:27:26.141749 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7187360e-55ef-4709-a8b5-4aa49ebdf70e-util" (OuterVolumeSpecName: "util") pod "7187360e-55ef-4709-a8b5-4aa49ebdf70e" (UID: "7187360e-55ef-4709-a8b5-4aa49ebdf70e"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:27:26 crc kubenswrapper[4790]: I1124 13:27:26.228351 4790 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7187360e-55ef-4709-a8b5-4aa49ebdf70e-util\") on node \"crc\" DevicePath \"\"" Nov 24 13:27:26 crc kubenswrapper[4790]: I1124 13:27:26.228419 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvfqc\" (UniqueName: \"kubernetes.io/projected/7187360e-55ef-4709-a8b5-4aa49ebdf70e-kube-api-access-kvfqc\") on node \"crc\" DevicePath \"\"" Nov 24 13:27:26 crc kubenswrapper[4790]: I1124 13:27:26.228441 4790 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7187360e-55ef-4709-a8b5-4aa49ebdf70e-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:27:26 crc kubenswrapper[4790]: I1124 13:27:26.738353 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t" event={"ID":"7187360e-55ef-4709-a8b5-4aa49ebdf70e","Type":"ContainerDied","Data":"07839163e2b5f35a60106d6bd561998893081787d3149b15944e513f48716c66"} Nov 24 13:27:26 crc kubenswrapper[4790]: I1124 13:27:26.738421 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07839163e2b5f35a60106d6bd561998893081787d3149b15944e513f48716c66" Nov 24 13:27:26 crc kubenswrapper[4790]: I1124 13:27:26.738434 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t" Nov 24 13:27:36 crc kubenswrapper[4790]: I1124 13:27:36.760721 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-768d985d78-np9x5"] Nov 24 13:27:36 crc kubenswrapper[4790]: E1124 13:27:36.762016 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7187360e-55ef-4709-a8b5-4aa49ebdf70e" containerName="util" Nov 24 13:27:36 crc kubenswrapper[4790]: I1124 13:27:36.762037 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="7187360e-55ef-4709-a8b5-4aa49ebdf70e" containerName="util" Nov 24 13:27:36 crc kubenswrapper[4790]: E1124 13:27:36.762050 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7187360e-55ef-4709-a8b5-4aa49ebdf70e" containerName="extract" Nov 24 13:27:36 crc kubenswrapper[4790]: I1124 13:27:36.762058 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="7187360e-55ef-4709-a8b5-4aa49ebdf70e" containerName="extract" Nov 24 13:27:36 crc kubenswrapper[4790]: E1124 13:27:36.762072 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aef429a6-aa76-4776-a6e2-30096a5e7ccb" containerName="console" Nov 24 13:27:36 crc kubenswrapper[4790]: I1124 13:27:36.762082 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="aef429a6-aa76-4776-a6e2-30096a5e7ccb" containerName="console" Nov 24 13:27:36 crc kubenswrapper[4790]: E1124 13:27:36.762098 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7187360e-55ef-4709-a8b5-4aa49ebdf70e" containerName="pull" Nov 24 13:27:36 crc kubenswrapper[4790]: I1124 13:27:36.762106 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="7187360e-55ef-4709-a8b5-4aa49ebdf70e" containerName="pull" Nov 24 13:27:36 crc kubenswrapper[4790]: I1124 13:27:36.762238 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="aef429a6-aa76-4776-a6e2-30096a5e7ccb" containerName="console" Nov 24 13:27:36 crc kubenswrapper[4790]: I1124 13:27:36.762259 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="7187360e-55ef-4709-a8b5-4aa49ebdf70e" containerName="extract" Nov 24 13:27:36 crc kubenswrapper[4790]: I1124 13:27:36.762793 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-768d985d78-np9x5" Nov 24 13:27:36 crc kubenswrapper[4790]: I1124 13:27:36.766203 4790 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-fhd8v" Nov 24 13:27:36 crc kubenswrapper[4790]: I1124 13:27:36.766204 4790 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 24 13:27:36 crc kubenswrapper[4790]: I1124 13:27:36.766207 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 24 13:27:36 crc kubenswrapper[4790]: I1124 13:27:36.766207 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 24 13:27:36 crc kubenswrapper[4790]: I1124 13:27:36.766550 4790 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 24 13:27:36 crc kubenswrapper[4790]: I1124 13:27:36.785754 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/79948d0b-b705-4c1b-8d13-836359599770-webhook-cert\") pod \"metallb-operator-controller-manager-768d985d78-np9x5\" (UID: \"79948d0b-b705-4c1b-8d13-836359599770\") " pod="metallb-system/metallb-operator-controller-manager-768d985d78-np9x5" Nov 24 13:27:36 crc kubenswrapper[4790]: I1124 13:27:36.785821 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/79948d0b-b705-4c1b-8d13-836359599770-apiservice-cert\") pod \"metallb-operator-controller-manager-768d985d78-np9x5\" (UID: \"79948d0b-b705-4c1b-8d13-836359599770\") " pod="metallb-system/metallb-operator-controller-manager-768d985d78-np9x5" Nov 24 13:27:36 crc kubenswrapper[4790]: I1124 13:27:36.785950 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwkhd\" (UniqueName: \"kubernetes.io/projected/79948d0b-b705-4c1b-8d13-836359599770-kube-api-access-cwkhd\") pod \"metallb-operator-controller-manager-768d985d78-np9x5\" (UID: \"79948d0b-b705-4c1b-8d13-836359599770\") " pod="metallb-system/metallb-operator-controller-manager-768d985d78-np9x5" Nov 24 13:27:36 crc kubenswrapper[4790]: I1124 13:27:36.798217 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-768d985d78-np9x5"] Nov 24 13:27:36 crc kubenswrapper[4790]: I1124 13:27:36.887044 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/79948d0b-b705-4c1b-8d13-836359599770-webhook-cert\") pod \"metallb-operator-controller-manager-768d985d78-np9x5\" (UID: \"79948d0b-b705-4c1b-8d13-836359599770\") " pod="metallb-system/metallb-operator-controller-manager-768d985d78-np9x5" Nov 24 13:27:36 crc kubenswrapper[4790]: I1124 13:27:36.887123 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/79948d0b-b705-4c1b-8d13-836359599770-apiservice-cert\") pod \"metallb-operator-controller-manager-768d985d78-np9x5\" (UID: \"79948d0b-b705-4c1b-8d13-836359599770\") " pod="metallb-system/metallb-operator-controller-manager-768d985d78-np9x5" Nov 24 13:27:36 crc kubenswrapper[4790]: I1124 13:27:36.887167 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwkhd\" (UniqueName: \"kubernetes.io/projected/79948d0b-b705-4c1b-8d13-836359599770-kube-api-access-cwkhd\") pod \"metallb-operator-controller-manager-768d985d78-np9x5\" (UID: \"79948d0b-b705-4c1b-8d13-836359599770\") " pod="metallb-system/metallb-operator-controller-manager-768d985d78-np9x5" Nov 24 13:27:36 crc kubenswrapper[4790]: I1124 13:27:36.896644 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/79948d0b-b705-4c1b-8d13-836359599770-apiservice-cert\") pod \"metallb-operator-controller-manager-768d985d78-np9x5\" (UID: \"79948d0b-b705-4c1b-8d13-836359599770\") " pod="metallb-system/metallb-operator-controller-manager-768d985d78-np9x5" Nov 24 13:27:36 crc kubenswrapper[4790]: I1124 13:27:36.897346 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/79948d0b-b705-4c1b-8d13-836359599770-webhook-cert\") pod \"metallb-operator-controller-manager-768d985d78-np9x5\" (UID: \"79948d0b-b705-4c1b-8d13-836359599770\") " pod="metallb-system/metallb-operator-controller-manager-768d985d78-np9x5" Nov 24 13:27:36 crc kubenswrapper[4790]: I1124 13:27:36.906977 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwkhd\" (UniqueName: \"kubernetes.io/projected/79948d0b-b705-4c1b-8d13-836359599770-kube-api-access-cwkhd\") pod \"metallb-operator-controller-manager-768d985d78-np9x5\" (UID: \"79948d0b-b705-4c1b-8d13-836359599770\") " pod="metallb-system/metallb-operator-controller-manager-768d985d78-np9x5" Nov 24 13:27:37 crc kubenswrapper[4790]: I1124 13:27:37.082984 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-78c98cf574-zvq6f"] Nov 24 13:27:37 crc kubenswrapper[4790]: I1124 13:27:37.084155 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-78c98cf574-zvq6f" Nov 24 13:27:37 crc kubenswrapper[4790]: I1124 13:27:37.084954 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-768d985d78-np9x5" Nov 24 13:27:37 crc kubenswrapper[4790]: I1124 13:27:37.086616 4790 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 24 13:27:37 crc kubenswrapper[4790]: I1124 13:27:37.087402 4790 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-gfchx" Nov 24 13:27:37 crc kubenswrapper[4790]: I1124 13:27:37.090154 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1bdc2295-5ae4-4d34-969d-f0a5251ab904-apiservice-cert\") pod \"metallb-operator-webhook-server-78c98cf574-zvq6f\" (UID: \"1bdc2295-5ae4-4d34-969d-f0a5251ab904\") " pod="metallb-system/metallb-operator-webhook-server-78c98cf574-zvq6f" Nov 24 13:27:37 crc kubenswrapper[4790]: I1124 13:27:37.090245 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8n25\" (UniqueName: \"kubernetes.io/projected/1bdc2295-5ae4-4d34-969d-f0a5251ab904-kube-api-access-n8n25\") pod \"metallb-operator-webhook-server-78c98cf574-zvq6f\" (UID: \"1bdc2295-5ae4-4d34-969d-f0a5251ab904\") " pod="metallb-system/metallb-operator-webhook-server-78c98cf574-zvq6f" Nov 24 13:27:37 crc kubenswrapper[4790]: I1124 13:27:37.090297 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1bdc2295-5ae4-4d34-969d-f0a5251ab904-webhook-cert\") pod \"metallb-operator-webhook-server-78c98cf574-zvq6f\" (UID: \"1bdc2295-5ae4-4d34-969d-f0a5251ab904\") " pod="metallb-system/metallb-operator-webhook-server-78c98cf574-zvq6f" Nov 24 13:27:37 crc kubenswrapper[4790]: I1124 13:27:37.092030 4790 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 24 13:27:37 crc kubenswrapper[4790]: I1124 13:27:37.100156 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-78c98cf574-zvq6f"] Nov 24 13:27:37 crc kubenswrapper[4790]: I1124 13:27:37.191558 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8n25\" (UniqueName: \"kubernetes.io/projected/1bdc2295-5ae4-4d34-969d-f0a5251ab904-kube-api-access-n8n25\") pod \"metallb-operator-webhook-server-78c98cf574-zvq6f\" (UID: \"1bdc2295-5ae4-4d34-969d-f0a5251ab904\") " pod="metallb-system/metallb-operator-webhook-server-78c98cf574-zvq6f" Nov 24 13:27:37 crc kubenswrapper[4790]: I1124 13:27:37.192224 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1bdc2295-5ae4-4d34-969d-f0a5251ab904-webhook-cert\") pod \"metallb-operator-webhook-server-78c98cf574-zvq6f\" (UID: \"1bdc2295-5ae4-4d34-969d-f0a5251ab904\") " pod="metallb-system/metallb-operator-webhook-server-78c98cf574-zvq6f" Nov 24 13:27:37 crc kubenswrapper[4790]: I1124 13:27:37.192398 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1bdc2295-5ae4-4d34-969d-f0a5251ab904-apiservice-cert\") pod \"metallb-operator-webhook-server-78c98cf574-zvq6f\" (UID: \"1bdc2295-5ae4-4d34-969d-f0a5251ab904\") " pod="metallb-system/metallb-operator-webhook-server-78c98cf574-zvq6f" Nov 24 13:27:37 crc kubenswrapper[4790]: I1124 13:27:37.205266 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1bdc2295-5ae4-4d34-969d-f0a5251ab904-apiservice-cert\") pod \"metallb-operator-webhook-server-78c98cf574-zvq6f\" (UID: \"1bdc2295-5ae4-4d34-969d-f0a5251ab904\") " pod="metallb-system/metallb-operator-webhook-server-78c98cf574-zvq6f" Nov 24 13:27:37 crc kubenswrapper[4790]: I1124 13:27:37.206021 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1bdc2295-5ae4-4d34-969d-f0a5251ab904-webhook-cert\") pod \"metallb-operator-webhook-server-78c98cf574-zvq6f\" (UID: \"1bdc2295-5ae4-4d34-969d-f0a5251ab904\") " pod="metallb-system/metallb-operator-webhook-server-78c98cf574-zvq6f" Nov 24 13:27:37 crc kubenswrapper[4790]: I1124 13:27:37.213562 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8n25\" (UniqueName: \"kubernetes.io/projected/1bdc2295-5ae4-4d34-969d-f0a5251ab904-kube-api-access-n8n25\") pod \"metallb-operator-webhook-server-78c98cf574-zvq6f\" (UID: \"1bdc2295-5ae4-4d34-969d-f0a5251ab904\") " pod="metallb-system/metallb-operator-webhook-server-78c98cf574-zvq6f" Nov 24 13:27:37 crc kubenswrapper[4790]: I1124 13:27:37.403994 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-78c98cf574-zvq6f" Nov 24 13:27:37 crc kubenswrapper[4790]: I1124 13:27:37.409207 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-768d985d78-np9x5"] Nov 24 13:27:37 crc kubenswrapper[4790]: I1124 13:27:37.744069 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-78c98cf574-zvq6f"] Nov 24 13:27:37 crc kubenswrapper[4790]: W1124 13:27:37.756081 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1bdc2295_5ae4_4d34_969d_f0a5251ab904.slice/crio-41830ed44205d423266d970341a60431b3e7b87583556a9847b37490da4a38f1 WatchSource:0}: Error finding container 41830ed44205d423266d970341a60431b3e7b87583556a9847b37490da4a38f1: Status 404 returned error can't find the container with id 41830ed44205d423266d970341a60431b3e7b87583556a9847b37490da4a38f1 Nov 24 13:27:37 crc kubenswrapper[4790]: I1124 13:27:37.817937 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-78c98cf574-zvq6f" event={"ID":"1bdc2295-5ae4-4d34-969d-f0a5251ab904","Type":"ContainerStarted","Data":"41830ed44205d423266d970341a60431b3e7b87583556a9847b37490da4a38f1"} Nov 24 13:27:37 crc kubenswrapper[4790]: I1124 13:27:37.819870 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-768d985d78-np9x5" event={"ID":"79948d0b-b705-4c1b-8d13-836359599770","Type":"ContainerStarted","Data":"95fd82b2bbcd66dc2b106e608d47f614d0e6a0cb3a1a8ccb1504bfe6d33f4206"} Nov 24 13:27:42 crc kubenswrapper[4790]: I1124 13:27:42.865325 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-768d985d78-np9x5" event={"ID":"79948d0b-b705-4c1b-8d13-836359599770","Type":"ContainerStarted","Data":"e82c77b6a33b13f5333e57ffc5bdc0e04cc5728051cbf1a984137c1031f95f1a"} Nov 24 13:27:42 crc kubenswrapper[4790]: I1124 13:27:42.866283 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-768d985d78-np9x5" Nov 24 13:27:42 crc kubenswrapper[4790]: I1124 13:27:42.875043 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-78c98cf574-zvq6f" event={"ID":"1bdc2295-5ae4-4d34-969d-f0a5251ab904","Type":"ContainerStarted","Data":"ca7817049228def1350c52eec620d591b1a3b21506454fcc2e898bdd2bcc5019"} Nov 24 13:27:42 crc kubenswrapper[4790]: I1124 13:27:42.875193 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-78c98cf574-zvq6f" Nov 24 13:27:42 crc kubenswrapper[4790]: I1124 13:27:42.901485 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-768d985d78-np9x5" podStartSLOduration=2.263417793 podStartE2EDuration="6.901445712s" podCreationTimestamp="2025-11-24 13:27:36 +0000 UTC" firstStartedPulling="2025-11-24 13:27:37.447251008 +0000 UTC m=+905.827144670" lastFinishedPulling="2025-11-24 13:27:42.085278927 +0000 UTC m=+910.465172589" observedRunningTime="2025-11-24 13:27:42.888392029 +0000 UTC m=+911.268285692" watchObservedRunningTime="2025-11-24 13:27:42.901445712 +0000 UTC m=+911.281339374" Nov 24 13:27:42 crc kubenswrapper[4790]: I1124 13:27:42.916016 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-78c98cf574-zvq6f" podStartSLOduration=1.336102544 podStartE2EDuration="5.915991507s" podCreationTimestamp="2025-11-24 13:27:37 +0000 UTC" firstStartedPulling="2025-11-24 13:27:37.760550906 +0000 UTC m=+906.140444578" lastFinishedPulling="2025-11-24 13:27:42.340439879 +0000 UTC m=+910.720333541" observedRunningTime="2025-11-24 13:27:42.914785073 +0000 UTC m=+911.294678735" watchObservedRunningTime="2025-11-24 13:27:42.915991507 +0000 UTC m=+911.295885169" Nov 24 13:27:57 crc kubenswrapper[4790]: I1124 13:27:57.411511 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-78c98cf574-zvq6f" Nov 24 13:28:17 crc kubenswrapper[4790]: I1124 13:28:17.089478 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-768d985d78-np9x5" Nov 24 13:28:17 crc kubenswrapper[4790]: I1124 13:28:17.890872 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-bvwnn"] Nov 24 13:28:17 crc kubenswrapper[4790]: I1124 13:28:17.894236 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-bvwnn" Nov 24 13:28:17 crc kubenswrapper[4790]: I1124 13:28:17.896335 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-sdghv"] Nov 24 13:28:17 crc kubenswrapper[4790]: I1124 13:28:17.897113 4790 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-lm288" Nov 24 13:28:17 crc kubenswrapper[4790]: I1124 13:28:17.897437 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-sdghv" Nov 24 13:28:17 crc kubenswrapper[4790]: I1124 13:28:17.897704 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 24 13:28:17 crc kubenswrapper[4790]: I1124 13:28:17.900949 4790 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 24 13:28:17 crc kubenswrapper[4790]: I1124 13:28:17.900995 4790 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 24 13:28:17 crc kubenswrapper[4790]: I1124 13:28:17.920643 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-sdghv"] Nov 24 13:28:17 crc kubenswrapper[4790]: I1124 13:28:17.982161 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-nglb4"] Nov 24 13:28:17 crc kubenswrapper[4790]: I1124 13:28:17.983375 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-nglb4" Nov 24 13:28:17 crc kubenswrapper[4790]: I1124 13:28:17.985821 4790 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 24 13:28:17 crc kubenswrapper[4790]: I1124 13:28:17.986350 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 24 13:28:17 crc kubenswrapper[4790]: I1124 13:28:17.986747 4790 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 24 13:28:17 crc kubenswrapper[4790]: I1124 13:28:17.988405 4790 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-4qxvj" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.006350 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-fx9l5"] Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.007871 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-fx9l5" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.009664 4790 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.021778 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-fx9l5"] Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.038386 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8ddb656f-ec21-4a37-a89c-3c02fdda39e2-metrics-certs\") pod \"frr-k8s-bvwnn\" (UID: \"8ddb656f-ec21-4a37-a89c-3c02fdda39e2\") " pod="metallb-system/frr-k8s-bvwnn" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.038437 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsxgs\" (UniqueName: \"kubernetes.io/projected/8ddb656f-ec21-4a37-a89c-3c02fdda39e2-kube-api-access-jsxgs\") pod \"frr-k8s-bvwnn\" (UID: \"8ddb656f-ec21-4a37-a89c-3c02fdda39e2\") " pod="metallb-system/frr-k8s-bvwnn" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.038504 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hc5vk\" (UniqueName: \"kubernetes.io/projected/47d9b626-df79-437b-a1af-5c584ff33f11-kube-api-access-hc5vk\") pod \"frr-k8s-webhook-server-6998585d5-sdghv\" (UID: \"47d9b626-df79-437b-a1af-5c584ff33f11\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-sdghv" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.038564 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/8ddb656f-ec21-4a37-a89c-3c02fdda39e2-frr-sockets\") pod \"frr-k8s-bvwnn\" (UID: \"8ddb656f-ec21-4a37-a89c-3c02fdda39e2\") " pod="metallb-system/frr-k8s-bvwnn" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.038585 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/8ddb656f-ec21-4a37-a89c-3c02fdda39e2-frr-startup\") pod \"frr-k8s-bvwnn\" (UID: \"8ddb656f-ec21-4a37-a89c-3c02fdda39e2\") " pod="metallb-system/frr-k8s-bvwnn" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.038604 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/8ddb656f-ec21-4a37-a89c-3c02fdda39e2-metrics\") pod \"frr-k8s-bvwnn\" (UID: \"8ddb656f-ec21-4a37-a89c-3c02fdda39e2\") " pod="metallb-system/frr-k8s-bvwnn" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.038626 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/8ddb656f-ec21-4a37-a89c-3c02fdda39e2-frr-conf\") pod \"frr-k8s-bvwnn\" (UID: \"8ddb656f-ec21-4a37-a89c-3c02fdda39e2\") " pod="metallb-system/frr-k8s-bvwnn" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.038649 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/47d9b626-df79-437b-a1af-5c584ff33f11-cert\") pod \"frr-k8s-webhook-server-6998585d5-sdghv\" (UID: \"47d9b626-df79-437b-a1af-5c584ff33f11\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-sdghv" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.038671 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/8ddb656f-ec21-4a37-a89c-3c02fdda39e2-reloader\") pod \"frr-k8s-bvwnn\" (UID: \"8ddb656f-ec21-4a37-a89c-3c02fdda39e2\") " pod="metallb-system/frr-k8s-bvwnn" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.140527 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/20b558ce-130d-4100-af6e-d2f4d58cafe3-metrics-certs\") pod \"controller-6c7b4b5f48-fx9l5\" (UID: \"20b558ce-130d-4100-af6e-d2f4d58cafe3\") " pod="metallb-system/controller-6c7b4b5f48-fx9l5" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.140598 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/8ddb656f-ec21-4a37-a89c-3c02fdda39e2-frr-sockets\") pod \"frr-k8s-bvwnn\" (UID: \"8ddb656f-ec21-4a37-a89c-3c02fdda39e2\") " pod="metallb-system/frr-k8s-bvwnn" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.140628 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/8ddb656f-ec21-4a37-a89c-3c02fdda39e2-metrics\") pod \"frr-k8s-bvwnn\" (UID: \"8ddb656f-ec21-4a37-a89c-3c02fdda39e2\") " pod="metallb-system/frr-k8s-bvwnn" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.140647 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/8ddb656f-ec21-4a37-a89c-3c02fdda39e2-frr-startup\") pod \"frr-k8s-bvwnn\" (UID: \"8ddb656f-ec21-4a37-a89c-3c02fdda39e2\") " pod="metallb-system/frr-k8s-bvwnn" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.140669 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/8ddb656f-ec21-4a37-a89c-3c02fdda39e2-frr-conf\") pod \"frr-k8s-bvwnn\" (UID: \"8ddb656f-ec21-4a37-a89c-3c02fdda39e2\") " pod="metallb-system/frr-k8s-bvwnn" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.140697 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/47d9b626-df79-437b-a1af-5c584ff33f11-cert\") pod \"frr-k8s-webhook-server-6998585d5-sdghv\" (UID: \"47d9b626-df79-437b-a1af-5c584ff33f11\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-sdghv" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.140720 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/8ddb656f-ec21-4a37-a89c-3c02fdda39e2-reloader\") pod \"frr-k8s-bvwnn\" (UID: \"8ddb656f-ec21-4a37-a89c-3c02fdda39e2\") " pod="metallb-system/frr-k8s-bvwnn" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.140761 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d-metallb-excludel2\") pod \"speaker-nglb4\" (UID: \"ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d\") " pod="metallb-system/speaker-nglb4" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.140786 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d-metrics-certs\") pod \"speaker-nglb4\" (UID: \"ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d\") " pod="metallb-system/speaker-nglb4" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.140809 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8ddb656f-ec21-4a37-a89c-3c02fdda39e2-metrics-certs\") pod \"frr-k8s-bvwnn\" (UID: \"8ddb656f-ec21-4a37-a89c-3c02fdda39e2\") " pod="metallb-system/frr-k8s-bvwnn" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.140844 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsxgs\" (UniqueName: \"kubernetes.io/projected/8ddb656f-ec21-4a37-a89c-3c02fdda39e2-kube-api-access-jsxgs\") pod \"frr-k8s-bvwnn\" (UID: \"8ddb656f-ec21-4a37-a89c-3c02fdda39e2\") " pod="metallb-system/frr-k8s-bvwnn" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.140908 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d-memberlist\") pod \"speaker-nglb4\" (UID: \"ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d\") " pod="metallb-system/speaker-nglb4" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.140951 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hc5vk\" (UniqueName: \"kubernetes.io/projected/47d9b626-df79-437b-a1af-5c584ff33f11-kube-api-access-hc5vk\") pod \"frr-k8s-webhook-server-6998585d5-sdghv\" (UID: \"47d9b626-df79-437b-a1af-5c584ff33f11\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-sdghv" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.140980 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lj6pm\" (UniqueName: \"kubernetes.io/projected/ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d-kube-api-access-lj6pm\") pod \"speaker-nglb4\" (UID: \"ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d\") " pod="metallb-system/speaker-nglb4" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.141009 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkx2j\" (UniqueName: \"kubernetes.io/projected/20b558ce-130d-4100-af6e-d2f4d58cafe3-kube-api-access-dkx2j\") pod \"controller-6c7b4b5f48-fx9l5\" (UID: \"20b558ce-130d-4100-af6e-d2f4d58cafe3\") " pod="metallb-system/controller-6c7b4b5f48-fx9l5" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.141051 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b558ce-130d-4100-af6e-d2f4d58cafe3-cert\") pod \"controller-6c7b4b5f48-fx9l5\" (UID: \"20b558ce-130d-4100-af6e-d2f4d58cafe3\") " pod="metallb-system/controller-6c7b4b5f48-fx9l5" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.141855 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/8ddb656f-ec21-4a37-a89c-3c02fdda39e2-frr-sockets\") pod \"frr-k8s-bvwnn\" (UID: \"8ddb656f-ec21-4a37-a89c-3c02fdda39e2\") " pod="metallb-system/frr-k8s-bvwnn" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.142204 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/8ddb656f-ec21-4a37-a89c-3c02fdda39e2-metrics\") pod \"frr-k8s-bvwnn\" (UID: \"8ddb656f-ec21-4a37-a89c-3c02fdda39e2\") " pod="metallb-system/frr-k8s-bvwnn" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.143157 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/8ddb656f-ec21-4a37-a89c-3c02fdda39e2-frr-startup\") pod \"frr-k8s-bvwnn\" (UID: \"8ddb656f-ec21-4a37-a89c-3c02fdda39e2\") " pod="metallb-system/frr-k8s-bvwnn" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.143385 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/8ddb656f-ec21-4a37-a89c-3c02fdda39e2-frr-conf\") pod \"frr-k8s-bvwnn\" (UID: \"8ddb656f-ec21-4a37-a89c-3c02fdda39e2\") " pod="metallb-system/frr-k8s-bvwnn" Nov 24 13:28:18 crc kubenswrapper[4790]: E1124 13:28:18.143470 4790 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Nov 24 13:28:18 crc kubenswrapper[4790]: E1124 13:28:18.143519 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/47d9b626-df79-437b-a1af-5c584ff33f11-cert podName:47d9b626-df79-437b-a1af-5c584ff33f11 nodeName:}" failed. No retries permitted until 2025-11-24 13:28:18.643502554 +0000 UTC m=+947.023396226 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/47d9b626-df79-437b-a1af-5c584ff33f11-cert") pod "frr-k8s-webhook-server-6998585d5-sdghv" (UID: "47d9b626-df79-437b-a1af-5c584ff33f11") : secret "frr-k8s-webhook-server-cert" not found Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.144032 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/8ddb656f-ec21-4a37-a89c-3c02fdda39e2-reloader\") pod \"frr-k8s-bvwnn\" (UID: \"8ddb656f-ec21-4a37-a89c-3c02fdda39e2\") " pod="metallb-system/frr-k8s-bvwnn" Nov 24 13:28:18 crc kubenswrapper[4790]: E1124 13:28:18.144133 4790 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Nov 24 13:28:18 crc kubenswrapper[4790]: E1124 13:28:18.144167 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8ddb656f-ec21-4a37-a89c-3c02fdda39e2-metrics-certs podName:8ddb656f-ec21-4a37-a89c-3c02fdda39e2 nodeName:}" failed. No retries permitted until 2025-11-24 13:28:18.644154932 +0000 UTC m=+947.024048594 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8ddb656f-ec21-4a37-a89c-3c02fdda39e2-metrics-certs") pod "frr-k8s-bvwnn" (UID: "8ddb656f-ec21-4a37-a89c-3c02fdda39e2") : secret "frr-k8s-certs-secret" not found Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.164601 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsxgs\" (UniqueName: \"kubernetes.io/projected/8ddb656f-ec21-4a37-a89c-3c02fdda39e2-kube-api-access-jsxgs\") pod \"frr-k8s-bvwnn\" (UID: \"8ddb656f-ec21-4a37-a89c-3c02fdda39e2\") " pod="metallb-system/frr-k8s-bvwnn" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.170046 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hc5vk\" (UniqueName: \"kubernetes.io/projected/47d9b626-df79-437b-a1af-5c584ff33f11-kube-api-access-hc5vk\") pod \"frr-k8s-webhook-server-6998585d5-sdghv\" (UID: \"47d9b626-df79-437b-a1af-5c584ff33f11\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-sdghv" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.242918 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d-memberlist\") pod \"speaker-nglb4\" (UID: \"ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d\") " pod="metallb-system/speaker-nglb4" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.242991 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lj6pm\" (UniqueName: \"kubernetes.io/projected/ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d-kube-api-access-lj6pm\") pod \"speaker-nglb4\" (UID: \"ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d\") " pod="metallb-system/speaker-nglb4" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.243017 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkx2j\" (UniqueName: \"kubernetes.io/projected/20b558ce-130d-4100-af6e-d2f4d58cafe3-kube-api-access-dkx2j\") pod \"controller-6c7b4b5f48-fx9l5\" (UID: \"20b558ce-130d-4100-af6e-d2f4d58cafe3\") " pod="metallb-system/controller-6c7b4b5f48-fx9l5" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.243061 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b558ce-130d-4100-af6e-d2f4d58cafe3-cert\") pod \"controller-6c7b4b5f48-fx9l5\" (UID: \"20b558ce-130d-4100-af6e-d2f4d58cafe3\") " pod="metallb-system/controller-6c7b4b5f48-fx9l5" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.243083 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/20b558ce-130d-4100-af6e-d2f4d58cafe3-metrics-certs\") pod \"controller-6c7b4b5f48-fx9l5\" (UID: \"20b558ce-130d-4100-af6e-d2f4d58cafe3\") " pod="metallb-system/controller-6c7b4b5f48-fx9l5" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.243137 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d-metallb-excludel2\") pod \"speaker-nglb4\" (UID: \"ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d\") " pod="metallb-system/speaker-nglb4" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.243156 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d-metrics-certs\") pod \"speaker-nglb4\" (UID: \"ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d\") " pod="metallb-system/speaker-nglb4" Nov 24 13:28:18 crc kubenswrapper[4790]: E1124 13:28:18.244318 4790 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 24 13:28:18 crc kubenswrapper[4790]: E1124 13:28:18.244430 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d-memberlist podName:ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d nodeName:}" failed. No retries permitted until 2025-11-24 13:28:18.744405488 +0000 UTC m=+947.124299150 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d-memberlist") pod "speaker-nglb4" (UID: "ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d") : secret "metallb-memberlist" not found Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.245333 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d-metallb-excludel2\") pod \"speaker-nglb4\" (UID: \"ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d\") " pod="metallb-system/speaker-nglb4" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.248610 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d-metrics-certs\") pod \"speaker-nglb4\" (UID: \"ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d\") " pod="metallb-system/speaker-nglb4" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.248652 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/20b558ce-130d-4100-af6e-d2f4d58cafe3-metrics-certs\") pod \"controller-6c7b4b5f48-fx9l5\" (UID: \"20b558ce-130d-4100-af6e-d2f4d58cafe3\") " pod="metallb-system/controller-6c7b4b5f48-fx9l5" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.254449 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b558ce-130d-4100-af6e-d2f4d58cafe3-cert\") pod \"controller-6c7b4b5f48-fx9l5\" (UID: \"20b558ce-130d-4100-af6e-d2f4d58cafe3\") " pod="metallb-system/controller-6c7b4b5f48-fx9l5" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.261517 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lj6pm\" (UniqueName: \"kubernetes.io/projected/ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d-kube-api-access-lj6pm\") pod \"speaker-nglb4\" (UID: \"ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d\") " pod="metallb-system/speaker-nglb4" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.265507 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkx2j\" (UniqueName: \"kubernetes.io/projected/20b558ce-130d-4100-af6e-d2f4d58cafe3-kube-api-access-dkx2j\") pod \"controller-6c7b4b5f48-fx9l5\" (UID: \"20b558ce-130d-4100-af6e-d2f4d58cafe3\") " pod="metallb-system/controller-6c7b4b5f48-fx9l5" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.326449 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-fx9l5" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.537090 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-fx9l5"] Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.649190 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/47d9b626-df79-437b-a1af-5c584ff33f11-cert\") pod \"frr-k8s-webhook-server-6998585d5-sdghv\" (UID: \"47d9b626-df79-437b-a1af-5c584ff33f11\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-sdghv" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.649796 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8ddb656f-ec21-4a37-a89c-3c02fdda39e2-metrics-certs\") pod \"frr-k8s-bvwnn\" (UID: \"8ddb656f-ec21-4a37-a89c-3c02fdda39e2\") " pod="metallb-system/frr-k8s-bvwnn" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.655495 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8ddb656f-ec21-4a37-a89c-3c02fdda39e2-metrics-certs\") pod \"frr-k8s-bvwnn\" (UID: \"8ddb656f-ec21-4a37-a89c-3c02fdda39e2\") " pod="metallb-system/frr-k8s-bvwnn" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.655660 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/47d9b626-df79-437b-a1af-5c584ff33f11-cert\") pod \"frr-k8s-webhook-server-6998585d5-sdghv\" (UID: \"47d9b626-df79-437b-a1af-5c584ff33f11\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-sdghv" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.751152 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d-memberlist\") pod \"speaker-nglb4\" (UID: \"ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d\") " pod="metallb-system/speaker-nglb4" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.754644 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d-memberlist\") pod \"speaker-nglb4\" (UID: \"ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d\") " pod="metallb-system/speaker-nglb4" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.816171 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-bvwnn" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.826008 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-sdghv" Nov 24 13:28:18 crc kubenswrapper[4790]: I1124 13:28:18.900392 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-nglb4" Nov 24 13:28:18 crc kubenswrapper[4790]: W1124 13:28:18.920745 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podddc6a4d3_6419_4d05_a3a1_4dc3b3f79e5d.slice/crio-51f8da55989c950b349e31686e0984f5bb2ad572021dbfd03b689640fd545c8b WatchSource:0}: Error finding container 51f8da55989c950b349e31686e0984f5bb2ad572021dbfd03b689640fd545c8b: Status 404 returned error can't find the container with id 51f8da55989c950b349e31686e0984f5bb2ad572021dbfd03b689640fd545c8b Nov 24 13:28:19 crc kubenswrapper[4790]: I1124 13:28:19.059067 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-sdghv"] Nov 24 13:28:19 crc kubenswrapper[4790]: W1124 13:28:19.072421 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47d9b626_df79_437b_a1af_5c584ff33f11.slice/crio-33f3ce19cb79fa8b0d4712f56b638cda13b655563e55063b6672c5179e8b2435 WatchSource:0}: Error finding container 33f3ce19cb79fa8b0d4712f56b638cda13b655563e55063b6672c5179e8b2435: Status 404 returned error can't find the container with id 33f3ce19cb79fa8b0d4712f56b638cda13b655563e55063b6672c5179e8b2435 Nov 24 13:28:19 crc kubenswrapper[4790]: I1124 13:28:19.096559 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-fx9l5" event={"ID":"20b558ce-130d-4100-af6e-d2f4d58cafe3","Type":"ContainerStarted","Data":"91b02e09231cda7675590db18fd6ccf25ac1288aedd010ee603886338c493db7"} Nov 24 13:28:19 crc kubenswrapper[4790]: I1124 13:28:19.096603 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-fx9l5" event={"ID":"20b558ce-130d-4100-af6e-d2f4d58cafe3","Type":"ContainerStarted","Data":"e00c93125af9ad00e9398de9d59bbc2f052a0d14d279a17164d8fd87fb703a2d"} Nov 24 13:28:19 crc kubenswrapper[4790]: I1124 13:28:19.096614 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-fx9l5" event={"ID":"20b558ce-130d-4100-af6e-d2f4d58cafe3","Type":"ContainerStarted","Data":"d87363506e29ae93eeeb23bac58f27bdf7f87d12bc9bdf774742d3129668eeaf"} Nov 24 13:28:19 crc kubenswrapper[4790]: I1124 13:28:19.096649 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-fx9l5" Nov 24 13:28:19 crc kubenswrapper[4790]: I1124 13:28:19.099246 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bvwnn" event={"ID":"8ddb656f-ec21-4a37-a89c-3c02fdda39e2","Type":"ContainerStarted","Data":"00c4fa8e5f7c4fbae5e51a9f8ca9bc4815fbb84804d8c67b8f3538ccd738935e"} Nov 24 13:28:19 crc kubenswrapper[4790]: I1124 13:28:19.100555 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-sdghv" event={"ID":"47d9b626-df79-437b-a1af-5c584ff33f11","Type":"ContainerStarted","Data":"33f3ce19cb79fa8b0d4712f56b638cda13b655563e55063b6672c5179e8b2435"} Nov 24 13:28:19 crc kubenswrapper[4790]: I1124 13:28:19.101995 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-nglb4" event={"ID":"ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d","Type":"ContainerStarted","Data":"51f8da55989c950b349e31686e0984f5bb2ad572021dbfd03b689640fd545c8b"} Nov 24 13:28:19 crc kubenswrapper[4790]: I1124 13:28:19.114695 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-fx9l5" podStartSLOduration=2.114676317 podStartE2EDuration="2.114676317s" podCreationTimestamp="2025-11-24 13:28:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:28:19.114190923 +0000 UTC m=+947.494084595" watchObservedRunningTime="2025-11-24 13:28:19.114676317 +0000 UTC m=+947.494569999" Nov 24 13:28:20 crc kubenswrapper[4790]: I1124 13:28:20.112907 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-nglb4" event={"ID":"ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d","Type":"ContainerStarted","Data":"14c030de84121e335248cc654c335a44955d5795848e6d57b41361366948420d"} Nov 24 13:28:20 crc kubenswrapper[4790]: I1124 13:28:20.113380 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-nglb4" Nov 24 13:28:20 crc kubenswrapper[4790]: I1124 13:28:20.113393 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-nglb4" event={"ID":"ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d","Type":"ContainerStarted","Data":"fe0f62973d39bab0cb313278d4acef1e22b15681950b2cdbea1dad48ed99ce97"} Nov 24 13:28:20 crc kubenswrapper[4790]: I1124 13:28:20.130679 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-nglb4" podStartSLOduration=3.130660085 podStartE2EDuration="3.130660085s" podCreationTimestamp="2025-11-24 13:28:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:28:20.128565227 +0000 UTC m=+948.508458919" watchObservedRunningTime="2025-11-24 13:28:20.130660085 +0000 UTC m=+948.510553747" Nov 24 13:28:27 crc kubenswrapper[4790]: I1124 13:28:27.184160 4790 generic.go:334] "Generic (PLEG): container finished" podID="8ddb656f-ec21-4a37-a89c-3c02fdda39e2" containerID="89c63071d3b01a515a319f298390345fa3877d4e0b9868de56ba14fa38057071" exitCode=0 Nov 24 13:28:27 crc kubenswrapper[4790]: I1124 13:28:27.184216 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bvwnn" event={"ID":"8ddb656f-ec21-4a37-a89c-3c02fdda39e2","Type":"ContainerDied","Data":"89c63071d3b01a515a319f298390345fa3877d4e0b9868de56ba14fa38057071"} Nov 24 13:28:27 crc kubenswrapper[4790]: I1124 13:28:27.186967 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-sdghv" event={"ID":"47d9b626-df79-437b-a1af-5c584ff33f11","Type":"ContainerStarted","Data":"bcd1bdbb937585c67cccab0fea4ab729c00c2a148f0cbd32bba99db64ccc6d61"} Nov 24 13:28:27 crc kubenswrapper[4790]: I1124 13:28:27.187115 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-sdghv" Nov 24 13:28:27 crc kubenswrapper[4790]: I1124 13:28:27.226670 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-sdghv" podStartSLOduration=3.242632968 podStartE2EDuration="10.226651852s" podCreationTimestamp="2025-11-24 13:28:17 +0000 UTC" firstStartedPulling="2025-11-24 13:28:19.074667155 +0000 UTC m=+947.454560807" lastFinishedPulling="2025-11-24 13:28:26.058686029 +0000 UTC m=+954.438579691" observedRunningTime="2025-11-24 13:28:27.223308829 +0000 UTC m=+955.603202491" watchObservedRunningTime="2025-11-24 13:28:27.226651852 +0000 UTC m=+955.606545514" Nov 24 13:28:28 crc kubenswrapper[4790]: I1124 13:28:28.194767 4790 generic.go:334] "Generic (PLEG): container finished" podID="8ddb656f-ec21-4a37-a89c-3c02fdda39e2" containerID="f386c61377114d246a19e767890a5b59c5ba7dbdcbf4ceb47dfa2806b50a0904" exitCode=0 Nov 24 13:28:28 crc kubenswrapper[4790]: I1124 13:28:28.194932 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bvwnn" event={"ID":"8ddb656f-ec21-4a37-a89c-3c02fdda39e2","Type":"ContainerDied","Data":"f386c61377114d246a19e767890a5b59c5ba7dbdcbf4ceb47dfa2806b50a0904"} Nov 24 13:28:28 crc kubenswrapper[4790]: I1124 13:28:28.335098 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-fx9l5" Nov 24 13:28:29 crc kubenswrapper[4790]: I1124 13:28:29.205584 4790 generic.go:334] "Generic (PLEG): container finished" podID="8ddb656f-ec21-4a37-a89c-3c02fdda39e2" containerID="cff42d10288bc74fe4889ab2ff6659ca37cad22e77d4bf2170c4749cf2362ff9" exitCode=0 Nov 24 13:28:29 crc kubenswrapper[4790]: I1124 13:28:29.205637 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bvwnn" event={"ID":"8ddb656f-ec21-4a37-a89c-3c02fdda39e2","Type":"ContainerDied","Data":"cff42d10288bc74fe4889ab2ff6659ca37cad22e77d4bf2170c4749cf2362ff9"} Nov 24 13:28:30 crc kubenswrapper[4790]: I1124 13:28:30.215133 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bvwnn" event={"ID":"8ddb656f-ec21-4a37-a89c-3c02fdda39e2","Type":"ContainerStarted","Data":"63cd789773a25c736106155e9af8835e4bd9d41bb4d73fbf82d25941be19f39a"} Nov 24 13:28:30 crc kubenswrapper[4790]: I1124 13:28:30.215429 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bvwnn" event={"ID":"8ddb656f-ec21-4a37-a89c-3c02fdda39e2","Type":"ContainerStarted","Data":"e743655a08aea01a831a9c670af6da92f265b226c67ccdbf8f45c0da88ead6c3"} Nov 24 13:28:30 crc kubenswrapper[4790]: I1124 13:28:30.215438 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bvwnn" event={"ID":"8ddb656f-ec21-4a37-a89c-3c02fdda39e2","Type":"ContainerStarted","Data":"11bd37ce546303339b88cb52b1d89b86a6bf1de9d715723c946432fd561276c8"} Nov 24 13:28:30 crc kubenswrapper[4790]: I1124 13:28:30.215447 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bvwnn" event={"ID":"8ddb656f-ec21-4a37-a89c-3c02fdda39e2","Type":"ContainerStarted","Data":"783ac601abd13f5525b13b869b6e81dc964f501e51c622b2693bd80712f88f34"} Nov 24 13:28:31 crc kubenswrapper[4790]: I1124 13:28:31.227442 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bvwnn" event={"ID":"8ddb656f-ec21-4a37-a89c-3c02fdda39e2","Type":"ContainerStarted","Data":"c94eb01c600b879a1a7eff7ee7572a82eb73df7bb46a7acdb9e27e5539c19e91"} Nov 24 13:28:31 crc kubenswrapper[4790]: I1124 13:28:31.227477 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bvwnn" event={"ID":"8ddb656f-ec21-4a37-a89c-3c02fdda39e2","Type":"ContainerStarted","Data":"18557c6badc61a7971a18faf02fa031d6a231a91a64ce0f7daaf1fb987eb2b9a"} Nov 24 13:28:31 crc kubenswrapper[4790]: I1124 13:28:31.228019 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-bvwnn" Nov 24 13:28:31 crc kubenswrapper[4790]: I1124 13:28:31.257666 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-bvwnn" podStartSLOduration=7.185345892 podStartE2EDuration="14.25765094s" podCreationTimestamp="2025-11-24 13:28:17 +0000 UTC" firstStartedPulling="2025-11-24 13:28:18.963709881 +0000 UTC m=+947.343603533" lastFinishedPulling="2025-11-24 13:28:26.036014919 +0000 UTC m=+954.415908581" observedRunningTime="2025-11-24 13:28:31.255198602 +0000 UTC m=+959.635092374" watchObservedRunningTime="2025-11-24 13:28:31.25765094 +0000 UTC m=+959.637544602" Nov 24 13:28:33 crc kubenswrapper[4790]: I1124 13:28:33.816446 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-bvwnn" Nov 24 13:28:33 crc kubenswrapper[4790]: I1124 13:28:33.861895 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-bvwnn" Nov 24 13:28:38 crc kubenswrapper[4790]: I1124 13:28:38.833474 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-sdghv" Nov 24 13:28:38 crc kubenswrapper[4790]: I1124 13:28:38.904534 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-nglb4" Nov 24 13:28:40 crc kubenswrapper[4790]: I1124 13:28:40.418294 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9"] Nov 24 13:28:40 crc kubenswrapper[4790]: I1124 13:28:40.419476 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9" Nov 24 13:28:40 crc kubenswrapper[4790]: I1124 13:28:40.422299 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 13:28:40 crc kubenswrapper[4790]: I1124 13:28:40.438524 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9"] Nov 24 13:28:40 crc kubenswrapper[4790]: I1124 13:28:40.467166 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77dvz\" (UniqueName: \"kubernetes.io/projected/47d9c197-a7d1-4977-b580-508aa507f368-kube-api-access-77dvz\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9\" (UID: \"47d9c197-a7d1-4977-b580-508aa507f368\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9" Nov 24 13:28:40 crc kubenswrapper[4790]: I1124 13:28:40.467247 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/47d9c197-a7d1-4977-b580-508aa507f368-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9\" (UID: \"47d9c197-a7d1-4977-b580-508aa507f368\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9" Nov 24 13:28:40 crc kubenswrapper[4790]: I1124 13:28:40.467440 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/47d9c197-a7d1-4977-b580-508aa507f368-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9\" (UID: \"47d9c197-a7d1-4977-b580-508aa507f368\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9" Nov 24 13:28:40 crc kubenswrapper[4790]: I1124 13:28:40.568547 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/47d9c197-a7d1-4977-b580-508aa507f368-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9\" (UID: \"47d9c197-a7d1-4977-b580-508aa507f368\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9" Nov 24 13:28:40 crc kubenswrapper[4790]: I1124 13:28:40.568645 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/47d9c197-a7d1-4977-b580-508aa507f368-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9\" (UID: \"47d9c197-a7d1-4977-b580-508aa507f368\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9" Nov 24 13:28:40 crc kubenswrapper[4790]: I1124 13:28:40.568690 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77dvz\" (UniqueName: \"kubernetes.io/projected/47d9c197-a7d1-4977-b580-508aa507f368-kube-api-access-77dvz\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9\" (UID: \"47d9c197-a7d1-4977-b580-508aa507f368\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9" Nov 24 13:28:40 crc kubenswrapper[4790]: I1124 13:28:40.569108 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/47d9c197-a7d1-4977-b580-508aa507f368-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9\" (UID: \"47d9c197-a7d1-4977-b580-508aa507f368\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9" Nov 24 13:28:40 crc kubenswrapper[4790]: I1124 13:28:40.569206 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/47d9c197-a7d1-4977-b580-508aa507f368-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9\" (UID: \"47d9c197-a7d1-4977-b580-508aa507f368\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9" Nov 24 13:28:40 crc kubenswrapper[4790]: I1124 13:28:40.590851 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77dvz\" (UniqueName: \"kubernetes.io/projected/47d9c197-a7d1-4977-b580-508aa507f368-kube-api-access-77dvz\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9\" (UID: \"47d9c197-a7d1-4977-b580-508aa507f368\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9" Nov 24 13:28:40 crc kubenswrapper[4790]: I1124 13:28:40.737123 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9" Nov 24 13:28:41 crc kubenswrapper[4790]: I1124 13:28:41.120597 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9"] Nov 24 13:28:41 crc kubenswrapper[4790]: I1124 13:28:41.281676 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9" event={"ID":"47d9c197-a7d1-4977-b580-508aa507f368","Type":"ContainerStarted","Data":"9dc1814e6621af75092d626ce4b8133689f2df3843961417fb2d3cb6787d001a"} Nov 24 13:28:42 crc kubenswrapper[4790]: I1124 13:28:42.287893 4790 generic.go:334] "Generic (PLEG): container finished" podID="47d9c197-a7d1-4977-b580-508aa507f368" containerID="d0ab8698f5f0cac8ef395cf1361d123ba064e0a45c7acf075a7d6abb39ac554a" exitCode=0 Nov 24 13:28:42 crc kubenswrapper[4790]: I1124 13:28:42.287938 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9" event={"ID":"47d9c197-a7d1-4977-b580-508aa507f368","Type":"ContainerDied","Data":"d0ab8698f5f0cac8ef395cf1361d123ba064e0a45c7acf075a7d6abb39ac554a"} Nov 24 13:28:45 crc kubenswrapper[4790]: I1124 13:28:45.311566 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9" event={"ID":"47d9c197-a7d1-4977-b580-508aa507f368","Type":"ContainerStarted","Data":"12feb8ca97c1f47acb2246269f656df5330469fc10c180d73ce5f78d2b8ff985"} Nov 24 13:28:46 crc kubenswrapper[4790]: I1124 13:28:46.318841 4790 generic.go:334] "Generic (PLEG): container finished" podID="47d9c197-a7d1-4977-b580-508aa507f368" containerID="12feb8ca97c1f47acb2246269f656df5330469fc10c180d73ce5f78d2b8ff985" exitCode=0 Nov 24 13:28:46 crc kubenswrapper[4790]: I1124 13:28:46.322509 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9" event={"ID":"47d9c197-a7d1-4977-b580-508aa507f368","Type":"ContainerDied","Data":"12feb8ca97c1f47acb2246269f656df5330469fc10c180d73ce5f78d2b8ff985"} Nov 24 13:28:47 crc kubenswrapper[4790]: I1124 13:28:47.325859 4790 generic.go:334] "Generic (PLEG): container finished" podID="47d9c197-a7d1-4977-b580-508aa507f368" containerID="da3c6a52afc8c4c7e2bbc78dd2e8326908ab37b986b6f4790f678443bf297c62" exitCode=0 Nov 24 13:28:47 crc kubenswrapper[4790]: I1124 13:28:47.325945 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9" event={"ID":"47d9c197-a7d1-4977-b580-508aa507f368","Type":"ContainerDied","Data":"da3c6a52afc8c4c7e2bbc78dd2e8326908ab37b986b6f4790f678443bf297c62"} Nov 24 13:28:48 crc kubenswrapper[4790]: I1124 13:28:48.545123 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9" Nov 24 13:28:48 crc kubenswrapper[4790]: I1124 13:28:48.592394 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/47d9c197-a7d1-4977-b580-508aa507f368-bundle\") pod \"47d9c197-a7d1-4977-b580-508aa507f368\" (UID: \"47d9c197-a7d1-4977-b580-508aa507f368\") " Nov 24 13:28:48 crc kubenswrapper[4790]: I1124 13:28:48.592499 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/47d9c197-a7d1-4977-b580-508aa507f368-util\") pod \"47d9c197-a7d1-4977-b580-508aa507f368\" (UID: \"47d9c197-a7d1-4977-b580-508aa507f368\") " Nov 24 13:28:48 crc kubenswrapper[4790]: I1124 13:28:48.592541 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77dvz\" (UniqueName: \"kubernetes.io/projected/47d9c197-a7d1-4977-b580-508aa507f368-kube-api-access-77dvz\") pod \"47d9c197-a7d1-4977-b580-508aa507f368\" (UID: \"47d9c197-a7d1-4977-b580-508aa507f368\") " Nov 24 13:28:48 crc kubenswrapper[4790]: I1124 13:28:48.593569 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47d9c197-a7d1-4977-b580-508aa507f368-bundle" (OuterVolumeSpecName: "bundle") pod "47d9c197-a7d1-4977-b580-508aa507f368" (UID: "47d9c197-a7d1-4977-b580-508aa507f368"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:28:48 crc kubenswrapper[4790]: I1124 13:28:48.597520 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47d9c197-a7d1-4977-b580-508aa507f368-kube-api-access-77dvz" (OuterVolumeSpecName: "kube-api-access-77dvz") pod "47d9c197-a7d1-4977-b580-508aa507f368" (UID: "47d9c197-a7d1-4977-b580-508aa507f368"). InnerVolumeSpecName "kube-api-access-77dvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:28:48 crc kubenswrapper[4790]: I1124 13:28:48.603562 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47d9c197-a7d1-4977-b580-508aa507f368-util" (OuterVolumeSpecName: "util") pod "47d9c197-a7d1-4977-b580-508aa507f368" (UID: "47d9c197-a7d1-4977-b580-508aa507f368"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:28:48 crc kubenswrapper[4790]: I1124 13:28:48.694437 4790 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/47d9c197-a7d1-4977-b580-508aa507f368-util\") on node \"crc\" DevicePath \"\"" Nov 24 13:28:48 crc kubenswrapper[4790]: I1124 13:28:48.694480 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77dvz\" (UniqueName: \"kubernetes.io/projected/47d9c197-a7d1-4977-b580-508aa507f368-kube-api-access-77dvz\") on node \"crc\" DevicePath \"\"" Nov 24 13:28:48 crc kubenswrapper[4790]: I1124 13:28:48.694492 4790 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/47d9c197-a7d1-4977-b580-508aa507f368-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:28:48 crc kubenswrapper[4790]: I1124 13:28:48.819193 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-bvwnn" Nov 24 13:28:49 crc kubenswrapper[4790]: I1124 13:28:49.338245 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9" event={"ID":"47d9c197-a7d1-4977-b580-508aa507f368","Type":"ContainerDied","Data":"9dc1814e6621af75092d626ce4b8133689f2df3843961417fb2d3cb6787d001a"} Nov 24 13:28:49 crc kubenswrapper[4790]: I1124 13:28:49.338314 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9dc1814e6621af75092d626ce4b8133689f2df3843961417fb2d3cb6787d001a" Nov 24 13:28:49 crc kubenswrapper[4790]: I1124 13:28:49.338317 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9" Nov 24 13:28:53 crc kubenswrapper[4790]: I1124 13:28:53.339204 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-wz49g"] Nov 24 13:28:53 crc kubenswrapper[4790]: E1124 13:28:53.341801 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47d9c197-a7d1-4977-b580-508aa507f368" containerName="pull" Nov 24 13:28:53 crc kubenswrapper[4790]: I1124 13:28:53.341922 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="47d9c197-a7d1-4977-b580-508aa507f368" containerName="pull" Nov 24 13:28:53 crc kubenswrapper[4790]: E1124 13:28:53.342006 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47d9c197-a7d1-4977-b580-508aa507f368" containerName="extract" Nov 24 13:28:53 crc kubenswrapper[4790]: I1124 13:28:53.342064 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="47d9c197-a7d1-4977-b580-508aa507f368" containerName="extract" Nov 24 13:28:53 crc kubenswrapper[4790]: E1124 13:28:53.342140 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47d9c197-a7d1-4977-b580-508aa507f368" containerName="util" Nov 24 13:28:53 crc kubenswrapper[4790]: I1124 13:28:53.342197 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="47d9c197-a7d1-4977-b580-508aa507f368" containerName="util" Nov 24 13:28:53 crc kubenswrapper[4790]: I1124 13:28:53.342368 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="47d9c197-a7d1-4977-b580-508aa507f368" containerName="extract" Nov 24 13:28:53 crc kubenswrapper[4790]: I1124 13:28:53.343204 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-wz49g" Nov 24 13:28:53 crc kubenswrapper[4790]: I1124 13:28:53.348622 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Nov 24 13:28:53 crc kubenswrapper[4790]: I1124 13:28:53.370740 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Nov 24 13:28:53 crc kubenswrapper[4790]: I1124 13:28:53.370747 4790 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-9vx6w" Nov 24 13:28:53 crc kubenswrapper[4790]: I1124 13:28:53.381761 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-wz49g"] Nov 24 13:28:53 crc kubenswrapper[4790]: I1124 13:28:53.472953 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2s24\" (UniqueName: \"kubernetes.io/projected/33873117-c431-4700-b40a-bb8901fef180-kube-api-access-w2s24\") pod \"cert-manager-operator-controller-manager-64cf6dff88-wz49g\" (UID: \"33873117-c431-4700-b40a-bb8901fef180\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-wz49g" Nov 24 13:28:53 crc kubenswrapper[4790]: I1124 13:28:53.473037 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/33873117-c431-4700-b40a-bb8901fef180-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-wz49g\" (UID: \"33873117-c431-4700-b40a-bb8901fef180\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-wz49g" Nov 24 13:28:53 crc kubenswrapper[4790]: I1124 13:28:53.574983 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/33873117-c431-4700-b40a-bb8901fef180-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-wz49g\" (UID: \"33873117-c431-4700-b40a-bb8901fef180\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-wz49g" Nov 24 13:28:53 crc kubenswrapper[4790]: I1124 13:28:53.575112 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2s24\" (UniqueName: \"kubernetes.io/projected/33873117-c431-4700-b40a-bb8901fef180-kube-api-access-w2s24\") pod \"cert-manager-operator-controller-manager-64cf6dff88-wz49g\" (UID: \"33873117-c431-4700-b40a-bb8901fef180\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-wz49g" Nov 24 13:28:53 crc kubenswrapper[4790]: I1124 13:28:53.575563 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/33873117-c431-4700-b40a-bb8901fef180-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-wz49g\" (UID: \"33873117-c431-4700-b40a-bb8901fef180\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-wz49g" Nov 24 13:28:53 crc kubenswrapper[4790]: I1124 13:28:53.596455 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2s24\" (UniqueName: \"kubernetes.io/projected/33873117-c431-4700-b40a-bb8901fef180-kube-api-access-w2s24\") pod \"cert-manager-operator-controller-manager-64cf6dff88-wz49g\" (UID: \"33873117-c431-4700-b40a-bb8901fef180\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-wz49g" Nov 24 13:28:53 crc kubenswrapper[4790]: I1124 13:28:53.670419 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-wz49g" Nov 24 13:28:54 crc kubenswrapper[4790]: I1124 13:28:54.125700 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-wz49g"] Nov 24 13:28:54 crc kubenswrapper[4790]: I1124 13:28:54.391012 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-wz49g" event={"ID":"33873117-c431-4700-b40a-bb8901fef180","Type":"ContainerStarted","Data":"d4abb4809f922caf73b1ee90746ff1b39a6e3f291cc61fd37de6003f5a095a6c"} Nov 24 13:29:01 crc kubenswrapper[4790]: I1124 13:29:01.437868 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-wz49g" event={"ID":"33873117-c431-4700-b40a-bb8901fef180","Type":"ContainerStarted","Data":"2e8cd0c7d382ccd24d426f03887c7157678124adb6bba28f9173c5d31c5eed2e"} Nov 24 13:29:01 crc kubenswrapper[4790]: I1124 13:29:01.469818 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-wz49g" podStartSLOduration=2.0410464 podStartE2EDuration="8.469795313s" podCreationTimestamp="2025-11-24 13:28:53 +0000 UTC" firstStartedPulling="2025-11-24 13:28:54.134531268 +0000 UTC m=+982.514424930" lastFinishedPulling="2025-11-24 13:29:00.563280181 +0000 UTC m=+988.943173843" observedRunningTime="2025-11-24 13:29:01.468507082 +0000 UTC m=+989.848400744" watchObservedRunningTime="2025-11-24 13:29:01.469795313 +0000 UTC m=+989.849688975" Nov 24 13:29:03 crc kubenswrapper[4790]: I1124 13:29:03.772999 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-dm8pw"] Nov 24 13:29:03 crc kubenswrapper[4790]: I1124 13:29:03.774035 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-dm8pw" Nov 24 13:29:03 crc kubenswrapper[4790]: I1124 13:29:03.776672 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 24 13:29:03 crc kubenswrapper[4790]: I1124 13:29:03.776803 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 24 13:29:03 crc kubenswrapper[4790]: I1124 13:29:03.776925 4790 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-7snp5" Nov 24 13:29:03 crc kubenswrapper[4790]: I1124 13:29:03.782530 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-dm8pw"] Nov 24 13:29:03 crc kubenswrapper[4790]: I1124 13:29:03.911413 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cdbadd01-32e5-48d1-bf16-68939d60bd02-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-dm8pw\" (UID: \"cdbadd01-32e5-48d1-bf16-68939d60bd02\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-dm8pw" Nov 24 13:29:03 crc kubenswrapper[4790]: I1124 13:29:03.911468 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqvkw\" (UniqueName: \"kubernetes.io/projected/cdbadd01-32e5-48d1-bf16-68939d60bd02-kube-api-access-tqvkw\") pod \"cert-manager-webhook-f4fb5df64-dm8pw\" (UID: \"cdbadd01-32e5-48d1-bf16-68939d60bd02\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-dm8pw" Nov 24 13:29:04 crc kubenswrapper[4790]: I1124 13:29:04.013179 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cdbadd01-32e5-48d1-bf16-68939d60bd02-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-dm8pw\" (UID: \"cdbadd01-32e5-48d1-bf16-68939d60bd02\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-dm8pw" Nov 24 13:29:04 crc kubenswrapper[4790]: I1124 13:29:04.013257 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqvkw\" (UniqueName: \"kubernetes.io/projected/cdbadd01-32e5-48d1-bf16-68939d60bd02-kube-api-access-tqvkw\") pod \"cert-manager-webhook-f4fb5df64-dm8pw\" (UID: \"cdbadd01-32e5-48d1-bf16-68939d60bd02\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-dm8pw" Nov 24 13:29:04 crc kubenswrapper[4790]: I1124 13:29:04.032397 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cdbadd01-32e5-48d1-bf16-68939d60bd02-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-dm8pw\" (UID: \"cdbadd01-32e5-48d1-bf16-68939d60bd02\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-dm8pw" Nov 24 13:29:04 crc kubenswrapper[4790]: I1124 13:29:04.032950 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqvkw\" (UniqueName: \"kubernetes.io/projected/cdbadd01-32e5-48d1-bf16-68939d60bd02-kube-api-access-tqvkw\") pod \"cert-manager-webhook-f4fb5df64-dm8pw\" (UID: \"cdbadd01-32e5-48d1-bf16-68939d60bd02\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-dm8pw" Nov 24 13:29:04 crc kubenswrapper[4790]: I1124 13:29:04.088716 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-dm8pw" Nov 24 13:29:04 crc kubenswrapper[4790]: I1124 13:29:04.570636 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-dm8pw"] Nov 24 13:29:05 crc kubenswrapper[4790]: I1124 13:29:05.460581 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-dm8pw" event={"ID":"cdbadd01-32e5-48d1-bf16-68939d60bd02","Type":"ContainerStarted","Data":"567c75ae1431bbd5840f1f903ba24e73dd7f390b0e9932d15b730f3205a90336"} Nov 24 13:29:06 crc kubenswrapper[4790]: I1124 13:29:06.949181 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-gz7ct"] Nov 24 13:29:06 crc kubenswrapper[4790]: I1124 13:29:06.950084 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-gz7ct" Nov 24 13:29:06 crc kubenswrapper[4790]: I1124 13:29:06.959242 4790 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-c97gj" Nov 24 13:29:06 crc kubenswrapper[4790]: I1124 13:29:06.967432 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-gz7ct"] Nov 24 13:29:07 crc kubenswrapper[4790]: I1124 13:29:07.076604 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kr5zb\" (UniqueName: \"kubernetes.io/projected/462f7ce3-f481-4d1b-b466-ffdbc9038f3a-kube-api-access-kr5zb\") pod \"cert-manager-cainjector-855d9ccff4-gz7ct\" (UID: \"462f7ce3-f481-4d1b-b466-ffdbc9038f3a\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-gz7ct" Nov 24 13:29:07 crc kubenswrapper[4790]: I1124 13:29:07.076679 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/462f7ce3-f481-4d1b-b466-ffdbc9038f3a-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-gz7ct\" (UID: \"462f7ce3-f481-4d1b-b466-ffdbc9038f3a\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-gz7ct" Nov 24 13:29:07 crc kubenswrapper[4790]: I1124 13:29:07.178732 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kr5zb\" (UniqueName: \"kubernetes.io/projected/462f7ce3-f481-4d1b-b466-ffdbc9038f3a-kube-api-access-kr5zb\") pod \"cert-manager-cainjector-855d9ccff4-gz7ct\" (UID: \"462f7ce3-f481-4d1b-b466-ffdbc9038f3a\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-gz7ct" Nov 24 13:29:07 crc kubenswrapper[4790]: I1124 13:29:07.178811 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/462f7ce3-f481-4d1b-b466-ffdbc9038f3a-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-gz7ct\" (UID: \"462f7ce3-f481-4d1b-b466-ffdbc9038f3a\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-gz7ct" Nov 24 13:29:07 crc kubenswrapper[4790]: I1124 13:29:07.201129 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/462f7ce3-f481-4d1b-b466-ffdbc9038f3a-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-gz7ct\" (UID: \"462f7ce3-f481-4d1b-b466-ffdbc9038f3a\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-gz7ct" Nov 24 13:29:07 crc kubenswrapper[4790]: I1124 13:29:07.210184 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kr5zb\" (UniqueName: \"kubernetes.io/projected/462f7ce3-f481-4d1b-b466-ffdbc9038f3a-kube-api-access-kr5zb\") pod \"cert-manager-cainjector-855d9ccff4-gz7ct\" (UID: \"462f7ce3-f481-4d1b-b466-ffdbc9038f3a\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-gz7ct" Nov 24 13:29:07 crc kubenswrapper[4790]: I1124 13:29:07.280353 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-gz7ct" Nov 24 13:29:07 crc kubenswrapper[4790]: I1124 13:29:07.484683 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-gz7ct"] Nov 24 13:29:07 crc kubenswrapper[4790]: W1124 13:29:07.521003 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod462f7ce3_f481_4d1b_b466_ffdbc9038f3a.slice/crio-07252be97c94378d73e211238d694f33595ba3ee3ecb2529154abb3b2e9d64b1 WatchSource:0}: Error finding container 07252be97c94378d73e211238d694f33595ba3ee3ecb2529154abb3b2e9d64b1: Status 404 returned error can't find the container with id 07252be97c94378d73e211238d694f33595ba3ee3ecb2529154abb3b2e9d64b1 Nov 24 13:29:08 crc kubenswrapper[4790]: I1124 13:29:08.479777 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-gz7ct" event={"ID":"462f7ce3-f481-4d1b-b466-ffdbc9038f3a","Type":"ContainerStarted","Data":"07252be97c94378d73e211238d694f33595ba3ee3ecb2529154abb3b2e9d64b1"} Nov 24 13:29:12 crc kubenswrapper[4790]: I1124 13:29:12.504089 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-gz7ct" event={"ID":"462f7ce3-f481-4d1b-b466-ffdbc9038f3a","Type":"ContainerStarted","Data":"e4d5af409297e64ab521a90503de0785a331e9bef20c5ff212fbbc2f0d4e4096"} Nov 24 13:29:12 crc kubenswrapper[4790]: I1124 13:29:12.505680 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-dm8pw" event={"ID":"cdbadd01-32e5-48d1-bf16-68939d60bd02","Type":"ContainerStarted","Data":"286148904ee7c64b88db43acb32ef5ca7b4865daadec710cbd006da9def5d022"} Nov 24 13:29:12 crc kubenswrapper[4790]: I1124 13:29:12.505761 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-f4fb5df64-dm8pw" Nov 24 13:29:12 crc kubenswrapper[4790]: I1124 13:29:12.521262 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-855d9ccff4-gz7ct" podStartSLOduration=2.38680101 podStartE2EDuration="6.521241061s" podCreationTimestamp="2025-11-24 13:29:06 +0000 UTC" firstStartedPulling="2025-11-24 13:29:07.523606845 +0000 UTC m=+995.903500507" lastFinishedPulling="2025-11-24 13:29:11.658046886 +0000 UTC m=+1000.037940558" observedRunningTime="2025-11-24 13:29:12.517105861 +0000 UTC m=+1000.896999533" watchObservedRunningTime="2025-11-24 13:29:12.521241061 +0000 UTC m=+1000.901134723" Nov 24 13:29:12 crc kubenswrapper[4790]: I1124 13:29:12.537164 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-f4fb5df64-dm8pw" podStartSLOduration=2.445782641 podStartE2EDuration="9.537142695s" podCreationTimestamp="2025-11-24 13:29:03 +0000 UTC" firstStartedPulling="2025-11-24 13:29:04.580448715 +0000 UTC m=+992.960342377" lastFinishedPulling="2025-11-24 13:29:11.671808769 +0000 UTC m=+1000.051702431" observedRunningTime="2025-11-24 13:29:12.533403645 +0000 UTC m=+1000.913297307" watchObservedRunningTime="2025-11-24 13:29:12.537142695 +0000 UTC m=+1000.917036357" Nov 24 13:29:13 crc kubenswrapper[4790]: I1124 13:29:13.938330 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:29:13 crc kubenswrapper[4790]: I1124 13:29:13.939066 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:29:19 crc kubenswrapper[4790]: I1124 13:29:19.093450 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-f4fb5df64-dm8pw" Nov 24 13:29:22 crc kubenswrapper[4790]: I1124 13:29:22.716164 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-86cb77c54b-q2m9q"] Nov 24 13:29:22 crc kubenswrapper[4790]: I1124 13:29:22.717323 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-q2m9q" Nov 24 13:29:22 crc kubenswrapper[4790]: I1124 13:29:22.726101 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-q2m9q"] Nov 24 13:29:22 crc kubenswrapper[4790]: I1124 13:29:22.728041 4790 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-96n6b" Nov 24 13:29:22 crc kubenswrapper[4790]: I1124 13:29:22.813342 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mt4z2\" (UniqueName: \"kubernetes.io/projected/4da891b8-efa8-4c31-921a-878c9b78c8a6-kube-api-access-mt4z2\") pod \"cert-manager-86cb77c54b-q2m9q\" (UID: \"4da891b8-efa8-4c31-921a-878c9b78c8a6\") " pod="cert-manager/cert-manager-86cb77c54b-q2m9q" Nov 24 13:29:22 crc kubenswrapper[4790]: I1124 13:29:22.813632 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4da891b8-efa8-4c31-921a-878c9b78c8a6-bound-sa-token\") pod \"cert-manager-86cb77c54b-q2m9q\" (UID: \"4da891b8-efa8-4c31-921a-878c9b78c8a6\") " pod="cert-manager/cert-manager-86cb77c54b-q2m9q" Nov 24 13:29:22 crc kubenswrapper[4790]: I1124 13:29:22.915324 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mt4z2\" (UniqueName: \"kubernetes.io/projected/4da891b8-efa8-4c31-921a-878c9b78c8a6-kube-api-access-mt4z2\") pod \"cert-manager-86cb77c54b-q2m9q\" (UID: \"4da891b8-efa8-4c31-921a-878c9b78c8a6\") " pod="cert-manager/cert-manager-86cb77c54b-q2m9q" Nov 24 13:29:22 crc kubenswrapper[4790]: I1124 13:29:22.915368 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4da891b8-efa8-4c31-921a-878c9b78c8a6-bound-sa-token\") pod \"cert-manager-86cb77c54b-q2m9q\" (UID: \"4da891b8-efa8-4c31-921a-878c9b78c8a6\") " pod="cert-manager/cert-manager-86cb77c54b-q2m9q" Nov 24 13:29:22 crc kubenswrapper[4790]: I1124 13:29:22.934400 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4da891b8-efa8-4c31-921a-878c9b78c8a6-bound-sa-token\") pod \"cert-manager-86cb77c54b-q2m9q\" (UID: \"4da891b8-efa8-4c31-921a-878c9b78c8a6\") " pod="cert-manager/cert-manager-86cb77c54b-q2m9q" Nov 24 13:29:22 crc kubenswrapper[4790]: I1124 13:29:22.934412 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mt4z2\" (UniqueName: \"kubernetes.io/projected/4da891b8-efa8-4c31-921a-878c9b78c8a6-kube-api-access-mt4z2\") pod \"cert-manager-86cb77c54b-q2m9q\" (UID: \"4da891b8-efa8-4c31-921a-878c9b78c8a6\") " pod="cert-manager/cert-manager-86cb77c54b-q2m9q" Nov 24 13:29:23 crc kubenswrapper[4790]: I1124 13:29:23.040921 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-q2m9q" Nov 24 13:29:23 crc kubenswrapper[4790]: I1124 13:29:23.433574 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-q2m9q"] Nov 24 13:29:23 crc kubenswrapper[4790]: I1124 13:29:23.574984 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-q2m9q" event={"ID":"4da891b8-efa8-4c31-921a-878c9b78c8a6","Type":"ContainerStarted","Data":"1e165c46b30669701ea2ffbf30e078cfa6860d98abf0cdf54c3bc27facdcb6dd"} Nov 24 13:29:24 crc kubenswrapper[4790]: I1124 13:29:24.583030 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-q2m9q" event={"ID":"4da891b8-efa8-4c31-921a-878c9b78c8a6","Type":"ContainerStarted","Data":"e996e729ee1c982a410e6daaeee8627b5599bbe43b85f979d008dab846d20c32"} Nov 24 13:29:24 crc kubenswrapper[4790]: I1124 13:29:24.599709 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-86cb77c54b-q2m9q" podStartSLOduration=2.599691836 podStartE2EDuration="2.599691836s" podCreationTimestamp="2025-11-24 13:29:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:29:24.598259042 +0000 UTC m=+1012.978152714" watchObservedRunningTime="2025-11-24 13:29:24.599691836 +0000 UTC m=+1012.979585498" Nov 24 13:29:32 crc kubenswrapper[4790]: I1124 13:29:32.867737 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-vkxbs"] Nov 24 13:29:32 crc kubenswrapper[4790]: I1124 13:29:32.868955 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-vkxbs" Nov 24 13:29:32 crc kubenswrapper[4790]: I1124 13:29:32.873506 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 24 13:29:32 crc kubenswrapper[4790]: I1124 13:29:32.873566 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 24 13:29:32 crc kubenswrapper[4790]: I1124 13:29:32.873851 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-tbhfc" Nov 24 13:29:32 crc kubenswrapper[4790]: I1124 13:29:32.930866 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-vkxbs"] Nov 24 13:29:33 crc kubenswrapper[4790]: I1124 13:29:33.039984 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jxjw\" (UniqueName: \"kubernetes.io/projected/382216c5-b05f-4e86-b5c2-eb93a948c4e5-kube-api-access-2jxjw\") pod \"openstack-operator-index-vkxbs\" (UID: \"382216c5-b05f-4e86-b5c2-eb93a948c4e5\") " pod="openstack-operators/openstack-operator-index-vkxbs" Nov 24 13:29:33 crc kubenswrapper[4790]: I1124 13:29:33.141268 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jxjw\" (UniqueName: \"kubernetes.io/projected/382216c5-b05f-4e86-b5c2-eb93a948c4e5-kube-api-access-2jxjw\") pod \"openstack-operator-index-vkxbs\" (UID: \"382216c5-b05f-4e86-b5c2-eb93a948c4e5\") " pod="openstack-operators/openstack-operator-index-vkxbs" Nov 24 13:29:33 crc kubenswrapper[4790]: I1124 13:29:33.167185 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jxjw\" (UniqueName: \"kubernetes.io/projected/382216c5-b05f-4e86-b5c2-eb93a948c4e5-kube-api-access-2jxjw\") pod \"openstack-operator-index-vkxbs\" (UID: \"382216c5-b05f-4e86-b5c2-eb93a948c4e5\") " pod="openstack-operators/openstack-operator-index-vkxbs" Nov 24 13:29:33 crc kubenswrapper[4790]: I1124 13:29:33.193292 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-vkxbs" Nov 24 13:29:33 crc kubenswrapper[4790]: I1124 13:29:33.584150 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-vkxbs"] Nov 24 13:29:33 crc kubenswrapper[4790]: I1124 13:29:33.640554 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-vkxbs" event={"ID":"382216c5-b05f-4e86-b5c2-eb93a948c4e5","Type":"ContainerStarted","Data":"0dbdae1a60e981931b8975b68254cd168dfd592b4df6dd9a15a2f237e60f3fd8"} Nov 24 13:29:36 crc kubenswrapper[4790]: I1124 13:29:36.662545 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-vkxbs" event={"ID":"382216c5-b05f-4e86-b5c2-eb93a948c4e5","Type":"ContainerStarted","Data":"cfaa55d42f640e831818f4f41aecfe78f6dd7dc579ddcb12e0c88a849f4611b2"} Nov 24 13:29:36 crc kubenswrapper[4790]: I1124 13:29:36.685823 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-vkxbs" podStartSLOduration=2.694424314 podStartE2EDuration="4.685807756s" podCreationTimestamp="2025-11-24 13:29:32 +0000 UTC" firstStartedPulling="2025-11-24 13:29:33.592211578 +0000 UTC m=+1021.972105250" lastFinishedPulling="2025-11-24 13:29:35.58359503 +0000 UTC m=+1023.963488692" observedRunningTime="2025-11-24 13:29:36.68305267 +0000 UTC m=+1025.062946342" watchObservedRunningTime="2025-11-24 13:29:36.685807756 +0000 UTC m=+1025.065701418" Nov 24 13:29:43 crc kubenswrapper[4790]: I1124 13:29:43.194467 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-vkxbs" Nov 24 13:29:43 crc kubenswrapper[4790]: I1124 13:29:43.194783 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-vkxbs" Nov 24 13:29:43 crc kubenswrapper[4790]: I1124 13:29:43.219103 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-vkxbs" Nov 24 13:29:43 crc kubenswrapper[4790]: I1124 13:29:43.732277 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-vkxbs" Nov 24 13:29:43 crc kubenswrapper[4790]: I1124 13:29:43.938847 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:29:43 crc kubenswrapper[4790]: I1124 13:29:43.939385 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:29:48 crc kubenswrapper[4790]: I1124 13:29:48.693682 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch"] Nov 24 13:29:48 crc kubenswrapper[4790]: I1124 13:29:48.695107 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch" Nov 24 13:29:48 crc kubenswrapper[4790]: I1124 13:29:48.697135 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-mrg9r" Nov 24 13:29:48 crc kubenswrapper[4790]: I1124 13:29:48.703813 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch"] Nov 24 13:29:48 crc kubenswrapper[4790]: I1124 13:29:48.866363 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pr5k\" (UniqueName: \"kubernetes.io/projected/8c370b97-cb4b-410e-b8cb-fd5b807de698-kube-api-access-2pr5k\") pod \"f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch\" (UID: \"8c370b97-cb4b-410e-b8cb-fd5b807de698\") " pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch" Nov 24 13:29:48 crc kubenswrapper[4790]: I1124 13:29:48.866441 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8c370b97-cb4b-410e-b8cb-fd5b807de698-util\") pod \"f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch\" (UID: \"8c370b97-cb4b-410e-b8cb-fd5b807de698\") " pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch" Nov 24 13:29:48 crc kubenswrapper[4790]: I1124 13:29:48.866480 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8c370b97-cb4b-410e-b8cb-fd5b807de698-bundle\") pod \"f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch\" (UID: \"8c370b97-cb4b-410e-b8cb-fd5b807de698\") " pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch" Nov 24 13:29:48 crc kubenswrapper[4790]: I1124 13:29:48.967936 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8c370b97-cb4b-410e-b8cb-fd5b807de698-util\") pod \"f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch\" (UID: \"8c370b97-cb4b-410e-b8cb-fd5b807de698\") " pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch" Nov 24 13:29:48 crc kubenswrapper[4790]: I1124 13:29:48.968358 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8c370b97-cb4b-410e-b8cb-fd5b807de698-bundle\") pod \"f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch\" (UID: \"8c370b97-cb4b-410e-b8cb-fd5b807de698\") " pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch" Nov 24 13:29:48 crc kubenswrapper[4790]: I1124 13:29:48.968487 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pr5k\" (UniqueName: \"kubernetes.io/projected/8c370b97-cb4b-410e-b8cb-fd5b807de698-kube-api-access-2pr5k\") pod \"f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch\" (UID: \"8c370b97-cb4b-410e-b8cb-fd5b807de698\") " pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch" Nov 24 13:29:48 crc kubenswrapper[4790]: I1124 13:29:48.968413 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8c370b97-cb4b-410e-b8cb-fd5b807de698-util\") pod \"f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch\" (UID: \"8c370b97-cb4b-410e-b8cb-fd5b807de698\") " pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch" Nov 24 13:29:48 crc kubenswrapper[4790]: I1124 13:29:48.968778 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8c370b97-cb4b-410e-b8cb-fd5b807de698-bundle\") pod \"f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch\" (UID: \"8c370b97-cb4b-410e-b8cb-fd5b807de698\") " pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch" Nov 24 13:29:48 crc kubenswrapper[4790]: I1124 13:29:48.986404 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pr5k\" (UniqueName: \"kubernetes.io/projected/8c370b97-cb4b-410e-b8cb-fd5b807de698-kube-api-access-2pr5k\") pod \"f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch\" (UID: \"8c370b97-cb4b-410e-b8cb-fd5b807de698\") " pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch" Nov 24 13:29:49 crc kubenswrapper[4790]: I1124 13:29:49.015244 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch" Nov 24 13:29:49 crc kubenswrapper[4790]: I1124 13:29:49.215076 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch"] Nov 24 13:29:49 crc kubenswrapper[4790]: I1124 13:29:49.741000 4790 generic.go:334] "Generic (PLEG): container finished" podID="8c370b97-cb4b-410e-b8cb-fd5b807de698" containerID="35cda6917adbc2f03b395a95cda1a5a114a59a54cf2f07319f76ab272bae1461" exitCode=0 Nov 24 13:29:49 crc kubenswrapper[4790]: I1124 13:29:49.741065 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch" event={"ID":"8c370b97-cb4b-410e-b8cb-fd5b807de698","Type":"ContainerDied","Data":"35cda6917adbc2f03b395a95cda1a5a114a59a54cf2f07319f76ab272bae1461"} Nov 24 13:29:49 crc kubenswrapper[4790]: I1124 13:29:49.741325 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch" event={"ID":"8c370b97-cb4b-410e-b8cb-fd5b807de698","Type":"ContainerStarted","Data":"5feadf31a5c682b4f22cc51538e7f689b3d24867cc3bf595fa6f94343758416e"} Nov 24 13:29:50 crc kubenswrapper[4790]: I1124 13:29:50.750730 4790 generic.go:334] "Generic (PLEG): container finished" podID="8c370b97-cb4b-410e-b8cb-fd5b807de698" containerID="873515ef653f09ea6c537ae7b91ebf78abd84d6b66caf49d81a4f68f548f4615" exitCode=0 Nov 24 13:29:50 crc kubenswrapper[4790]: I1124 13:29:50.750808 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch" event={"ID":"8c370b97-cb4b-410e-b8cb-fd5b807de698","Type":"ContainerDied","Data":"873515ef653f09ea6c537ae7b91ebf78abd84d6b66caf49d81a4f68f548f4615"} Nov 24 13:29:51 crc kubenswrapper[4790]: I1124 13:29:51.762051 4790 generic.go:334] "Generic (PLEG): container finished" podID="8c370b97-cb4b-410e-b8cb-fd5b807de698" containerID="b904b0f297920eb1bd38527998a4708695c18ea9e94907cfd6d801f0cca3becf" exitCode=0 Nov 24 13:29:51 crc kubenswrapper[4790]: I1124 13:29:51.762110 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch" event={"ID":"8c370b97-cb4b-410e-b8cb-fd5b807de698","Type":"ContainerDied","Data":"b904b0f297920eb1bd38527998a4708695c18ea9e94907cfd6d801f0cca3becf"} Nov 24 13:29:52 crc kubenswrapper[4790]: I1124 13:29:52.996373 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch" Nov 24 13:29:53 crc kubenswrapper[4790]: I1124 13:29:53.124829 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8c370b97-cb4b-410e-b8cb-fd5b807de698-bundle\") pod \"8c370b97-cb4b-410e-b8cb-fd5b807de698\" (UID: \"8c370b97-cb4b-410e-b8cb-fd5b807de698\") " Nov 24 13:29:53 crc kubenswrapper[4790]: I1124 13:29:53.125166 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pr5k\" (UniqueName: \"kubernetes.io/projected/8c370b97-cb4b-410e-b8cb-fd5b807de698-kube-api-access-2pr5k\") pod \"8c370b97-cb4b-410e-b8cb-fd5b807de698\" (UID: \"8c370b97-cb4b-410e-b8cb-fd5b807de698\") " Nov 24 13:29:53 crc kubenswrapper[4790]: I1124 13:29:53.125251 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8c370b97-cb4b-410e-b8cb-fd5b807de698-util\") pod \"8c370b97-cb4b-410e-b8cb-fd5b807de698\" (UID: \"8c370b97-cb4b-410e-b8cb-fd5b807de698\") " Nov 24 13:29:53 crc kubenswrapper[4790]: I1124 13:29:53.128439 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c370b97-cb4b-410e-b8cb-fd5b807de698-bundle" (OuterVolumeSpecName: "bundle") pod "8c370b97-cb4b-410e-b8cb-fd5b807de698" (UID: "8c370b97-cb4b-410e-b8cb-fd5b807de698"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:29:53 crc kubenswrapper[4790]: I1124 13:29:53.130869 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c370b97-cb4b-410e-b8cb-fd5b807de698-kube-api-access-2pr5k" (OuterVolumeSpecName: "kube-api-access-2pr5k") pod "8c370b97-cb4b-410e-b8cb-fd5b807de698" (UID: "8c370b97-cb4b-410e-b8cb-fd5b807de698"). InnerVolumeSpecName "kube-api-access-2pr5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:29:53 crc kubenswrapper[4790]: I1124 13:29:53.138753 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c370b97-cb4b-410e-b8cb-fd5b807de698-util" (OuterVolumeSpecName: "util") pod "8c370b97-cb4b-410e-b8cb-fd5b807de698" (UID: "8c370b97-cb4b-410e-b8cb-fd5b807de698"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:29:53 crc kubenswrapper[4790]: I1124 13:29:53.227738 4790 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8c370b97-cb4b-410e-b8cb-fd5b807de698-util\") on node \"crc\" DevicePath \"\"" Nov 24 13:29:53 crc kubenswrapper[4790]: I1124 13:29:53.227776 4790 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8c370b97-cb4b-410e-b8cb-fd5b807de698-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:29:53 crc kubenswrapper[4790]: I1124 13:29:53.227796 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pr5k\" (UniqueName: \"kubernetes.io/projected/8c370b97-cb4b-410e-b8cb-fd5b807de698-kube-api-access-2pr5k\") on node \"crc\" DevicePath \"\"" Nov 24 13:29:53 crc kubenswrapper[4790]: I1124 13:29:53.776031 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch" event={"ID":"8c370b97-cb4b-410e-b8cb-fd5b807de698","Type":"ContainerDied","Data":"5feadf31a5c682b4f22cc51538e7f689b3d24867cc3bf595fa6f94343758416e"} Nov 24 13:29:53 crc kubenswrapper[4790]: I1124 13:29:53.776123 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5feadf31a5c682b4f22cc51538e7f689b3d24867cc3bf595fa6f94343758416e" Nov 24 13:29:53 crc kubenswrapper[4790]: I1124 13:29:53.776080 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch" Nov 24 13:29:56 crc kubenswrapper[4790]: I1124 13:29:56.802114 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-866b74d8b7-t2h2s"] Nov 24 13:29:56 crc kubenswrapper[4790]: E1124 13:29:56.803327 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c370b97-cb4b-410e-b8cb-fd5b807de698" containerName="pull" Nov 24 13:29:56 crc kubenswrapper[4790]: I1124 13:29:56.803395 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c370b97-cb4b-410e-b8cb-fd5b807de698" containerName="pull" Nov 24 13:29:56 crc kubenswrapper[4790]: E1124 13:29:56.803486 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c370b97-cb4b-410e-b8cb-fd5b807de698" containerName="extract" Nov 24 13:29:56 crc kubenswrapper[4790]: I1124 13:29:56.803537 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c370b97-cb4b-410e-b8cb-fd5b807de698" containerName="extract" Nov 24 13:29:56 crc kubenswrapper[4790]: E1124 13:29:56.803596 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c370b97-cb4b-410e-b8cb-fd5b807de698" containerName="util" Nov 24 13:29:56 crc kubenswrapper[4790]: I1124 13:29:56.803645 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c370b97-cb4b-410e-b8cb-fd5b807de698" containerName="util" Nov 24 13:29:56 crc kubenswrapper[4790]: I1124 13:29:56.803806 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c370b97-cb4b-410e-b8cb-fd5b807de698" containerName="extract" Nov 24 13:29:56 crc kubenswrapper[4790]: I1124 13:29:56.804422 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-866b74d8b7-t2h2s" Nov 24 13:29:56 crc kubenswrapper[4790]: I1124 13:29:56.806519 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-t4nd2" Nov 24 13:29:56 crc kubenswrapper[4790]: I1124 13:29:56.846547 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-866b74d8b7-t2h2s"] Nov 24 13:29:56 crc kubenswrapper[4790]: I1124 13:29:56.980152 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvzxw\" (UniqueName: \"kubernetes.io/projected/123e2f50-7f37-422b-a3ca-91252d609541-kube-api-access-nvzxw\") pod \"openstack-operator-controller-operator-866b74d8b7-t2h2s\" (UID: \"123e2f50-7f37-422b-a3ca-91252d609541\") " pod="openstack-operators/openstack-operator-controller-operator-866b74d8b7-t2h2s" Nov 24 13:29:57 crc kubenswrapper[4790]: I1124 13:29:57.081201 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvzxw\" (UniqueName: \"kubernetes.io/projected/123e2f50-7f37-422b-a3ca-91252d609541-kube-api-access-nvzxw\") pod \"openstack-operator-controller-operator-866b74d8b7-t2h2s\" (UID: \"123e2f50-7f37-422b-a3ca-91252d609541\") " pod="openstack-operators/openstack-operator-controller-operator-866b74d8b7-t2h2s" Nov 24 13:29:57 crc kubenswrapper[4790]: I1124 13:29:57.099918 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvzxw\" (UniqueName: \"kubernetes.io/projected/123e2f50-7f37-422b-a3ca-91252d609541-kube-api-access-nvzxw\") pod \"openstack-operator-controller-operator-866b74d8b7-t2h2s\" (UID: \"123e2f50-7f37-422b-a3ca-91252d609541\") " pod="openstack-operators/openstack-operator-controller-operator-866b74d8b7-t2h2s" Nov 24 13:29:57 crc kubenswrapper[4790]: I1124 13:29:57.120754 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-866b74d8b7-t2h2s" Nov 24 13:29:57 crc kubenswrapper[4790]: I1124 13:29:57.361572 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-866b74d8b7-t2h2s"] Nov 24 13:29:57 crc kubenswrapper[4790]: I1124 13:29:57.887746 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-866b74d8b7-t2h2s" event={"ID":"123e2f50-7f37-422b-a3ca-91252d609541","Type":"ContainerStarted","Data":"ae47e4965eb39f408fc1bc73f9243c0d2fb65dee51d27410e98b90aa77ebcdbb"} Nov 24 13:30:00 crc kubenswrapper[4790]: I1124 13:30:00.130651 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399850-dhxbs"] Nov 24 13:30:00 crc kubenswrapper[4790]: I1124 13:30:00.131594 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399850-dhxbs" Nov 24 13:30:00 crc kubenswrapper[4790]: I1124 13:30:00.134271 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 13:30:00 crc kubenswrapper[4790]: I1124 13:30:00.135709 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 13:30:00 crc kubenswrapper[4790]: I1124 13:30:00.150777 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399850-dhxbs"] Nov 24 13:30:00 crc kubenswrapper[4790]: I1124 13:30:00.251940 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9ed9d0d9-aa9f-44ec-918a-59231503a4ba-secret-volume\") pod \"collect-profiles-29399850-dhxbs\" (UID: \"9ed9d0d9-aa9f-44ec-918a-59231503a4ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399850-dhxbs" Nov 24 13:30:00 crc kubenswrapper[4790]: I1124 13:30:00.252143 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sczj5\" (UniqueName: \"kubernetes.io/projected/9ed9d0d9-aa9f-44ec-918a-59231503a4ba-kube-api-access-sczj5\") pod \"collect-profiles-29399850-dhxbs\" (UID: \"9ed9d0d9-aa9f-44ec-918a-59231503a4ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399850-dhxbs" Nov 24 13:30:00 crc kubenswrapper[4790]: I1124 13:30:00.252195 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9ed9d0d9-aa9f-44ec-918a-59231503a4ba-config-volume\") pod \"collect-profiles-29399850-dhxbs\" (UID: \"9ed9d0d9-aa9f-44ec-918a-59231503a4ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399850-dhxbs" Nov 24 13:30:00 crc kubenswrapper[4790]: I1124 13:30:00.354292 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9ed9d0d9-aa9f-44ec-918a-59231503a4ba-secret-volume\") pod \"collect-profiles-29399850-dhxbs\" (UID: \"9ed9d0d9-aa9f-44ec-918a-59231503a4ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399850-dhxbs" Nov 24 13:30:00 crc kubenswrapper[4790]: I1124 13:30:00.354387 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sczj5\" (UniqueName: \"kubernetes.io/projected/9ed9d0d9-aa9f-44ec-918a-59231503a4ba-kube-api-access-sczj5\") pod \"collect-profiles-29399850-dhxbs\" (UID: \"9ed9d0d9-aa9f-44ec-918a-59231503a4ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399850-dhxbs" Nov 24 13:30:00 crc kubenswrapper[4790]: I1124 13:30:00.354430 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9ed9d0d9-aa9f-44ec-918a-59231503a4ba-config-volume\") pod \"collect-profiles-29399850-dhxbs\" (UID: \"9ed9d0d9-aa9f-44ec-918a-59231503a4ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399850-dhxbs" Nov 24 13:30:00 crc kubenswrapper[4790]: I1124 13:30:00.355602 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9ed9d0d9-aa9f-44ec-918a-59231503a4ba-config-volume\") pod \"collect-profiles-29399850-dhxbs\" (UID: \"9ed9d0d9-aa9f-44ec-918a-59231503a4ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399850-dhxbs" Nov 24 13:30:00 crc kubenswrapper[4790]: I1124 13:30:00.363347 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9ed9d0d9-aa9f-44ec-918a-59231503a4ba-secret-volume\") pod \"collect-profiles-29399850-dhxbs\" (UID: \"9ed9d0d9-aa9f-44ec-918a-59231503a4ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399850-dhxbs" Nov 24 13:30:00 crc kubenswrapper[4790]: I1124 13:30:00.379395 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sczj5\" (UniqueName: \"kubernetes.io/projected/9ed9d0d9-aa9f-44ec-918a-59231503a4ba-kube-api-access-sczj5\") pod \"collect-profiles-29399850-dhxbs\" (UID: \"9ed9d0d9-aa9f-44ec-918a-59231503a4ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399850-dhxbs" Nov 24 13:30:00 crc kubenswrapper[4790]: I1124 13:30:00.458715 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399850-dhxbs" Nov 24 13:30:01 crc kubenswrapper[4790]: I1124 13:30:01.653588 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399850-dhxbs"] Nov 24 13:30:01 crc kubenswrapper[4790]: W1124 13:30:01.666515 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9ed9d0d9_aa9f_44ec_918a_59231503a4ba.slice/crio-59c4bec115b49f588ae285b6eec9317f7662fcf1e144c4001ca568514e7a0ffd WatchSource:0}: Error finding container 59c4bec115b49f588ae285b6eec9317f7662fcf1e144c4001ca568514e7a0ffd: Status 404 returned error can't find the container with id 59c4bec115b49f588ae285b6eec9317f7662fcf1e144c4001ca568514e7a0ffd Nov 24 13:30:01 crc kubenswrapper[4790]: I1124 13:30:01.918691 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-866b74d8b7-t2h2s" event={"ID":"123e2f50-7f37-422b-a3ca-91252d609541","Type":"ContainerStarted","Data":"7c4d4cd7c2c133616b2d79cf73fee4582f0862597cb83941c07b6eb038e26d04"} Nov 24 13:30:01 crc kubenswrapper[4790]: I1124 13:30:01.926216 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399850-dhxbs" event={"ID":"9ed9d0d9-aa9f-44ec-918a-59231503a4ba","Type":"ContainerStarted","Data":"59c4bec115b49f588ae285b6eec9317f7662fcf1e144c4001ca568514e7a0ffd"} Nov 24 13:30:02 crc kubenswrapper[4790]: I1124 13:30:02.934478 4790 generic.go:334] "Generic (PLEG): container finished" podID="9ed9d0d9-aa9f-44ec-918a-59231503a4ba" containerID="b53df9baaaa61a1ffd04ee843ecc8ca5fc1c09e5b75dae56c5c6b1bb8b96c4b5" exitCode=0 Nov 24 13:30:02 crc kubenswrapper[4790]: I1124 13:30:02.934555 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399850-dhxbs" event={"ID":"9ed9d0d9-aa9f-44ec-918a-59231503a4ba","Type":"ContainerDied","Data":"b53df9baaaa61a1ffd04ee843ecc8ca5fc1c09e5b75dae56c5c6b1bb8b96c4b5"} Nov 24 13:30:03 crc kubenswrapper[4790]: I1124 13:30:03.943952 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-866b74d8b7-t2h2s" event={"ID":"123e2f50-7f37-422b-a3ca-91252d609541","Type":"ContainerStarted","Data":"c29b262a9af3afee2f189cb203c25ed86ade6d135a4f94dcf5e709f88590250c"} Nov 24 13:30:03 crc kubenswrapper[4790]: I1124 13:30:03.975846 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-866b74d8b7-t2h2s" podStartSLOduration=1.614791989 podStartE2EDuration="7.975827903s" podCreationTimestamp="2025-11-24 13:29:56 +0000 UTC" firstStartedPulling="2025-11-24 13:29:57.375771836 +0000 UTC m=+1045.755665498" lastFinishedPulling="2025-11-24 13:30:03.73680775 +0000 UTC m=+1052.116701412" observedRunningTime="2025-11-24 13:30:03.972772354 +0000 UTC m=+1052.352666036" watchObservedRunningTime="2025-11-24 13:30:03.975827903 +0000 UTC m=+1052.355721565" Nov 24 13:30:04 crc kubenswrapper[4790]: I1124 13:30:04.231476 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399850-dhxbs" Nov 24 13:30:04 crc kubenswrapper[4790]: I1124 13:30:04.311756 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9ed9d0d9-aa9f-44ec-918a-59231503a4ba-config-volume\") pod \"9ed9d0d9-aa9f-44ec-918a-59231503a4ba\" (UID: \"9ed9d0d9-aa9f-44ec-918a-59231503a4ba\") " Nov 24 13:30:04 crc kubenswrapper[4790]: I1124 13:30:04.311848 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sczj5\" (UniqueName: \"kubernetes.io/projected/9ed9d0d9-aa9f-44ec-918a-59231503a4ba-kube-api-access-sczj5\") pod \"9ed9d0d9-aa9f-44ec-918a-59231503a4ba\" (UID: \"9ed9d0d9-aa9f-44ec-918a-59231503a4ba\") " Nov 24 13:30:04 crc kubenswrapper[4790]: I1124 13:30:04.311988 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9ed9d0d9-aa9f-44ec-918a-59231503a4ba-secret-volume\") pod \"9ed9d0d9-aa9f-44ec-918a-59231503a4ba\" (UID: \"9ed9d0d9-aa9f-44ec-918a-59231503a4ba\") " Nov 24 13:30:04 crc kubenswrapper[4790]: I1124 13:30:04.312284 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ed9d0d9-aa9f-44ec-918a-59231503a4ba-config-volume" (OuterVolumeSpecName: "config-volume") pod "9ed9d0d9-aa9f-44ec-918a-59231503a4ba" (UID: "9ed9d0d9-aa9f-44ec-918a-59231503a4ba"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:30:04 crc kubenswrapper[4790]: I1124 13:30:04.319462 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ed9d0d9-aa9f-44ec-918a-59231503a4ba-kube-api-access-sczj5" (OuterVolumeSpecName: "kube-api-access-sczj5") pod "9ed9d0d9-aa9f-44ec-918a-59231503a4ba" (UID: "9ed9d0d9-aa9f-44ec-918a-59231503a4ba"). InnerVolumeSpecName "kube-api-access-sczj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:30:04 crc kubenswrapper[4790]: I1124 13:30:04.326068 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ed9d0d9-aa9f-44ec-918a-59231503a4ba-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9ed9d0d9-aa9f-44ec-918a-59231503a4ba" (UID: "9ed9d0d9-aa9f-44ec-918a-59231503a4ba"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:30:04 crc kubenswrapper[4790]: I1124 13:30:04.413843 4790 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9ed9d0d9-aa9f-44ec-918a-59231503a4ba-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 13:30:04 crc kubenswrapper[4790]: I1124 13:30:04.413903 4790 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9ed9d0d9-aa9f-44ec-918a-59231503a4ba-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 13:30:04 crc kubenswrapper[4790]: I1124 13:30:04.413916 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sczj5\" (UniqueName: \"kubernetes.io/projected/9ed9d0d9-aa9f-44ec-918a-59231503a4ba-kube-api-access-sczj5\") on node \"crc\" DevicePath \"\"" Nov 24 13:30:04 crc kubenswrapper[4790]: I1124 13:30:04.955113 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399850-dhxbs" event={"ID":"9ed9d0d9-aa9f-44ec-918a-59231503a4ba","Type":"ContainerDied","Data":"59c4bec115b49f588ae285b6eec9317f7662fcf1e144c4001ca568514e7a0ffd"} Nov 24 13:30:04 crc kubenswrapper[4790]: I1124 13:30:04.955523 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59c4bec115b49f588ae285b6eec9317f7662fcf1e144c4001ca568514e7a0ffd" Nov 24 13:30:04 crc kubenswrapper[4790]: I1124 13:30:04.955550 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-866b74d8b7-t2h2s" Nov 24 13:30:04 crc kubenswrapper[4790]: I1124 13:30:04.955168 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399850-dhxbs" Nov 24 13:30:07 crc kubenswrapper[4790]: I1124 13:30:07.124549 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-866b74d8b7-t2h2s" Nov 24 13:30:13 crc kubenswrapper[4790]: I1124 13:30:13.939232 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:30:13 crc kubenswrapper[4790]: I1124 13:30:13.939829 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:30:13 crc kubenswrapper[4790]: I1124 13:30:13.939907 4790 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 13:30:13 crc kubenswrapper[4790]: I1124 13:30:13.940632 4790 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e635b1d9a632257918b8c7ec939010f4c4491014f07d31a3ef3de514129b90d0"} pod="openshift-machine-config-operator/machine-config-daemon-xz49t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 13:30:13 crc kubenswrapper[4790]: I1124 13:30:13.940692 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" containerID="cri-o://e635b1d9a632257918b8c7ec939010f4c4491014f07d31a3ef3de514129b90d0" gracePeriod=600 Nov 24 13:30:15 crc kubenswrapper[4790]: I1124 13:30:15.018754 4790 generic.go:334] "Generic (PLEG): container finished" podID="0d73b133-48f0-455f-8f6a-742e633f631a" containerID="e635b1d9a632257918b8c7ec939010f4c4491014f07d31a3ef3de514129b90d0" exitCode=0 Nov 24 13:30:15 crc kubenswrapper[4790]: I1124 13:30:15.018822 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerDied","Data":"e635b1d9a632257918b8c7ec939010f4c4491014f07d31a3ef3de514129b90d0"} Nov 24 13:30:15 crc kubenswrapper[4790]: I1124 13:30:15.019051 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"56e28eeaa0ab287a217413e69cba64d88c1ede93557f55581b2a1331d38d609a"} Nov 24 13:30:15 crc kubenswrapper[4790]: I1124 13:30:15.019074 4790 scope.go:117] "RemoveContainer" containerID="039103a6742b65a25f073d9f23e4fb704fe4fd769ff785303a1ff06bb7cfe2d2" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.162933 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-75fb479bcc-9wl9p"] Nov 24 13:30:24 crc kubenswrapper[4790]: E1124 13:30:24.163689 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ed9d0d9-aa9f-44ec-918a-59231503a4ba" containerName="collect-profiles" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.163703 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ed9d0d9-aa9f-44ec-918a-59231503a4ba" containerName="collect-profiles" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.163837 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ed9d0d9-aa9f-44ec-918a-59231503a4ba" containerName="collect-profiles" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.164441 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-9wl9p" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.169974 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-7tmtz" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.179052 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6498cbf48f-44wlb"] Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.180969 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-44wlb" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.182934 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-mhmpl" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.188558 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-75fb479bcc-9wl9p"] Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.191899 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6498cbf48f-44wlb"] Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.203806 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-767ccfd65f-h8nkv"] Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.205145 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-h8nkv" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.207535 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-6x2xj" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.222242 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-767ccfd65f-h8nkv"] Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.231551 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-7969689c84-wq7t5"] Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.232569 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7969689c84-wq7t5" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.240268 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-l99dv" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.245664 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7969689c84-wq7t5"] Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.251490 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-56f54d6746-bzh6h"] Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.252445 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-bzh6h" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.260023 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-85phr" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.290079 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-598f69df5d-95l27"] Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.291398 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-95l27" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.293546 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpm25\" (UniqueName: \"kubernetes.io/projected/b2b9b9e9-800e-4213-8605-058d91412276-kube-api-access-bpm25\") pod \"designate-operator-controller-manager-767ccfd65f-h8nkv\" (UID: \"b2b9b9e9-800e-4213-8605-058d91412276\") " pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-h8nkv" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.293609 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l22cc\" (UniqueName: \"kubernetes.io/projected/d39e8176-e052-4a0e-9d16-e66b6312fbfc-kube-api-access-l22cc\") pod \"glance-operator-controller-manager-7969689c84-wq7t5\" (UID: \"d39e8176-e052-4a0e-9d16-e66b6312fbfc\") " pod="openstack-operators/glance-operator-controller-manager-7969689c84-wq7t5" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.293637 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w889m\" (UniqueName: \"kubernetes.io/projected/b55ba4e8-3d0a-4746-817c-feb1a210b01e-kube-api-access-w889m\") pod \"barbican-operator-controller-manager-75fb479bcc-9wl9p\" (UID: \"b55ba4e8-3d0a-4746-817c-feb1a210b01e\") " pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-9wl9p" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.293662 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpcgg\" (UniqueName: \"kubernetes.io/projected/49de2fb5-2dde-400e-9ae4-d36dfe196ff1-kube-api-access-kpcgg\") pod \"heat-operator-controller-manager-56f54d6746-bzh6h\" (UID: \"49de2fb5-2dde-400e-9ae4-d36dfe196ff1\") " pod="openstack-operators/heat-operator-controller-manager-56f54d6746-bzh6h" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.293708 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5bkn\" (UniqueName: \"kubernetes.io/projected/09e60927-110e-4cda-b7e3-6f47e70f2388-kube-api-access-d5bkn\") pod \"cinder-operator-controller-manager-6498cbf48f-44wlb\" (UID: \"09e60927-110e-4cda-b7e3-6f47e70f2388\") " pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-44wlb" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.295564 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-kvnjk" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.295702 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-56f54d6746-bzh6h"] Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.306602 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-6dd8864d7c-fxplj"] Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.307862 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-fxplj" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.312588 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-598f69df5d-95l27"] Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.318812 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-x5cjh" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.319055 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.334679 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-99b499f4-ntfjb"] Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.335752 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6dd8864d7c-fxplj"] Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.335777 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7454b96578-4bndq"] Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.336804 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-4bndq" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.337387 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-ntfjb" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.343197 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-mmf7c" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.343341 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-j6v6v" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.344544 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-99b499f4-ntfjb"] Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.353331 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7454b96578-4bndq"] Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.361563 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-58f887965d-qvbnj"] Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.363271 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58f887965d-qvbnj" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.367570 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-2mgz9" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.376805 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-54b5986bb8-wzvsq"] Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.389797 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-wzvsq" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.394516 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnn2c\" (UniqueName: \"kubernetes.io/projected/8b819ac2-ef31-494d-b4cf-882b602feb70-kube-api-access-fnn2c\") pod \"horizon-operator-controller-manager-598f69df5d-95l27\" (UID: \"8b819ac2-ef31-494d-b4cf-882b602feb70\") " pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-95l27" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.394567 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jp2d9\" (UniqueName: \"kubernetes.io/projected/2c64049e-cf92-40ac-84d9-e2724c334c3f-kube-api-access-jp2d9\") pod \"keystone-operator-controller-manager-7454b96578-4bndq\" (UID: \"2c64049e-cf92-40ac-84d9-e2724c334c3f\") " pod="openstack-operators/keystone-operator-controller-manager-7454b96578-4bndq" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.394591 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkq4b\" (UniqueName: \"kubernetes.io/projected/7000614b-01cb-4878-b85f-b5a1f8ef91b2-kube-api-access-mkq4b\") pod \"infra-operator-controller-manager-6dd8864d7c-fxplj\" (UID: \"7000614b-01cb-4878-b85f-b5a1f8ef91b2\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-fxplj" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.394615 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpm25\" (UniqueName: \"kubernetes.io/projected/b2b9b9e9-800e-4213-8605-058d91412276-kube-api-access-bpm25\") pod \"designate-operator-controller-manager-767ccfd65f-h8nkv\" (UID: \"b2b9b9e9-800e-4213-8605-058d91412276\") " pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-h8nkv" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.394642 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l22cc\" (UniqueName: \"kubernetes.io/projected/d39e8176-e052-4a0e-9d16-e66b6312fbfc-kube-api-access-l22cc\") pod \"glance-operator-controller-manager-7969689c84-wq7t5\" (UID: \"d39e8176-e052-4a0e-9d16-e66b6312fbfc\") " pod="openstack-operators/glance-operator-controller-manager-7969689c84-wq7t5" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.394666 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w889m\" (UniqueName: \"kubernetes.io/projected/b55ba4e8-3d0a-4746-817c-feb1a210b01e-kube-api-access-w889m\") pod \"barbican-operator-controller-manager-75fb479bcc-9wl9p\" (UID: \"b55ba4e8-3d0a-4746-817c-feb1a210b01e\") " pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-9wl9p" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.394689 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpcgg\" (UniqueName: \"kubernetes.io/projected/49de2fb5-2dde-400e-9ae4-d36dfe196ff1-kube-api-access-kpcgg\") pod \"heat-operator-controller-manager-56f54d6746-bzh6h\" (UID: \"49de2fb5-2dde-400e-9ae4-d36dfe196ff1\") " pod="openstack-operators/heat-operator-controller-manager-56f54d6746-bzh6h" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.394719 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx4t9\" (UniqueName: \"kubernetes.io/projected/3dc6a5ba-d447-40fc-bd8b-640e4b49990b-kube-api-access-xx4t9\") pod \"manila-operator-controller-manager-58f887965d-qvbnj\" (UID: \"3dc6a5ba-d447-40fc-bd8b-640e4b49990b\") " pod="openstack-operators/manila-operator-controller-manager-58f887965d-qvbnj" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.394748 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7000614b-01cb-4878-b85f-b5a1f8ef91b2-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-fxplj\" (UID: \"7000614b-01cb-4878-b85f-b5a1f8ef91b2\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-fxplj" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.394776 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srgtj\" (UniqueName: \"kubernetes.io/projected/52519a2b-7c33-425f-9115-6fd4876ba60d-kube-api-access-srgtj\") pod \"ironic-operator-controller-manager-99b499f4-ntfjb\" (UID: \"52519a2b-7c33-425f-9115-6fd4876ba60d\") " pod="openstack-operators/ironic-operator-controller-manager-99b499f4-ntfjb" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.394797 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5bkn\" (UniqueName: \"kubernetes.io/projected/09e60927-110e-4cda-b7e3-6f47e70f2388-kube-api-access-d5bkn\") pod \"cinder-operator-controller-manager-6498cbf48f-44wlb\" (UID: \"09e60927-110e-4cda-b7e3-6f47e70f2388\") " pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-44wlb" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.395073 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-pbwv6" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.445361 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpm25\" (UniqueName: \"kubernetes.io/projected/b2b9b9e9-800e-4213-8605-058d91412276-kube-api-access-bpm25\") pod \"designate-operator-controller-manager-767ccfd65f-h8nkv\" (UID: \"b2b9b9e9-800e-4213-8605-058d91412276\") " pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-h8nkv" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.454948 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58f887965d-qvbnj"] Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.468713 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w889m\" (UniqueName: \"kubernetes.io/projected/b55ba4e8-3d0a-4746-817c-feb1a210b01e-kube-api-access-w889m\") pod \"barbican-operator-controller-manager-75fb479bcc-9wl9p\" (UID: \"b55ba4e8-3d0a-4746-817c-feb1a210b01e\") " pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-9wl9p" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.472576 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-54b5986bb8-wzvsq"] Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.485566 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5bkn\" (UniqueName: \"kubernetes.io/projected/09e60927-110e-4cda-b7e3-6f47e70f2388-kube-api-access-d5bkn\") pod \"cinder-operator-controller-manager-6498cbf48f-44wlb\" (UID: \"09e60927-110e-4cda-b7e3-6f47e70f2388\") " pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-44wlb" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.486110 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpcgg\" (UniqueName: \"kubernetes.io/projected/49de2fb5-2dde-400e-9ae4-d36dfe196ff1-kube-api-access-kpcgg\") pod \"heat-operator-controller-manager-56f54d6746-bzh6h\" (UID: \"49de2fb5-2dde-400e-9ae4-d36dfe196ff1\") " pod="openstack-operators/heat-operator-controller-manager-56f54d6746-bzh6h" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.497968 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78bd47f458-d8gzf"] Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.521188 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx4t9\" (UniqueName: \"kubernetes.io/projected/3dc6a5ba-d447-40fc-bd8b-640e4b49990b-kube-api-access-xx4t9\") pod \"manila-operator-controller-manager-58f887965d-qvbnj\" (UID: \"3dc6a5ba-d447-40fc-bd8b-640e4b49990b\") " pod="openstack-operators/manila-operator-controller-manager-58f887965d-qvbnj" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.521249 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7000614b-01cb-4878-b85f-b5a1f8ef91b2-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-fxplj\" (UID: \"7000614b-01cb-4878-b85f-b5a1f8ef91b2\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-fxplj" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.521327 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srgtj\" (UniqueName: \"kubernetes.io/projected/52519a2b-7c33-425f-9115-6fd4876ba60d-kube-api-access-srgtj\") pod \"ironic-operator-controller-manager-99b499f4-ntfjb\" (UID: \"52519a2b-7c33-425f-9115-6fd4876ba60d\") " pod="openstack-operators/ironic-operator-controller-manager-99b499f4-ntfjb" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.521414 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnn2c\" (UniqueName: \"kubernetes.io/projected/8b819ac2-ef31-494d-b4cf-882b602feb70-kube-api-access-fnn2c\") pod \"horizon-operator-controller-manager-598f69df5d-95l27\" (UID: \"8b819ac2-ef31-494d-b4cf-882b602feb70\") " pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-95l27" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.521512 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jp2d9\" (UniqueName: \"kubernetes.io/projected/2c64049e-cf92-40ac-84d9-e2724c334c3f-kube-api-access-jp2d9\") pod \"keystone-operator-controller-manager-7454b96578-4bndq\" (UID: \"2c64049e-cf92-40ac-84d9-e2724c334c3f\") " pod="openstack-operators/keystone-operator-controller-manager-7454b96578-4bndq" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.521544 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkq4b\" (UniqueName: \"kubernetes.io/projected/7000614b-01cb-4878-b85f-b5a1f8ef91b2-kube-api-access-mkq4b\") pod \"infra-operator-controller-manager-6dd8864d7c-fxplj\" (UID: \"7000614b-01cb-4878-b85f-b5a1f8ef91b2\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-fxplj" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.521603 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsrnt\" (UniqueName: \"kubernetes.io/projected/4c1c13a1-3285-45f5-888b-d9abb0008661-kube-api-access-vsrnt\") pod \"mariadb-operator-controller-manager-54b5986bb8-wzvsq\" (UID: \"4c1c13a1-3285-45f5-888b-d9abb0008661\") " pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-wzvsq" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.526781 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-9wl9p" Nov 24 13:30:24 crc kubenswrapper[4790]: E1124 13:30:24.530706 4790 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 24 13:30:24 crc kubenswrapper[4790]: E1124 13:30:24.531148 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7000614b-01cb-4878-b85f-b5a1f8ef91b2-cert podName:7000614b-01cb-4878-b85f-b5a1f8ef91b2 nodeName:}" failed. No retries permitted until 2025-11-24 13:30:25.03094949 +0000 UTC m=+1073.410843152 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7000614b-01cb-4878-b85f-b5a1f8ef91b2-cert") pod "infra-operator-controller-manager-6dd8864d7c-fxplj" (UID: "7000614b-01cb-4878-b85f-b5a1f8ef91b2") : secret "infra-operator-webhook-server-cert" not found Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.547567 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-44wlb" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.602664 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-bzh6h" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.603400 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-d8gzf" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.603445 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-h8nkv" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.604541 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l22cc\" (UniqueName: \"kubernetes.io/projected/d39e8176-e052-4a0e-9d16-e66b6312fbfc-kube-api-access-l22cc\") pod \"glance-operator-controller-manager-7969689c84-wq7t5\" (UID: \"d39e8176-e052-4a0e-9d16-e66b6312fbfc\") " pod="openstack-operators/glance-operator-controller-manager-7969689c84-wq7t5" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.617693 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78bd47f458-d8gzf"] Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.629434 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-pvx49" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.635645 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsrnt\" (UniqueName: \"kubernetes.io/projected/4c1c13a1-3285-45f5-888b-d9abb0008661-kube-api-access-vsrnt\") pod \"mariadb-operator-controller-manager-54b5986bb8-wzvsq\" (UID: \"4c1c13a1-3285-45f5-888b-d9abb0008661\") " pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-wzvsq" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.651902 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srgtj\" (UniqueName: \"kubernetes.io/projected/52519a2b-7c33-425f-9115-6fd4876ba60d-kube-api-access-srgtj\") pod \"ironic-operator-controller-manager-99b499f4-ntfjb\" (UID: \"52519a2b-7c33-425f-9115-6fd4876ba60d\") " pod="openstack-operators/ironic-operator-controller-manager-99b499f4-ntfjb" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.659692 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jp2d9\" (UniqueName: \"kubernetes.io/projected/2c64049e-cf92-40ac-84d9-e2724c334c3f-kube-api-access-jp2d9\") pod \"keystone-operator-controller-manager-7454b96578-4bndq\" (UID: \"2c64049e-cf92-40ac-84d9-e2724c334c3f\") " pod="openstack-operators/keystone-operator-controller-manager-7454b96578-4bndq" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.666783 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx4t9\" (UniqueName: \"kubernetes.io/projected/3dc6a5ba-d447-40fc-bd8b-640e4b49990b-kube-api-access-xx4t9\") pod \"manila-operator-controller-manager-58f887965d-qvbnj\" (UID: \"3dc6a5ba-d447-40fc-bd8b-640e4b49990b\") " pod="openstack-operators/manila-operator-controller-manager-58f887965d-qvbnj" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.670010 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-4bndq" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.671976 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnn2c\" (UniqueName: \"kubernetes.io/projected/8b819ac2-ef31-494d-b4cf-882b602feb70-kube-api-access-fnn2c\") pod \"horizon-operator-controller-manager-598f69df5d-95l27\" (UID: \"8b819ac2-ef31-494d-b4cf-882b602feb70\") " pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-95l27" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.678802 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-cfbb9c588-gkdz7"] Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.680332 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-ntfjb" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.681979 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-gkdz7" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.687009 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsrnt\" (UniqueName: \"kubernetes.io/projected/4c1c13a1-3285-45f5-888b-d9abb0008661-kube-api-access-vsrnt\") pod \"mariadb-operator-controller-manager-54b5986bb8-wzvsq\" (UID: \"4c1c13a1-3285-45f5-888b-d9abb0008661\") " pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-wzvsq" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.688391 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58f887965d-qvbnj" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.688951 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-9gnsv"] Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.690122 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-9gnsv" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.691589 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-9jnmb" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.696763 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-fj884" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.700166 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkq4b\" (UniqueName: \"kubernetes.io/projected/7000614b-01cb-4878-b85f-b5a1f8ef91b2-kube-api-access-mkq4b\") pod \"infra-operator-controller-manager-6dd8864d7c-fxplj\" (UID: \"7000614b-01cb-4878-b85f-b5a1f8ef91b2\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-fxplj" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.702236 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-cfbb9c588-gkdz7"] Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.728380 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-wzvsq" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.738038 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kk5wd\" (UniqueName: \"kubernetes.io/projected/dc60f94f-01f9-4179-a0ce-67112913da75-kube-api-access-kk5wd\") pod \"octavia-operator-controller-manager-54cfbf4c7d-9gnsv\" (UID: \"dc60f94f-01f9-4179-a0ce-67112913da75\") " pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-9gnsv" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.738204 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b64h4\" (UniqueName: \"kubernetes.io/projected/e25e339c-c102-44f0-b4a4-15f0ce87b38c-kube-api-access-b64h4\") pod \"nova-operator-controller-manager-cfbb9c588-gkdz7\" (UID: \"e25e339c-c102-44f0-b4a4-15f0ce87b38c\") " pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-gkdz7" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.738319 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nd8r\" (UniqueName: \"kubernetes.io/projected/0c8ec99a-dfcc-414d-85c9-f005c476919d-kube-api-access-8nd8r\") pod \"neutron-operator-controller-manager-78bd47f458-d8gzf\" (UID: \"0c8ec99a-dfcc-414d-85c9-f005c476919d\") " pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-d8gzf" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.755830 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-9gnsv"] Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.829981 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-jg2n4"] Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.831430 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-jg2n4" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.838396 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.839746 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nd8r\" (UniqueName: \"kubernetes.io/projected/0c8ec99a-dfcc-414d-85c9-f005c476919d-kube-api-access-8nd8r\") pod \"neutron-operator-controller-manager-78bd47f458-d8gzf\" (UID: \"0c8ec99a-dfcc-414d-85c9-f005c476919d\") " pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-d8gzf" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.839850 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kk5wd\" (UniqueName: \"kubernetes.io/projected/dc60f94f-01f9-4179-a0ce-67112913da75-kube-api-access-kk5wd\") pod \"octavia-operator-controller-manager-54cfbf4c7d-9gnsv\" (UID: \"dc60f94f-01f9-4179-a0ce-67112913da75\") " pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-9gnsv" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.839912 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b64h4\" (UniqueName: \"kubernetes.io/projected/e25e339c-c102-44f0-b4a4-15f0ce87b38c-kube-api-access-b64h4\") pod \"nova-operator-controller-manager-cfbb9c588-gkdz7\" (UID: \"e25e339c-c102-44f0-b4a4-15f0ce87b38c\") " pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-gkdz7" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.853824 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-2tfdc" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.863118 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b797b8dff-c8xtz"] Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.864209 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-c8xtz" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.866462 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7969689c84-wq7t5" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.884848 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b64h4\" (UniqueName: \"kubernetes.io/projected/e25e339c-c102-44f0-b4a4-15f0ce87b38c-kube-api-access-b64h4\") pod \"nova-operator-controller-manager-cfbb9c588-gkdz7\" (UID: \"e25e339c-c102-44f0-b4a4-15f0ce87b38c\") " pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-gkdz7" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.889567 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nd8r\" (UniqueName: \"kubernetes.io/projected/0c8ec99a-dfcc-414d-85c9-f005c476919d-kube-api-access-8nd8r\") pod \"neutron-operator-controller-manager-78bd47f458-d8gzf\" (UID: \"0c8ec99a-dfcc-414d-85c9-f005c476919d\") " pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-d8gzf" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.891325 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-qccfv" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.893239 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54fc5f65b7-ld798"] Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.907623 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-ld798" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.908173 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-95l27" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.910835 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kk5wd\" (UniqueName: \"kubernetes.io/projected/dc60f94f-01f9-4179-a0ce-67112913da75-kube-api-access-kk5wd\") pod \"octavia-operator-controller-manager-54cfbf4c7d-9gnsv\" (UID: \"dc60f94f-01f9-4179-a0ce-67112913da75\") " pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-9gnsv" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.917308 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-v98n2" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.930931 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-d656998f4-lfvdf"] Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.943848 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c861ea1f-e001-4f01-8e0d-12e97b7628cc-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-jg2n4\" (UID: \"c861ea1f-e001-4f01-8e0d-12e97b7628cc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-jg2n4" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.943933 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jz4hd\" (UniqueName: \"kubernetes.io/projected/c861ea1f-e001-4f01-8e0d-12e97b7628cc-kube-api-access-jz4hd\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-jg2n4\" (UID: \"c861ea1f-e001-4f01-8e0d-12e97b7628cc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-jg2n4" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.944008 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnn6k\" (UniqueName: \"kubernetes.io/projected/cb16570b-3e90-4b2e-996b-d0cd5c0a0259-kube-api-access-bnn6k\") pod \"placement-operator-controller-manager-5b797b8dff-c8xtz\" (UID: \"cb16570b-3e90-4b2e-996b-d0cd5c0a0259\") " pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-c8xtz" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.944035 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5824b\" (UniqueName: \"kubernetes.io/projected/84281187-e5c6-4b57-807a-ae5637942260-kube-api-access-5824b\") pod \"ovn-operator-controller-manager-54fc5f65b7-ld798\" (UID: \"84281187-e5c6-4b57-807a-ae5637942260\") " pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-ld798" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.964088 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d656998f4-lfvdf" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.983721 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-d8gzf" Nov 24 13:30:24 crc kubenswrapper[4790]: I1124 13:30:24.988339 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-44nrf" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.022479 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-jg2n4"] Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.032230 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-gkdz7" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.048972 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54fc5f65b7-ld798"] Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.050064 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnn6k\" (UniqueName: \"kubernetes.io/projected/cb16570b-3e90-4b2e-996b-d0cd5c0a0259-kube-api-access-bnn6k\") pod \"placement-operator-controller-manager-5b797b8dff-c8xtz\" (UID: \"cb16570b-3e90-4b2e-996b-d0cd5c0a0259\") " pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-c8xtz" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.050095 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5824b\" (UniqueName: \"kubernetes.io/projected/84281187-e5c6-4b57-807a-ae5637942260-kube-api-access-5824b\") pod \"ovn-operator-controller-manager-54fc5f65b7-ld798\" (UID: \"84281187-e5c6-4b57-807a-ae5637942260\") " pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-ld798" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.050193 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5h8bg\" (UniqueName: \"kubernetes.io/projected/69644679-f0fb-4c09-abec-8e4084140f05-kube-api-access-5h8bg\") pod \"swift-operator-controller-manager-d656998f4-lfvdf\" (UID: \"69644679-f0fb-4c09-abec-8e4084140f05\") " pod="openstack-operators/swift-operator-controller-manager-d656998f4-lfvdf" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.050228 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7000614b-01cb-4878-b85f-b5a1f8ef91b2-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-fxplj\" (UID: \"7000614b-01cb-4878-b85f-b5a1f8ef91b2\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-fxplj" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.050252 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c861ea1f-e001-4f01-8e0d-12e97b7628cc-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-jg2n4\" (UID: \"c861ea1f-e001-4f01-8e0d-12e97b7628cc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-jg2n4" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.050272 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jz4hd\" (UniqueName: \"kubernetes.io/projected/c861ea1f-e001-4f01-8e0d-12e97b7628cc-kube-api-access-jz4hd\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-jg2n4\" (UID: \"c861ea1f-e001-4f01-8e0d-12e97b7628cc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-jg2n4" Nov 24 13:30:25 crc kubenswrapper[4790]: E1124 13:30:25.051789 4790 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 13:30:25 crc kubenswrapper[4790]: E1124 13:30:25.051844 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c861ea1f-e001-4f01-8e0d-12e97b7628cc-cert podName:c861ea1f-e001-4f01-8e0d-12e97b7628cc nodeName:}" failed. No retries permitted until 2025-11-24 13:30:25.551828388 +0000 UTC m=+1073.931722040 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c861ea1f-e001-4f01-8e0d-12e97b7628cc-cert") pod "openstack-baremetal-operator-controller-manager-8c7444f48-jg2n4" (UID: "c861ea1f-e001-4f01-8e0d-12e97b7628cc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.052339 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-9gnsv" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.078571 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7000614b-01cb-4878-b85f-b5a1f8ef91b2-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-fxplj\" (UID: \"7000614b-01cb-4878-b85f-b5a1f8ef91b2\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-fxplj" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.107091 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jz4hd\" (UniqueName: \"kubernetes.io/projected/c861ea1f-e001-4f01-8e0d-12e97b7628cc-kube-api-access-jz4hd\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-jg2n4\" (UID: \"c861ea1f-e001-4f01-8e0d-12e97b7628cc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-jg2n4" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.109681 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnn6k\" (UniqueName: \"kubernetes.io/projected/cb16570b-3e90-4b2e-996b-d0cd5c0a0259-kube-api-access-bnn6k\") pod \"placement-operator-controller-manager-5b797b8dff-c8xtz\" (UID: \"cb16570b-3e90-4b2e-996b-d0cd5c0a0259\") " pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-c8xtz" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.114203 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5824b\" (UniqueName: \"kubernetes.io/projected/84281187-e5c6-4b57-807a-ae5637942260-kube-api-access-5824b\") pod \"ovn-operator-controller-manager-54fc5f65b7-ld798\" (UID: \"84281187-e5c6-4b57-807a-ae5637942260\") " pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-ld798" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.120412 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-fg9nc"] Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.133280 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-fg9nc" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.141330 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-59gl2" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.168446 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5h8bg\" (UniqueName: \"kubernetes.io/projected/69644679-f0fb-4c09-abec-8e4084140f05-kube-api-access-5h8bg\") pod \"swift-operator-controller-manager-d656998f4-lfvdf\" (UID: \"69644679-f0fb-4c09-abec-8e4084140f05\") " pod="openstack-operators/swift-operator-controller-manager-d656998f4-lfvdf" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.168528 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9llch\" (UniqueName: \"kubernetes.io/projected/1e565309-31eb-4278-98fd-5a4321d5eb10-kube-api-access-9llch\") pod \"telemetry-operator-controller-manager-6d4bf84b58-fg9nc\" (UID: \"1e565309-31eb-4278-98fd-5a4321d5eb10\") " pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-fg9nc" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.173000 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b797b8dff-c8xtz"] Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.217036 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5h8bg\" (UniqueName: \"kubernetes.io/projected/69644679-f0fb-4c09-abec-8e4084140f05-kube-api-access-5h8bg\") pod \"swift-operator-controller-manager-d656998f4-lfvdf\" (UID: \"69644679-f0fb-4c09-abec-8e4084140f05\") " pod="openstack-operators/swift-operator-controller-manager-d656998f4-lfvdf" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.231673 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-b4c496f69-8k88p"] Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.233038 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-b4c496f69-8k88p" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.240165 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d656998f4-lfvdf"] Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.245963 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-gxw8g" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.258588 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-fxplj" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.269521 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs5wp\" (UniqueName: \"kubernetes.io/projected/5e93acb0-367e-4f6c-b13e-530d41db7851-kube-api-access-xs5wp\") pod \"test-operator-controller-manager-b4c496f69-8k88p\" (UID: \"5e93acb0-367e-4f6c-b13e-530d41db7851\") " pod="openstack-operators/test-operator-controller-manager-b4c496f69-8k88p" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.269629 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9llch\" (UniqueName: \"kubernetes.io/projected/1e565309-31eb-4278-98fd-5a4321d5eb10-kube-api-access-9llch\") pod \"telemetry-operator-controller-manager-6d4bf84b58-fg9nc\" (UID: \"1e565309-31eb-4278-98fd-5a4321d5eb10\") " pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-fg9nc" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.286583 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-b4c496f69-8k88p"] Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.294558 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-c8xtz" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.295018 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-fg9nc"] Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.308497 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9llch\" (UniqueName: \"kubernetes.io/projected/1e565309-31eb-4278-98fd-5a4321d5eb10-kube-api-access-9llch\") pod \"telemetry-operator-controller-manager-6d4bf84b58-fg9nc\" (UID: \"1e565309-31eb-4278-98fd-5a4321d5eb10\") " pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-fg9nc" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.315300 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8c6448b9f-cddlm"] Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.316982 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-cddlm" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.320370 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-ld798" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.322698 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8c6448b9f-cddlm"] Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.332278 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-fpgtf" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.337593 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d656998f4-lfvdf" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.375391 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xs5wp\" (UniqueName: \"kubernetes.io/projected/5e93acb0-367e-4f6c-b13e-530d41db7851-kube-api-access-xs5wp\") pod \"test-operator-controller-manager-b4c496f69-8k88p\" (UID: \"5e93acb0-367e-4f6c-b13e-530d41db7851\") " pod="openstack-operators/test-operator-controller-manager-b4c496f69-8k88p" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.375510 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lfpv\" (UniqueName: \"kubernetes.io/projected/e85f1bfb-0153-40e0-b54b-728171cb39f0-kube-api-access-7lfpv\") pod \"watcher-operator-controller-manager-8c6448b9f-cddlm\" (UID: \"e85f1bfb-0153-40e0-b54b-728171cb39f0\") " pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-cddlm" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.451772 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xs5wp\" (UniqueName: \"kubernetes.io/projected/5e93acb0-367e-4f6c-b13e-530d41db7851-kube-api-access-xs5wp\") pod \"test-operator-controller-manager-b4c496f69-8k88p\" (UID: \"5e93acb0-367e-4f6c-b13e-530d41db7851\") " pod="openstack-operators/test-operator-controller-manager-b4c496f69-8k88p" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.504350 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lfpv\" (UniqueName: \"kubernetes.io/projected/e85f1bfb-0153-40e0-b54b-728171cb39f0-kube-api-access-7lfpv\") pod \"watcher-operator-controller-manager-8c6448b9f-cddlm\" (UID: \"e85f1bfb-0153-40e0-b54b-728171cb39f0\") " pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-cddlm" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.532465 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7547bd9b9-skhbm"] Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.534103 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7547bd9b9-skhbm" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.534560 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-fg9nc" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.541563 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-q9nbf" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.542007 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.571098 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7547bd9b9-skhbm"] Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.573757 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lfpv\" (UniqueName: \"kubernetes.io/projected/e85f1bfb-0153-40e0-b54b-728171cb39f0-kube-api-access-7lfpv\") pod \"watcher-operator-controller-manager-8c6448b9f-cddlm\" (UID: \"e85f1bfb-0153-40e0-b54b-728171cb39f0\") " pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-cddlm" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.587218 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-b4c496f69-8k88p" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.597423 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-6dwzt"] Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.598825 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-6dwzt" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.603590 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-6dwzt"] Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.604602 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-vqqn9" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.605768 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7c72ca69-273b-4c92-a5f7-54c2b968621b-cert\") pod \"openstack-operator-controller-manager-7547bd9b9-skhbm\" (UID: \"7c72ca69-273b-4c92-a5f7-54c2b968621b\") " pod="openstack-operators/openstack-operator-controller-manager-7547bd9b9-skhbm" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.605807 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c861ea1f-e001-4f01-8e0d-12e97b7628cc-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-jg2n4\" (UID: \"c861ea1f-e001-4f01-8e0d-12e97b7628cc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-jg2n4" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.605827 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggk6h\" (UniqueName: \"kubernetes.io/projected/7c72ca69-273b-4c92-a5f7-54c2b968621b-kube-api-access-ggk6h\") pod \"openstack-operator-controller-manager-7547bd9b9-skhbm\" (UID: \"7c72ca69-273b-4c92-a5f7-54c2b968621b\") " pod="openstack-operators/openstack-operator-controller-manager-7547bd9b9-skhbm" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.610237 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c861ea1f-e001-4f01-8e0d-12e97b7628cc-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-jg2n4\" (UID: \"c861ea1f-e001-4f01-8e0d-12e97b7628cc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-jg2n4" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.623034 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-75fb479bcc-9wl9p"] Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.636896 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-56f54d6746-bzh6h"] Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.663760 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-cddlm" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.707639 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7c72ca69-273b-4c92-a5f7-54c2b968621b-cert\") pod \"openstack-operator-controller-manager-7547bd9b9-skhbm\" (UID: \"7c72ca69-273b-4c92-a5f7-54c2b968621b\") " pod="openstack-operators/openstack-operator-controller-manager-7547bd9b9-skhbm" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.707710 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggk6h\" (UniqueName: \"kubernetes.io/projected/7c72ca69-273b-4c92-a5f7-54c2b968621b-kube-api-access-ggk6h\") pod \"openstack-operator-controller-manager-7547bd9b9-skhbm\" (UID: \"7c72ca69-273b-4c92-a5f7-54c2b968621b\") " pod="openstack-operators/openstack-operator-controller-manager-7547bd9b9-skhbm" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.708121 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nl85k\" (UniqueName: \"kubernetes.io/projected/8e1bedf8-78bf-4558-91e6-4228e9765356-kube-api-access-nl85k\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-6dwzt\" (UID: \"8e1bedf8-78bf-4558-91e6-4228e9765356\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-6dwzt" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.737128 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggk6h\" (UniqueName: \"kubernetes.io/projected/7c72ca69-273b-4c92-a5f7-54c2b968621b-kube-api-access-ggk6h\") pod \"openstack-operator-controller-manager-7547bd9b9-skhbm\" (UID: \"7c72ca69-273b-4c92-a5f7-54c2b968621b\") " pod="openstack-operators/openstack-operator-controller-manager-7547bd9b9-skhbm" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.737732 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7c72ca69-273b-4c92-a5f7-54c2b968621b-cert\") pod \"openstack-operator-controller-manager-7547bd9b9-skhbm\" (UID: \"7c72ca69-273b-4c92-a5f7-54c2b968621b\") " pod="openstack-operators/openstack-operator-controller-manager-7547bd9b9-skhbm" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.759038 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7547bd9b9-skhbm" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.810348 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nl85k\" (UniqueName: \"kubernetes.io/projected/8e1bedf8-78bf-4558-91e6-4228e9765356-kube-api-access-nl85k\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-6dwzt\" (UID: \"8e1bedf8-78bf-4558-91e6-4228e9765356\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-6dwzt" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.837741 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-jg2n4" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.843927 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nl85k\" (UniqueName: \"kubernetes.io/projected/8e1bedf8-78bf-4558-91e6-4228e9765356-kube-api-access-nl85k\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-6dwzt\" (UID: \"8e1bedf8-78bf-4558-91e6-4228e9765356\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-6dwzt" Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.889918 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-767ccfd65f-h8nkv"] Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.907699 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6498cbf48f-44wlb"] Nov 24 13:30:25 crc kubenswrapper[4790]: I1124 13:30:25.952176 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7454b96578-4bndq"] Nov 24 13:30:25 crc kubenswrapper[4790]: W1124 13:30:25.969351 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09e60927_110e_4cda_b7e3_6f47e70f2388.slice/crio-379ad3fc516369cbb1c446612cc3dce74998d9e32593b29c6932d24d9393ee54 WatchSource:0}: Error finding container 379ad3fc516369cbb1c446612cc3dce74998d9e32593b29c6932d24d9393ee54: Status 404 returned error can't find the container with id 379ad3fc516369cbb1c446612cc3dce74998d9e32593b29c6932d24d9393ee54 Nov 24 13:30:26 crc kubenswrapper[4790]: W1124 13:30:26.000354 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c64049e_cf92_40ac_84d9_e2724c334c3f.slice/crio-e79d54eac32eef0409e2a8321a81aede67273c950d02ff34e1b949282b3f57cd WatchSource:0}: Error finding container e79d54eac32eef0409e2a8321a81aede67273c950d02ff34e1b949282b3f57cd: Status 404 returned error can't find the container with id e79d54eac32eef0409e2a8321a81aede67273c950d02ff34e1b949282b3f57cd Nov 24 13:30:26 crc kubenswrapper[4790]: I1124 13:30:26.039992 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-99b499f4-ntfjb"] Nov 24 13:30:26 crc kubenswrapper[4790]: I1124 13:30:26.064593 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-54b5986bb8-wzvsq"] Nov 24 13:30:26 crc kubenswrapper[4790]: I1124 13:30:26.078168 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-6dwzt" Nov 24 13:30:26 crc kubenswrapper[4790]: I1124 13:30:26.133989 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-h8nkv" event={"ID":"b2b9b9e9-800e-4213-8605-058d91412276","Type":"ContainerStarted","Data":"bb41974a21102903f35c8dd9323c6b3fbdb098fe89665c3c515c600f3e683f4e"} Nov 24 13:30:26 crc kubenswrapper[4790]: I1124 13:30:26.144426 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-4bndq" event={"ID":"2c64049e-cf92-40ac-84d9-e2724c334c3f","Type":"ContainerStarted","Data":"e79d54eac32eef0409e2a8321a81aede67273c950d02ff34e1b949282b3f57cd"} Nov 24 13:30:26 crc kubenswrapper[4790]: I1124 13:30:26.175834 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-bzh6h" event={"ID":"49de2fb5-2dde-400e-9ae4-d36dfe196ff1","Type":"ContainerStarted","Data":"6ee7e60c8c02273732ee063ae1b95b254a3b80d8bc273f41f0c71bfa6762d263"} Nov 24 13:30:26 crc kubenswrapper[4790]: I1124 13:30:26.180749 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-ntfjb" event={"ID":"52519a2b-7c33-425f-9115-6fd4876ba60d","Type":"ContainerStarted","Data":"219f3e70c23cadcdcabfac2c888dffd71725926ced42f2a5cb79fedc6d67611f"} Nov 24 13:30:26 crc kubenswrapper[4790]: I1124 13:30:26.182551 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-9wl9p" event={"ID":"b55ba4e8-3d0a-4746-817c-feb1a210b01e","Type":"ContainerStarted","Data":"8e2b758057c2a9f461f9fdf0a6f132d5b1d9b897b4ca784f21ec530019f6db3a"} Nov 24 13:30:26 crc kubenswrapper[4790]: I1124 13:30:26.184243 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-44wlb" event={"ID":"09e60927-110e-4cda-b7e3-6f47e70f2388","Type":"ContainerStarted","Data":"379ad3fc516369cbb1c446612cc3dce74998d9e32593b29c6932d24d9393ee54"} Nov 24 13:30:26 crc kubenswrapper[4790]: I1124 13:30:26.310184 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-598f69df5d-95l27"] Nov 24 13:30:26 crc kubenswrapper[4790]: I1124 13:30:26.338839 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6dd8864d7c-fxplj"] Nov 24 13:30:26 crc kubenswrapper[4790]: I1124 13:30:26.338899 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7969689c84-wq7t5"] Nov 24 13:30:26 crc kubenswrapper[4790]: I1124 13:30:26.454012 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78bd47f458-d8gzf"] Nov 24 13:30:26 crc kubenswrapper[4790]: I1124 13:30:26.498634 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-9gnsv"] Nov 24 13:30:26 crc kubenswrapper[4790]: W1124 13:30:26.529231 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddc60f94f_01f9_4179_a0ce_67112913da75.slice/crio-16f07ad656e565489ddd0d1119ddfb96600f71532f8e8d2e104e25320ca07194 WatchSource:0}: Error finding container 16f07ad656e565489ddd0d1119ddfb96600f71532f8e8d2e104e25320ca07194: Status 404 returned error can't find the container with id 16f07ad656e565489ddd0d1119ddfb96600f71532f8e8d2e104e25320ca07194 Nov 24 13:30:26 crc kubenswrapper[4790]: I1124 13:30:26.537116 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58f887965d-qvbnj"] Nov 24 13:30:26 crc kubenswrapper[4790]: I1124 13:30:26.615275 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-fg9nc"] Nov 24 13:30:26 crc kubenswrapper[4790]: I1124 13:30:26.628906 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-cfbb9c588-gkdz7"] Nov 24 13:30:26 crc kubenswrapper[4790]: E1124 13:30:26.633001 4790 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-b64h4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-cfbb9c588-gkdz7_openstack-operators(e25e339c-c102-44f0-b4a4-15f0ce87b38c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 13:30:26 crc kubenswrapper[4790]: I1124 13:30:26.851395 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8c6448b9f-cddlm"] Nov 24 13:30:26 crc kubenswrapper[4790]: I1124 13:30:26.868316 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-b4c496f69-8k88p"] Nov 24 13:30:26 crc kubenswrapper[4790]: W1124 13:30:26.873507 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode85f1bfb_0153_40e0_b54b_728171cb39f0.slice/crio-05ca497d180f177fcabdcb269c7d903369f9847d5e9324357eced1881c6b7be2 WatchSource:0}: Error finding container 05ca497d180f177fcabdcb269c7d903369f9847d5e9324357eced1881c6b7be2: Status 404 returned error can't find the container with id 05ca497d180f177fcabdcb269c7d903369f9847d5e9324357eced1881c6b7be2 Nov 24 13:30:26 crc kubenswrapper[4790]: I1124 13:30:26.876088 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d656998f4-lfvdf"] Nov 24 13:30:26 crc kubenswrapper[4790]: I1124 13:30:26.880735 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b797b8dff-c8xtz"] Nov 24 13:30:26 crc kubenswrapper[4790]: I1124 13:30:26.885426 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54fc5f65b7-ld798"] Nov 24 13:30:26 crc kubenswrapper[4790]: I1124 13:30:26.892159 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-jg2n4"] Nov 24 13:30:26 crc kubenswrapper[4790]: I1124 13:30:26.894537 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7547bd9b9-skhbm"] Nov 24 13:30:26 crc kubenswrapper[4790]: I1124 13:30:26.897826 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-6dwzt"] Nov 24 13:30:26 crc kubenswrapper[4790]: W1124 13:30:26.908449 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod84281187_e5c6_4b57_807a_ae5637942260.slice/crio-6fac1fdc49a67b1cfa4196ea1c88754a8a54940a03a370d09321731e17ad8c4f WatchSource:0}: Error finding container 6fac1fdc49a67b1cfa4196ea1c88754a8a54940a03a370d09321731e17ad8c4f: Status 404 returned error can't find the container with id 6fac1fdc49a67b1cfa4196ea1c88754a8a54940a03a370d09321731e17ad8c4f Nov 24 13:30:26 crc kubenswrapper[4790]: W1124 13:30:26.912490 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb16570b_3e90_4b2e_996b_d0cd5c0a0259.slice/crio-7bffe5a6101e6b14469d250147e758e9fb40fdd482fd2ef1a8fc8fe52f9afb62 WatchSource:0}: Error finding container 7bffe5a6101e6b14469d250147e758e9fb40fdd482fd2ef1a8fc8fe52f9afb62: Status 404 returned error can't find the container with id 7bffe5a6101e6b14469d250147e758e9fb40fdd482fd2ef1a8fc8fe52f9afb62 Nov 24 13:30:26 crc kubenswrapper[4790]: E1124 13:30:26.912538 4790 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5824b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-54fc5f65b7-ld798_openstack-operators(84281187-e5c6-4b57-807a-ae5637942260): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 13:30:26 crc kubenswrapper[4790]: W1124 13:30:26.913075 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c72ca69_273b_4c92_a5f7_54c2b968621b.slice/crio-b549349e166d89a901290ff1fec6ee2fa7610715fabae8ccc207edc23a4b2e86 WatchSource:0}: Error finding container b549349e166d89a901290ff1fec6ee2fa7610715fabae8ccc207edc23a4b2e86: Status 404 returned error can't find the container with id b549349e166d89a901290ff1fec6ee2fa7610715fabae8ccc207edc23a4b2e86 Nov 24 13:30:26 crc kubenswrapper[4790]: E1124 13:30:26.914159 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-gkdz7" podUID="e25e339c-c102-44f0-b4a4-15f0ce87b38c" Nov 24 13:30:26 crc kubenswrapper[4790]: W1124 13:30:26.918838 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc861ea1f_e001_4f01_8e0d_12e97b7628cc.slice/crio-6eb45d5a45d9e2391963856b4c50184b8a4d601b55eef570625da28ff1872ceb WatchSource:0}: Error finding container 6eb45d5a45d9e2391963856b4c50184b8a4d601b55eef570625da28ff1872ceb: Status 404 returned error can't find the container with id 6eb45d5a45d9e2391963856b4c50184b8a4d601b55eef570625da28ff1872ceb Nov 24 13:30:26 crc kubenswrapper[4790]: E1124 13:30:26.921104 4790 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bnn6k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-5b797b8dff-c8xtz_openstack-operators(cb16570b-3e90-4b2e-996b-d0cd5c0a0259): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 13:30:26 crc kubenswrapper[4790]: W1124 13:30:26.925013 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e93acb0_367e_4f6c_b13e_530d41db7851.slice/crio-d94c8f5393ffabf2a8db4192e8d00a96c569ed5b636b3f901b3672ded1ec599a WatchSource:0}: Error finding container d94c8f5393ffabf2a8db4192e8d00a96c569ed5b636b3f901b3672ded1ec599a: Status 404 returned error can't find the container with id d94c8f5393ffabf2a8db4192e8d00a96c569ed5b636b3f901b3672ded1ec599a Nov 24 13:30:26 crc kubenswrapper[4790]: E1124 13:30:26.930182 4790 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xs5wp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-b4c496f69-8k88p_openstack-operators(5e93acb0-367e-4f6c-b13e-530d41db7851): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 13:30:26 crc kubenswrapper[4790]: E1124 13:30:26.931038 4790 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_API_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_PROC_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-processor:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jz4hd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-8c7444f48-jg2n4_openstack-operators(c861ea1f-e001-4f01-8e0d-12e97b7628cc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 13:30:26 crc kubenswrapper[4790]: E1124 13:30:26.950496 4790 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nl85k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-6dwzt_openstack-operators(8e1bedf8-78bf-4558-91e6-4228e9765356): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 13:30:26 crc kubenswrapper[4790]: E1124 13:30:26.951907 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-6dwzt" podUID="8e1bedf8-78bf-4558-91e6-4228e9765356" Nov 24 13:30:27 crc kubenswrapper[4790]: I1124 13:30:27.192653 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-95l27" event={"ID":"8b819ac2-ef31-494d-b4cf-882b602feb70","Type":"ContainerStarted","Data":"11967cd8e9c19dbc699afcd9bd5ef226f7ce298872314a97f86f49355645aba1"} Nov 24 13:30:27 crc kubenswrapper[4790]: I1124 13:30:27.198792 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-b4c496f69-8k88p" event={"ID":"5e93acb0-367e-4f6c-b13e-530d41db7851","Type":"ContainerStarted","Data":"d94c8f5393ffabf2a8db4192e8d00a96c569ed5b636b3f901b3672ded1ec599a"} Nov 24 13:30:27 crc kubenswrapper[4790]: I1124 13:30:27.216250 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-9gnsv" event={"ID":"dc60f94f-01f9-4179-a0ce-67112913da75","Type":"ContainerStarted","Data":"16f07ad656e565489ddd0d1119ddfb96600f71532f8e8d2e104e25320ca07194"} Nov 24 13:30:27 crc kubenswrapper[4790]: I1124 13:30:27.226505 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-cddlm" event={"ID":"e85f1bfb-0153-40e0-b54b-728171cb39f0","Type":"ContainerStarted","Data":"05ca497d180f177fcabdcb269c7d903369f9847d5e9324357eced1881c6b7be2"} Nov 24 13:30:27 crc kubenswrapper[4790]: I1124 13:30:27.230226 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58f887965d-qvbnj" event={"ID":"3dc6a5ba-d447-40fc-bd8b-640e4b49990b","Type":"ContainerStarted","Data":"13c5ce3ab613f10a0d904274ab4865676e75fc0f171876303a623eac4e5e78fe"} Nov 24 13:30:27 crc kubenswrapper[4790]: I1124 13:30:27.231517 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-wzvsq" event={"ID":"4c1c13a1-3285-45f5-888b-d9abb0008661","Type":"ContainerStarted","Data":"303d3b81d37ccc0857175ffb22818ce2020a0ce11b8ee4541ff90ba024d7bcbd"} Nov 24 13:30:27 crc kubenswrapper[4790]: E1124 13:30:27.235567 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-b4c496f69-8k88p" podUID="5e93acb0-367e-4f6c-b13e-530d41db7851" Nov 24 13:30:27 crc kubenswrapper[4790]: I1124 13:30:27.236416 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-d8gzf" event={"ID":"0c8ec99a-dfcc-414d-85c9-f005c476919d","Type":"ContainerStarted","Data":"ee5be357d400ec3c0f468437fa84ad98a4d2b3d9d0c634f873c46b31656b6ada"} Nov 24 13:30:27 crc kubenswrapper[4790]: I1124 13:30:27.240467 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-6dwzt" event={"ID":"8e1bedf8-78bf-4558-91e6-4228e9765356","Type":"ContainerStarted","Data":"804ad961112176f58a4765c79dffe53445a2606aad41bc81d43fd3aa0e69dcba"} Nov 24 13:30:27 crc kubenswrapper[4790]: E1124 13:30:27.242438 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-6dwzt" podUID="8e1bedf8-78bf-4558-91e6-4228e9765356" Nov 24 13:30:27 crc kubenswrapper[4790]: I1124 13:30:27.246410 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-jg2n4" event={"ID":"c861ea1f-e001-4f01-8e0d-12e97b7628cc","Type":"ContainerStarted","Data":"6eb45d5a45d9e2391963856b4c50184b8a4d601b55eef570625da28ff1872ceb"} Nov 24 13:30:27 crc kubenswrapper[4790]: I1124 13:30:27.254465 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-ld798" event={"ID":"84281187-e5c6-4b57-807a-ae5637942260","Type":"ContainerStarted","Data":"6fac1fdc49a67b1cfa4196ea1c88754a8a54940a03a370d09321731e17ad8c4f"} Nov 24 13:30:27 crc kubenswrapper[4790]: I1124 13:30:27.256374 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-c8xtz" event={"ID":"cb16570b-3e90-4b2e-996b-d0cd5c0a0259","Type":"ContainerStarted","Data":"7bffe5a6101e6b14469d250147e758e9fb40fdd482fd2ef1a8fc8fe52f9afb62"} Nov 24 13:30:27 crc kubenswrapper[4790]: I1124 13:30:27.262790 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-gkdz7" event={"ID":"e25e339c-c102-44f0-b4a4-15f0ce87b38c","Type":"ContainerStarted","Data":"ceef9be8c8732806429b36a5e0fd91ee2691b8c3c3188fb6026bac6b85e04d33"} Nov 24 13:30:27 crc kubenswrapper[4790]: I1124 13:30:27.262826 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-gkdz7" event={"ID":"e25e339c-c102-44f0-b4a4-15f0ce87b38c","Type":"ContainerStarted","Data":"9171a8a53c7cc2a75031c031129d5b137088a340b5109f4b99337186afdc6fa5"} Nov 24 13:30:27 crc kubenswrapper[4790]: E1124 13:30:27.265632 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\"" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-gkdz7" podUID="e25e339c-c102-44f0-b4a4-15f0ce87b38c" Nov 24 13:30:27 crc kubenswrapper[4790]: I1124 13:30:27.267318 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d656998f4-lfvdf" event={"ID":"69644679-f0fb-4c09-abec-8e4084140f05","Type":"ContainerStarted","Data":"63075b51560b91a156a861dda6c43cadb3869df01a499dbe4e900df8b2db16f6"} Nov 24 13:30:27 crc kubenswrapper[4790]: I1124 13:30:27.275805 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-fxplj" event={"ID":"7000614b-01cb-4878-b85f-b5a1f8ef91b2","Type":"ContainerStarted","Data":"a260102132b49ca2cc9901a1806f4f587ea8213699586f6f706f733874933631"} Nov 24 13:30:27 crc kubenswrapper[4790]: I1124 13:30:27.297162 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-fg9nc" event={"ID":"1e565309-31eb-4278-98fd-5a4321d5eb10","Type":"ContainerStarted","Data":"496b0d2bae217642b1338fbac57ed2691bf2cbe97e67e6923cda6a51b14dda09"} Nov 24 13:30:27 crc kubenswrapper[4790]: E1124 13:30:27.326799 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-c8xtz" podUID="cb16570b-3e90-4b2e-996b-d0cd5c0a0259" Nov 24 13:30:27 crc kubenswrapper[4790]: I1124 13:30:27.329690 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7969689c84-wq7t5" event={"ID":"d39e8176-e052-4a0e-9d16-e66b6312fbfc","Type":"ContainerStarted","Data":"2430da42e4a137ac4dbe1c038ec04c90d4be4bba957861ae9182a4f2670ec74a"} Nov 24 13:30:27 crc kubenswrapper[4790]: I1124 13:30:27.333023 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7547bd9b9-skhbm" event={"ID":"7c72ca69-273b-4c92-a5f7-54c2b968621b","Type":"ContainerStarted","Data":"b549349e166d89a901290ff1fec6ee2fa7610715fabae8ccc207edc23a4b2e86"} Nov 24 13:30:27 crc kubenswrapper[4790]: E1124 13:30:27.383221 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-jg2n4" podUID="c861ea1f-e001-4f01-8e0d-12e97b7628cc" Nov 24 13:30:27 crc kubenswrapper[4790]: E1124 13:30:27.528762 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-ld798" podUID="84281187-e5c6-4b57-807a-ae5637942260" Nov 24 13:30:28 crc kubenswrapper[4790]: I1124 13:30:28.385519 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-b4c496f69-8k88p" event={"ID":"5e93acb0-367e-4f6c-b13e-530d41db7851","Type":"ContainerStarted","Data":"8a753c2a119724eb3c4fb8b6b8e5f3cf4e3572ccb19878e2aa03009881f5b281"} Nov 24 13:30:28 crc kubenswrapper[4790]: E1124 13:30:28.387788 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\"" pod="openstack-operators/test-operator-controller-manager-b4c496f69-8k88p" podUID="5e93acb0-367e-4f6c-b13e-530d41db7851" Nov 24 13:30:28 crc kubenswrapper[4790]: I1124 13:30:28.393446 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-jg2n4" event={"ID":"c861ea1f-e001-4f01-8e0d-12e97b7628cc","Type":"ContainerStarted","Data":"71f945d1c633f7e4fef7184021c0cf977d206391185228f07d17d42765015da7"} Nov 24 13:30:28 crc kubenswrapper[4790]: E1124 13:30:28.397340 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-jg2n4" podUID="c861ea1f-e001-4f01-8e0d-12e97b7628cc" Nov 24 13:30:28 crc kubenswrapper[4790]: I1124 13:30:28.402129 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-ld798" event={"ID":"84281187-e5c6-4b57-807a-ae5637942260","Type":"ContainerStarted","Data":"fe5d0230a8510115ebfe6c1f826d184671c9cd44682675fb9f8dc62bb5692361"} Nov 24 13:30:28 crc kubenswrapper[4790]: E1124 13:30:28.403853 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-ld798" podUID="84281187-e5c6-4b57-807a-ae5637942260" Nov 24 13:30:28 crc kubenswrapper[4790]: I1124 13:30:28.406534 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-c8xtz" event={"ID":"cb16570b-3e90-4b2e-996b-d0cd5c0a0259","Type":"ContainerStarted","Data":"b2724d958e393225995f9edf43f7aed0199b6510c326fc1706b7f53d58cc4461"} Nov 24 13:30:28 crc kubenswrapper[4790]: E1124 13:30:28.408187 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-c8xtz" podUID="cb16570b-3e90-4b2e-996b-d0cd5c0a0259" Nov 24 13:30:28 crc kubenswrapper[4790]: I1124 13:30:28.438795 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7547bd9b9-skhbm" event={"ID":"7c72ca69-273b-4c92-a5f7-54c2b968621b","Type":"ContainerStarted","Data":"938e69531046f0fae589bef74a8c1e6c82a2165651e0f63ba3a22bc9c32f4b25"} Nov 24 13:30:28 crc kubenswrapper[4790]: I1124 13:30:28.438858 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7547bd9b9-skhbm" event={"ID":"7c72ca69-273b-4c92-a5f7-54c2b968621b","Type":"ContainerStarted","Data":"0eaaac8dc5cedcfe96c644600c1dbda0a63b55456df8d0549ed9859057119fbd"} Nov 24 13:30:28 crc kubenswrapper[4790]: E1124 13:30:28.441748 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\"" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-gkdz7" podUID="e25e339c-c102-44f0-b4a4-15f0ce87b38c" Nov 24 13:30:28 crc kubenswrapper[4790]: E1124 13:30:28.460903 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-6dwzt" podUID="8e1bedf8-78bf-4558-91e6-4228e9765356" Nov 24 13:30:28 crc kubenswrapper[4790]: I1124 13:30:28.557549 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-7547bd9b9-skhbm" podStartSLOduration=3.557530888 podStartE2EDuration="3.557530888s" podCreationTimestamp="2025-11-24 13:30:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:30:28.55657596 +0000 UTC m=+1076.936469622" watchObservedRunningTime="2025-11-24 13:30:28.557530888 +0000 UTC m=+1076.937424550" Nov 24 13:30:29 crc kubenswrapper[4790]: I1124 13:30:29.452123 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-7547bd9b9-skhbm" Nov 24 13:30:29 crc kubenswrapper[4790]: E1124 13:30:29.453468 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-c8xtz" podUID="cb16570b-3e90-4b2e-996b-d0cd5c0a0259" Nov 24 13:30:29 crc kubenswrapper[4790]: E1124 13:30:29.454626 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-jg2n4" podUID="c861ea1f-e001-4f01-8e0d-12e97b7628cc" Nov 24 13:30:29 crc kubenswrapper[4790]: E1124 13:30:29.456430 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\"" pod="openstack-operators/test-operator-controller-manager-b4c496f69-8k88p" podUID="5e93acb0-367e-4f6c-b13e-530d41db7851" Nov 24 13:30:29 crc kubenswrapper[4790]: E1124 13:30:29.456525 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-ld798" podUID="84281187-e5c6-4b57-807a-ae5637942260" Nov 24 13:30:35 crc kubenswrapper[4790]: I1124 13:30:35.771090 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-7547bd9b9-skhbm" Nov 24 13:30:39 crc kubenswrapper[4790]: E1124 13:30:39.126340 4790 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0" Nov 24 13:30:39 crc kubenswrapper[4790]: E1124 13:30:39.127313 4790 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5h8bg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-d656998f4-lfvdf_openstack-operators(69644679-f0fb-4c09-abec-8e4084140f05): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:30:40 crc kubenswrapper[4790]: E1124 13:30:40.032203 4790 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:b749a5dd8bc718875c3f5e81b38d54d003be77ab92de4a3e9f9595566496a58a" Nov 24 13:30:40 crc kubenswrapper[4790]: E1124 13:30:40.032509 4790 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:b749a5dd8bc718875c3f5e81b38d54d003be77ab92de4a3e9f9595566496a58a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xx4t9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-58f887965d-qvbnj_openstack-operators(3dc6a5ba-d447-40fc-bd8b-640e4b49990b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:30:40 crc kubenswrapper[4790]: E1124 13:30:40.504260 4790 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:70cce55bcf89468c5d468ca2fc317bfc3dc5f2bef1c502df9faca2eb1293ede7" Nov 24 13:30:40 crc kubenswrapper[4790]: E1124 13:30:40.504508 4790 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:70cce55bcf89468c5d468ca2fc317bfc3dc5f2bef1c502df9faca2eb1293ede7,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-w889m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-75fb479bcc-9wl9p_openstack-operators(b55ba4e8-3d0a-4746-817c-feb1a210b01e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:30:41 crc kubenswrapper[4790]: E1124 13:30:41.923381 4790 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f" Nov 24 13:30:41 crc kubenswrapper[4790]: E1124 13:30:41.923596 4790 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7lfpv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-8c6448b9f-cddlm_openstack-operators(e85f1bfb-0153-40e0-b54b-728171cb39f0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:30:42 crc kubenswrapper[4790]: E1124 13:30:42.208902 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-d656998f4-lfvdf" podUID="69644679-f0fb-4c09-abec-8e4084140f05" Nov 24 13:30:42 crc kubenswrapper[4790]: E1124 13:30:42.224571 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-9wl9p" podUID="b55ba4e8-3d0a-4746-817c-feb1a210b01e" Nov 24 13:30:42 crc kubenswrapper[4790]: E1124 13:30:42.265672 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-58f887965d-qvbnj" podUID="3dc6a5ba-d447-40fc-bd8b-640e4b49990b" Nov 24 13:30:42 crc kubenswrapper[4790]: I1124 13:30:42.561001 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7969689c84-wq7t5" event={"ID":"d39e8176-e052-4a0e-9d16-e66b6312fbfc","Type":"ContainerStarted","Data":"4202ced4100e1c48db6befa8c98a6bdd61bc4eb25f2b81bc0ecb25f3b8f774c4"} Nov 24 13:30:42 crc kubenswrapper[4790]: I1124 13:30:42.562203 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-d8gzf" event={"ID":"0c8ec99a-dfcc-414d-85c9-f005c476919d","Type":"ContainerStarted","Data":"7d0917cb36eebeccc71bbc5b67c560b8c18e917e43ed233d29c99eca50b54ba2"} Nov 24 13:30:42 crc kubenswrapper[4790]: I1124 13:30:42.569132 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d656998f4-lfvdf" event={"ID":"69644679-f0fb-4c09-abec-8e4084140f05","Type":"ContainerStarted","Data":"92fbb081fe139838dd8f40527fe41dcfdc35de55713f5dc63ef30a051806b43b"} Nov 24 13:30:42 crc kubenswrapper[4790]: E1124 13:30:42.570480 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0\\\"\"" pod="openstack-operators/swift-operator-controller-manager-d656998f4-lfvdf" podUID="69644679-f0fb-4c09-abec-8e4084140f05" Nov 24 13:30:42 crc kubenswrapper[4790]: I1124 13:30:42.572343 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58f887965d-qvbnj" event={"ID":"3dc6a5ba-d447-40fc-bd8b-640e4b49990b","Type":"ContainerStarted","Data":"fb7dfdd5e84c69fdf4ade8d41669f10feac87d4a371b9adc346937b48632ea90"} Nov 24 13:30:42 crc kubenswrapper[4790]: E1124 13:30:42.573566 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:b749a5dd8bc718875c3f5e81b38d54d003be77ab92de4a3e9f9595566496a58a\\\"\"" pod="openstack-operators/manila-operator-controller-manager-58f887965d-qvbnj" podUID="3dc6a5ba-d447-40fc-bd8b-640e4b49990b" Nov 24 13:30:42 crc kubenswrapper[4790]: I1124 13:30:42.575791 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-fg9nc" event={"ID":"1e565309-31eb-4278-98fd-5a4321d5eb10","Type":"ContainerStarted","Data":"34e1e8960f81db5e50423ad7997ef6e40ad58d9a133c5ad6d24772d7c270ecaa"} Nov 24 13:30:42 crc kubenswrapper[4790]: I1124 13:30:42.576745 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-4bndq" event={"ID":"2c64049e-cf92-40ac-84d9-e2724c334c3f","Type":"ContainerStarted","Data":"4a074d42fe46f666ecba5e5e6a8a008f528b63e23f23c7ea975fec624410a38a"} Nov 24 13:30:42 crc kubenswrapper[4790]: I1124 13:30:42.578252 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-9wl9p" event={"ID":"b55ba4e8-3d0a-4746-817c-feb1a210b01e","Type":"ContainerStarted","Data":"7107799b417af5213c437fd34c89e90cf0e29bb013cbe293eee64ffd202a8f84"} Nov 24 13:30:42 crc kubenswrapper[4790]: E1124 13:30:42.579473 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/barbican-operator@sha256:70cce55bcf89468c5d468ca2fc317bfc3dc5f2bef1c502df9faca2eb1293ede7\\\"\"" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-9wl9p" podUID="b55ba4e8-3d0a-4746-817c-feb1a210b01e" Nov 24 13:30:43 crc kubenswrapper[4790]: I1124 13:30:43.585735 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-h8nkv" event={"ID":"b2b9b9e9-800e-4213-8605-058d91412276","Type":"ContainerStarted","Data":"2de5f2618453ba4545772eb32ab625ace5fb9877e8cbdfc8d9274f874ea4a79a"} Nov 24 13:30:43 crc kubenswrapper[4790]: I1124 13:30:43.587769 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-bzh6h" event={"ID":"49de2fb5-2dde-400e-9ae4-d36dfe196ff1","Type":"ContainerStarted","Data":"01103089c8665334e67ce5f59003b70f3fcde7796d0e2a864f3fcf5189d344de"} Nov 24 13:30:43 crc kubenswrapper[4790]: I1124 13:30:43.589425 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-wzvsq" event={"ID":"4c1c13a1-3285-45f5-888b-d9abb0008661","Type":"ContainerStarted","Data":"d524b808d5f430b54a53394c5b830df08c5dfc07771d81465db373cc7b8da435"} Nov 24 13:30:43 crc kubenswrapper[4790]: I1124 13:30:43.592128 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-44wlb" event={"ID":"09e60927-110e-4cda-b7e3-6f47e70f2388","Type":"ContainerStarted","Data":"3823de70dbaef054d0689cab6ec646ef271b12df9206e6ac6837a224ad4d8a93"} Nov 24 13:30:43 crc kubenswrapper[4790]: E1124 13:30:43.594057 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/barbican-operator@sha256:70cce55bcf89468c5d468ca2fc317bfc3dc5f2bef1c502df9faca2eb1293ede7\\\"\"" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-9wl9p" podUID="b55ba4e8-3d0a-4746-817c-feb1a210b01e" Nov 24 13:30:43 crc kubenswrapper[4790]: E1124 13:30:43.594425 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0\\\"\"" pod="openstack-operators/swift-operator-controller-manager-d656998f4-lfvdf" podUID="69644679-f0fb-4c09-abec-8e4084140f05" Nov 24 13:30:43 crc kubenswrapper[4790]: E1124 13:30:43.594873 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:b749a5dd8bc718875c3f5e81b38d54d003be77ab92de4a3e9f9595566496a58a\\\"\"" pod="openstack-operators/manila-operator-controller-manager-58f887965d-qvbnj" podUID="3dc6a5ba-d447-40fc-bd8b-640e4b49990b" Nov 24 13:30:44 crc kubenswrapper[4790]: I1124 13:30:44.630445 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-ntfjb" event={"ID":"52519a2b-7c33-425f-9115-6fd4876ba60d","Type":"ContainerStarted","Data":"397d0b657401ec5cb8b443bc8148faac5dc013bb98c7a6f6cbed0473a7f7ef37"} Nov 24 13:30:44 crc kubenswrapper[4790]: I1124 13:30:44.632894 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-9gnsv" event={"ID":"dc60f94f-01f9-4179-a0ce-67112913da75","Type":"ContainerStarted","Data":"378ceb017c8942a96c40d4f7078da12bcf024039a648e153045f374499c891ab"} Nov 24 13:30:44 crc kubenswrapper[4790]: I1124 13:30:44.634005 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-95l27" event={"ID":"8b819ac2-ef31-494d-b4cf-882b602feb70","Type":"ContainerStarted","Data":"94717ca96e7fdc09dcbbeee621fa0acfd069438c933a69b10770e042ba2630d5"} Nov 24 13:30:44 crc kubenswrapper[4790]: I1124 13:30:44.635288 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-fxplj" event={"ID":"7000614b-01cb-4878-b85f-b5a1f8ef91b2","Type":"ContainerStarted","Data":"afeb91744493aa11ae7c5b29435d21ec8f54f5480f512bb6b82a724f33dc9dcb"} Nov 24 13:30:45 crc kubenswrapper[4790]: E1124 13:30:45.352715 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-cddlm" podUID="e85f1bfb-0153-40e0-b54b-728171cb39f0" Nov 24 13:30:45 crc kubenswrapper[4790]: I1124 13:30:45.644206 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-cddlm" event={"ID":"e85f1bfb-0153-40e0-b54b-728171cb39f0","Type":"ContainerStarted","Data":"d51ed2d8625c61e7f7f1bba91af7b814258c9c5dd87ea130fda9105f3d127135"} Nov 24 13:30:45 crc kubenswrapper[4790]: E1124 13:30:45.647557 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-cddlm" podUID="e85f1bfb-0153-40e0-b54b-728171cb39f0" Nov 24 13:30:46 crc kubenswrapper[4790]: I1124 13:30:46.652546 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-4bndq" event={"ID":"2c64049e-cf92-40ac-84d9-e2724c334c3f","Type":"ContainerStarted","Data":"24af3aed94f943e9d56596d5993e99013125a3ca6ec71571b0c5f0dac36eafa1"} Nov 24 13:30:46 crc kubenswrapper[4790]: E1124 13:30:46.653809 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-cddlm" podUID="e85f1bfb-0153-40e0-b54b-728171cb39f0" Nov 24 13:30:46 crc kubenswrapper[4790]: I1124 13:30:46.683591 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-4bndq" podStartSLOduration=6.80414353 podStartE2EDuration="22.683573631s" podCreationTimestamp="2025-11-24 13:30:24 +0000 UTC" firstStartedPulling="2025-11-24 13:30:26.005141618 +0000 UTC m=+1074.385035280" lastFinishedPulling="2025-11-24 13:30:41.884571719 +0000 UTC m=+1090.264465381" observedRunningTime="2025-11-24 13:30:46.66839166 +0000 UTC m=+1095.048285332" watchObservedRunningTime="2025-11-24 13:30:46.683573631 +0000 UTC m=+1095.063467293" Nov 24 13:30:47 crc kubenswrapper[4790]: I1124 13:30:47.661345 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-95l27" event={"ID":"8b819ac2-ef31-494d-b4cf-882b602feb70","Type":"ContainerStarted","Data":"b35041e5bf114e959f987f08f8b279832c2906e31ef843d274baf5a5f4fce29f"} Nov 24 13:30:47 crc kubenswrapper[4790]: I1124 13:30:47.662222 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-95l27" Nov 24 13:30:47 crc kubenswrapper[4790]: I1124 13:30:47.665517 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-wzvsq" event={"ID":"4c1c13a1-3285-45f5-888b-d9abb0008661","Type":"ContainerStarted","Data":"3b8568ba84cd88073bbfc8d40367aa89915353996797088479eeeddb0a5e73fc"} Nov 24 13:30:47 crc kubenswrapper[4790]: I1124 13:30:47.665592 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-wzvsq" Nov 24 13:30:47 crc kubenswrapper[4790]: I1124 13:30:47.670641 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-ntfjb" event={"ID":"52519a2b-7c33-425f-9115-6fd4876ba60d","Type":"ContainerStarted","Data":"2eebc5e118b8b61e8c2181ef430633d8e790772958410a631871e50b6872a10c"} Nov 24 13:30:47 crc kubenswrapper[4790]: I1124 13:30:47.671277 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-ntfjb" Nov 24 13:30:47 crc kubenswrapper[4790]: I1124 13:30:47.673971 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-gkdz7" event={"ID":"e25e339c-c102-44f0-b4a4-15f0ce87b38c","Type":"ContainerStarted","Data":"691a582bd267faece9840985e4cb8316898ba4d1b70f1ff01500bd052190b9af"} Nov 24 13:30:47 crc kubenswrapper[4790]: I1124 13:30:47.674579 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-gkdz7" Nov 24 13:30:47 crc kubenswrapper[4790]: I1124 13:30:47.683105 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-95l27" podStartSLOduration=8.113065105 podStartE2EDuration="23.683082791s" podCreationTimestamp="2025-11-24 13:30:24 +0000 UTC" firstStartedPulling="2025-11-24 13:30:26.336742636 +0000 UTC m=+1074.716636298" lastFinishedPulling="2025-11-24 13:30:41.906760322 +0000 UTC m=+1090.286653984" observedRunningTime="2025-11-24 13:30:47.677040095 +0000 UTC m=+1096.056933757" watchObservedRunningTime="2025-11-24 13:30:47.683082791 +0000 UTC m=+1096.062976453" Nov 24 13:30:47 crc kubenswrapper[4790]: I1124 13:30:47.686445 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-44wlb" event={"ID":"09e60927-110e-4cda-b7e3-6f47e70f2388","Type":"ContainerStarted","Data":"ff6f08fa37309d04aedb5ac16efacc6095db32e5318cdc214dc46467124636bd"} Nov 24 13:30:47 crc kubenswrapper[4790]: I1124 13:30:47.687032 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-44wlb" Nov 24 13:30:47 crc kubenswrapper[4790]: I1124 13:30:47.689948 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-d8gzf" event={"ID":"0c8ec99a-dfcc-414d-85c9-f005c476919d","Type":"ContainerStarted","Data":"6c6e8419a9f2ae648accf6b7e3b04ac62d54b00d454b3e723bf9c2b94ea0a6eb"} Nov 24 13:30:47 crc kubenswrapper[4790]: I1124 13:30:47.690225 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-d8gzf" Nov 24 13:30:47 crc kubenswrapper[4790]: I1124 13:30:47.694427 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-d8gzf" Nov 24 13:30:47 crc kubenswrapper[4790]: I1124 13:30:47.704136 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-gkdz7" podStartSLOduration=3.161600871 podStartE2EDuration="23.704115221s" podCreationTimestamp="2025-11-24 13:30:24 +0000 UTC" firstStartedPulling="2025-11-24 13:30:26.632792263 +0000 UTC m=+1075.012685925" lastFinishedPulling="2025-11-24 13:30:47.175306573 +0000 UTC m=+1095.555200275" observedRunningTime="2025-11-24 13:30:47.700436414 +0000 UTC m=+1096.080330076" watchObservedRunningTime="2025-11-24 13:30:47.704115221 +0000 UTC m=+1096.084008883" Nov 24 13:30:47 crc kubenswrapper[4790]: I1124 13:30:47.711384 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-bzh6h" event={"ID":"49de2fb5-2dde-400e-9ae4-d36dfe196ff1","Type":"ContainerStarted","Data":"e17930f4070e53886e855ce9684c2f1a2f94e7318f7ea900b8c26f5b58dbe760"} Nov 24 13:30:47 crc kubenswrapper[4790]: I1124 13:30:47.711660 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-4bndq" Nov 24 13:30:47 crc kubenswrapper[4790]: I1124 13:30:47.712150 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-bzh6h" Nov 24 13:30:47 crc kubenswrapper[4790]: I1124 13:30:47.716618 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-4bndq" Nov 24 13:30:47 crc kubenswrapper[4790]: I1124 13:30:47.726238 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-wzvsq" podStartSLOduration=7.961693084 podStartE2EDuration="23.726220372s" podCreationTimestamp="2025-11-24 13:30:24 +0000 UTC" firstStartedPulling="2025-11-24 13:30:26.122917754 +0000 UTC m=+1074.502811416" lastFinishedPulling="2025-11-24 13:30:41.887445042 +0000 UTC m=+1090.267338704" observedRunningTime="2025-11-24 13:30:47.725699487 +0000 UTC m=+1096.105593149" watchObservedRunningTime="2025-11-24 13:30:47.726220372 +0000 UTC m=+1096.106114024" Nov 24 13:30:47 crc kubenswrapper[4790]: I1124 13:30:47.761334 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-ntfjb" podStartSLOduration=7.902280382 podStartE2EDuration="23.7613087s" podCreationTimestamp="2025-11-24 13:30:24 +0000 UTC" firstStartedPulling="2025-11-24 13:30:26.047768595 +0000 UTC m=+1074.427662247" lastFinishedPulling="2025-11-24 13:30:41.906796903 +0000 UTC m=+1090.286690565" observedRunningTime="2025-11-24 13:30:47.755329226 +0000 UTC m=+1096.135222888" watchObservedRunningTime="2025-11-24 13:30:47.7613087 +0000 UTC m=+1096.141202352" Nov 24 13:30:47 crc kubenswrapper[4790]: I1124 13:30:47.830300 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-44wlb" podStartSLOduration=7.920265852 podStartE2EDuration="23.83027981s" podCreationTimestamp="2025-11-24 13:30:24 +0000 UTC" firstStartedPulling="2025-11-24 13:30:25.976630431 +0000 UTC m=+1074.356524093" lastFinishedPulling="2025-11-24 13:30:41.886644389 +0000 UTC m=+1090.266538051" observedRunningTime="2025-11-24 13:30:47.82682283 +0000 UTC m=+1096.206716492" watchObservedRunningTime="2025-11-24 13:30:47.83027981 +0000 UTC m=+1096.210173472" Nov 24 13:30:47 crc kubenswrapper[4790]: I1124 13:30:47.888832 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-bzh6h" podStartSLOduration=7.651782275 podStartE2EDuration="23.888814038s" podCreationTimestamp="2025-11-24 13:30:24 +0000 UTC" firstStartedPulling="2025-11-24 13:30:25.648465023 +0000 UTC m=+1074.028358685" lastFinishedPulling="2025-11-24 13:30:41.885496786 +0000 UTC m=+1090.265390448" observedRunningTime="2025-11-24 13:30:47.867358276 +0000 UTC m=+1096.247251948" watchObservedRunningTime="2025-11-24 13:30:47.888814038 +0000 UTC m=+1096.268707700" Nov 24 13:30:48 crc kubenswrapper[4790]: I1124 13:30:48.719808 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-ld798" event={"ID":"84281187-e5c6-4b57-807a-ae5637942260","Type":"ContainerStarted","Data":"e221040bc09239151d02d50f4cfde32d4d151c854669627144aec925ff30c297"} Nov 24 13:30:48 crc kubenswrapper[4790]: I1124 13:30:48.720252 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-ld798" Nov 24 13:30:48 crc kubenswrapper[4790]: I1124 13:30:48.721523 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-h8nkv" event={"ID":"b2b9b9e9-800e-4213-8605-058d91412276","Type":"ContainerStarted","Data":"e73e0909e6310e60dd12d4e845566970fb23a306235ff351927c04e2500f0e2b"} Nov 24 13:30:48 crc kubenswrapper[4790]: I1124 13:30:48.721773 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-h8nkv" Nov 24 13:30:48 crc kubenswrapper[4790]: I1124 13:30:48.723047 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-fxplj" event={"ID":"7000614b-01cb-4878-b85f-b5a1f8ef91b2","Type":"ContainerStarted","Data":"5bf3348e8016027dc4e9f63667821e3fde082b47a5deca62af544c4d120be617"} Nov 24 13:30:48 crc kubenswrapper[4790]: I1124 13:30:48.723183 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-fxplj" Nov 24 13:30:48 crc kubenswrapper[4790]: I1124 13:30:48.724176 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-6dwzt" event={"ID":"8e1bedf8-78bf-4558-91e6-4228e9765356","Type":"ContainerStarted","Data":"77c9dae0b7fdc0c5b7e2aa1902a9604cd947caa767b771693c61190d0a353f62"} Nov 24 13:30:48 crc kubenswrapper[4790]: I1124 13:30:48.724253 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-h8nkv" Nov 24 13:30:48 crc kubenswrapper[4790]: I1124 13:30:48.725508 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-jg2n4" event={"ID":"c861ea1f-e001-4f01-8e0d-12e97b7628cc","Type":"ContainerStarted","Data":"a4353fb2d7bd61029edd606ecb718c2f40afa00ad43386d38759428baa7fbfe6"} Nov 24 13:30:48 crc kubenswrapper[4790]: I1124 13:30:48.725646 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-jg2n4" Nov 24 13:30:48 crc kubenswrapper[4790]: I1124 13:30:48.726823 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-c8xtz" event={"ID":"cb16570b-3e90-4b2e-996b-d0cd5c0a0259","Type":"ContainerStarted","Data":"eba5162d3ce8bc2cd1b83b6e59a32360ae00bce7a847de2654a62cf86992cdd2"} Nov 24 13:30:48 crc kubenswrapper[4790]: I1124 13:30:48.727049 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-c8xtz" Nov 24 13:30:48 crc kubenswrapper[4790]: I1124 13:30:48.728145 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7969689c84-wq7t5" event={"ID":"d39e8176-e052-4a0e-9d16-e66b6312fbfc","Type":"ContainerStarted","Data":"191da6decb899e8302fb04c03d28aff6ef7402e012979ea60e633839c6acd873"} Nov 24 13:30:48 crc kubenswrapper[4790]: I1124 13:30:48.728304 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-7969689c84-wq7t5" Nov 24 13:30:48 crc kubenswrapper[4790]: I1124 13:30:48.729496 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-9gnsv" event={"ID":"dc60f94f-01f9-4179-a0ce-67112913da75","Type":"ContainerStarted","Data":"466d71876e83346102b903deda06721536cf0963f3654b58f80a238bae4c9ba4"} Nov 24 13:30:48 crc kubenswrapper[4790]: I1124 13:30:48.729623 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-9gnsv" Nov 24 13:30:48 crc kubenswrapper[4790]: I1124 13:30:48.729954 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-7969689c84-wq7t5" Nov 24 13:30:48 crc kubenswrapper[4790]: I1124 13:30:48.730867 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-fg9nc" event={"ID":"1e565309-31eb-4278-98fd-5a4321d5eb10","Type":"ContainerStarted","Data":"cd34493cb58109915b7e9a657004646307d4441bed14dc50c7d29bc49cbdf26c"} Nov 24 13:30:48 crc kubenswrapper[4790]: I1124 13:30:48.731058 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-fg9nc" Nov 24 13:30:48 crc kubenswrapper[4790]: I1124 13:30:48.732262 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-fg9nc" Nov 24 13:30:48 crc kubenswrapper[4790]: I1124 13:30:48.732477 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-b4c496f69-8k88p" event={"ID":"5e93acb0-367e-4f6c-b13e-530d41db7851","Type":"ContainerStarted","Data":"d0248d50f1efc4643b61de1c622e2829e168288c669651ce28a969360b2224a0"} Nov 24 13:30:48 crc kubenswrapper[4790]: I1124 13:30:48.734103 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-wzvsq" Nov 24 13:30:48 crc kubenswrapper[4790]: I1124 13:30:48.735660 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-bzh6h" Nov 24 13:30:48 crc kubenswrapper[4790]: I1124 13:30:48.735693 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-44wlb" Nov 24 13:30:48 crc kubenswrapper[4790]: I1124 13:30:48.742694 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-ld798" podStartSLOduration=4.479717721 podStartE2EDuration="24.742674373s" podCreationTimestamp="2025-11-24 13:30:24 +0000 UTC" firstStartedPulling="2025-11-24 13:30:26.91233175 +0000 UTC m=+1075.292225412" lastFinishedPulling="2025-11-24 13:30:47.175288402 +0000 UTC m=+1095.555182064" observedRunningTime="2025-11-24 13:30:48.740462899 +0000 UTC m=+1097.120356581" watchObservedRunningTime="2025-11-24 13:30:48.742674373 +0000 UTC m=+1097.122568045" Nov 24 13:30:48 crc kubenswrapper[4790]: I1124 13:30:48.744097 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-d8gzf" podStartSLOduration=9.331160794 podStartE2EDuration="24.744087514s" podCreationTimestamp="2025-11-24 13:30:24 +0000 UTC" firstStartedPulling="2025-11-24 13:30:26.494679987 +0000 UTC m=+1074.874573649" lastFinishedPulling="2025-11-24 13:30:41.907606707 +0000 UTC m=+1090.287500369" observedRunningTime="2025-11-24 13:30:47.888376505 +0000 UTC m=+1096.268270167" watchObservedRunningTime="2025-11-24 13:30:48.744087514 +0000 UTC m=+1097.123981186" Nov 24 13:30:48 crc kubenswrapper[4790]: I1124 13:30:48.756820 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-7969689c84-wq7t5" podStartSLOduration=9.231817153 podStartE2EDuration="24.756799193s" podCreationTimestamp="2025-11-24 13:30:24 +0000 UTC" firstStartedPulling="2025-11-24 13:30:26.358866588 +0000 UTC m=+1074.738760250" lastFinishedPulling="2025-11-24 13:30:41.883848628 +0000 UTC m=+1090.263742290" observedRunningTime="2025-11-24 13:30:48.754533937 +0000 UTC m=+1097.134427599" watchObservedRunningTime="2025-11-24 13:30:48.756799193 +0000 UTC m=+1097.136692865" Nov 24 13:30:48 crc kubenswrapper[4790]: I1124 13:30:48.818591 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-fxplj" podStartSLOduration=9.248188157 podStartE2EDuration="24.818573524s" podCreationTimestamp="2025-11-24 13:30:24 +0000 UTC" firstStartedPulling="2025-11-24 13:30:26.336807618 +0000 UTC m=+1074.716701280" lastFinishedPulling="2025-11-24 13:30:41.907192985 +0000 UTC m=+1090.287086647" observedRunningTime="2025-11-24 13:30:48.818143342 +0000 UTC m=+1097.198037004" watchObservedRunningTime="2025-11-24 13:30:48.818573524 +0000 UTC m=+1097.198467186" Nov 24 13:30:48 crc kubenswrapper[4790]: I1124 13:30:48.859998 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-jg2n4" podStartSLOduration=4.615202641 podStartE2EDuration="24.859956595s" podCreationTimestamp="2025-11-24 13:30:24 +0000 UTC" firstStartedPulling="2025-11-24 13:30:26.930418225 +0000 UTC m=+1075.310311887" lastFinishedPulling="2025-11-24 13:30:47.175172179 +0000 UTC m=+1095.555065841" observedRunningTime="2025-11-24 13:30:48.851654474 +0000 UTC m=+1097.231548136" watchObservedRunningTime="2025-11-24 13:30:48.859956595 +0000 UTC m=+1097.239850267" Nov 24 13:30:48 crc kubenswrapper[4790]: I1124 13:30:48.871539 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-6dwzt" podStartSLOduration=3.646647323 podStartE2EDuration="23.87151952s" podCreationTimestamp="2025-11-24 13:30:25 +0000 UTC" firstStartedPulling="2025-11-24 13:30:26.950275171 +0000 UTC m=+1075.330168833" lastFinishedPulling="2025-11-24 13:30:47.175147368 +0000 UTC m=+1095.555041030" observedRunningTime="2025-11-24 13:30:48.869772729 +0000 UTC m=+1097.249666401" watchObservedRunningTime="2025-11-24 13:30:48.87151952 +0000 UTC m=+1097.251413182" Nov 24 13:30:48 crc kubenswrapper[4790]: I1124 13:30:48.920714 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-fg9nc" podStartSLOduration=9.610819266 podStartE2EDuration="24.920694796s" podCreationTimestamp="2025-11-24 13:30:24 +0000 UTC" firstStartedPulling="2025-11-24 13:30:26.574989937 +0000 UTC m=+1074.954883599" lastFinishedPulling="2025-11-24 13:30:41.884865467 +0000 UTC m=+1090.264759129" observedRunningTime="2025-11-24 13:30:48.915372882 +0000 UTC m=+1097.295266544" watchObservedRunningTime="2025-11-24 13:30:48.920694796 +0000 UTC m=+1097.300588458" Nov 24 13:30:48 crc kubenswrapper[4790]: I1124 13:30:48.951005 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-b4c496f69-8k88p" podStartSLOduration=5.7251729749999996 podStartE2EDuration="24.950982205s" podCreationTimestamp="2025-11-24 13:30:24 +0000 UTC" firstStartedPulling="2025-11-24 13:30:26.930007503 +0000 UTC m=+1075.309901165" lastFinishedPulling="2025-11-24 13:30:46.155816723 +0000 UTC m=+1094.535710395" observedRunningTime="2025-11-24 13:30:48.949288266 +0000 UTC m=+1097.329181928" watchObservedRunningTime="2025-11-24 13:30:48.950982205 +0000 UTC m=+1097.330875887" Nov 24 13:30:49 crc kubenswrapper[4790]: I1124 13:30:48.991696 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-c8xtz" podStartSLOduration=4.852831812 podStartE2EDuration="24.991662615s" podCreationTimestamp="2025-11-24 13:30:24 +0000 UTC" firstStartedPulling="2025-11-24 13:30:26.920718493 +0000 UTC m=+1075.300612155" lastFinishedPulling="2025-11-24 13:30:47.059549296 +0000 UTC m=+1095.439442958" observedRunningTime="2025-11-24 13:30:48.974370613 +0000 UTC m=+1097.354264295" watchObservedRunningTime="2025-11-24 13:30:48.991662615 +0000 UTC m=+1097.371556277" Nov 24 13:30:49 crc kubenswrapper[4790]: I1124 13:30:48.994169 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-9gnsv" podStartSLOduration=9.716390758 podStartE2EDuration="24.994159127s" podCreationTimestamp="2025-11-24 13:30:24 +0000 UTC" firstStartedPulling="2025-11-24 13:30:26.607429918 +0000 UTC m=+1074.987323580" lastFinishedPulling="2025-11-24 13:30:41.885198287 +0000 UTC m=+1090.265091949" observedRunningTime="2025-11-24 13:30:48.989575094 +0000 UTC m=+1097.369468756" watchObservedRunningTime="2025-11-24 13:30:48.994159127 +0000 UTC m=+1097.374052789" Nov 24 13:30:49 crc kubenswrapper[4790]: I1124 13:30:49.020898 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-h8nkv" podStartSLOduration=9.09037374 podStartE2EDuration="25.020856462s" podCreationTimestamp="2025-11-24 13:30:24 +0000 UTC" firstStartedPulling="2025-11-24 13:30:25.954113998 +0000 UTC m=+1074.334007660" lastFinishedPulling="2025-11-24 13:30:41.88459672 +0000 UTC m=+1090.264490382" observedRunningTime="2025-11-24 13:30:49.012258252 +0000 UTC m=+1097.392151914" watchObservedRunningTime="2025-11-24 13:30:49.020856462 +0000 UTC m=+1097.400750144" Nov 24 13:30:49 crc kubenswrapper[4790]: I1124 13:30:49.742486 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-9gnsv" Nov 24 13:30:49 crc kubenswrapper[4790]: I1124 13:30:49.743240 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-ntfjb" Nov 24 13:30:49 crc kubenswrapper[4790]: I1124 13:30:49.745248 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-fxplj" Nov 24 13:30:54 crc kubenswrapper[4790]: I1124 13:30:54.916524 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-95l27" Nov 24 13:30:55 crc kubenswrapper[4790]: I1124 13:30:55.036671 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-gkdz7" Nov 24 13:30:55 crc kubenswrapper[4790]: I1124 13:30:55.301762 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-c8xtz" Nov 24 13:30:55 crc kubenswrapper[4790]: I1124 13:30:55.324059 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-ld798" Nov 24 13:30:55 crc kubenswrapper[4790]: I1124 13:30:55.587984 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-b4c496f69-8k88p" Nov 24 13:30:55 crc kubenswrapper[4790]: I1124 13:30:55.591390 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-b4c496f69-8k88p" Nov 24 13:30:55 crc kubenswrapper[4790]: I1124 13:30:55.845263 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-jg2n4" Nov 24 13:30:56 crc kubenswrapper[4790]: I1124 13:30:56.317562 4790 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 13:30:57 crc kubenswrapper[4790]: I1124 13:30:57.811664 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-9wl9p" event={"ID":"b55ba4e8-3d0a-4746-817c-feb1a210b01e","Type":"ContainerStarted","Data":"c72b1633640d0810633fcbde8a9fa6d40494262174f6cdcf05e4b672b5a84d00"} Nov 24 13:30:57 crc kubenswrapper[4790]: I1124 13:30:57.812279 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-9wl9p" Nov 24 13:30:57 crc kubenswrapper[4790]: I1124 13:30:57.835593 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-9wl9p" podStartSLOduration=2.688787177 podStartE2EDuration="33.835578811s" podCreationTimestamp="2025-11-24 13:30:24 +0000 UTC" firstStartedPulling="2025-11-24 13:30:25.719499993 +0000 UTC m=+1074.099393655" lastFinishedPulling="2025-11-24 13:30:56.866291617 +0000 UTC m=+1105.246185289" observedRunningTime="2025-11-24 13:30:57.830824309 +0000 UTC m=+1106.210717971" watchObservedRunningTime="2025-11-24 13:30:57.835578811 +0000 UTC m=+1106.215472473" Nov 24 13:30:58 crc kubenswrapper[4790]: I1124 13:30:58.823169 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d656998f4-lfvdf" event={"ID":"69644679-f0fb-4c09-abec-8e4084140f05","Type":"ContainerStarted","Data":"4ef132795f2d95eb5f1f86d97ec3fcd4d77a03dde2e8fce801a0ee0edc358f66"} Nov 24 13:30:58 crc kubenswrapper[4790]: I1124 13:30:58.823595 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-d656998f4-lfvdf" Nov 24 13:30:58 crc kubenswrapper[4790]: I1124 13:30:58.850476 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-d656998f4-lfvdf" podStartSLOduration=4.018209222 podStartE2EDuration="34.850455549s" podCreationTimestamp="2025-11-24 13:30:24 +0000 UTC" firstStartedPulling="2025-11-24 13:30:26.897325215 +0000 UTC m=+1075.277218877" lastFinishedPulling="2025-11-24 13:30:57.729571542 +0000 UTC m=+1106.109465204" observedRunningTime="2025-11-24 13:30:58.842290543 +0000 UTC m=+1107.222184245" watchObservedRunningTime="2025-11-24 13:30:58.850455549 +0000 UTC m=+1107.230349221" Nov 24 13:30:59 crc kubenswrapper[4790]: I1124 13:30:59.833325 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58f887965d-qvbnj" event={"ID":"3dc6a5ba-d447-40fc-bd8b-640e4b49990b","Type":"ContainerStarted","Data":"35154761b62aa38c7caa02d5d6d357ea993f13aefcf19fa514e417877c645171"} Nov 24 13:30:59 crc kubenswrapper[4790]: I1124 13:30:59.834073 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-58f887965d-qvbnj" Nov 24 13:30:59 crc kubenswrapper[4790]: I1124 13:30:59.856002 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-58f887965d-qvbnj" podStartSLOduration=3.50127921 podStartE2EDuration="35.855975627s" podCreationTimestamp="2025-11-24 13:30:24 +0000 UTC" firstStartedPulling="2025-11-24 13:30:26.5375136 +0000 UTC m=+1074.917407262" lastFinishedPulling="2025-11-24 13:30:58.892210017 +0000 UTC m=+1107.272103679" observedRunningTime="2025-11-24 13:30:59.852046338 +0000 UTC m=+1108.231940020" watchObservedRunningTime="2025-11-24 13:30:59.855975627 +0000 UTC m=+1108.235869289" Nov 24 13:31:04 crc kubenswrapper[4790]: I1124 13:31:04.536223 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-9wl9p" Nov 24 13:31:04 crc kubenswrapper[4790]: I1124 13:31:04.693507 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-58f887965d-qvbnj" Nov 24 13:31:05 crc kubenswrapper[4790]: I1124 13:31:05.342136 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-d656998f4-lfvdf" Nov 24 13:31:12 crc kubenswrapper[4790]: I1124 13:31:12.953948 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-cddlm" event={"ID":"e85f1bfb-0153-40e0-b54b-728171cb39f0","Type":"ContainerStarted","Data":"f9bb0e46eebf3de8e540789897a856b76e64b2c6d54ffb8caf82625b33f586c2"} Nov 24 13:31:12 crc kubenswrapper[4790]: I1124 13:31:12.955127 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-cddlm" Nov 24 13:31:25 crc kubenswrapper[4790]: I1124 13:31:25.666721 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-cddlm" Nov 24 13:31:25 crc kubenswrapper[4790]: I1124 13:31:25.686645 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-cddlm" podStartSLOduration=15.972316048 podStartE2EDuration="1m1.686626773s" podCreationTimestamp="2025-11-24 13:30:24 +0000 UTC" firstStartedPulling="2025-11-24 13:30:26.892611378 +0000 UTC m=+1075.272505040" lastFinishedPulling="2025-11-24 13:31:12.606922103 +0000 UTC m=+1120.986815765" observedRunningTime="2025-11-24 13:31:12.970511054 +0000 UTC m=+1121.350404716" watchObservedRunningTime="2025-11-24 13:31:25.686626773 +0000 UTC m=+1134.066520435" Nov 24 13:31:42 crc kubenswrapper[4790]: I1124 13:31:42.178922 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-78kb8"] Nov 24 13:31:42 crc kubenswrapper[4790]: I1124 13:31:42.181302 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-78kb8" Nov 24 13:31:42 crc kubenswrapper[4790]: I1124 13:31:42.185484 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 24 13:31:42 crc kubenswrapper[4790]: I1124 13:31:42.185506 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 24 13:31:42 crc kubenswrapper[4790]: I1124 13:31:42.185528 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-5wkx8" Nov 24 13:31:42 crc kubenswrapper[4790]: I1124 13:31:42.185658 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 24 13:31:42 crc kubenswrapper[4790]: I1124 13:31:42.191123 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-78kb8"] Nov 24 13:31:42 crc kubenswrapper[4790]: I1124 13:31:42.267291 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-z68d4"] Nov 24 13:31:42 crc kubenswrapper[4790]: I1124 13:31:42.268730 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-z68d4" Nov 24 13:31:42 crc kubenswrapper[4790]: I1124 13:31:42.275875 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 24 13:31:42 crc kubenswrapper[4790]: I1124 13:31:42.281396 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-z68d4"] Nov 24 13:31:42 crc kubenswrapper[4790]: I1124 13:31:42.292642 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d2804d3-2309-417d-ae57-d48815681344-config\") pod \"dnsmasq-dns-675f4bcbfc-78kb8\" (UID: \"8d2804d3-2309-417d-ae57-d48815681344\") " pod="openstack/dnsmasq-dns-675f4bcbfc-78kb8" Nov 24 13:31:42 crc kubenswrapper[4790]: I1124 13:31:42.292710 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvfb2\" (UniqueName: \"kubernetes.io/projected/8d2804d3-2309-417d-ae57-d48815681344-kube-api-access-nvfb2\") pod \"dnsmasq-dns-675f4bcbfc-78kb8\" (UID: \"8d2804d3-2309-417d-ae57-d48815681344\") " pod="openstack/dnsmasq-dns-675f4bcbfc-78kb8" Nov 24 13:31:42 crc kubenswrapper[4790]: I1124 13:31:42.393815 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvfb2\" (UniqueName: \"kubernetes.io/projected/8d2804d3-2309-417d-ae57-d48815681344-kube-api-access-nvfb2\") pod \"dnsmasq-dns-675f4bcbfc-78kb8\" (UID: \"8d2804d3-2309-417d-ae57-d48815681344\") " pod="openstack/dnsmasq-dns-675f4bcbfc-78kb8" Nov 24 13:31:42 crc kubenswrapper[4790]: I1124 13:31:42.393954 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd9d10cd-1e7e-4032-b88e-65ff25726653-config\") pod \"dnsmasq-dns-78dd6ddcc-z68d4\" (UID: \"cd9d10cd-1e7e-4032-b88e-65ff25726653\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z68d4" Nov 24 13:31:42 crc kubenswrapper[4790]: I1124 13:31:42.394005 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d2804d3-2309-417d-ae57-d48815681344-config\") pod \"dnsmasq-dns-675f4bcbfc-78kb8\" (UID: \"8d2804d3-2309-417d-ae57-d48815681344\") " pod="openstack/dnsmasq-dns-675f4bcbfc-78kb8" Nov 24 13:31:42 crc kubenswrapper[4790]: I1124 13:31:42.394039 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd9d10cd-1e7e-4032-b88e-65ff25726653-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-z68d4\" (UID: \"cd9d10cd-1e7e-4032-b88e-65ff25726653\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z68d4" Nov 24 13:31:42 crc kubenswrapper[4790]: I1124 13:31:42.394092 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jsn2\" (UniqueName: \"kubernetes.io/projected/cd9d10cd-1e7e-4032-b88e-65ff25726653-kube-api-access-8jsn2\") pod \"dnsmasq-dns-78dd6ddcc-z68d4\" (UID: \"cd9d10cd-1e7e-4032-b88e-65ff25726653\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z68d4" Nov 24 13:31:42 crc kubenswrapper[4790]: I1124 13:31:42.395284 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d2804d3-2309-417d-ae57-d48815681344-config\") pod \"dnsmasq-dns-675f4bcbfc-78kb8\" (UID: \"8d2804d3-2309-417d-ae57-d48815681344\") " pod="openstack/dnsmasq-dns-675f4bcbfc-78kb8" Nov 24 13:31:42 crc kubenswrapper[4790]: I1124 13:31:42.422222 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvfb2\" (UniqueName: \"kubernetes.io/projected/8d2804d3-2309-417d-ae57-d48815681344-kube-api-access-nvfb2\") pod \"dnsmasq-dns-675f4bcbfc-78kb8\" (UID: \"8d2804d3-2309-417d-ae57-d48815681344\") " pod="openstack/dnsmasq-dns-675f4bcbfc-78kb8" Nov 24 13:31:42 crc kubenswrapper[4790]: I1124 13:31:42.495413 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd9d10cd-1e7e-4032-b88e-65ff25726653-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-z68d4\" (UID: \"cd9d10cd-1e7e-4032-b88e-65ff25726653\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z68d4" Nov 24 13:31:42 crc kubenswrapper[4790]: I1124 13:31:42.495479 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jsn2\" (UniqueName: \"kubernetes.io/projected/cd9d10cd-1e7e-4032-b88e-65ff25726653-kube-api-access-8jsn2\") pod \"dnsmasq-dns-78dd6ddcc-z68d4\" (UID: \"cd9d10cd-1e7e-4032-b88e-65ff25726653\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z68d4" Nov 24 13:31:42 crc kubenswrapper[4790]: I1124 13:31:42.495588 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd9d10cd-1e7e-4032-b88e-65ff25726653-config\") pod \"dnsmasq-dns-78dd6ddcc-z68d4\" (UID: \"cd9d10cd-1e7e-4032-b88e-65ff25726653\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z68d4" Nov 24 13:31:42 crc kubenswrapper[4790]: I1124 13:31:42.496400 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd9d10cd-1e7e-4032-b88e-65ff25726653-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-z68d4\" (UID: \"cd9d10cd-1e7e-4032-b88e-65ff25726653\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z68d4" Nov 24 13:31:42 crc kubenswrapper[4790]: I1124 13:31:42.496812 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd9d10cd-1e7e-4032-b88e-65ff25726653-config\") pod \"dnsmasq-dns-78dd6ddcc-z68d4\" (UID: \"cd9d10cd-1e7e-4032-b88e-65ff25726653\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z68d4" Nov 24 13:31:42 crc kubenswrapper[4790]: I1124 13:31:42.512692 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jsn2\" (UniqueName: \"kubernetes.io/projected/cd9d10cd-1e7e-4032-b88e-65ff25726653-kube-api-access-8jsn2\") pod \"dnsmasq-dns-78dd6ddcc-z68d4\" (UID: \"cd9d10cd-1e7e-4032-b88e-65ff25726653\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z68d4" Nov 24 13:31:42 crc kubenswrapper[4790]: I1124 13:31:42.518334 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-78kb8" Nov 24 13:31:42 crc kubenswrapper[4790]: I1124 13:31:42.593260 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-z68d4" Nov 24 13:31:42 crc kubenswrapper[4790]: I1124 13:31:42.949301 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-78kb8"] Nov 24 13:31:43 crc kubenswrapper[4790]: I1124 13:31:43.023018 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-z68d4"] Nov 24 13:31:43 crc kubenswrapper[4790]: W1124 13:31:43.025139 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd9d10cd_1e7e_4032_b88e_65ff25726653.slice/crio-952e74f44e641a8c993b420668997543fa9d5752a1f9827864866574222114f1 WatchSource:0}: Error finding container 952e74f44e641a8c993b420668997543fa9d5752a1f9827864866574222114f1: Status 404 returned error can't find the container with id 952e74f44e641a8c993b420668997543fa9d5752a1f9827864866574222114f1 Nov 24 13:31:43 crc kubenswrapper[4790]: I1124 13:31:43.195412 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-z68d4" event={"ID":"cd9d10cd-1e7e-4032-b88e-65ff25726653","Type":"ContainerStarted","Data":"952e74f44e641a8c993b420668997543fa9d5752a1f9827864866574222114f1"} Nov 24 13:31:43 crc kubenswrapper[4790]: I1124 13:31:43.198975 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-78kb8" event={"ID":"8d2804d3-2309-417d-ae57-d48815681344","Type":"ContainerStarted","Data":"30299be8f96ecb5f137ad702cf37a873181b7af9cee609dd0d136310baf22d47"} Nov 24 13:31:44 crc kubenswrapper[4790]: I1124 13:31:44.417728 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-78kb8"] Nov 24 13:31:44 crc kubenswrapper[4790]: I1124 13:31:44.454998 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-q9w8p"] Nov 24 13:31:44 crc kubenswrapper[4790]: I1124 13:31:44.457222 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-q9w8p" Nov 24 13:31:44 crc kubenswrapper[4790]: I1124 13:31:44.484785 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-q9w8p"] Nov 24 13:31:44 crc kubenswrapper[4790]: I1124 13:31:44.524471 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7fcc2c7-c415-4bd7-9b8a-367d9292985d-config\") pod \"dnsmasq-dns-666b6646f7-q9w8p\" (UID: \"d7fcc2c7-c415-4bd7-9b8a-367d9292985d\") " pod="openstack/dnsmasq-dns-666b6646f7-q9w8p" Nov 24 13:31:44 crc kubenswrapper[4790]: I1124 13:31:44.524562 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlm72\" (UniqueName: \"kubernetes.io/projected/d7fcc2c7-c415-4bd7-9b8a-367d9292985d-kube-api-access-nlm72\") pod \"dnsmasq-dns-666b6646f7-q9w8p\" (UID: \"d7fcc2c7-c415-4bd7-9b8a-367d9292985d\") " pod="openstack/dnsmasq-dns-666b6646f7-q9w8p" Nov 24 13:31:44 crc kubenswrapper[4790]: I1124 13:31:44.524896 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7fcc2c7-c415-4bd7-9b8a-367d9292985d-dns-svc\") pod \"dnsmasq-dns-666b6646f7-q9w8p\" (UID: \"d7fcc2c7-c415-4bd7-9b8a-367d9292985d\") " pod="openstack/dnsmasq-dns-666b6646f7-q9w8p" Nov 24 13:31:44 crc kubenswrapper[4790]: I1124 13:31:44.626797 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7fcc2c7-c415-4bd7-9b8a-367d9292985d-dns-svc\") pod \"dnsmasq-dns-666b6646f7-q9w8p\" (UID: \"d7fcc2c7-c415-4bd7-9b8a-367d9292985d\") " pod="openstack/dnsmasq-dns-666b6646f7-q9w8p" Nov 24 13:31:44 crc kubenswrapper[4790]: I1124 13:31:44.626909 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7fcc2c7-c415-4bd7-9b8a-367d9292985d-config\") pod \"dnsmasq-dns-666b6646f7-q9w8p\" (UID: \"d7fcc2c7-c415-4bd7-9b8a-367d9292985d\") " pod="openstack/dnsmasq-dns-666b6646f7-q9w8p" Nov 24 13:31:44 crc kubenswrapper[4790]: I1124 13:31:44.626958 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlm72\" (UniqueName: \"kubernetes.io/projected/d7fcc2c7-c415-4bd7-9b8a-367d9292985d-kube-api-access-nlm72\") pod \"dnsmasq-dns-666b6646f7-q9w8p\" (UID: \"d7fcc2c7-c415-4bd7-9b8a-367d9292985d\") " pod="openstack/dnsmasq-dns-666b6646f7-q9w8p" Nov 24 13:31:44 crc kubenswrapper[4790]: I1124 13:31:44.628462 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7fcc2c7-c415-4bd7-9b8a-367d9292985d-config\") pod \"dnsmasq-dns-666b6646f7-q9w8p\" (UID: \"d7fcc2c7-c415-4bd7-9b8a-367d9292985d\") " pod="openstack/dnsmasq-dns-666b6646f7-q9w8p" Nov 24 13:31:44 crc kubenswrapper[4790]: I1124 13:31:44.628646 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7fcc2c7-c415-4bd7-9b8a-367d9292985d-dns-svc\") pod \"dnsmasq-dns-666b6646f7-q9w8p\" (UID: \"d7fcc2c7-c415-4bd7-9b8a-367d9292985d\") " pod="openstack/dnsmasq-dns-666b6646f7-q9w8p" Nov 24 13:31:44 crc kubenswrapper[4790]: I1124 13:31:44.653950 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlm72\" (UniqueName: \"kubernetes.io/projected/d7fcc2c7-c415-4bd7-9b8a-367d9292985d-kube-api-access-nlm72\") pod \"dnsmasq-dns-666b6646f7-q9w8p\" (UID: \"d7fcc2c7-c415-4bd7-9b8a-367d9292985d\") " pod="openstack/dnsmasq-dns-666b6646f7-q9w8p" Nov 24 13:31:44 crc kubenswrapper[4790]: I1124 13:31:44.784916 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-q9w8p" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.126136 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-z68d4"] Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.164630 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-q8jn7"] Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.172384 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-q8jn7" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.181034 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-q8jn7"] Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.243805 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8trg\" (UniqueName: \"kubernetes.io/projected/dc896e16-9261-443a-86d7-c45c2fbef3d8-kube-api-access-k8trg\") pod \"dnsmasq-dns-57d769cc4f-q8jn7\" (UID: \"dc896e16-9261-443a-86d7-c45c2fbef3d8\") " pod="openstack/dnsmasq-dns-57d769cc4f-q8jn7" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.243959 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc896e16-9261-443a-86d7-c45c2fbef3d8-config\") pod \"dnsmasq-dns-57d769cc4f-q8jn7\" (UID: \"dc896e16-9261-443a-86d7-c45c2fbef3d8\") " pod="openstack/dnsmasq-dns-57d769cc4f-q8jn7" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.244005 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc896e16-9261-443a-86d7-c45c2fbef3d8-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-q8jn7\" (UID: \"dc896e16-9261-443a-86d7-c45c2fbef3d8\") " pod="openstack/dnsmasq-dns-57d769cc4f-q8jn7" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.345976 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc896e16-9261-443a-86d7-c45c2fbef3d8-config\") pod \"dnsmasq-dns-57d769cc4f-q8jn7\" (UID: \"dc896e16-9261-443a-86d7-c45c2fbef3d8\") " pod="openstack/dnsmasq-dns-57d769cc4f-q8jn7" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.346101 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc896e16-9261-443a-86d7-c45c2fbef3d8-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-q8jn7\" (UID: \"dc896e16-9261-443a-86d7-c45c2fbef3d8\") " pod="openstack/dnsmasq-dns-57d769cc4f-q8jn7" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.346215 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8trg\" (UniqueName: \"kubernetes.io/projected/dc896e16-9261-443a-86d7-c45c2fbef3d8-kube-api-access-k8trg\") pod \"dnsmasq-dns-57d769cc4f-q8jn7\" (UID: \"dc896e16-9261-443a-86d7-c45c2fbef3d8\") " pod="openstack/dnsmasq-dns-57d769cc4f-q8jn7" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.348229 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc896e16-9261-443a-86d7-c45c2fbef3d8-config\") pod \"dnsmasq-dns-57d769cc4f-q8jn7\" (UID: \"dc896e16-9261-443a-86d7-c45c2fbef3d8\") " pod="openstack/dnsmasq-dns-57d769cc4f-q8jn7" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.348399 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc896e16-9261-443a-86d7-c45c2fbef3d8-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-q8jn7\" (UID: \"dc896e16-9261-443a-86d7-c45c2fbef3d8\") " pod="openstack/dnsmasq-dns-57d769cc4f-q8jn7" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.348511 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-q9w8p"] Nov 24 13:31:45 crc kubenswrapper[4790]: W1124 13:31:45.368560 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7fcc2c7_c415_4bd7_9b8a_367d9292985d.slice/crio-ed718cdd9c23dc552e7effeffc08c94bc60b4cb77793724f157797f859b4a049 WatchSource:0}: Error finding container ed718cdd9c23dc552e7effeffc08c94bc60b4cb77793724f157797f859b4a049: Status 404 returned error can't find the container with id ed718cdd9c23dc552e7effeffc08c94bc60b4cb77793724f157797f859b4a049 Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.382716 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8trg\" (UniqueName: \"kubernetes.io/projected/dc896e16-9261-443a-86d7-c45c2fbef3d8-kube-api-access-k8trg\") pod \"dnsmasq-dns-57d769cc4f-q8jn7\" (UID: \"dc896e16-9261-443a-86d7-c45c2fbef3d8\") " pod="openstack/dnsmasq-dns-57d769cc4f-q8jn7" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.498632 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-q8jn7" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.586096 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.587446 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.590305 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.590755 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.590969 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.591116 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.590965 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.591347 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-4qhgl" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.591095 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.607085 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.651541 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/39508987-8655-46ed-861b-e1448652ddf6-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.651591 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/39508987-8655-46ed-861b-e1448652ddf6-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.651642 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/39508987-8655-46ed-861b-e1448652ddf6-config-data\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.651659 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/39508987-8655-46ed-861b-e1448652ddf6-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.651679 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/39508987-8655-46ed-861b-e1448652ddf6-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.651699 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbsrv\" (UniqueName: \"kubernetes.io/projected/39508987-8655-46ed-861b-e1448652ddf6-kube-api-access-vbsrv\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.652384 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/39508987-8655-46ed-861b-e1448652ddf6-pod-info\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.652424 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/39508987-8655-46ed-861b-e1448652ddf6-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.652446 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/39508987-8655-46ed-861b-e1448652ddf6-server-conf\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.652465 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/39508987-8655-46ed-861b-e1448652ddf6-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.652491 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.755204 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/39508987-8655-46ed-861b-e1448652ddf6-config-data\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.755259 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/39508987-8655-46ed-861b-e1448652ddf6-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.755278 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/39508987-8655-46ed-861b-e1448652ddf6-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.755300 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbsrv\" (UniqueName: \"kubernetes.io/projected/39508987-8655-46ed-861b-e1448652ddf6-kube-api-access-vbsrv\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.755328 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/39508987-8655-46ed-861b-e1448652ddf6-pod-info\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.755360 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/39508987-8655-46ed-861b-e1448652ddf6-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.755386 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/39508987-8655-46ed-861b-e1448652ddf6-server-conf\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.755408 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/39508987-8655-46ed-861b-e1448652ddf6-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.755427 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.755469 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/39508987-8655-46ed-861b-e1448652ddf6-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.755498 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/39508987-8655-46ed-861b-e1448652ddf6-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.756670 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/39508987-8655-46ed-861b-e1448652ddf6-config-data\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.757665 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/39508987-8655-46ed-861b-e1448652ddf6-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.758520 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/39508987-8655-46ed-861b-e1448652ddf6-server-conf\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.758905 4790 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.759027 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/39508987-8655-46ed-861b-e1448652ddf6-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.759264 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/39508987-8655-46ed-861b-e1448652ddf6-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.762451 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/39508987-8655-46ed-861b-e1448652ddf6-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.763870 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/39508987-8655-46ed-861b-e1448652ddf6-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.765542 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/39508987-8655-46ed-861b-e1448652ddf6-pod-info\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.767561 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/39508987-8655-46ed-861b-e1448652ddf6-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.772410 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbsrv\" (UniqueName: \"kubernetes.io/projected/39508987-8655-46ed-861b-e1448652ddf6-kube-api-access-vbsrv\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.846391 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:45 crc kubenswrapper[4790]: I1124 13:31:45.928761 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.015832 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-q8jn7"] Nov 24 13:31:46 crc kubenswrapper[4790]: W1124 13:31:46.035097 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc896e16_9261_443a_86d7_c45c2fbef3d8.slice/crio-f22da277f6bb1ae7c6aaa1c1bbd36bd0193310ac36911af86462d2947ac445e5 WatchSource:0}: Error finding container f22da277f6bb1ae7c6aaa1c1bbd36bd0193310ac36911af86462d2947ac445e5: Status 404 returned error can't find the container with id f22da277f6bb1ae7c6aaa1c1bbd36bd0193310ac36911af86462d2947ac445e5 Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.258414 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-q8jn7" event={"ID":"dc896e16-9261-443a-86d7-c45c2fbef3d8","Type":"ContainerStarted","Data":"f22da277f6bb1ae7c6aaa1c1bbd36bd0193310ac36911af86462d2947ac445e5"} Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.260693 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-q9w8p" event={"ID":"d7fcc2c7-c415-4bd7-9b8a-367d9292985d","Type":"ContainerStarted","Data":"ed718cdd9c23dc552e7effeffc08c94bc60b4cb77793724f157797f859b4a049"} Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.300641 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.302252 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.305661 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.306242 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.313197 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.313206 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.313387 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.313670 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.313738 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-69sqx" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.333424 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.363501 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.363565 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/85e04b92-ae17-4eb6-a736-395c5d44f563-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.363585 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/85e04b92-ae17-4eb6-a736-395c5d44f563-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.363622 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/85e04b92-ae17-4eb6-a736-395c5d44f563-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.363655 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/85e04b92-ae17-4eb6-a736-395c5d44f563-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.363673 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zxlr\" (UniqueName: \"kubernetes.io/projected/85e04b92-ae17-4eb6-a736-395c5d44f563-kube-api-access-7zxlr\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.363726 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/85e04b92-ae17-4eb6-a736-395c5d44f563-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.363750 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/85e04b92-ae17-4eb6-a736-395c5d44f563-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.363772 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/85e04b92-ae17-4eb6-a736-395c5d44f563-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.363808 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/85e04b92-ae17-4eb6-a736-395c5d44f563-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.363831 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/85e04b92-ae17-4eb6-a736-395c5d44f563-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.401738 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 13:31:46 crc kubenswrapper[4790]: W1124 13:31:46.420859 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod39508987_8655_46ed_861b_e1448652ddf6.slice/crio-51ec4a5ea0e39374fa24505f9849fedadea6ce01615a1254dbd15f98bbefcab0 WatchSource:0}: Error finding container 51ec4a5ea0e39374fa24505f9849fedadea6ce01615a1254dbd15f98bbefcab0: Status 404 returned error can't find the container with id 51ec4a5ea0e39374fa24505f9849fedadea6ce01615a1254dbd15f98bbefcab0 Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.465058 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/85e04b92-ae17-4eb6-a736-395c5d44f563-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.465118 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/85e04b92-ae17-4eb6-a736-395c5d44f563-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.465143 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zxlr\" (UniqueName: \"kubernetes.io/projected/85e04b92-ae17-4eb6-a736-395c5d44f563-kube-api-access-7zxlr\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.465181 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/85e04b92-ae17-4eb6-a736-395c5d44f563-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.465229 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/85e04b92-ae17-4eb6-a736-395c5d44f563-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.465247 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/85e04b92-ae17-4eb6-a736-395c5d44f563-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.465277 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/85e04b92-ae17-4eb6-a736-395c5d44f563-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.465302 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/85e04b92-ae17-4eb6-a736-395c5d44f563-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.465395 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.465424 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/85e04b92-ae17-4eb6-a736-395c5d44f563-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.465450 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/85e04b92-ae17-4eb6-a736-395c5d44f563-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.466295 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/85e04b92-ae17-4eb6-a736-395c5d44f563-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.466441 4790 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.466561 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/85e04b92-ae17-4eb6-a736-395c5d44f563-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.466802 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/85e04b92-ae17-4eb6-a736-395c5d44f563-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.466907 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/85e04b92-ae17-4eb6-a736-395c5d44f563-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.466929 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/85e04b92-ae17-4eb6-a736-395c5d44f563-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.472237 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/85e04b92-ae17-4eb6-a736-395c5d44f563-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.472488 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/85e04b92-ae17-4eb6-a736-395c5d44f563-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.472586 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/85e04b92-ae17-4eb6-a736-395c5d44f563-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.483711 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/85e04b92-ae17-4eb6-a736-395c5d44f563-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.487504 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zxlr\" (UniqueName: \"kubernetes.io/projected/85e04b92-ae17-4eb6-a736-395c5d44f563-kube-api-access-7zxlr\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.490653 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:46 crc kubenswrapper[4790]: I1124 13:31:46.632192 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.146238 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 13:31:47 crc kubenswrapper[4790]: W1124 13:31:47.216622 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod85e04b92_ae17_4eb6_a736_395c5d44f563.slice/crio-bc9f2037bb2695552eaa4fead244895c9466a76a2ae724a8942e06cc71d82ae3 WatchSource:0}: Error finding container bc9f2037bb2695552eaa4fead244895c9466a76a2ae724a8942e06cc71d82ae3: Status 404 returned error can't find the container with id bc9f2037bb2695552eaa4fead244895c9466a76a2ae724a8942e06cc71d82ae3 Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.271366 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"85e04b92-ae17-4eb6-a736-395c5d44f563","Type":"ContainerStarted","Data":"bc9f2037bb2695552eaa4fead244895c9466a76a2ae724a8942e06cc71d82ae3"} Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.280393 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"39508987-8655-46ed-861b-e1448652ddf6","Type":"ContainerStarted","Data":"51ec4a5ea0e39374fa24505f9849fedadea6ce01615a1254dbd15f98bbefcab0"} Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.621267 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.622825 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.628954 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.629014 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.629194 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.629425 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-27xc2" Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.637952 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.643560 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.682791 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a434802-b18e-4d14-8415-4879eaa44191-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"5a434802-b18e-4d14-8415-4879eaa44191\") " pod="openstack/openstack-galera-0" Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.682835 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5a434802-b18e-4d14-8415-4879eaa44191-config-data-generated\") pod \"openstack-galera-0\" (UID: \"5a434802-b18e-4d14-8415-4879eaa44191\") " pod="openstack/openstack-galera-0" Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.682854 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a434802-b18e-4d14-8415-4879eaa44191-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"5a434802-b18e-4d14-8415-4879eaa44191\") " pod="openstack/openstack-galera-0" Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.682900 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llwqf\" (UniqueName: \"kubernetes.io/projected/5a434802-b18e-4d14-8415-4879eaa44191-kube-api-access-llwqf\") pod \"openstack-galera-0\" (UID: \"5a434802-b18e-4d14-8415-4879eaa44191\") " pod="openstack/openstack-galera-0" Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.682931 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a434802-b18e-4d14-8415-4879eaa44191-operator-scripts\") pod \"openstack-galera-0\" (UID: \"5a434802-b18e-4d14-8415-4879eaa44191\") " pod="openstack/openstack-galera-0" Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.682954 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"5a434802-b18e-4d14-8415-4879eaa44191\") " pod="openstack/openstack-galera-0" Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.683008 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5a434802-b18e-4d14-8415-4879eaa44191-config-data-default\") pod \"openstack-galera-0\" (UID: \"5a434802-b18e-4d14-8415-4879eaa44191\") " pod="openstack/openstack-galera-0" Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.683030 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5a434802-b18e-4d14-8415-4879eaa44191-kolla-config\") pod \"openstack-galera-0\" (UID: \"5a434802-b18e-4d14-8415-4879eaa44191\") " pod="openstack/openstack-galera-0" Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.784316 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a434802-b18e-4d14-8415-4879eaa44191-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"5a434802-b18e-4d14-8415-4879eaa44191\") " pod="openstack/openstack-galera-0" Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.784368 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5a434802-b18e-4d14-8415-4879eaa44191-config-data-generated\") pod \"openstack-galera-0\" (UID: \"5a434802-b18e-4d14-8415-4879eaa44191\") " pod="openstack/openstack-galera-0" Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.784390 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a434802-b18e-4d14-8415-4879eaa44191-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"5a434802-b18e-4d14-8415-4879eaa44191\") " pod="openstack/openstack-galera-0" Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.784415 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llwqf\" (UniqueName: \"kubernetes.io/projected/5a434802-b18e-4d14-8415-4879eaa44191-kube-api-access-llwqf\") pod \"openstack-galera-0\" (UID: \"5a434802-b18e-4d14-8415-4879eaa44191\") " pod="openstack/openstack-galera-0" Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.784453 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a434802-b18e-4d14-8415-4879eaa44191-operator-scripts\") pod \"openstack-galera-0\" (UID: \"5a434802-b18e-4d14-8415-4879eaa44191\") " pod="openstack/openstack-galera-0" Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.784486 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"5a434802-b18e-4d14-8415-4879eaa44191\") " pod="openstack/openstack-galera-0" Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.784562 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5a434802-b18e-4d14-8415-4879eaa44191-config-data-default\") pod \"openstack-galera-0\" (UID: \"5a434802-b18e-4d14-8415-4879eaa44191\") " pod="openstack/openstack-galera-0" Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.784593 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5a434802-b18e-4d14-8415-4879eaa44191-kolla-config\") pod \"openstack-galera-0\" (UID: \"5a434802-b18e-4d14-8415-4879eaa44191\") " pod="openstack/openstack-galera-0" Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.785458 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5a434802-b18e-4d14-8415-4879eaa44191-kolla-config\") pod \"openstack-galera-0\" (UID: \"5a434802-b18e-4d14-8415-4879eaa44191\") " pod="openstack/openstack-galera-0" Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.790360 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a434802-b18e-4d14-8415-4879eaa44191-operator-scripts\") pod \"openstack-galera-0\" (UID: \"5a434802-b18e-4d14-8415-4879eaa44191\") " pod="openstack/openstack-galera-0" Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.792090 4790 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"5a434802-b18e-4d14-8415-4879eaa44191\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/openstack-galera-0" Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.793748 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5a434802-b18e-4d14-8415-4879eaa44191-config-data-default\") pod \"openstack-galera-0\" (UID: \"5a434802-b18e-4d14-8415-4879eaa44191\") " pod="openstack/openstack-galera-0" Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.796562 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a434802-b18e-4d14-8415-4879eaa44191-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"5a434802-b18e-4d14-8415-4879eaa44191\") " pod="openstack/openstack-galera-0" Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.797039 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5a434802-b18e-4d14-8415-4879eaa44191-config-data-generated\") pod \"openstack-galera-0\" (UID: \"5a434802-b18e-4d14-8415-4879eaa44191\") " pod="openstack/openstack-galera-0" Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.806022 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a434802-b18e-4d14-8415-4879eaa44191-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"5a434802-b18e-4d14-8415-4879eaa44191\") " pod="openstack/openstack-galera-0" Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.809272 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llwqf\" (UniqueName: \"kubernetes.io/projected/5a434802-b18e-4d14-8415-4879eaa44191-kube-api-access-llwqf\") pod \"openstack-galera-0\" (UID: \"5a434802-b18e-4d14-8415-4879eaa44191\") " pod="openstack/openstack-galera-0" Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.834579 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"5a434802-b18e-4d14-8415-4879eaa44191\") " pod="openstack/openstack-galera-0" Nov 24 13:31:47 crc kubenswrapper[4790]: I1124 13:31:47.955253 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 13:31:48 crc kubenswrapper[4790]: I1124 13:31:48.654439 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 13:31:48 crc kubenswrapper[4790]: I1124 13:31:48.952815 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 13:31:48 crc kubenswrapper[4790]: I1124 13:31:48.967933 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 13:31:48 crc kubenswrapper[4790]: I1124 13:31:48.968070 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 13:31:48 crc kubenswrapper[4790]: I1124 13:31:48.972515 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-4f54z" Nov 24 13:31:48 crc kubenswrapper[4790]: I1124 13:31:48.974324 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 24 13:31:48 crc kubenswrapper[4790]: I1124 13:31:48.974511 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 24 13:31:48 crc kubenswrapper[4790]: I1124 13:31:48.974688 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.114846 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9c866267-37e8-48d7-83ca-53e3cbf8b762-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"9c866267-37e8-48d7-83ca-53e3cbf8b762\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.114960 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"9c866267-37e8-48d7-83ca-53e3cbf8b762\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.114997 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9c866267-37e8-48d7-83ca-53e3cbf8b762-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"9c866267-37e8-48d7-83ca-53e3cbf8b762\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.115021 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c866267-37e8-48d7-83ca-53e3cbf8b762-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"9c866267-37e8-48d7-83ca-53e3cbf8b762\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.115052 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c866267-37e8-48d7-83ca-53e3cbf8b762-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"9c866267-37e8-48d7-83ca-53e3cbf8b762\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.115078 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c866267-37e8-48d7-83ca-53e3cbf8b762-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"9c866267-37e8-48d7-83ca-53e3cbf8b762\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.115129 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9c866267-37e8-48d7-83ca-53e3cbf8b762-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"9c866267-37e8-48d7-83ca-53e3cbf8b762\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.115228 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kb2fg\" (UniqueName: \"kubernetes.io/projected/9c866267-37e8-48d7-83ca-53e3cbf8b762-kube-api-access-kb2fg\") pod \"openstack-cell1-galera-0\" (UID: \"9c866267-37e8-48d7-83ca-53e3cbf8b762\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.215264 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.216630 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kb2fg\" (UniqueName: \"kubernetes.io/projected/9c866267-37e8-48d7-83ca-53e3cbf8b762-kube-api-access-kb2fg\") pod \"openstack-cell1-galera-0\" (UID: \"9c866267-37e8-48d7-83ca-53e3cbf8b762\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.216726 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9c866267-37e8-48d7-83ca-53e3cbf8b762-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"9c866267-37e8-48d7-83ca-53e3cbf8b762\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.216760 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"9c866267-37e8-48d7-83ca-53e3cbf8b762\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.216785 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9c866267-37e8-48d7-83ca-53e3cbf8b762-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"9c866267-37e8-48d7-83ca-53e3cbf8b762\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.216807 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c866267-37e8-48d7-83ca-53e3cbf8b762-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"9c866267-37e8-48d7-83ca-53e3cbf8b762\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.216834 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c866267-37e8-48d7-83ca-53e3cbf8b762-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"9c866267-37e8-48d7-83ca-53e3cbf8b762\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.216865 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c866267-37e8-48d7-83ca-53e3cbf8b762-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"9c866267-37e8-48d7-83ca-53e3cbf8b762\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.217252 4790 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"9c866267-37e8-48d7-83ca-53e3cbf8b762\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/openstack-cell1-galera-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.217953 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9c866267-37e8-48d7-83ca-53e3cbf8b762-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"9c866267-37e8-48d7-83ca-53e3cbf8b762\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.218445 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9c866267-37e8-48d7-83ca-53e3cbf8b762-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"9c866267-37e8-48d7-83ca-53e3cbf8b762\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.219069 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9c866267-37e8-48d7-83ca-53e3cbf8b762-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"9c866267-37e8-48d7-83ca-53e3cbf8b762\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.216644 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.224065 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c866267-37e8-48d7-83ca-53e3cbf8b762-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"9c866267-37e8-48d7-83ca-53e3cbf8b762\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.225043 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9c866267-37e8-48d7-83ca-53e3cbf8b762-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"9c866267-37e8-48d7-83ca-53e3cbf8b762\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.228506 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c866267-37e8-48d7-83ca-53e3cbf8b762-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"9c866267-37e8-48d7-83ca-53e3cbf8b762\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.229662 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c866267-37e8-48d7-83ca-53e3cbf8b762-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"9c866267-37e8-48d7-83ca-53e3cbf8b762\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.236474 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-xzglq" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.236612 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kb2fg\" (UniqueName: \"kubernetes.io/projected/9c866267-37e8-48d7-83ca-53e3cbf8b762-kube-api-access-kb2fg\") pod \"openstack-cell1-galera-0\" (UID: \"9c866267-37e8-48d7-83ca-53e3cbf8b762\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.237931 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.238206 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.240080 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.298800 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"9c866267-37e8-48d7-83ca-53e3cbf8b762\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.306744 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5a434802-b18e-4d14-8415-4879eaa44191","Type":"ContainerStarted","Data":"0f939fb4108c2d62d2ca6653bb9bf2cfe97165eb15af2979d0b39e7b7325bd4f"} Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.308139 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.321229 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56d67de4-df9a-44fa-92b2-cbb168c4ae87-config-data\") pod \"memcached-0\" (UID: \"56d67de4-df9a-44fa-92b2-cbb168c4ae87\") " pod="openstack/memcached-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.321272 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56d67de4-df9a-44fa-92b2-cbb168c4ae87-combined-ca-bundle\") pod \"memcached-0\" (UID: \"56d67de4-df9a-44fa-92b2-cbb168c4ae87\") " pod="openstack/memcached-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.321310 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pt66d\" (UniqueName: \"kubernetes.io/projected/56d67de4-df9a-44fa-92b2-cbb168c4ae87-kube-api-access-pt66d\") pod \"memcached-0\" (UID: \"56d67de4-df9a-44fa-92b2-cbb168c4ae87\") " pod="openstack/memcached-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.321348 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/56d67de4-df9a-44fa-92b2-cbb168c4ae87-kolla-config\") pod \"memcached-0\" (UID: \"56d67de4-df9a-44fa-92b2-cbb168c4ae87\") " pod="openstack/memcached-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.321367 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/56d67de4-df9a-44fa-92b2-cbb168c4ae87-memcached-tls-certs\") pod \"memcached-0\" (UID: \"56d67de4-df9a-44fa-92b2-cbb168c4ae87\") " pod="openstack/memcached-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.426282 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56d67de4-df9a-44fa-92b2-cbb168c4ae87-config-data\") pod \"memcached-0\" (UID: \"56d67de4-df9a-44fa-92b2-cbb168c4ae87\") " pod="openstack/memcached-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.426957 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56d67de4-df9a-44fa-92b2-cbb168c4ae87-combined-ca-bundle\") pod \"memcached-0\" (UID: \"56d67de4-df9a-44fa-92b2-cbb168c4ae87\") " pod="openstack/memcached-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.427126 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pt66d\" (UniqueName: \"kubernetes.io/projected/56d67de4-df9a-44fa-92b2-cbb168c4ae87-kube-api-access-pt66d\") pod \"memcached-0\" (UID: \"56d67de4-df9a-44fa-92b2-cbb168c4ae87\") " pod="openstack/memcached-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.428136 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/56d67de4-df9a-44fa-92b2-cbb168c4ae87-kolla-config\") pod \"memcached-0\" (UID: \"56d67de4-df9a-44fa-92b2-cbb168c4ae87\") " pod="openstack/memcached-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.428183 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/56d67de4-df9a-44fa-92b2-cbb168c4ae87-memcached-tls-certs\") pod \"memcached-0\" (UID: \"56d67de4-df9a-44fa-92b2-cbb168c4ae87\") " pod="openstack/memcached-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.429068 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56d67de4-df9a-44fa-92b2-cbb168c4ae87-config-data\") pod \"memcached-0\" (UID: \"56d67de4-df9a-44fa-92b2-cbb168c4ae87\") " pod="openstack/memcached-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.429904 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/56d67de4-df9a-44fa-92b2-cbb168c4ae87-kolla-config\") pod \"memcached-0\" (UID: \"56d67de4-df9a-44fa-92b2-cbb168c4ae87\") " pod="openstack/memcached-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.435923 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/56d67de4-df9a-44fa-92b2-cbb168c4ae87-memcached-tls-certs\") pod \"memcached-0\" (UID: \"56d67de4-df9a-44fa-92b2-cbb168c4ae87\") " pod="openstack/memcached-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.437003 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56d67de4-df9a-44fa-92b2-cbb168c4ae87-combined-ca-bundle\") pod \"memcached-0\" (UID: \"56d67de4-df9a-44fa-92b2-cbb168c4ae87\") " pod="openstack/memcached-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.445806 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pt66d\" (UniqueName: \"kubernetes.io/projected/56d67de4-df9a-44fa-92b2-cbb168c4ae87-kube-api-access-pt66d\") pod \"memcached-0\" (UID: \"56d67de4-df9a-44fa-92b2-cbb168c4ae87\") " pod="openstack/memcached-0" Nov 24 13:31:49 crc kubenswrapper[4790]: I1124 13:31:49.664413 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 13:31:50 crc kubenswrapper[4790]: I1124 13:31:50.723315 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 13:31:50 crc kubenswrapper[4790]: I1124 13:31:50.724432 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 13:31:50 crc kubenswrapper[4790]: I1124 13:31:50.728058 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-lbfjz" Nov 24 13:31:50 crc kubenswrapper[4790]: I1124 13:31:50.735595 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 13:31:50 crc kubenswrapper[4790]: I1124 13:31:50.773178 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8th6r\" (UniqueName: \"kubernetes.io/projected/d9726b87-983d-46fc-8880-7e19cbe1e35d-kube-api-access-8th6r\") pod \"kube-state-metrics-0\" (UID: \"d9726b87-983d-46fc-8880-7e19cbe1e35d\") " pod="openstack/kube-state-metrics-0" Nov 24 13:31:50 crc kubenswrapper[4790]: I1124 13:31:50.874750 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8th6r\" (UniqueName: \"kubernetes.io/projected/d9726b87-983d-46fc-8880-7e19cbe1e35d-kube-api-access-8th6r\") pod \"kube-state-metrics-0\" (UID: \"d9726b87-983d-46fc-8880-7e19cbe1e35d\") " pod="openstack/kube-state-metrics-0" Nov 24 13:31:50 crc kubenswrapper[4790]: I1124 13:31:50.909386 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8th6r\" (UniqueName: \"kubernetes.io/projected/d9726b87-983d-46fc-8880-7e19cbe1e35d-kube-api-access-8th6r\") pod \"kube-state-metrics-0\" (UID: \"d9726b87-983d-46fc-8880-7e19cbe1e35d\") " pod="openstack/kube-state-metrics-0" Nov 24 13:31:51 crc kubenswrapper[4790]: I1124 13:31:51.054744 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.208922 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-rkxr6"] Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.217391 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rkxr6" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.218733 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-28cwj"] Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.220490 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.220871 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-pdrbr" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.221004 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.223195 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-28cwj" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.231702 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rkxr6"] Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.259765 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-28cwj"] Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.321701 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8fa811d-92cf-46a2-a9fd-539130369eb1-ovn-controller-tls-certs\") pod \"ovn-controller-rkxr6\" (UID: \"c8fa811d-92cf-46a2-a9fd-539130369eb1\") " pod="openstack/ovn-controller-rkxr6" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.321740 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c8fa811d-92cf-46a2-a9fd-539130369eb1-var-run-ovn\") pod \"ovn-controller-rkxr6\" (UID: \"c8fa811d-92cf-46a2-a9fd-539130369eb1\") " pod="openstack/ovn-controller-rkxr6" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.321764 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wwpr\" (UniqueName: \"kubernetes.io/projected/c8fa811d-92cf-46a2-a9fd-539130369eb1-kube-api-access-2wwpr\") pod \"ovn-controller-rkxr6\" (UID: \"c8fa811d-92cf-46a2-a9fd-539130369eb1\") " pod="openstack/ovn-controller-rkxr6" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.321783 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4d6d5d88-adda-4679-80de-0df0e0847bcb-var-run\") pod \"ovn-controller-ovs-28cwj\" (UID: \"4d6d5d88-adda-4679-80de-0df0e0847bcb\") " pod="openstack/ovn-controller-ovs-28cwj" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.321896 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/4d6d5d88-adda-4679-80de-0df0e0847bcb-var-lib\") pod \"ovn-controller-ovs-28cwj\" (UID: \"4d6d5d88-adda-4679-80de-0df0e0847bcb\") " pod="openstack/ovn-controller-ovs-28cwj" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.321965 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c8fa811d-92cf-46a2-a9fd-539130369eb1-scripts\") pod \"ovn-controller-rkxr6\" (UID: \"c8fa811d-92cf-46a2-a9fd-539130369eb1\") " pod="openstack/ovn-controller-rkxr6" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.321997 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/4d6d5d88-adda-4679-80de-0df0e0847bcb-etc-ovs\") pod \"ovn-controller-ovs-28cwj\" (UID: \"4d6d5d88-adda-4679-80de-0df0e0847bcb\") " pod="openstack/ovn-controller-ovs-28cwj" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.322035 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c8fa811d-92cf-46a2-a9fd-539130369eb1-var-log-ovn\") pod \"ovn-controller-rkxr6\" (UID: \"c8fa811d-92cf-46a2-a9fd-539130369eb1\") " pod="openstack/ovn-controller-rkxr6" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.322052 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/4d6d5d88-adda-4679-80de-0df0e0847bcb-var-log\") pod \"ovn-controller-ovs-28cwj\" (UID: \"4d6d5d88-adda-4679-80de-0df0e0847bcb\") " pod="openstack/ovn-controller-ovs-28cwj" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.322074 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2g5hd\" (UniqueName: \"kubernetes.io/projected/4d6d5d88-adda-4679-80de-0df0e0847bcb-kube-api-access-2g5hd\") pod \"ovn-controller-ovs-28cwj\" (UID: \"4d6d5d88-adda-4679-80de-0df0e0847bcb\") " pod="openstack/ovn-controller-ovs-28cwj" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.322092 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c8fa811d-92cf-46a2-a9fd-539130369eb1-var-run\") pod \"ovn-controller-rkxr6\" (UID: \"c8fa811d-92cf-46a2-a9fd-539130369eb1\") " pod="openstack/ovn-controller-rkxr6" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.322113 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4d6d5d88-adda-4679-80de-0df0e0847bcb-scripts\") pod \"ovn-controller-ovs-28cwj\" (UID: \"4d6d5d88-adda-4679-80de-0df0e0847bcb\") " pod="openstack/ovn-controller-ovs-28cwj" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.322129 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8fa811d-92cf-46a2-a9fd-539130369eb1-combined-ca-bundle\") pod \"ovn-controller-rkxr6\" (UID: \"c8fa811d-92cf-46a2-a9fd-539130369eb1\") " pod="openstack/ovn-controller-rkxr6" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.425662 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8fa811d-92cf-46a2-a9fd-539130369eb1-ovn-controller-tls-certs\") pod \"ovn-controller-rkxr6\" (UID: \"c8fa811d-92cf-46a2-a9fd-539130369eb1\") " pod="openstack/ovn-controller-rkxr6" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.425731 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c8fa811d-92cf-46a2-a9fd-539130369eb1-var-run-ovn\") pod \"ovn-controller-rkxr6\" (UID: \"c8fa811d-92cf-46a2-a9fd-539130369eb1\") " pod="openstack/ovn-controller-rkxr6" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.425769 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wwpr\" (UniqueName: \"kubernetes.io/projected/c8fa811d-92cf-46a2-a9fd-539130369eb1-kube-api-access-2wwpr\") pod \"ovn-controller-rkxr6\" (UID: \"c8fa811d-92cf-46a2-a9fd-539130369eb1\") " pod="openstack/ovn-controller-rkxr6" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.425802 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4d6d5d88-adda-4679-80de-0df0e0847bcb-var-run\") pod \"ovn-controller-ovs-28cwj\" (UID: \"4d6d5d88-adda-4679-80de-0df0e0847bcb\") " pod="openstack/ovn-controller-ovs-28cwj" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.425840 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/4d6d5d88-adda-4679-80de-0df0e0847bcb-var-lib\") pod \"ovn-controller-ovs-28cwj\" (UID: \"4d6d5d88-adda-4679-80de-0df0e0847bcb\") " pod="openstack/ovn-controller-ovs-28cwj" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.425960 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c8fa811d-92cf-46a2-a9fd-539130369eb1-scripts\") pod \"ovn-controller-rkxr6\" (UID: \"c8fa811d-92cf-46a2-a9fd-539130369eb1\") " pod="openstack/ovn-controller-rkxr6" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.427139 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4d6d5d88-adda-4679-80de-0df0e0847bcb-var-run\") pod \"ovn-controller-ovs-28cwj\" (UID: \"4d6d5d88-adda-4679-80de-0df0e0847bcb\") " pod="openstack/ovn-controller-ovs-28cwj" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.430141 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c8fa811d-92cf-46a2-a9fd-539130369eb1-var-log-ovn\") pod \"ovn-controller-rkxr6\" (UID: \"c8fa811d-92cf-46a2-a9fd-539130369eb1\") " pod="openstack/ovn-controller-rkxr6" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.430175 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/4d6d5d88-adda-4679-80de-0df0e0847bcb-etc-ovs\") pod \"ovn-controller-ovs-28cwj\" (UID: \"4d6d5d88-adda-4679-80de-0df0e0847bcb\") " pod="openstack/ovn-controller-ovs-28cwj" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.430195 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/4d6d5d88-adda-4679-80de-0df0e0847bcb-var-log\") pod \"ovn-controller-ovs-28cwj\" (UID: \"4d6d5d88-adda-4679-80de-0df0e0847bcb\") " pod="openstack/ovn-controller-ovs-28cwj" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.430249 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2g5hd\" (UniqueName: \"kubernetes.io/projected/4d6d5d88-adda-4679-80de-0df0e0847bcb-kube-api-access-2g5hd\") pod \"ovn-controller-ovs-28cwj\" (UID: \"4d6d5d88-adda-4679-80de-0df0e0847bcb\") " pod="openstack/ovn-controller-ovs-28cwj" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.430274 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c8fa811d-92cf-46a2-a9fd-539130369eb1-var-run\") pod \"ovn-controller-rkxr6\" (UID: \"c8fa811d-92cf-46a2-a9fd-539130369eb1\") " pod="openstack/ovn-controller-rkxr6" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.430327 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4d6d5d88-adda-4679-80de-0df0e0847bcb-scripts\") pod \"ovn-controller-ovs-28cwj\" (UID: \"4d6d5d88-adda-4679-80de-0df0e0847bcb\") " pod="openstack/ovn-controller-ovs-28cwj" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.430343 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8fa811d-92cf-46a2-a9fd-539130369eb1-combined-ca-bundle\") pod \"ovn-controller-rkxr6\" (UID: \"c8fa811d-92cf-46a2-a9fd-539130369eb1\") " pod="openstack/ovn-controller-rkxr6" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.430368 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c8fa811d-92cf-46a2-a9fd-539130369eb1-var-run-ovn\") pod \"ovn-controller-rkxr6\" (UID: \"c8fa811d-92cf-46a2-a9fd-539130369eb1\") " pod="openstack/ovn-controller-rkxr6" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.430368 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/4d6d5d88-adda-4679-80de-0df0e0847bcb-var-lib\") pod \"ovn-controller-ovs-28cwj\" (UID: \"4d6d5d88-adda-4679-80de-0df0e0847bcb\") " pod="openstack/ovn-controller-ovs-28cwj" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.430581 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/4d6d5d88-adda-4679-80de-0df0e0847bcb-etc-ovs\") pod \"ovn-controller-ovs-28cwj\" (UID: \"4d6d5d88-adda-4679-80de-0df0e0847bcb\") " pod="openstack/ovn-controller-ovs-28cwj" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.431181 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/4d6d5d88-adda-4679-80de-0df0e0847bcb-var-log\") pod \"ovn-controller-ovs-28cwj\" (UID: \"4d6d5d88-adda-4679-80de-0df0e0847bcb\") " pod="openstack/ovn-controller-ovs-28cwj" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.431287 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c8fa811d-92cf-46a2-a9fd-539130369eb1-var-log-ovn\") pod \"ovn-controller-rkxr6\" (UID: \"c8fa811d-92cf-46a2-a9fd-539130369eb1\") " pod="openstack/ovn-controller-rkxr6" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.431344 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c8fa811d-92cf-46a2-a9fd-539130369eb1-scripts\") pod \"ovn-controller-rkxr6\" (UID: \"c8fa811d-92cf-46a2-a9fd-539130369eb1\") " pod="openstack/ovn-controller-rkxr6" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.431412 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c8fa811d-92cf-46a2-a9fd-539130369eb1-var-run\") pod \"ovn-controller-rkxr6\" (UID: \"c8fa811d-92cf-46a2-a9fd-539130369eb1\") " pod="openstack/ovn-controller-rkxr6" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.431545 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8fa811d-92cf-46a2-a9fd-539130369eb1-ovn-controller-tls-certs\") pod \"ovn-controller-rkxr6\" (UID: \"c8fa811d-92cf-46a2-a9fd-539130369eb1\") " pod="openstack/ovn-controller-rkxr6" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.433280 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4d6d5d88-adda-4679-80de-0df0e0847bcb-scripts\") pod \"ovn-controller-ovs-28cwj\" (UID: \"4d6d5d88-adda-4679-80de-0df0e0847bcb\") " pod="openstack/ovn-controller-ovs-28cwj" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.435271 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8fa811d-92cf-46a2-a9fd-539130369eb1-combined-ca-bundle\") pod \"ovn-controller-rkxr6\" (UID: \"c8fa811d-92cf-46a2-a9fd-539130369eb1\") " pod="openstack/ovn-controller-rkxr6" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.448953 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wwpr\" (UniqueName: \"kubernetes.io/projected/c8fa811d-92cf-46a2-a9fd-539130369eb1-kube-api-access-2wwpr\") pod \"ovn-controller-rkxr6\" (UID: \"c8fa811d-92cf-46a2-a9fd-539130369eb1\") " pod="openstack/ovn-controller-rkxr6" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.471526 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2g5hd\" (UniqueName: \"kubernetes.io/projected/4d6d5d88-adda-4679-80de-0df0e0847bcb-kube-api-access-2g5hd\") pod \"ovn-controller-ovs-28cwj\" (UID: \"4d6d5d88-adda-4679-80de-0df0e0847bcb\") " pod="openstack/ovn-controller-ovs-28cwj" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.560874 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rkxr6" Nov 24 13:31:53 crc kubenswrapper[4790]: I1124 13:31:53.573123 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-28cwj" Nov 24 13:31:55 crc kubenswrapper[4790]: I1124 13:31:55.674545 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 13:31:55 crc kubenswrapper[4790]: I1124 13:31:55.676061 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 13:31:55 crc kubenswrapper[4790]: I1124 13:31:55.678822 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 24 13:31:55 crc kubenswrapper[4790]: I1124 13:31:55.678926 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 24 13:31:55 crc kubenswrapper[4790]: I1124 13:31:55.679143 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 24 13:31:55 crc kubenswrapper[4790]: I1124 13:31:55.679317 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 24 13:31:55 crc kubenswrapper[4790]: I1124 13:31:55.679593 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-kbwnx" Nov 24 13:31:55 crc kubenswrapper[4790]: I1124 13:31:55.685336 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 13:31:55 crc kubenswrapper[4790]: I1124 13:31:55.773449 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4sxg\" (UniqueName: \"kubernetes.io/projected/08baac2a-300a-498f-8d0c-d1f261e6c8fc-kube-api-access-q4sxg\") pod \"ovsdbserver-nb-0\" (UID: \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:31:55 crc kubenswrapper[4790]: I1124 13:31:55.773549 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/08baac2a-300a-498f-8d0c-d1f261e6c8fc-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:31:55 crc kubenswrapper[4790]: I1124 13:31:55.773624 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:31:55 crc kubenswrapper[4790]: I1124 13:31:55.773691 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08baac2a-300a-498f-8d0c-d1f261e6c8fc-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:31:55 crc kubenswrapper[4790]: I1124 13:31:55.773998 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/08baac2a-300a-498f-8d0c-d1f261e6c8fc-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:31:55 crc kubenswrapper[4790]: I1124 13:31:55.774036 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/08baac2a-300a-498f-8d0c-d1f261e6c8fc-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:31:55 crc kubenswrapper[4790]: I1124 13:31:55.774182 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08baac2a-300a-498f-8d0c-d1f261e6c8fc-config\") pod \"ovsdbserver-nb-0\" (UID: \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:31:55 crc kubenswrapper[4790]: I1124 13:31:55.774211 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/08baac2a-300a-498f-8d0c-d1f261e6c8fc-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:31:55 crc kubenswrapper[4790]: I1124 13:31:55.875614 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/08baac2a-300a-498f-8d0c-d1f261e6c8fc-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:31:55 crc kubenswrapper[4790]: I1124 13:31:55.875651 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/08baac2a-300a-498f-8d0c-d1f261e6c8fc-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:31:55 crc kubenswrapper[4790]: I1124 13:31:55.875681 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08baac2a-300a-498f-8d0c-d1f261e6c8fc-config\") pod \"ovsdbserver-nb-0\" (UID: \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:31:55 crc kubenswrapper[4790]: I1124 13:31:55.875699 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/08baac2a-300a-498f-8d0c-d1f261e6c8fc-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:31:55 crc kubenswrapper[4790]: I1124 13:31:55.875742 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4sxg\" (UniqueName: \"kubernetes.io/projected/08baac2a-300a-498f-8d0c-d1f261e6c8fc-kube-api-access-q4sxg\") pod \"ovsdbserver-nb-0\" (UID: \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:31:55 crc kubenswrapper[4790]: I1124 13:31:55.875763 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/08baac2a-300a-498f-8d0c-d1f261e6c8fc-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:31:55 crc kubenswrapper[4790]: I1124 13:31:55.875784 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:31:55 crc kubenswrapper[4790]: I1124 13:31:55.875798 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08baac2a-300a-498f-8d0c-d1f261e6c8fc-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:31:55 crc kubenswrapper[4790]: I1124 13:31:55.876520 4790 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ovsdbserver-nb-0" Nov 24 13:31:55 crc kubenswrapper[4790]: I1124 13:31:55.876635 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/08baac2a-300a-498f-8d0c-d1f261e6c8fc-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:31:55 crc kubenswrapper[4790]: I1124 13:31:55.877082 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/08baac2a-300a-498f-8d0c-d1f261e6c8fc-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:31:55 crc kubenswrapper[4790]: I1124 13:31:55.877253 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08baac2a-300a-498f-8d0c-d1f261e6c8fc-config\") pod \"ovsdbserver-nb-0\" (UID: \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:31:55 crc kubenswrapper[4790]: I1124 13:31:55.884295 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/08baac2a-300a-498f-8d0c-d1f261e6c8fc-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:31:55 crc kubenswrapper[4790]: I1124 13:31:55.884615 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/08baac2a-300a-498f-8d0c-d1f261e6c8fc-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:31:55 crc kubenswrapper[4790]: I1124 13:31:55.888131 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08baac2a-300a-498f-8d0c-d1f261e6c8fc-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:31:55 crc kubenswrapper[4790]: I1124 13:31:55.905265 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4sxg\" (UniqueName: \"kubernetes.io/projected/08baac2a-300a-498f-8d0c-d1f261e6c8fc-kube-api-access-q4sxg\") pod \"ovsdbserver-nb-0\" (UID: \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:31:55 crc kubenswrapper[4790]: I1124 13:31:55.910468 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:31:56 crc kubenswrapper[4790]: I1124 13:31:56.006429 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 13:31:58 crc kubenswrapper[4790]: I1124 13:31:58.376414 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 13:31:58 crc kubenswrapper[4790]: I1124 13:31:58.378554 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 13:31:58 crc kubenswrapper[4790]: I1124 13:31:58.381726 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 24 13:31:58 crc kubenswrapper[4790]: I1124 13:31:58.381987 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-7b4h9" Nov 24 13:31:58 crc kubenswrapper[4790]: I1124 13:31:58.382112 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 24 13:31:58 crc kubenswrapper[4790]: I1124 13:31:58.382871 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 24 13:31:58 crc kubenswrapper[4790]: I1124 13:31:58.403717 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 13:31:58 crc kubenswrapper[4790]: I1124 13:31:58.434265 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/23e9fd30-8462-4b44-b9ae-11e3a367c814-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"23e9fd30-8462-4b44-b9ae-11e3a367c814\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:31:58 crc kubenswrapper[4790]: I1124 13:31:58.434347 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/23e9fd30-8462-4b44-b9ae-11e3a367c814-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"23e9fd30-8462-4b44-b9ae-11e3a367c814\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:31:58 crc kubenswrapper[4790]: I1124 13:31:58.434376 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnx7x\" (UniqueName: \"kubernetes.io/projected/23e9fd30-8462-4b44-b9ae-11e3a367c814-kube-api-access-bnx7x\") pod \"ovsdbserver-sb-0\" (UID: \"23e9fd30-8462-4b44-b9ae-11e3a367c814\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:31:58 crc kubenswrapper[4790]: I1124 13:31:58.434578 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/23e9fd30-8462-4b44-b9ae-11e3a367c814-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"23e9fd30-8462-4b44-b9ae-11e3a367c814\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:31:58 crc kubenswrapper[4790]: I1124 13:31:58.434792 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/23e9fd30-8462-4b44-b9ae-11e3a367c814-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"23e9fd30-8462-4b44-b9ae-11e3a367c814\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:31:58 crc kubenswrapper[4790]: I1124 13:31:58.434848 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"23e9fd30-8462-4b44-b9ae-11e3a367c814\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:31:58 crc kubenswrapper[4790]: I1124 13:31:58.434927 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23e9fd30-8462-4b44-b9ae-11e3a367c814-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"23e9fd30-8462-4b44-b9ae-11e3a367c814\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:31:58 crc kubenswrapper[4790]: I1124 13:31:58.434967 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23e9fd30-8462-4b44-b9ae-11e3a367c814-config\") pod \"ovsdbserver-sb-0\" (UID: \"23e9fd30-8462-4b44-b9ae-11e3a367c814\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:31:58 crc kubenswrapper[4790]: I1124 13:31:58.536730 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/23e9fd30-8462-4b44-b9ae-11e3a367c814-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"23e9fd30-8462-4b44-b9ae-11e3a367c814\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:31:58 crc kubenswrapper[4790]: I1124 13:31:58.536790 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnx7x\" (UniqueName: \"kubernetes.io/projected/23e9fd30-8462-4b44-b9ae-11e3a367c814-kube-api-access-bnx7x\") pod \"ovsdbserver-sb-0\" (UID: \"23e9fd30-8462-4b44-b9ae-11e3a367c814\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:31:58 crc kubenswrapper[4790]: I1124 13:31:58.536852 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/23e9fd30-8462-4b44-b9ae-11e3a367c814-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"23e9fd30-8462-4b44-b9ae-11e3a367c814\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:31:58 crc kubenswrapper[4790]: I1124 13:31:58.536985 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/23e9fd30-8462-4b44-b9ae-11e3a367c814-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"23e9fd30-8462-4b44-b9ae-11e3a367c814\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:31:58 crc kubenswrapper[4790]: I1124 13:31:58.537042 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"23e9fd30-8462-4b44-b9ae-11e3a367c814\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:31:58 crc kubenswrapper[4790]: I1124 13:31:58.537066 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23e9fd30-8462-4b44-b9ae-11e3a367c814-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"23e9fd30-8462-4b44-b9ae-11e3a367c814\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:31:58 crc kubenswrapper[4790]: I1124 13:31:58.537102 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23e9fd30-8462-4b44-b9ae-11e3a367c814-config\") pod \"ovsdbserver-sb-0\" (UID: \"23e9fd30-8462-4b44-b9ae-11e3a367c814\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:31:58 crc kubenswrapper[4790]: I1124 13:31:58.537135 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/23e9fd30-8462-4b44-b9ae-11e3a367c814-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"23e9fd30-8462-4b44-b9ae-11e3a367c814\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:31:58 crc kubenswrapper[4790]: I1124 13:31:58.537285 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/23e9fd30-8462-4b44-b9ae-11e3a367c814-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"23e9fd30-8462-4b44-b9ae-11e3a367c814\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:31:58 crc kubenswrapper[4790]: I1124 13:31:58.537511 4790 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"23e9fd30-8462-4b44-b9ae-11e3a367c814\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-sb-0" Nov 24 13:31:58 crc kubenswrapper[4790]: I1124 13:31:58.538821 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/23e9fd30-8462-4b44-b9ae-11e3a367c814-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"23e9fd30-8462-4b44-b9ae-11e3a367c814\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:31:58 crc kubenswrapper[4790]: I1124 13:31:58.539948 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23e9fd30-8462-4b44-b9ae-11e3a367c814-config\") pod \"ovsdbserver-sb-0\" (UID: \"23e9fd30-8462-4b44-b9ae-11e3a367c814\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:31:58 crc kubenswrapper[4790]: I1124 13:31:58.544398 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23e9fd30-8462-4b44-b9ae-11e3a367c814-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"23e9fd30-8462-4b44-b9ae-11e3a367c814\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:31:58 crc kubenswrapper[4790]: I1124 13:31:58.545427 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/23e9fd30-8462-4b44-b9ae-11e3a367c814-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"23e9fd30-8462-4b44-b9ae-11e3a367c814\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:31:58 crc kubenswrapper[4790]: I1124 13:31:58.545964 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/23e9fd30-8462-4b44-b9ae-11e3a367c814-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"23e9fd30-8462-4b44-b9ae-11e3a367c814\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:31:58 crc kubenswrapper[4790]: I1124 13:31:58.556695 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnx7x\" (UniqueName: \"kubernetes.io/projected/23e9fd30-8462-4b44-b9ae-11e3a367c814-kube-api-access-bnx7x\") pod \"ovsdbserver-sb-0\" (UID: \"23e9fd30-8462-4b44-b9ae-11e3a367c814\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:31:58 crc kubenswrapper[4790]: I1124 13:31:58.570274 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"23e9fd30-8462-4b44-b9ae-11e3a367c814\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:31:58 crc kubenswrapper[4790]: I1124 13:31:58.702760 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:05 crc kubenswrapper[4790]: E1124 13:32:05.593698 4790 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 24 13:32:05 crc kubenswrapper[4790]: E1124 13:32:05.594461 4790 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nvfb2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-78kb8_openstack(8d2804d3-2309-417d-ae57-d48815681344): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:32:05 crc kubenswrapper[4790]: E1124 13:32:05.595608 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-78kb8" podUID="8d2804d3-2309-417d-ae57-d48815681344" Nov 24 13:32:05 crc kubenswrapper[4790]: E1124 13:32:05.605674 4790 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 24 13:32:05 crc kubenswrapper[4790]: E1124 13:32:05.605945 4790 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8jsn2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-z68d4_openstack(cd9d10cd-1e7e-4032-b88e-65ff25726653): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:32:05 crc kubenswrapper[4790]: E1124 13:32:05.607417 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-z68d4" podUID="cd9d10cd-1e7e-4032-b88e-65ff25726653" Nov 24 13:32:05 crc kubenswrapper[4790]: E1124 13:32:05.647350 4790 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 24 13:32:05 crc kubenswrapper[4790]: E1124 13:32:05.647615 4790 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nlm72,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-q9w8p_openstack(d7fcc2c7-c415-4bd7-9b8a-367d9292985d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:32:05 crc kubenswrapper[4790]: E1124 13:32:05.650613 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-q9w8p" podUID="d7fcc2c7-c415-4bd7-9b8a-367d9292985d" Nov 24 13:32:05 crc kubenswrapper[4790]: E1124 13:32:05.713679 4790 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 24 13:32:05 crc kubenswrapper[4790]: E1124 13:32:05.713854 4790 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k8trg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-q8jn7_openstack(dc896e16-9261-443a-86d7-c45c2fbef3d8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:32:05 crc kubenswrapper[4790]: E1124 13:32:05.716088 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-q8jn7" podUID="dc896e16-9261-443a-86d7-c45c2fbef3d8" Nov 24 13:32:06 crc kubenswrapper[4790]: I1124 13:32:06.105479 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 24 13:32:06 crc kubenswrapper[4790]: W1124 13:32:06.140820 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9726b87_983d_46fc_8880_7e19cbe1e35d.slice/crio-5ef77fdaab4bbd04d6baec365f079e0bb59ea2d4e9c12b47ab434f3c80a17f74 WatchSource:0}: Error finding container 5ef77fdaab4bbd04d6baec365f079e0bb59ea2d4e9c12b47ab434f3c80a17f74: Status 404 returned error can't find the container with id 5ef77fdaab4bbd04d6baec365f079e0bb59ea2d4e9c12b47ab434f3c80a17f74 Nov 24 13:32:06 crc kubenswrapper[4790]: I1124 13:32:06.141594 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 13:32:06 crc kubenswrapper[4790]: I1124 13:32:06.150665 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 13:32:06 crc kubenswrapper[4790]: I1124 13:32:06.332347 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rkxr6"] Nov 24 13:32:06 crc kubenswrapper[4790]: W1124 13:32:06.340579 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8fa811d_92cf_46a2_a9fd_539130369eb1.slice/crio-48362faf8e69bd45114ae3862302df825a91d31f4faf109b679162750f6f967d WatchSource:0}: Error finding container 48362faf8e69bd45114ae3862302df825a91d31f4faf109b679162750f6f967d: Status 404 returned error can't find the container with id 48362faf8e69bd45114ae3862302df825a91d31f4faf109b679162750f6f967d Nov 24 13:32:06 crc kubenswrapper[4790]: I1124 13:32:06.443613 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 13:32:06 crc kubenswrapper[4790]: I1124 13:32:06.483173 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"56d67de4-df9a-44fa-92b2-cbb168c4ae87","Type":"ContainerStarted","Data":"5b0c3aba8fa6e8f3f97dd7758dadf64532bd3c89036783eb8c58ac1dff1546a6"} Nov 24 13:32:06 crc kubenswrapper[4790]: I1124 13:32:06.484485 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5a434802-b18e-4d14-8415-4879eaa44191","Type":"ContainerStarted","Data":"31d8225618eff635579b4cb575f6c144d5c9c01d087c36178d30076b744ae0fd"} Nov 24 13:32:06 crc kubenswrapper[4790]: I1124 13:32:06.487499 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"9c866267-37e8-48d7-83ca-53e3cbf8b762","Type":"ContainerStarted","Data":"ac2a207b8fd870760b58312e8f38e10ea17fb9654e5b83ea853f3eeae396f453"} Nov 24 13:32:06 crc kubenswrapper[4790]: I1124 13:32:06.487541 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"9c866267-37e8-48d7-83ca-53e3cbf8b762","Type":"ContainerStarted","Data":"9d54c6d89a4aa563bf6e3e34eb7285204e3ed1ccfd61c3285c7cc567678f7c48"} Nov 24 13:32:06 crc kubenswrapper[4790]: I1124 13:32:06.492292 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"23e9fd30-8462-4b44-b9ae-11e3a367c814","Type":"ContainerStarted","Data":"ff0a313f41ee312fd3dbafb956ceb2ec653b95ab2f5bd477fb801596d8544d55"} Nov 24 13:32:06 crc kubenswrapper[4790]: I1124 13:32:06.493444 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rkxr6" event={"ID":"c8fa811d-92cf-46a2-a9fd-539130369eb1","Type":"ContainerStarted","Data":"48362faf8e69bd45114ae3862302df825a91d31f4faf109b679162750f6f967d"} Nov 24 13:32:06 crc kubenswrapper[4790]: I1124 13:32:06.494909 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d9726b87-983d-46fc-8880-7e19cbe1e35d","Type":"ContainerStarted","Data":"5ef77fdaab4bbd04d6baec365f079e0bb59ea2d4e9c12b47ab434f3c80a17f74"} Nov 24 13:32:06 crc kubenswrapper[4790]: E1124 13:32:06.496753 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-q8jn7" podUID="dc896e16-9261-443a-86d7-c45c2fbef3d8" Nov 24 13:32:06 crc kubenswrapper[4790]: E1124 13:32:06.497049 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-q9w8p" podUID="d7fcc2c7-c415-4bd7-9b8a-367d9292985d" Nov 24 13:32:06 crc kubenswrapper[4790]: I1124 13:32:06.929833 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-78kb8" Nov 24 13:32:07 crc kubenswrapper[4790]: I1124 13:32:07.029951 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d2804d3-2309-417d-ae57-d48815681344-config\") pod \"8d2804d3-2309-417d-ae57-d48815681344\" (UID: \"8d2804d3-2309-417d-ae57-d48815681344\") " Nov 24 13:32:07 crc kubenswrapper[4790]: I1124 13:32:07.030040 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvfb2\" (UniqueName: \"kubernetes.io/projected/8d2804d3-2309-417d-ae57-d48815681344-kube-api-access-nvfb2\") pod \"8d2804d3-2309-417d-ae57-d48815681344\" (UID: \"8d2804d3-2309-417d-ae57-d48815681344\") " Nov 24 13:32:07 crc kubenswrapper[4790]: I1124 13:32:07.031309 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d2804d3-2309-417d-ae57-d48815681344-config" (OuterVolumeSpecName: "config") pod "8d2804d3-2309-417d-ae57-d48815681344" (UID: "8d2804d3-2309-417d-ae57-d48815681344"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:07 crc kubenswrapper[4790]: I1124 13:32:07.037921 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d2804d3-2309-417d-ae57-d48815681344-kube-api-access-nvfb2" (OuterVolumeSpecName: "kube-api-access-nvfb2") pod "8d2804d3-2309-417d-ae57-d48815681344" (UID: "8d2804d3-2309-417d-ae57-d48815681344"). InnerVolumeSpecName "kube-api-access-nvfb2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:32:07 crc kubenswrapper[4790]: I1124 13:32:07.046148 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-z68d4" Nov 24 13:32:07 crc kubenswrapper[4790]: I1124 13:32:07.081032 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-28cwj"] Nov 24 13:32:07 crc kubenswrapper[4790]: I1124 13:32:07.131382 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jsn2\" (UniqueName: \"kubernetes.io/projected/cd9d10cd-1e7e-4032-b88e-65ff25726653-kube-api-access-8jsn2\") pod \"cd9d10cd-1e7e-4032-b88e-65ff25726653\" (UID: \"cd9d10cd-1e7e-4032-b88e-65ff25726653\") " Nov 24 13:32:07 crc kubenswrapper[4790]: I1124 13:32:07.131446 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd9d10cd-1e7e-4032-b88e-65ff25726653-dns-svc\") pod \"cd9d10cd-1e7e-4032-b88e-65ff25726653\" (UID: \"cd9d10cd-1e7e-4032-b88e-65ff25726653\") " Nov 24 13:32:07 crc kubenswrapper[4790]: I1124 13:32:07.131667 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd9d10cd-1e7e-4032-b88e-65ff25726653-config\") pod \"cd9d10cd-1e7e-4032-b88e-65ff25726653\" (UID: \"cd9d10cd-1e7e-4032-b88e-65ff25726653\") " Nov 24 13:32:07 crc kubenswrapper[4790]: I1124 13:32:07.132154 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvfb2\" (UniqueName: \"kubernetes.io/projected/8d2804d3-2309-417d-ae57-d48815681344-kube-api-access-nvfb2\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:07 crc kubenswrapper[4790]: I1124 13:32:07.132188 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d2804d3-2309-417d-ae57-d48815681344-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:07 crc kubenswrapper[4790]: I1124 13:32:07.132047 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd9d10cd-1e7e-4032-b88e-65ff25726653-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cd9d10cd-1e7e-4032-b88e-65ff25726653" (UID: "cd9d10cd-1e7e-4032-b88e-65ff25726653"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:07 crc kubenswrapper[4790]: I1124 13:32:07.132264 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd9d10cd-1e7e-4032-b88e-65ff25726653-config" (OuterVolumeSpecName: "config") pod "cd9d10cd-1e7e-4032-b88e-65ff25726653" (UID: "cd9d10cd-1e7e-4032-b88e-65ff25726653"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:07 crc kubenswrapper[4790]: I1124 13:32:07.135258 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd9d10cd-1e7e-4032-b88e-65ff25726653-kube-api-access-8jsn2" (OuterVolumeSpecName: "kube-api-access-8jsn2") pod "cd9d10cd-1e7e-4032-b88e-65ff25726653" (UID: "cd9d10cd-1e7e-4032-b88e-65ff25726653"). InnerVolumeSpecName "kube-api-access-8jsn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:32:07 crc kubenswrapper[4790]: I1124 13:32:07.234326 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd9d10cd-1e7e-4032-b88e-65ff25726653-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:07 crc kubenswrapper[4790]: I1124 13:32:07.234362 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jsn2\" (UniqueName: \"kubernetes.io/projected/cd9d10cd-1e7e-4032-b88e-65ff25726653-kube-api-access-8jsn2\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:07 crc kubenswrapper[4790]: I1124 13:32:07.234374 4790 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd9d10cd-1e7e-4032-b88e-65ff25726653-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:07 crc kubenswrapper[4790]: I1124 13:32:07.319834 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 13:32:07 crc kubenswrapper[4790]: I1124 13:32:07.506844 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-78kb8" event={"ID":"8d2804d3-2309-417d-ae57-d48815681344","Type":"ContainerDied","Data":"30299be8f96ecb5f137ad702cf37a873181b7af9cee609dd0d136310baf22d47"} Nov 24 13:32:07 crc kubenswrapper[4790]: I1124 13:32:07.506941 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-78kb8" Nov 24 13:32:07 crc kubenswrapper[4790]: I1124 13:32:07.511433 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-28cwj" event={"ID":"4d6d5d88-adda-4679-80de-0df0e0847bcb","Type":"ContainerStarted","Data":"dbc82dc390d1fdd1005a9ed6190bb3b0ffc70585af9231cca01bec44931c045e"} Nov 24 13:32:07 crc kubenswrapper[4790]: I1124 13:32:07.514097 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"85e04b92-ae17-4eb6-a736-395c5d44f563","Type":"ContainerStarted","Data":"569c178276bd7deaa7ff29136e16e02f63f79ee3c9e59aae1ab276151057e59e"} Nov 24 13:32:07 crc kubenswrapper[4790]: I1124 13:32:07.518713 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-z68d4" event={"ID":"cd9d10cd-1e7e-4032-b88e-65ff25726653","Type":"ContainerDied","Data":"952e74f44e641a8c993b420668997543fa9d5752a1f9827864866574222114f1"} Nov 24 13:32:07 crc kubenswrapper[4790]: I1124 13:32:07.518726 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-z68d4" Nov 24 13:32:07 crc kubenswrapper[4790]: W1124 13:32:07.519776 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08baac2a_300a_498f_8d0c_d1f261e6c8fc.slice/crio-3b21aedb6f04264926dc7ff26b8325a37ce70a1f9a469e53fcd12df7aac97be8 WatchSource:0}: Error finding container 3b21aedb6f04264926dc7ff26b8325a37ce70a1f9a469e53fcd12df7aac97be8: Status 404 returned error can't find the container with id 3b21aedb6f04264926dc7ff26b8325a37ce70a1f9a469e53fcd12df7aac97be8 Nov 24 13:32:07 crc kubenswrapper[4790]: I1124 13:32:07.543925 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"39508987-8655-46ed-861b-e1448652ddf6","Type":"ContainerStarted","Data":"9ed4735bba7ba4c006885cc1b29d1af4db7e93db39ebe812c28f73a898f8fa68"} Nov 24 13:32:07 crc kubenswrapper[4790]: I1124 13:32:07.587079 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-78kb8"] Nov 24 13:32:07 crc kubenswrapper[4790]: I1124 13:32:07.592949 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-78kb8"] Nov 24 13:32:07 crc kubenswrapper[4790]: I1124 13:32:07.667551 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-z68d4"] Nov 24 13:32:07 crc kubenswrapper[4790]: I1124 13:32:07.694107 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-z68d4"] Nov 24 13:32:08 crc kubenswrapper[4790]: I1124 13:32:08.325961 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d2804d3-2309-417d-ae57-d48815681344" path="/var/lib/kubelet/pods/8d2804d3-2309-417d-ae57-d48815681344/volumes" Nov 24 13:32:08 crc kubenswrapper[4790]: I1124 13:32:08.326346 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd9d10cd-1e7e-4032-b88e-65ff25726653" path="/var/lib/kubelet/pods/cd9d10cd-1e7e-4032-b88e-65ff25726653/volumes" Nov 24 13:32:08 crc kubenswrapper[4790]: I1124 13:32:08.545660 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"08baac2a-300a-498f-8d0c-d1f261e6c8fc","Type":"ContainerStarted","Data":"3b21aedb6f04264926dc7ff26b8325a37ce70a1f9a469e53fcd12df7aac97be8"} Nov 24 13:32:09 crc kubenswrapper[4790]: I1124 13:32:09.554659 4790 generic.go:334] "Generic (PLEG): container finished" podID="5a434802-b18e-4d14-8415-4879eaa44191" containerID="31d8225618eff635579b4cb575f6c144d5c9c01d087c36178d30076b744ae0fd" exitCode=0 Nov 24 13:32:09 crc kubenswrapper[4790]: I1124 13:32:09.554700 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5a434802-b18e-4d14-8415-4879eaa44191","Type":"ContainerDied","Data":"31d8225618eff635579b4cb575f6c144d5c9c01d087c36178d30076b744ae0fd"} Nov 24 13:32:10 crc kubenswrapper[4790]: I1124 13:32:10.564607 4790 generic.go:334] "Generic (PLEG): container finished" podID="9c866267-37e8-48d7-83ca-53e3cbf8b762" containerID="ac2a207b8fd870760b58312e8f38e10ea17fb9654e5b83ea853f3eeae396f453" exitCode=0 Nov 24 13:32:10 crc kubenswrapper[4790]: I1124 13:32:10.564659 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"9c866267-37e8-48d7-83ca-53e3cbf8b762","Type":"ContainerDied","Data":"ac2a207b8fd870760b58312e8f38e10ea17fb9654e5b83ea853f3eeae396f453"} Nov 24 13:32:11 crc kubenswrapper[4790]: I1124 13:32:11.574268 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d9726b87-983d-46fc-8880-7e19cbe1e35d","Type":"ContainerStarted","Data":"9d976a5e89a22f94794138f6f993ae7529618b4e787e03ded2687af1b33877b2"} Nov 24 13:32:11 crc kubenswrapper[4790]: I1124 13:32:11.575850 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 24 13:32:11 crc kubenswrapper[4790]: I1124 13:32:11.577644 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5a434802-b18e-4d14-8415-4879eaa44191","Type":"ContainerStarted","Data":"3da69fd56ac70e4cd6a7e0c488e7105251f570721cf9c47f92c46924f483efc6"} Nov 24 13:32:11 crc kubenswrapper[4790]: I1124 13:32:11.590288 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"56d67de4-df9a-44fa-92b2-cbb168c4ae87","Type":"ContainerStarted","Data":"6501f4d3c19251ce46a87c27719040d3bcdb4ba9e7bcdb6dbb40bc282a09c44d"} Nov 24 13:32:11 crc kubenswrapper[4790]: I1124 13:32:11.591002 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 24 13:32:11 crc kubenswrapper[4790]: I1124 13:32:11.593513 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"9c866267-37e8-48d7-83ca-53e3cbf8b762","Type":"ContainerStarted","Data":"bdeaeaffd01a11cc6124ff6c3f7d882d3b5bdbc38615b718c3eb5ed5e5f217f3"} Nov 24 13:32:11 crc kubenswrapper[4790]: I1124 13:32:11.594011 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=16.629098277 podStartE2EDuration="21.59399852s" podCreationTimestamp="2025-11-24 13:31:50 +0000 UTC" firstStartedPulling="2025-11-24 13:32:06.145276223 +0000 UTC m=+1174.525169885" lastFinishedPulling="2025-11-24 13:32:11.110176466 +0000 UTC m=+1179.490070128" observedRunningTime="2025-11-24 13:32:11.589518796 +0000 UTC m=+1179.969412458" watchObservedRunningTime="2025-11-24 13:32:11.59399852 +0000 UTC m=+1179.973892192" Nov 24 13:32:11 crc kubenswrapper[4790]: I1124 13:32:11.599712 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"23e9fd30-8462-4b44-b9ae-11e3a367c814","Type":"ContainerStarted","Data":"995eda8d4390f50082b69ba23ac398c85f53e9011ae0253de271653fcc9abd18"} Nov 24 13:32:11 crc kubenswrapper[4790]: I1124 13:32:11.602595 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"08baac2a-300a-498f-8d0c-d1f261e6c8fc","Type":"ContainerStarted","Data":"c70103120e0d00167ed2d2d0209229ea06b056321fe8833b38fe79ae3cb43775"} Nov 24 13:32:11 crc kubenswrapper[4790]: I1124 13:32:11.606220 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rkxr6" event={"ID":"c8fa811d-92cf-46a2-a9fd-539130369eb1","Type":"ContainerStarted","Data":"3e36b6d6b9a369ac071cdab1b1d854f1f347521f301b072cb644c63a4b575d0c"} Nov 24 13:32:11 crc kubenswrapper[4790]: I1124 13:32:11.606729 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-rkxr6" Nov 24 13:32:11 crc kubenswrapper[4790]: I1124 13:32:11.610363 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-28cwj" event={"ID":"4d6d5d88-adda-4679-80de-0df0e0847bcb","Type":"ContainerStarted","Data":"d8259b93cb0ee7e3a787272e19de83ebfd12257522109ed7dfb63ae1d006ab09"} Nov 24 13:32:11 crc kubenswrapper[4790]: I1124 13:32:11.615248 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=8.666618262 podStartE2EDuration="25.615236229s" podCreationTimestamp="2025-11-24 13:31:46 +0000 UTC" firstStartedPulling="2025-11-24 13:31:48.669524581 +0000 UTC m=+1157.049418243" lastFinishedPulling="2025-11-24 13:32:05.618142538 +0000 UTC m=+1173.998036210" observedRunningTime="2025-11-24 13:32:11.610153908 +0000 UTC m=+1179.990047570" watchObservedRunningTime="2025-11-24 13:32:11.615236229 +0000 UTC m=+1179.995129891" Nov 24 13:32:11 crc kubenswrapper[4790]: I1124 13:32:11.633568 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-rkxr6" podStartSLOduration=14.394866259 podStartE2EDuration="18.633551117s" podCreationTimestamp="2025-11-24 13:31:53 +0000 UTC" firstStartedPulling="2025-11-24 13:32:06.394436001 +0000 UTC m=+1174.774329653" lastFinishedPulling="2025-11-24 13:32:10.633120849 +0000 UTC m=+1179.013014511" observedRunningTime="2025-11-24 13:32:11.625603666 +0000 UTC m=+1180.005497338" watchObservedRunningTime="2025-11-24 13:32:11.633551117 +0000 UTC m=+1180.013444779" Nov 24 13:32:11 crc kubenswrapper[4790]: I1124 13:32:11.668515 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=18.140338021 podStartE2EDuration="22.667348344s" podCreationTimestamp="2025-11-24 13:31:49 +0000 UTC" firstStartedPulling="2025-11-24 13:32:06.106130057 +0000 UTC m=+1174.486023719" lastFinishedPulling="2025-11-24 13:32:10.63314038 +0000 UTC m=+1179.013034042" observedRunningTime="2025-11-24 13:32:11.655281699 +0000 UTC m=+1180.035175351" watchObservedRunningTime="2025-11-24 13:32:11.667348344 +0000 UTC m=+1180.047242016" Nov 24 13:32:11 crc kubenswrapper[4790]: I1124 13:32:11.685108 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=24.685089636 podStartE2EDuration="24.685089636s" podCreationTimestamp="2025-11-24 13:31:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:32:11.677750672 +0000 UTC m=+1180.057644344" watchObservedRunningTime="2025-11-24 13:32:11.685089636 +0000 UTC m=+1180.064983298" Nov 24 13:32:12 crc kubenswrapper[4790]: I1124 13:32:12.620253 4790 generic.go:334] "Generic (PLEG): container finished" podID="4d6d5d88-adda-4679-80de-0df0e0847bcb" containerID="d8259b93cb0ee7e3a787272e19de83ebfd12257522109ed7dfb63ae1d006ab09" exitCode=0 Nov 24 13:32:12 crc kubenswrapper[4790]: I1124 13:32:12.620294 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-28cwj" event={"ID":"4d6d5d88-adda-4679-80de-0df0e0847bcb","Type":"ContainerDied","Data":"d8259b93cb0ee7e3a787272e19de83ebfd12257522109ed7dfb63ae1d006ab09"} Nov 24 13:32:17 crc kubenswrapper[4790]: I1124 13:32:17.956522 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 24 13:32:17 crc kubenswrapper[4790]: I1124 13:32:17.956988 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 24 13:32:18 crc kubenswrapper[4790]: E1124 13:32:18.417541 4790 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.69:40426->38.129.56.69:36625: write tcp 38.129.56.69:40426->38.129.56.69:36625: write: broken pipe Nov 24 13:32:18 crc kubenswrapper[4790]: I1124 13:32:18.669161 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-28cwj" event={"ID":"4d6d5d88-adda-4679-80de-0df0e0847bcb","Type":"ContainerStarted","Data":"e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3"} Nov 24 13:32:19 crc kubenswrapper[4790]: I1124 13:32:19.309934 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:19 crc kubenswrapper[4790]: I1124 13:32:19.310193 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:19 crc kubenswrapper[4790]: I1124 13:32:19.667408 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 24 13:32:20 crc kubenswrapper[4790]: I1124 13:32:20.687927 4790 generic.go:334] "Generic (PLEG): container finished" podID="dc896e16-9261-443a-86d7-c45c2fbef3d8" containerID="6980ad4fe960e033c715bd3155792b9fc55f15acbcc5bf7afda27567238e7feb" exitCode=0 Nov 24 13:32:20 crc kubenswrapper[4790]: I1124 13:32:20.688108 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-q8jn7" event={"ID":"dc896e16-9261-443a-86d7-c45c2fbef3d8","Type":"ContainerDied","Data":"6980ad4fe960e033c715bd3155792b9fc55f15acbcc5bf7afda27567238e7feb"} Nov 24 13:32:20 crc kubenswrapper[4790]: I1124 13:32:20.692743 4790 generic.go:334] "Generic (PLEG): container finished" podID="d7fcc2c7-c415-4bd7-9b8a-367d9292985d" containerID="0735e2434710622ef854e0dbf6a71c0acc25cea40f272117c0ed5ab3006d8ddd" exitCode=0 Nov 24 13:32:20 crc kubenswrapper[4790]: I1124 13:32:20.692804 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-q9w8p" event={"ID":"d7fcc2c7-c415-4bd7-9b8a-367d9292985d","Type":"ContainerDied","Data":"0735e2434710622ef854e0dbf6a71c0acc25cea40f272117c0ed5ab3006d8ddd"} Nov 24 13:32:20 crc kubenswrapper[4790]: I1124 13:32:20.696390 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-28cwj" event={"ID":"4d6d5d88-adda-4679-80de-0df0e0847bcb","Type":"ContainerStarted","Data":"46182cf796e680922c3d2c23de93f19051dee361d1797948d997345bb2652287"} Nov 24 13:32:20 crc kubenswrapper[4790]: I1124 13:32:20.696446 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-28cwj" Nov 24 13:32:20 crc kubenswrapper[4790]: I1124 13:32:20.696459 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-28cwj" Nov 24 13:32:20 crc kubenswrapper[4790]: I1124 13:32:20.699207 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"23e9fd30-8462-4b44-b9ae-11e3a367c814","Type":"ContainerStarted","Data":"72c394197db4eb126b656feee3931bef19393201c83eb96112678edb0aa5ebd7"} Nov 24 13:32:20 crc kubenswrapper[4790]: I1124 13:32:20.703857 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"08baac2a-300a-498f-8d0c-d1f261e6c8fc","Type":"ContainerStarted","Data":"500eaeafad5bdd0b7c31ca6862cf936a5fd98d2a297af4a1a569abfdeafe28a7"} Nov 24 13:32:20 crc kubenswrapper[4790]: I1124 13:32:20.742791 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=14.715922765 podStartE2EDuration="26.742774594s" podCreationTimestamp="2025-11-24 13:31:54 +0000 UTC" firstStartedPulling="2025-11-24 13:32:07.531069144 +0000 UTC m=+1175.910962806" lastFinishedPulling="2025-11-24 13:32:19.557920973 +0000 UTC m=+1187.937814635" observedRunningTime="2025-11-24 13:32:20.732501509 +0000 UTC m=+1189.112395171" watchObservedRunningTime="2025-11-24 13:32:20.742774594 +0000 UTC m=+1189.122668256" Nov 24 13:32:20 crc kubenswrapper[4790]: I1124 13:32:20.783350 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-28cwj" podStartSLOduration=24.034097369 podStartE2EDuration="27.783335048s" podCreationTimestamp="2025-11-24 13:31:53 +0000 UTC" firstStartedPulling="2025-11-24 13:32:07.260371838 +0000 UTC m=+1175.640265500" lastFinishedPulling="2025-11-24 13:32:11.009609517 +0000 UTC m=+1179.389503179" observedRunningTime="2025-11-24 13:32:20.779019769 +0000 UTC m=+1189.158913431" watchObservedRunningTime="2025-11-24 13:32:20.783335048 +0000 UTC m=+1189.163228710" Nov 24 13:32:20 crc kubenswrapper[4790]: I1124 13:32:20.804551 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=10.717104162 podStartE2EDuration="23.804533096s" podCreationTimestamp="2025-11-24 13:31:57 +0000 UTC" firstStartedPulling="2025-11-24 13:32:06.456266295 +0000 UTC m=+1174.836159957" lastFinishedPulling="2025-11-24 13:32:19.543695229 +0000 UTC m=+1187.923588891" observedRunningTime="2025-11-24 13:32:20.8007179 +0000 UTC m=+1189.180611562" watchObservedRunningTime="2025-11-24 13:32:20.804533096 +0000 UTC m=+1189.184426758" Nov 24 13:32:21 crc kubenswrapper[4790]: I1124 13:32:21.007003 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:21 crc kubenswrapper[4790]: I1124 13:32:21.066124 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 24 13:32:21 crc kubenswrapper[4790]: I1124 13:32:21.164531 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-q9w8p"] Nov 24 13:32:21 crc kubenswrapper[4790]: I1124 13:32:21.203547 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-4d4ck"] Nov 24 13:32:21 crc kubenswrapper[4790]: I1124 13:32:21.207687 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-4d4ck" Nov 24 13:32:21 crc kubenswrapper[4790]: I1124 13:32:21.218310 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-4d4ck"] Nov 24 13:32:21 crc kubenswrapper[4790]: I1124 13:32:21.357169 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f924c39a-8259-45d6-aa00-4cc41f14e7ee-config\") pod \"dnsmasq-dns-7cb5889db5-4d4ck\" (UID: \"f924c39a-8259-45d6-aa00-4cc41f14e7ee\") " pod="openstack/dnsmasq-dns-7cb5889db5-4d4ck" Nov 24 13:32:21 crc kubenswrapper[4790]: I1124 13:32:21.357362 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f924c39a-8259-45d6-aa00-4cc41f14e7ee-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-4d4ck\" (UID: \"f924c39a-8259-45d6-aa00-4cc41f14e7ee\") " pod="openstack/dnsmasq-dns-7cb5889db5-4d4ck" Nov 24 13:32:21 crc kubenswrapper[4790]: I1124 13:32:21.357459 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfztq\" (UniqueName: \"kubernetes.io/projected/f924c39a-8259-45d6-aa00-4cc41f14e7ee-kube-api-access-gfztq\") pod \"dnsmasq-dns-7cb5889db5-4d4ck\" (UID: \"f924c39a-8259-45d6-aa00-4cc41f14e7ee\") " pod="openstack/dnsmasq-dns-7cb5889db5-4d4ck" Nov 24 13:32:21 crc kubenswrapper[4790]: I1124 13:32:21.459254 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f924c39a-8259-45d6-aa00-4cc41f14e7ee-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-4d4ck\" (UID: \"f924c39a-8259-45d6-aa00-4cc41f14e7ee\") " pod="openstack/dnsmasq-dns-7cb5889db5-4d4ck" Nov 24 13:32:21 crc kubenswrapper[4790]: I1124 13:32:21.459338 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfztq\" (UniqueName: \"kubernetes.io/projected/f924c39a-8259-45d6-aa00-4cc41f14e7ee-kube-api-access-gfztq\") pod \"dnsmasq-dns-7cb5889db5-4d4ck\" (UID: \"f924c39a-8259-45d6-aa00-4cc41f14e7ee\") " pod="openstack/dnsmasq-dns-7cb5889db5-4d4ck" Nov 24 13:32:21 crc kubenswrapper[4790]: I1124 13:32:21.459392 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f924c39a-8259-45d6-aa00-4cc41f14e7ee-config\") pod \"dnsmasq-dns-7cb5889db5-4d4ck\" (UID: \"f924c39a-8259-45d6-aa00-4cc41f14e7ee\") " pod="openstack/dnsmasq-dns-7cb5889db5-4d4ck" Nov 24 13:32:21 crc kubenswrapper[4790]: I1124 13:32:21.460434 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f924c39a-8259-45d6-aa00-4cc41f14e7ee-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-4d4ck\" (UID: \"f924c39a-8259-45d6-aa00-4cc41f14e7ee\") " pod="openstack/dnsmasq-dns-7cb5889db5-4d4ck" Nov 24 13:32:21 crc kubenswrapper[4790]: I1124 13:32:21.460606 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f924c39a-8259-45d6-aa00-4cc41f14e7ee-config\") pod \"dnsmasq-dns-7cb5889db5-4d4ck\" (UID: \"f924c39a-8259-45d6-aa00-4cc41f14e7ee\") " pod="openstack/dnsmasq-dns-7cb5889db5-4d4ck" Nov 24 13:32:21 crc kubenswrapper[4790]: I1124 13:32:21.485789 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfztq\" (UniqueName: \"kubernetes.io/projected/f924c39a-8259-45d6-aa00-4cc41f14e7ee-kube-api-access-gfztq\") pod \"dnsmasq-dns-7cb5889db5-4d4ck\" (UID: \"f924c39a-8259-45d6-aa00-4cc41f14e7ee\") " pod="openstack/dnsmasq-dns-7cb5889db5-4d4ck" Nov 24 13:32:21 crc kubenswrapper[4790]: I1124 13:32:21.495989 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:21 crc kubenswrapper[4790]: I1124 13:32:21.536037 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-4d4ck" Nov 24 13:32:21 crc kubenswrapper[4790]: I1124 13:32:21.589146 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:21 crc kubenswrapper[4790]: I1124 13:32:21.724890 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-q8jn7" event={"ID":"dc896e16-9261-443a-86d7-c45c2fbef3d8","Type":"ContainerStarted","Data":"5b1ec6ba7c1716aedfa3c3f6c36a931e3a92d1ced968935496c33bcf7abe14d7"} Nov 24 13:32:21 crc kubenswrapper[4790]: I1124 13:32:21.725217 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-q8jn7" Nov 24 13:32:21 crc kubenswrapper[4790]: I1124 13:32:21.733017 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-q9w8p" event={"ID":"d7fcc2c7-c415-4bd7-9b8a-367d9292985d","Type":"ContainerStarted","Data":"2029256ebcfff8392e758225b7868aee953c152e054e7a004fb3794dbd8c5661"} Nov 24 13:32:21 crc kubenswrapper[4790]: I1124 13:32:21.733240 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-q9w8p" podUID="d7fcc2c7-c415-4bd7-9b8a-367d9292985d" containerName="dnsmasq-dns" containerID="cri-o://2029256ebcfff8392e758225b7868aee953c152e054e7a004fb3794dbd8c5661" gracePeriod=10 Nov 24 13:32:21 crc kubenswrapper[4790]: I1124 13:32:21.770222 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-q9w8p" podStartSLOduration=3.60550628 podStartE2EDuration="37.770207299s" podCreationTimestamp="2025-11-24 13:31:44 +0000 UTC" firstStartedPulling="2025-11-24 13:31:45.37900032 +0000 UTC m=+1153.758893982" lastFinishedPulling="2025-11-24 13:32:19.543701339 +0000 UTC m=+1187.923595001" observedRunningTime="2025-11-24 13:32:21.7680858 +0000 UTC m=+1190.147979462" watchObservedRunningTime="2025-11-24 13:32:21.770207299 +0000 UTC m=+1190.150100961" Nov 24 13:32:21 crc kubenswrapper[4790]: I1124 13:32:21.773214 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-q8jn7" podStartSLOduration=2.846914042 podStartE2EDuration="36.773204682s" podCreationTimestamp="2025-11-24 13:31:45 +0000 UTC" firstStartedPulling="2025-11-24 13:31:46.043289767 +0000 UTC m=+1154.423183429" lastFinishedPulling="2025-11-24 13:32:19.969580407 +0000 UTC m=+1188.349474069" observedRunningTime="2025-11-24 13:32:21.74679424 +0000 UTC m=+1190.126687902" watchObservedRunningTime="2025-11-24 13:32:21.773204682 +0000 UTC m=+1190.153098344" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.044400 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-4d4ck"] Nov 24 13:32:22 crc kubenswrapper[4790]: W1124 13:32:22.060618 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf924c39a_8259_45d6_aa00_4cc41f14e7ee.slice/crio-de4d1efcd3fe9c001bf095ca8a6688a900b6c37255742586bb026ec6ec054230 WatchSource:0}: Error finding container de4d1efcd3fe9c001bf095ca8a6688a900b6c37255742586bb026ec6ec054230: Status 404 returned error can't find the container with id de4d1efcd3fe9c001bf095ca8a6688a900b6c37255742586bb026ec6ec054230 Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.100814 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.158343 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-q9w8p" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.208375 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.282651 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7fcc2c7-c415-4bd7-9b8a-367d9292985d-dns-svc\") pod \"d7fcc2c7-c415-4bd7-9b8a-367d9292985d\" (UID: \"d7fcc2c7-c415-4bd7-9b8a-367d9292985d\") " Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.283088 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7fcc2c7-c415-4bd7-9b8a-367d9292985d-config\") pod \"d7fcc2c7-c415-4bd7-9b8a-367d9292985d\" (UID: \"d7fcc2c7-c415-4bd7-9b8a-367d9292985d\") " Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.283217 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlm72\" (UniqueName: \"kubernetes.io/projected/d7fcc2c7-c415-4bd7-9b8a-367d9292985d-kube-api-access-nlm72\") pod \"d7fcc2c7-c415-4bd7-9b8a-367d9292985d\" (UID: \"d7fcc2c7-c415-4bd7-9b8a-367d9292985d\") " Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.287871 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7fcc2c7-c415-4bd7-9b8a-367d9292985d-kube-api-access-nlm72" (OuterVolumeSpecName: "kube-api-access-nlm72") pod "d7fcc2c7-c415-4bd7-9b8a-367d9292985d" (UID: "d7fcc2c7-c415-4bd7-9b8a-367d9292985d"). InnerVolumeSpecName "kube-api-access-nlm72". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.326837 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7fcc2c7-c415-4bd7-9b8a-367d9292985d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d7fcc2c7-c415-4bd7-9b8a-367d9292985d" (UID: "d7fcc2c7-c415-4bd7-9b8a-367d9292985d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.341219 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Nov 24 13:32:22 crc kubenswrapper[4790]: E1124 13:32:22.341655 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7fcc2c7-c415-4bd7-9b8a-367d9292985d" containerName="dnsmasq-dns" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.341678 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7fcc2c7-c415-4bd7-9b8a-367d9292985d" containerName="dnsmasq-dns" Nov 24 13:32:22 crc kubenswrapper[4790]: E1124 13:32:22.341716 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7fcc2c7-c415-4bd7-9b8a-367d9292985d" containerName="init" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.341724 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7fcc2c7-c415-4bd7-9b8a-367d9292985d" containerName="init" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.341994 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7fcc2c7-c415-4bd7-9b8a-367d9292985d" containerName="dnsmasq-dns" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.348673 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7fcc2c7-c415-4bd7-9b8a-367d9292985d-config" (OuterVolumeSpecName: "config") pod "d7fcc2c7-c415-4bd7-9b8a-367d9292985d" (UID: "d7fcc2c7-c415-4bd7-9b8a-367d9292985d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.355792 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.355981 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.360458 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.364087 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.364212 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.365329 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-9pbzx" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.384626 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlm72\" (UniqueName: \"kubernetes.io/projected/d7fcc2c7-c415-4bd7-9b8a-367d9292985d-kube-api-access-nlm72\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.384658 4790 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7fcc2c7-c415-4bd7-9b8a-367d9292985d-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.384669 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7fcc2c7-c415-4bd7-9b8a-367d9292985d-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.486022 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/63ce5126-cdab-4cdc-968a-acc26c6caf87-cache\") pod \"swift-storage-0\" (UID: \"63ce5126-cdab-4cdc-968a-acc26c6caf87\") " pod="openstack/swift-storage-0" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.486097 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5jk4\" (UniqueName: \"kubernetes.io/projected/63ce5126-cdab-4cdc-968a-acc26c6caf87-kube-api-access-n5jk4\") pod \"swift-storage-0\" (UID: \"63ce5126-cdab-4cdc-968a-acc26c6caf87\") " pod="openstack/swift-storage-0" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.486142 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/63ce5126-cdab-4cdc-968a-acc26c6caf87-etc-swift\") pod \"swift-storage-0\" (UID: \"63ce5126-cdab-4cdc-968a-acc26c6caf87\") " pod="openstack/swift-storage-0" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.486498 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"63ce5126-cdab-4cdc-968a-acc26c6caf87\") " pod="openstack/swift-storage-0" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.486685 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/63ce5126-cdab-4cdc-968a-acc26c6caf87-lock\") pod \"swift-storage-0\" (UID: \"63ce5126-cdab-4cdc-968a-acc26c6caf87\") " pod="openstack/swift-storage-0" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.552373 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-48flt"] Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.553749 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-48flt" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.555351 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.556300 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.556511 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.561438 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-48flt"] Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.588723 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/63ce5126-cdab-4cdc-968a-acc26c6caf87-etc-swift\") pod \"swift-storage-0\" (UID: \"63ce5126-cdab-4cdc-968a-acc26c6caf87\") " pod="openstack/swift-storage-0" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.588797 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"63ce5126-cdab-4cdc-968a-acc26c6caf87\") " pod="openstack/swift-storage-0" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.588841 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/63ce5126-cdab-4cdc-968a-acc26c6caf87-lock\") pod \"swift-storage-0\" (UID: \"63ce5126-cdab-4cdc-968a-acc26c6caf87\") " pod="openstack/swift-storage-0" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.588860 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/63ce5126-cdab-4cdc-968a-acc26c6caf87-cache\") pod \"swift-storage-0\" (UID: \"63ce5126-cdab-4cdc-968a-acc26c6caf87\") " pod="openstack/swift-storage-0" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.588926 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5jk4\" (UniqueName: \"kubernetes.io/projected/63ce5126-cdab-4cdc-968a-acc26c6caf87-kube-api-access-n5jk4\") pod \"swift-storage-0\" (UID: \"63ce5126-cdab-4cdc-968a-acc26c6caf87\") " pod="openstack/swift-storage-0" Nov 24 13:32:22 crc kubenswrapper[4790]: E1124 13:32:22.588957 4790 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 13:32:22 crc kubenswrapper[4790]: E1124 13:32:22.588985 4790 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 13:32:22 crc kubenswrapper[4790]: E1124 13:32:22.589055 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/63ce5126-cdab-4cdc-968a-acc26c6caf87-etc-swift podName:63ce5126-cdab-4cdc-968a-acc26c6caf87 nodeName:}" failed. No retries permitted until 2025-11-24 13:32:23.089031452 +0000 UTC m=+1191.468925194 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/63ce5126-cdab-4cdc-968a-acc26c6caf87-etc-swift") pod "swift-storage-0" (UID: "63ce5126-cdab-4cdc-968a-acc26c6caf87") : configmap "swift-ring-files" not found Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.589150 4790 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"63ce5126-cdab-4cdc-968a-acc26c6caf87\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/swift-storage-0" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.589384 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/63ce5126-cdab-4cdc-968a-acc26c6caf87-lock\") pod \"swift-storage-0\" (UID: \"63ce5126-cdab-4cdc-968a-acc26c6caf87\") " pod="openstack/swift-storage-0" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.589825 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/63ce5126-cdab-4cdc-968a-acc26c6caf87-cache\") pod \"swift-storage-0\" (UID: \"63ce5126-cdab-4cdc-968a-acc26c6caf87\") " pod="openstack/swift-storage-0" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.641042 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5jk4\" (UniqueName: \"kubernetes.io/projected/63ce5126-cdab-4cdc-968a-acc26c6caf87-kube-api-access-n5jk4\") pod \"swift-storage-0\" (UID: \"63ce5126-cdab-4cdc-968a-acc26c6caf87\") " pod="openstack/swift-storage-0" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.652336 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"63ce5126-cdab-4cdc-968a-acc26c6caf87\") " pod="openstack/swift-storage-0" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.690716 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87e08198-d2d0-4608-87f3-02d832fe7c55-combined-ca-bundle\") pod \"swift-ring-rebalance-48flt\" (UID: \"87e08198-d2d0-4608-87f3-02d832fe7c55\") " pod="openstack/swift-ring-rebalance-48flt" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.690816 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/87e08198-d2d0-4608-87f3-02d832fe7c55-ring-data-devices\") pod \"swift-ring-rebalance-48flt\" (UID: \"87e08198-d2d0-4608-87f3-02d832fe7c55\") " pod="openstack/swift-ring-rebalance-48flt" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.690846 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-485t4\" (UniqueName: \"kubernetes.io/projected/87e08198-d2d0-4608-87f3-02d832fe7c55-kube-api-access-485t4\") pod \"swift-ring-rebalance-48flt\" (UID: \"87e08198-d2d0-4608-87f3-02d832fe7c55\") " pod="openstack/swift-ring-rebalance-48flt" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.690874 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/87e08198-d2d0-4608-87f3-02d832fe7c55-swiftconf\") pod \"swift-ring-rebalance-48flt\" (UID: \"87e08198-d2d0-4608-87f3-02d832fe7c55\") " pod="openstack/swift-ring-rebalance-48flt" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.690922 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/87e08198-d2d0-4608-87f3-02d832fe7c55-scripts\") pod \"swift-ring-rebalance-48flt\" (UID: \"87e08198-d2d0-4608-87f3-02d832fe7c55\") " pod="openstack/swift-ring-rebalance-48flt" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.690958 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/87e08198-d2d0-4608-87f3-02d832fe7c55-dispersionconf\") pod \"swift-ring-rebalance-48flt\" (UID: \"87e08198-d2d0-4608-87f3-02d832fe7c55\") " pod="openstack/swift-ring-rebalance-48flt" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.690982 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/87e08198-d2d0-4608-87f3-02d832fe7c55-etc-swift\") pod \"swift-ring-rebalance-48flt\" (UID: \"87e08198-d2d0-4608-87f3-02d832fe7c55\") " pod="openstack/swift-ring-rebalance-48flt" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.703125 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.760489 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.764306 4790 generic.go:334] "Generic (PLEG): container finished" podID="f924c39a-8259-45d6-aa00-4cc41f14e7ee" containerID="66e7fe0a6042aa8f4790b3a22b5fd027de4aef544a4fdf42731e2c840ffd4cf2" exitCode=0 Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.764388 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-4d4ck" event={"ID":"f924c39a-8259-45d6-aa00-4cc41f14e7ee","Type":"ContainerDied","Data":"66e7fe0a6042aa8f4790b3a22b5fd027de4aef544a4fdf42731e2c840ffd4cf2"} Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.764442 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-4d4ck" event={"ID":"f924c39a-8259-45d6-aa00-4cc41f14e7ee","Type":"ContainerStarted","Data":"de4d1efcd3fe9c001bf095ca8a6688a900b6c37255742586bb026ec6ec054230"} Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.766919 4790 generic.go:334] "Generic (PLEG): container finished" podID="d7fcc2c7-c415-4bd7-9b8a-367d9292985d" containerID="2029256ebcfff8392e758225b7868aee953c152e054e7a004fb3794dbd8c5661" exitCode=0 Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.766983 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-q9w8p" event={"ID":"d7fcc2c7-c415-4bd7-9b8a-367d9292985d","Type":"ContainerDied","Data":"2029256ebcfff8392e758225b7868aee953c152e054e7a004fb3794dbd8c5661"} Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.767024 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-q9w8p" event={"ID":"d7fcc2c7-c415-4bd7-9b8a-367d9292985d","Type":"ContainerDied","Data":"ed718cdd9c23dc552e7effeffc08c94bc60b4cb77793724f157797f859b4a049"} Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.767045 4790 scope.go:117] "RemoveContainer" containerID="2029256ebcfff8392e758225b7868aee953c152e054e7a004fb3794dbd8c5661" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.766966 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-q9w8p" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.769137 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.791912 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/87e08198-d2d0-4608-87f3-02d832fe7c55-dispersionconf\") pod \"swift-ring-rebalance-48flt\" (UID: \"87e08198-d2d0-4608-87f3-02d832fe7c55\") " pod="openstack/swift-ring-rebalance-48flt" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.791974 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/87e08198-d2d0-4608-87f3-02d832fe7c55-etc-swift\") pod \"swift-ring-rebalance-48flt\" (UID: \"87e08198-d2d0-4608-87f3-02d832fe7c55\") " pod="openstack/swift-ring-rebalance-48flt" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.792014 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87e08198-d2d0-4608-87f3-02d832fe7c55-combined-ca-bundle\") pod \"swift-ring-rebalance-48flt\" (UID: \"87e08198-d2d0-4608-87f3-02d832fe7c55\") " pod="openstack/swift-ring-rebalance-48flt" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.792089 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/87e08198-d2d0-4608-87f3-02d832fe7c55-ring-data-devices\") pod \"swift-ring-rebalance-48flt\" (UID: \"87e08198-d2d0-4608-87f3-02d832fe7c55\") " pod="openstack/swift-ring-rebalance-48flt" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.792129 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-485t4\" (UniqueName: \"kubernetes.io/projected/87e08198-d2d0-4608-87f3-02d832fe7c55-kube-api-access-485t4\") pod \"swift-ring-rebalance-48flt\" (UID: \"87e08198-d2d0-4608-87f3-02d832fe7c55\") " pod="openstack/swift-ring-rebalance-48flt" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.792161 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/87e08198-d2d0-4608-87f3-02d832fe7c55-swiftconf\") pod \"swift-ring-rebalance-48flt\" (UID: \"87e08198-d2d0-4608-87f3-02d832fe7c55\") " pod="openstack/swift-ring-rebalance-48flt" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.792201 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/87e08198-d2d0-4608-87f3-02d832fe7c55-scripts\") pod \"swift-ring-rebalance-48flt\" (UID: \"87e08198-d2d0-4608-87f3-02d832fe7c55\") " pod="openstack/swift-ring-rebalance-48flt" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.792869 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/87e08198-d2d0-4608-87f3-02d832fe7c55-etc-swift\") pod \"swift-ring-rebalance-48flt\" (UID: \"87e08198-d2d0-4608-87f3-02d832fe7c55\") " pod="openstack/swift-ring-rebalance-48flt" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.793257 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/87e08198-d2d0-4608-87f3-02d832fe7c55-scripts\") pod \"swift-ring-rebalance-48flt\" (UID: \"87e08198-d2d0-4608-87f3-02d832fe7c55\") " pod="openstack/swift-ring-rebalance-48flt" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.793289 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/87e08198-d2d0-4608-87f3-02d832fe7c55-ring-data-devices\") pod \"swift-ring-rebalance-48flt\" (UID: \"87e08198-d2d0-4608-87f3-02d832fe7c55\") " pod="openstack/swift-ring-rebalance-48flt" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.800690 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87e08198-d2d0-4608-87f3-02d832fe7c55-combined-ca-bundle\") pod \"swift-ring-rebalance-48flt\" (UID: \"87e08198-d2d0-4608-87f3-02d832fe7c55\") " pod="openstack/swift-ring-rebalance-48flt" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.812371 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/87e08198-d2d0-4608-87f3-02d832fe7c55-swiftconf\") pod \"swift-ring-rebalance-48flt\" (UID: \"87e08198-d2d0-4608-87f3-02d832fe7c55\") " pod="openstack/swift-ring-rebalance-48flt" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.813330 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/87e08198-d2d0-4608-87f3-02d832fe7c55-dispersionconf\") pod \"swift-ring-rebalance-48flt\" (UID: \"87e08198-d2d0-4608-87f3-02d832fe7c55\") " pod="openstack/swift-ring-rebalance-48flt" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.813636 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-485t4\" (UniqueName: \"kubernetes.io/projected/87e08198-d2d0-4608-87f3-02d832fe7c55-kube-api-access-485t4\") pod \"swift-ring-rebalance-48flt\" (UID: \"87e08198-d2d0-4608-87f3-02d832fe7c55\") " pod="openstack/swift-ring-rebalance-48flt" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.835867 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.861835 4790 scope.go:117] "RemoveContainer" containerID="0735e2434710622ef854e0dbf6a71c0acc25cea40f272117c0ed5ab3006d8ddd" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.920179 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-q9w8p"] Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.932386 4790 scope.go:117] "RemoveContainer" containerID="2029256ebcfff8392e758225b7868aee953c152e054e7a004fb3794dbd8c5661" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.932768 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-q9w8p"] Nov 24 13:32:22 crc kubenswrapper[4790]: E1124 13:32:22.932817 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2029256ebcfff8392e758225b7868aee953c152e054e7a004fb3794dbd8c5661\": container with ID starting with 2029256ebcfff8392e758225b7868aee953c152e054e7a004fb3794dbd8c5661 not found: ID does not exist" containerID="2029256ebcfff8392e758225b7868aee953c152e054e7a004fb3794dbd8c5661" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.932864 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2029256ebcfff8392e758225b7868aee953c152e054e7a004fb3794dbd8c5661"} err="failed to get container status \"2029256ebcfff8392e758225b7868aee953c152e054e7a004fb3794dbd8c5661\": rpc error: code = NotFound desc = could not find container \"2029256ebcfff8392e758225b7868aee953c152e054e7a004fb3794dbd8c5661\": container with ID starting with 2029256ebcfff8392e758225b7868aee953c152e054e7a004fb3794dbd8c5661 not found: ID does not exist" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.932955 4790 scope.go:117] "RemoveContainer" containerID="0735e2434710622ef854e0dbf6a71c0acc25cea40f272117c0ed5ab3006d8ddd" Nov 24 13:32:22 crc kubenswrapper[4790]: E1124 13:32:22.933222 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0735e2434710622ef854e0dbf6a71c0acc25cea40f272117c0ed5ab3006d8ddd\": container with ID starting with 0735e2434710622ef854e0dbf6a71c0acc25cea40f272117c0ed5ab3006d8ddd not found: ID does not exist" containerID="0735e2434710622ef854e0dbf6a71c0acc25cea40f272117c0ed5ab3006d8ddd" Nov 24 13:32:22 crc kubenswrapper[4790]: I1124 13:32:22.933248 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0735e2434710622ef854e0dbf6a71c0acc25cea40f272117c0ed5ab3006d8ddd"} err="failed to get container status \"0735e2434710622ef854e0dbf6a71c0acc25cea40f272117c0ed5ab3006d8ddd\": rpc error: code = NotFound desc = could not find container \"0735e2434710622ef854e0dbf6a71c0acc25cea40f272117c0ed5ab3006d8ddd\": container with ID starting with 0735e2434710622ef854e0dbf6a71c0acc25cea40f272117c0ed5ab3006d8ddd not found: ID does not exist" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.006662 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.013028 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-48flt" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.074390 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.100155 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/63ce5126-cdab-4cdc-968a-acc26c6caf87-etc-swift\") pod \"swift-storage-0\" (UID: \"63ce5126-cdab-4cdc-968a-acc26c6caf87\") " pod="openstack/swift-storage-0" Nov 24 13:32:23 crc kubenswrapper[4790]: E1124 13:32:23.100361 4790 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 13:32:23 crc kubenswrapper[4790]: E1124 13:32:23.100377 4790 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 13:32:23 crc kubenswrapper[4790]: E1124 13:32:23.100419 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/63ce5126-cdab-4cdc-968a-acc26c6caf87-etc-swift podName:63ce5126-cdab-4cdc-968a-acc26c6caf87 nodeName:}" failed. No retries permitted until 2025-11-24 13:32:24.10040597 +0000 UTC m=+1192.480299622 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/63ce5126-cdab-4cdc-968a-acc26c6caf87-etc-swift") pod "swift-storage-0" (UID: "63ce5126-cdab-4cdc-968a-acc26c6caf87") : configmap "swift-ring-files" not found Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.114768 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-q8jn7"] Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.135328 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8cc7fc4dc-5d55x"] Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.140658 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8cc7fc4dc-5d55x" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.146558 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8cc7fc4dc-5d55x"] Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.148841 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.206420 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81f66d48-47b6-4a1c-afd1-12ce651dda1f-dns-svc\") pod \"dnsmasq-dns-8cc7fc4dc-5d55x\" (UID: \"81f66d48-47b6-4a1c-afd1-12ce651dda1f\") " pod="openstack/dnsmasq-dns-8cc7fc4dc-5d55x" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.206467 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49vwz\" (UniqueName: \"kubernetes.io/projected/81f66d48-47b6-4a1c-afd1-12ce651dda1f-kube-api-access-49vwz\") pod \"dnsmasq-dns-8cc7fc4dc-5d55x\" (UID: \"81f66d48-47b6-4a1c-afd1-12ce651dda1f\") " pod="openstack/dnsmasq-dns-8cc7fc4dc-5d55x" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.206542 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81f66d48-47b6-4a1c-afd1-12ce651dda1f-config\") pod \"dnsmasq-dns-8cc7fc4dc-5d55x\" (UID: \"81f66d48-47b6-4a1c-afd1-12ce651dda1f\") " pod="openstack/dnsmasq-dns-8cc7fc4dc-5d55x" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.206986 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/81f66d48-47b6-4a1c-afd1-12ce651dda1f-ovsdbserver-sb\") pod \"dnsmasq-dns-8cc7fc4dc-5d55x\" (UID: \"81f66d48-47b6-4a1c-afd1-12ce651dda1f\") " pod="openstack/dnsmasq-dns-8cc7fc4dc-5d55x" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.217451 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-z8gjd"] Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.219081 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-z8gjd" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.224374 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.237802 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-z8gjd"] Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.309225 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9c5a116-6438-47b5-8462-207eb656b553-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-z8gjd\" (UID: \"a9c5a116-6438-47b5-8462-207eb656b553\") " pod="openstack/ovn-controller-metrics-z8gjd" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.309529 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9c5a116-6438-47b5-8462-207eb656b553-config\") pod \"ovn-controller-metrics-z8gjd\" (UID: \"a9c5a116-6438-47b5-8462-207eb656b553\") " pod="openstack/ovn-controller-metrics-z8gjd" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.309580 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/81f66d48-47b6-4a1c-afd1-12ce651dda1f-ovsdbserver-sb\") pod \"dnsmasq-dns-8cc7fc4dc-5d55x\" (UID: \"81f66d48-47b6-4a1c-afd1-12ce651dda1f\") " pod="openstack/dnsmasq-dns-8cc7fc4dc-5d55x" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.309615 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbtgd\" (UniqueName: \"kubernetes.io/projected/a9c5a116-6438-47b5-8462-207eb656b553-kube-api-access-kbtgd\") pod \"ovn-controller-metrics-z8gjd\" (UID: \"a9c5a116-6438-47b5-8462-207eb656b553\") " pod="openstack/ovn-controller-metrics-z8gjd" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.309694 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81f66d48-47b6-4a1c-afd1-12ce651dda1f-dns-svc\") pod \"dnsmasq-dns-8cc7fc4dc-5d55x\" (UID: \"81f66d48-47b6-4a1c-afd1-12ce651dda1f\") " pod="openstack/dnsmasq-dns-8cc7fc4dc-5d55x" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.309728 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49vwz\" (UniqueName: \"kubernetes.io/projected/81f66d48-47b6-4a1c-afd1-12ce651dda1f-kube-api-access-49vwz\") pod \"dnsmasq-dns-8cc7fc4dc-5d55x\" (UID: \"81f66d48-47b6-4a1c-afd1-12ce651dda1f\") " pod="openstack/dnsmasq-dns-8cc7fc4dc-5d55x" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.309759 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a9c5a116-6438-47b5-8462-207eb656b553-ovn-rundir\") pod \"ovn-controller-metrics-z8gjd\" (UID: \"a9c5a116-6438-47b5-8462-207eb656b553\") " pod="openstack/ovn-controller-metrics-z8gjd" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.309782 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a9c5a116-6438-47b5-8462-207eb656b553-ovs-rundir\") pod \"ovn-controller-metrics-z8gjd\" (UID: \"a9c5a116-6438-47b5-8462-207eb656b553\") " pod="openstack/ovn-controller-metrics-z8gjd" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.309815 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9c5a116-6438-47b5-8462-207eb656b553-combined-ca-bundle\") pod \"ovn-controller-metrics-z8gjd\" (UID: \"a9c5a116-6438-47b5-8462-207eb656b553\") " pod="openstack/ovn-controller-metrics-z8gjd" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.309933 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81f66d48-47b6-4a1c-afd1-12ce651dda1f-config\") pod \"dnsmasq-dns-8cc7fc4dc-5d55x\" (UID: \"81f66d48-47b6-4a1c-afd1-12ce651dda1f\") " pod="openstack/dnsmasq-dns-8cc7fc4dc-5d55x" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.310584 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81f66d48-47b6-4a1c-afd1-12ce651dda1f-dns-svc\") pod \"dnsmasq-dns-8cc7fc4dc-5d55x\" (UID: \"81f66d48-47b6-4a1c-afd1-12ce651dda1f\") " pod="openstack/dnsmasq-dns-8cc7fc4dc-5d55x" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.310849 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/81f66d48-47b6-4a1c-afd1-12ce651dda1f-ovsdbserver-sb\") pod \"dnsmasq-dns-8cc7fc4dc-5d55x\" (UID: \"81f66d48-47b6-4a1c-afd1-12ce651dda1f\") " pod="openstack/dnsmasq-dns-8cc7fc4dc-5d55x" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.310865 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81f66d48-47b6-4a1c-afd1-12ce651dda1f-config\") pod \"dnsmasq-dns-8cc7fc4dc-5d55x\" (UID: \"81f66d48-47b6-4a1c-afd1-12ce651dda1f\") " pod="openstack/dnsmasq-dns-8cc7fc4dc-5d55x" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.328655 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49vwz\" (UniqueName: \"kubernetes.io/projected/81f66d48-47b6-4a1c-afd1-12ce651dda1f-kube-api-access-49vwz\") pod \"dnsmasq-dns-8cc7fc4dc-5d55x\" (UID: \"81f66d48-47b6-4a1c-afd1-12ce651dda1f\") " pod="openstack/dnsmasq-dns-8cc7fc4dc-5d55x" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.411069 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9c5a116-6438-47b5-8462-207eb656b553-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-z8gjd\" (UID: \"a9c5a116-6438-47b5-8462-207eb656b553\") " pod="openstack/ovn-controller-metrics-z8gjd" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.411120 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9c5a116-6438-47b5-8462-207eb656b553-config\") pod \"ovn-controller-metrics-z8gjd\" (UID: \"a9c5a116-6438-47b5-8462-207eb656b553\") " pod="openstack/ovn-controller-metrics-z8gjd" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.411165 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbtgd\" (UniqueName: \"kubernetes.io/projected/a9c5a116-6438-47b5-8462-207eb656b553-kube-api-access-kbtgd\") pod \"ovn-controller-metrics-z8gjd\" (UID: \"a9c5a116-6438-47b5-8462-207eb656b553\") " pod="openstack/ovn-controller-metrics-z8gjd" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.411224 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a9c5a116-6438-47b5-8462-207eb656b553-ovn-rundir\") pod \"ovn-controller-metrics-z8gjd\" (UID: \"a9c5a116-6438-47b5-8462-207eb656b553\") " pod="openstack/ovn-controller-metrics-z8gjd" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.411249 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a9c5a116-6438-47b5-8462-207eb656b553-ovs-rundir\") pod \"ovn-controller-metrics-z8gjd\" (UID: \"a9c5a116-6438-47b5-8462-207eb656b553\") " pod="openstack/ovn-controller-metrics-z8gjd" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.411265 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9c5a116-6438-47b5-8462-207eb656b553-combined-ca-bundle\") pod \"ovn-controller-metrics-z8gjd\" (UID: \"a9c5a116-6438-47b5-8462-207eb656b553\") " pod="openstack/ovn-controller-metrics-z8gjd" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.413010 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a9c5a116-6438-47b5-8462-207eb656b553-ovs-rundir\") pod \"ovn-controller-metrics-z8gjd\" (UID: \"a9c5a116-6438-47b5-8462-207eb656b553\") " pod="openstack/ovn-controller-metrics-z8gjd" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.413010 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a9c5a116-6438-47b5-8462-207eb656b553-ovn-rundir\") pod \"ovn-controller-metrics-z8gjd\" (UID: \"a9c5a116-6438-47b5-8462-207eb656b553\") " pod="openstack/ovn-controller-metrics-z8gjd" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.413962 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9c5a116-6438-47b5-8462-207eb656b553-config\") pod \"ovn-controller-metrics-z8gjd\" (UID: \"a9c5a116-6438-47b5-8462-207eb656b553\") " pod="openstack/ovn-controller-metrics-z8gjd" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.415411 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9c5a116-6438-47b5-8462-207eb656b553-combined-ca-bundle\") pod \"ovn-controller-metrics-z8gjd\" (UID: \"a9c5a116-6438-47b5-8462-207eb656b553\") " pod="openstack/ovn-controller-metrics-z8gjd" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.416030 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9c5a116-6438-47b5-8462-207eb656b553-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-z8gjd\" (UID: \"a9c5a116-6438-47b5-8462-207eb656b553\") " pod="openstack/ovn-controller-metrics-z8gjd" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.428676 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbtgd\" (UniqueName: \"kubernetes.io/projected/a9c5a116-6438-47b5-8462-207eb656b553-kube-api-access-kbtgd\") pod \"ovn-controller-metrics-z8gjd\" (UID: \"a9c5a116-6438-47b5-8462-207eb656b553\") " pod="openstack/ovn-controller-metrics-z8gjd" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.463257 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8cc7fc4dc-5d55x" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.538504 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-z8gjd" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.558349 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-4d4ck"] Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.567391 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-48flt"] Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.601157 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-b4bt2"] Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.602683 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-b4bt2" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.605305 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.619914 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-b4bt2"] Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.715526 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5bde9a3b-eafa-454f-b6f2-af094d594752-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-b4bt2\" (UID: \"5bde9a3b-eafa-454f-b6f2-af094d594752\") " pod="openstack/dnsmasq-dns-b8fbc5445-b4bt2" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.715630 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wm5qx\" (UniqueName: \"kubernetes.io/projected/5bde9a3b-eafa-454f-b6f2-af094d594752-kube-api-access-wm5qx\") pod \"dnsmasq-dns-b8fbc5445-b4bt2\" (UID: \"5bde9a3b-eafa-454f-b6f2-af094d594752\") " pod="openstack/dnsmasq-dns-b8fbc5445-b4bt2" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.715652 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bde9a3b-eafa-454f-b6f2-af094d594752-config\") pod \"dnsmasq-dns-b8fbc5445-b4bt2\" (UID: \"5bde9a3b-eafa-454f-b6f2-af094d594752\") " pod="openstack/dnsmasq-dns-b8fbc5445-b4bt2" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.715768 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5bde9a3b-eafa-454f-b6f2-af094d594752-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-b4bt2\" (UID: \"5bde9a3b-eafa-454f-b6f2-af094d594752\") " pod="openstack/dnsmasq-dns-b8fbc5445-b4bt2" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.715828 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5bde9a3b-eafa-454f-b6f2-af094d594752-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-b4bt2\" (UID: \"5bde9a3b-eafa-454f-b6f2-af094d594752\") " pod="openstack/dnsmasq-dns-b8fbc5445-b4bt2" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.817867 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5bde9a3b-eafa-454f-b6f2-af094d594752-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-b4bt2\" (UID: \"5bde9a3b-eafa-454f-b6f2-af094d594752\") " pod="openstack/dnsmasq-dns-b8fbc5445-b4bt2" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.817975 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5bde9a3b-eafa-454f-b6f2-af094d594752-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-b4bt2\" (UID: \"5bde9a3b-eafa-454f-b6f2-af094d594752\") " pod="openstack/dnsmasq-dns-b8fbc5445-b4bt2" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.818008 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5bde9a3b-eafa-454f-b6f2-af094d594752-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-b4bt2\" (UID: \"5bde9a3b-eafa-454f-b6f2-af094d594752\") " pod="openstack/dnsmasq-dns-b8fbc5445-b4bt2" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.818061 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wm5qx\" (UniqueName: \"kubernetes.io/projected/5bde9a3b-eafa-454f-b6f2-af094d594752-kube-api-access-wm5qx\") pod \"dnsmasq-dns-b8fbc5445-b4bt2\" (UID: \"5bde9a3b-eafa-454f-b6f2-af094d594752\") " pod="openstack/dnsmasq-dns-b8fbc5445-b4bt2" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.818086 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bde9a3b-eafa-454f-b6f2-af094d594752-config\") pod \"dnsmasq-dns-b8fbc5445-b4bt2\" (UID: \"5bde9a3b-eafa-454f-b6f2-af094d594752\") " pod="openstack/dnsmasq-dns-b8fbc5445-b4bt2" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.818760 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5bde9a3b-eafa-454f-b6f2-af094d594752-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-b4bt2\" (UID: \"5bde9a3b-eafa-454f-b6f2-af094d594752\") " pod="openstack/dnsmasq-dns-b8fbc5445-b4bt2" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.819102 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bde9a3b-eafa-454f-b6f2-af094d594752-config\") pod \"dnsmasq-dns-b8fbc5445-b4bt2\" (UID: \"5bde9a3b-eafa-454f-b6f2-af094d594752\") " pod="openstack/dnsmasq-dns-b8fbc5445-b4bt2" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.819305 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5bde9a3b-eafa-454f-b6f2-af094d594752-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-b4bt2\" (UID: \"5bde9a3b-eafa-454f-b6f2-af094d594752\") " pod="openstack/dnsmasq-dns-b8fbc5445-b4bt2" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.819519 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-48flt" event={"ID":"87e08198-d2d0-4608-87f3-02d832fe7c55","Type":"ContainerStarted","Data":"cf9fcff8296cbb4ced33aad8d54fb43654949ffd0d30ffca394d175e5e4723f0"} Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.819935 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5bde9a3b-eafa-454f-b6f2-af094d594752-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-b4bt2\" (UID: \"5bde9a3b-eafa-454f-b6f2-af094d594752\") " pod="openstack/dnsmasq-dns-b8fbc5445-b4bt2" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.822551 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-q8jn7" podUID="dc896e16-9261-443a-86d7-c45c2fbef3d8" containerName="dnsmasq-dns" containerID="cri-o://5b1ec6ba7c1716aedfa3c3f6c36a931e3a92d1ced968935496c33bcf7abe14d7" gracePeriod=10 Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.823402 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-4d4ck" event={"ID":"f924c39a-8259-45d6-aa00-4cc41f14e7ee","Type":"ContainerStarted","Data":"e2a955b6a87db6036e41d12bbbbf6162cfd89d2b0bb6d235a325aafe02fc51ca"} Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.823429 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cb5889db5-4d4ck" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.853455 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wm5qx\" (UniqueName: \"kubernetes.io/projected/5bde9a3b-eafa-454f-b6f2-af094d594752-kube-api-access-wm5qx\") pod \"dnsmasq-dns-b8fbc5445-b4bt2\" (UID: \"5bde9a3b-eafa-454f-b6f2-af094d594752\") " pod="openstack/dnsmasq-dns-b8fbc5445-b4bt2" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.858220 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cb5889db5-4d4ck" podStartSLOduration=2.85820359 podStartE2EDuration="2.85820359s" podCreationTimestamp="2025-11-24 13:32:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:32:23.845038795 +0000 UTC m=+1192.224932457" watchObservedRunningTime="2025-11-24 13:32:23.85820359 +0000 UTC m=+1192.238097252" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.880359 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:23 crc kubenswrapper[4790]: I1124 13:32:23.979508 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8cc7fc4dc-5d55x"] Nov 24 13:32:23 crc kubenswrapper[4790]: W1124 13:32:23.985412 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81f66d48_47b6_4a1c_afd1_12ce651dda1f.slice/crio-edb7a9441d0c953eaf381d64c78881b291536f48e18dc01c346fe734a68b6c1a WatchSource:0}: Error finding container edb7a9441d0c953eaf381d64c78881b291536f48e18dc01c346fe734a68b6c1a: Status 404 returned error can't find the container with id edb7a9441d0c953eaf381d64c78881b291536f48e18dc01c346fe734a68b6c1a Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.002025 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-b4bt2" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.128389 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/63ce5126-cdab-4cdc-968a-acc26c6caf87-etc-swift\") pod \"swift-storage-0\" (UID: \"63ce5126-cdab-4cdc-968a-acc26c6caf87\") " pod="openstack/swift-storage-0" Nov 24 13:32:24 crc kubenswrapper[4790]: E1124 13:32:24.128518 4790 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 13:32:24 crc kubenswrapper[4790]: E1124 13:32:24.128984 4790 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 13:32:24 crc kubenswrapper[4790]: E1124 13:32:24.129054 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/63ce5126-cdab-4cdc-968a-acc26c6caf87-etc-swift podName:63ce5126-cdab-4cdc-968a-acc26c6caf87 nodeName:}" failed. No retries permitted until 2025-11-24 13:32:26.129034209 +0000 UTC m=+1194.508927871 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/63ce5126-cdab-4cdc-968a-acc26c6caf87-etc-swift") pod "swift-storage-0" (UID: "63ce5126-cdab-4cdc-968a-acc26c6caf87") : configmap "swift-ring-files" not found Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.143625 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.144865 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.148352 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.148532 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-8n8hr" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.150651 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.151182 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.176799 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.201338 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-z8gjd"] Nov 24 13:32:24 crc kubenswrapper[4790]: W1124 13:32:24.202500 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda9c5a116_6438_47b5_8462_207eb656b553.slice/crio-4822f6e99ac9559176292b80d7e95de938d66bc14872d70aaf7e01d84ccedf49 WatchSource:0}: Error finding container 4822f6e99ac9559176292b80d7e95de938d66bc14872d70aaf7e01d84ccedf49: Status 404 returned error can't find the container with id 4822f6e99ac9559176292b80d7e95de938d66bc14872d70aaf7e01d84ccedf49 Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.229285 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-q8jn7" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.230579 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3fa97ac-7189-4828-8b95-664e852dea8e-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a3fa97ac-7189-4828-8b95-664e852dea8e\") " pod="openstack/ovn-northd-0" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.230614 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a3fa97ac-7189-4828-8b95-664e852dea8e-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a3fa97ac-7189-4828-8b95-664e852dea8e\") " pod="openstack/ovn-northd-0" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.230647 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a3fa97ac-7189-4828-8b95-664e852dea8e-scripts\") pod \"ovn-northd-0\" (UID: \"a3fa97ac-7189-4828-8b95-664e852dea8e\") " pod="openstack/ovn-northd-0" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.230690 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96dm4\" (UniqueName: \"kubernetes.io/projected/a3fa97ac-7189-4828-8b95-664e852dea8e-kube-api-access-96dm4\") pod \"ovn-northd-0\" (UID: \"a3fa97ac-7189-4828-8b95-664e852dea8e\") " pod="openstack/ovn-northd-0" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.230736 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3fa97ac-7189-4828-8b95-664e852dea8e-config\") pod \"ovn-northd-0\" (UID: \"a3fa97ac-7189-4828-8b95-664e852dea8e\") " pod="openstack/ovn-northd-0" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.230753 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3fa97ac-7189-4828-8b95-664e852dea8e-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a3fa97ac-7189-4828-8b95-664e852dea8e\") " pod="openstack/ovn-northd-0" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.230775 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3fa97ac-7189-4828-8b95-664e852dea8e-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a3fa97ac-7189-4828-8b95-664e852dea8e\") " pod="openstack/ovn-northd-0" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.290246 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-b4bt2"] Nov 24 13:32:24 crc kubenswrapper[4790]: W1124 13:32:24.297061 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5bde9a3b_eafa_454f_b6f2_af094d594752.slice/crio-565093f3d6e5aae5cfe76c48ac351d7692a9092157b87fc00681a0de3cccd850 WatchSource:0}: Error finding container 565093f3d6e5aae5cfe76c48ac351d7692a9092157b87fc00681a0de3cccd850: Status 404 returned error can't find the container with id 565093f3d6e5aae5cfe76c48ac351d7692a9092157b87fc00681a0de3cccd850 Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.324652 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7fcc2c7-c415-4bd7-9b8a-367d9292985d" path="/var/lib/kubelet/pods/d7fcc2c7-c415-4bd7-9b8a-367d9292985d/volumes" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.331480 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc896e16-9261-443a-86d7-c45c2fbef3d8-config\") pod \"dc896e16-9261-443a-86d7-c45c2fbef3d8\" (UID: \"dc896e16-9261-443a-86d7-c45c2fbef3d8\") " Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.331583 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc896e16-9261-443a-86d7-c45c2fbef3d8-dns-svc\") pod \"dc896e16-9261-443a-86d7-c45c2fbef3d8\" (UID: \"dc896e16-9261-443a-86d7-c45c2fbef3d8\") " Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.331728 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8trg\" (UniqueName: \"kubernetes.io/projected/dc896e16-9261-443a-86d7-c45c2fbef3d8-kube-api-access-k8trg\") pod \"dc896e16-9261-443a-86d7-c45c2fbef3d8\" (UID: \"dc896e16-9261-443a-86d7-c45c2fbef3d8\") " Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.332005 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3fa97ac-7189-4828-8b95-664e852dea8e-config\") pod \"ovn-northd-0\" (UID: \"a3fa97ac-7189-4828-8b95-664e852dea8e\") " pod="openstack/ovn-northd-0" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.332064 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3fa97ac-7189-4828-8b95-664e852dea8e-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a3fa97ac-7189-4828-8b95-664e852dea8e\") " pod="openstack/ovn-northd-0" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.332089 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3fa97ac-7189-4828-8b95-664e852dea8e-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a3fa97ac-7189-4828-8b95-664e852dea8e\") " pod="openstack/ovn-northd-0" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.332139 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3fa97ac-7189-4828-8b95-664e852dea8e-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a3fa97ac-7189-4828-8b95-664e852dea8e\") " pod="openstack/ovn-northd-0" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.332167 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a3fa97ac-7189-4828-8b95-664e852dea8e-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a3fa97ac-7189-4828-8b95-664e852dea8e\") " pod="openstack/ovn-northd-0" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.332244 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a3fa97ac-7189-4828-8b95-664e852dea8e-scripts\") pod \"ovn-northd-0\" (UID: \"a3fa97ac-7189-4828-8b95-664e852dea8e\") " pod="openstack/ovn-northd-0" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.332329 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96dm4\" (UniqueName: \"kubernetes.io/projected/a3fa97ac-7189-4828-8b95-664e852dea8e-kube-api-access-96dm4\") pod \"ovn-northd-0\" (UID: \"a3fa97ac-7189-4828-8b95-664e852dea8e\") " pod="openstack/ovn-northd-0" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.333368 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a3fa97ac-7189-4828-8b95-664e852dea8e-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a3fa97ac-7189-4828-8b95-664e852dea8e\") " pod="openstack/ovn-northd-0" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.335332 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc896e16-9261-443a-86d7-c45c2fbef3d8-kube-api-access-k8trg" (OuterVolumeSpecName: "kube-api-access-k8trg") pod "dc896e16-9261-443a-86d7-c45c2fbef3d8" (UID: "dc896e16-9261-443a-86d7-c45c2fbef3d8"). InnerVolumeSpecName "kube-api-access-k8trg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.335654 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a3fa97ac-7189-4828-8b95-664e852dea8e-scripts\") pod \"ovn-northd-0\" (UID: \"a3fa97ac-7189-4828-8b95-664e852dea8e\") " pod="openstack/ovn-northd-0" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.337387 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3fa97ac-7189-4828-8b95-664e852dea8e-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a3fa97ac-7189-4828-8b95-664e852dea8e\") " pod="openstack/ovn-northd-0" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.337473 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3fa97ac-7189-4828-8b95-664e852dea8e-config\") pod \"ovn-northd-0\" (UID: \"a3fa97ac-7189-4828-8b95-664e852dea8e\") " pod="openstack/ovn-northd-0" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.339388 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3fa97ac-7189-4828-8b95-664e852dea8e-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a3fa97ac-7189-4828-8b95-664e852dea8e\") " pod="openstack/ovn-northd-0" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.340251 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3fa97ac-7189-4828-8b95-664e852dea8e-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a3fa97ac-7189-4828-8b95-664e852dea8e\") " pod="openstack/ovn-northd-0" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.348805 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96dm4\" (UniqueName: \"kubernetes.io/projected/a3fa97ac-7189-4828-8b95-664e852dea8e-kube-api-access-96dm4\") pod \"ovn-northd-0\" (UID: \"a3fa97ac-7189-4828-8b95-664e852dea8e\") " pod="openstack/ovn-northd-0" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.378205 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc896e16-9261-443a-86d7-c45c2fbef3d8-config" (OuterVolumeSpecName: "config") pod "dc896e16-9261-443a-86d7-c45c2fbef3d8" (UID: "dc896e16-9261-443a-86d7-c45c2fbef3d8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.384807 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc896e16-9261-443a-86d7-c45c2fbef3d8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dc896e16-9261-443a-86d7-c45c2fbef3d8" (UID: "dc896e16-9261-443a-86d7-c45c2fbef3d8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.434223 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8trg\" (UniqueName: \"kubernetes.io/projected/dc896e16-9261-443a-86d7-c45c2fbef3d8-kube-api-access-k8trg\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.434260 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc896e16-9261-443a-86d7-c45c2fbef3d8-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.434274 4790 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc896e16-9261-443a-86d7-c45c2fbef3d8-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.478039 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.842767 4790 generic.go:334] "Generic (PLEG): container finished" podID="5bde9a3b-eafa-454f-b6f2-af094d594752" containerID="78514cecd06fc8a736fc2455d2533f997ed9388d52795cf38e89cb7ee03442f8" exitCode=0 Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.842856 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-b4bt2" event={"ID":"5bde9a3b-eafa-454f-b6f2-af094d594752","Type":"ContainerDied","Data":"78514cecd06fc8a736fc2455d2533f997ed9388d52795cf38e89cb7ee03442f8"} Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.843142 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-b4bt2" event={"ID":"5bde9a3b-eafa-454f-b6f2-af094d594752","Type":"ContainerStarted","Data":"565093f3d6e5aae5cfe76c48ac351d7692a9092157b87fc00681a0de3cccd850"} Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.845646 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-z8gjd" event={"ID":"a9c5a116-6438-47b5-8462-207eb656b553","Type":"ContainerStarted","Data":"c6902baf408c3891cff01a43b41f9f037b6b29b99555835a2a635032bcee8c85"} Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.845694 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-z8gjd" event={"ID":"a9c5a116-6438-47b5-8462-207eb656b553","Type":"ContainerStarted","Data":"4822f6e99ac9559176292b80d7e95de938d66bc14872d70aaf7e01d84ccedf49"} Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.848252 4790 generic.go:334] "Generic (PLEG): container finished" podID="dc896e16-9261-443a-86d7-c45c2fbef3d8" containerID="5b1ec6ba7c1716aedfa3c3f6c36a931e3a92d1ced968935496c33bcf7abe14d7" exitCode=0 Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.848289 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-q8jn7" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.848359 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-q8jn7" event={"ID":"dc896e16-9261-443a-86d7-c45c2fbef3d8","Type":"ContainerDied","Data":"5b1ec6ba7c1716aedfa3c3f6c36a931e3a92d1ced968935496c33bcf7abe14d7"} Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.848387 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-q8jn7" event={"ID":"dc896e16-9261-443a-86d7-c45c2fbef3d8","Type":"ContainerDied","Data":"f22da277f6bb1ae7c6aaa1c1bbd36bd0193310ac36911af86462d2947ac445e5"} Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.848435 4790 scope.go:117] "RemoveContainer" containerID="5b1ec6ba7c1716aedfa3c3f6c36a931e3a92d1ced968935496c33bcf7abe14d7" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.874690 4790 generic.go:334] "Generic (PLEG): container finished" podID="81f66d48-47b6-4a1c-afd1-12ce651dda1f" containerID="36419dcb1166d10a4271c8f86e7b0e58aa5b800327848c11a2786cf78cbde7f7" exitCode=0 Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.876231 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8cc7fc4dc-5d55x" event={"ID":"81f66d48-47b6-4a1c-afd1-12ce651dda1f","Type":"ContainerDied","Data":"36419dcb1166d10a4271c8f86e7b0e58aa5b800327848c11a2786cf78cbde7f7"} Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.876289 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8cc7fc4dc-5d55x" event={"ID":"81f66d48-47b6-4a1c-afd1-12ce651dda1f","Type":"ContainerStarted","Data":"edb7a9441d0c953eaf381d64c78881b291536f48e18dc01c346fe734a68b6c1a"} Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.876735 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cb5889db5-4d4ck" podUID="f924c39a-8259-45d6-aa00-4cc41f14e7ee" containerName="dnsmasq-dns" containerID="cri-o://e2a955b6a87db6036e41d12bbbbf6162cfd89d2b0bb6d235a325aafe02fc51ca" gracePeriod=10 Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.888795 4790 scope.go:117] "RemoveContainer" containerID="6980ad4fe960e033c715bd3155792b9fc55f15acbcc5bf7afda27567238e7feb" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.896819 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-z8gjd" podStartSLOduration=1.896797485 podStartE2EDuration="1.896797485s" podCreationTimestamp="2025-11-24 13:32:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:32:24.883348852 +0000 UTC m=+1193.263242524" watchObservedRunningTime="2025-11-24 13:32:24.896797485 +0000 UTC m=+1193.276691147" Nov 24 13:32:24 crc kubenswrapper[4790]: I1124 13:32:24.976201 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 13:32:25 crc kubenswrapper[4790]: I1124 13:32:25.082668 4790 scope.go:117] "RemoveContainer" containerID="5b1ec6ba7c1716aedfa3c3f6c36a931e3a92d1ced968935496c33bcf7abe14d7" Nov 24 13:32:25 crc kubenswrapper[4790]: E1124 13:32:25.083605 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b1ec6ba7c1716aedfa3c3f6c36a931e3a92d1ced968935496c33bcf7abe14d7\": container with ID starting with 5b1ec6ba7c1716aedfa3c3f6c36a931e3a92d1ced968935496c33bcf7abe14d7 not found: ID does not exist" containerID="5b1ec6ba7c1716aedfa3c3f6c36a931e3a92d1ced968935496c33bcf7abe14d7" Nov 24 13:32:25 crc kubenswrapper[4790]: I1124 13:32:25.083650 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b1ec6ba7c1716aedfa3c3f6c36a931e3a92d1ced968935496c33bcf7abe14d7"} err="failed to get container status \"5b1ec6ba7c1716aedfa3c3f6c36a931e3a92d1ced968935496c33bcf7abe14d7\": rpc error: code = NotFound desc = could not find container \"5b1ec6ba7c1716aedfa3c3f6c36a931e3a92d1ced968935496c33bcf7abe14d7\": container with ID starting with 5b1ec6ba7c1716aedfa3c3f6c36a931e3a92d1ced968935496c33bcf7abe14d7 not found: ID does not exist" Nov 24 13:32:25 crc kubenswrapper[4790]: I1124 13:32:25.083681 4790 scope.go:117] "RemoveContainer" containerID="6980ad4fe960e033c715bd3155792b9fc55f15acbcc5bf7afda27567238e7feb" Nov 24 13:32:25 crc kubenswrapper[4790]: E1124 13:32:25.085403 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6980ad4fe960e033c715bd3155792b9fc55f15acbcc5bf7afda27567238e7feb\": container with ID starting with 6980ad4fe960e033c715bd3155792b9fc55f15acbcc5bf7afda27567238e7feb not found: ID does not exist" containerID="6980ad4fe960e033c715bd3155792b9fc55f15acbcc5bf7afda27567238e7feb" Nov 24 13:32:25 crc kubenswrapper[4790]: I1124 13:32:25.085430 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6980ad4fe960e033c715bd3155792b9fc55f15acbcc5bf7afda27567238e7feb"} err="failed to get container status \"6980ad4fe960e033c715bd3155792b9fc55f15acbcc5bf7afda27567238e7feb\": rpc error: code = NotFound desc = could not find container \"6980ad4fe960e033c715bd3155792b9fc55f15acbcc5bf7afda27567238e7feb\": container with ID starting with 6980ad4fe960e033c715bd3155792b9fc55f15acbcc5bf7afda27567238e7feb not found: ID does not exist" Nov 24 13:32:25 crc kubenswrapper[4790]: I1124 13:32:25.103861 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-q8jn7"] Nov 24 13:32:25 crc kubenswrapper[4790]: I1124 13:32:25.112678 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-q8jn7"] Nov 24 13:32:25 crc kubenswrapper[4790]: I1124 13:32:25.884607 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8cc7fc4dc-5d55x" event={"ID":"81f66d48-47b6-4a1c-afd1-12ce651dda1f","Type":"ContainerStarted","Data":"d1247e35b9ee7e3f73a2cb7b18e8833595b28040ca68121c03b967719171cc2c"} Nov 24 13:32:25 crc kubenswrapper[4790]: I1124 13:32:25.885047 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8cc7fc4dc-5d55x" Nov 24 13:32:25 crc kubenswrapper[4790]: I1124 13:32:25.890195 4790 generic.go:334] "Generic (PLEG): container finished" podID="f924c39a-8259-45d6-aa00-4cc41f14e7ee" containerID="e2a955b6a87db6036e41d12bbbbf6162cfd89d2b0bb6d235a325aafe02fc51ca" exitCode=0 Nov 24 13:32:25 crc kubenswrapper[4790]: I1124 13:32:25.890314 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-4d4ck" event={"ID":"f924c39a-8259-45d6-aa00-4cc41f14e7ee","Type":"ContainerDied","Data":"e2a955b6a87db6036e41d12bbbbf6162cfd89d2b0bb6d235a325aafe02fc51ca"} Nov 24 13:32:25 crc kubenswrapper[4790]: I1124 13:32:25.892213 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-b4bt2" event={"ID":"5bde9a3b-eafa-454f-b6f2-af094d594752","Type":"ContainerStarted","Data":"f33e9fc8e484ad15ca70bba7e44986f290e2e29ff65185a8d0930c4eb8085c4a"} Nov 24 13:32:25 crc kubenswrapper[4790]: I1124 13:32:25.892899 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-b4bt2" Nov 24 13:32:25 crc kubenswrapper[4790]: I1124 13:32:25.894301 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a3fa97ac-7189-4828-8b95-664e852dea8e","Type":"ContainerStarted","Data":"e10229356b9c9fecc124d25b09f6663da5e7a6d25dcf71f7791d4f45c918d2f0"} Nov 24 13:32:25 crc kubenswrapper[4790]: I1124 13:32:25.909081 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8cc7fc4dc-5d55x" podStartSLOduration=2.909063481 podStartE2EDuration="2.909063481s" podCreationTimestamp="2025-11-24 13:32:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:32:25.90327006 +0000 UTC m=+1194.283163722" watchObservedRunningTime="2025-11-24 13:32:25.909063481 +0000 UTC m=+1194.288957143" Nov 24 13:32:25 crc kubenswrapper[4790]: I1124 13:32:25.926378 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-b4bt2" podStartSLOduration=2.92636289 podStartE2EDuration="2.92636289s" podCreationTimestamp="2025-11-24 13:32:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:32:25.922663808 +0000 UTC m=+1194.302557480" watchObservedRunningTime="2025-11-24 13:32:25.92636289 +0000 UTC m=+1194.306256552" Nov 24 13:32:26 crc kubenswrapper[4790]: I1124 13:32:26.163429 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/63ce5126-cdab-4cdc-968a-acc26c6caf87-etc-swift\") pod \"swift-storage-0\" (UID: \"63ce5126-cdab-4cdc-968a-acc26c6caf87\") " pod="openstack/swift-storage-0" Nov 24 13:32:26 crc kubenswrapper[4790]: E1124 13:32:26.163650 4790 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 13:32:26 crc kubenswrapper[4790]: E1124 13:32:26.163668 4790 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 13:32:26 crc kubenswrapper[4790]: E1124 13:32:26.163730 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/63ce5126-cdab-4cdc-968a-acc26c6caf87-etc-swift podName:63ce5126-cdab-4cdc-968a-acc26c6caf87 nodeName:}" failed. No retries permitted until 2025-11-24 13:32:30.163711341 +0000 UTC m=+1198.543605003 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/63ce5126-cdab-4cdc-968a-acc26c6caf87-etc-swift") pod "swift-storage-0" (UID: "63ce5126-cdab-4cdc-968a-acc26c6caf87") : configmap "swift-ring-files" not found Nov 24 13:32:26 crc kubenswrapper[4790]: I1124 13:32:26.327988 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc896e16-9261-443a-86d7-c45c2fbef3d8" path="/var/lib/kubelet/pods/dc896e16-9261-443a-86d7-c45c2fbef3d8/volumes" Nov 24 13:32:26 crc kubenswrapper[4790]: I1124 13:32:26.907248 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-4d4ck" event={"ID":"f924c39a-8259-45d6-aa00-4cc41f14e7ee","Type":"ContainerDied","Data":"de4d1efcd3fe9c001bf095ca8a6688a900b6c37255742586bb026ec6ec054230"} Nov 24 13:32:26 crc kubenswrapper[4790]: I1124 13:32:26.907338 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de4d1efcd3fe9c001bf095ca8a6688a900b6c37255742586bb026ec6ec054230" Nov 24 13:32:26 crc kubenswrapper[4790]: I1124 13:32:26.963040 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-4d4ck" Nov 24 13:32:27 crc kubenswrapper[4790]: I1124 13:32:27.077999 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f924c39a-8259-45d6-aa00-4cc41f14e7ee-dns-svc\") pod \"f924c39a-8259-45d6-aa00-4cc41f14e7ee\" (UID: \"f924c39a-8259-45d6-aa00-4cc41f14e7ee\") " Nov 24 13:32:27 crc kubenswrapper[4790]: I1124 13:32:27.078546 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfztq\" (UniqueName: \"kubernetes.io/projected/f924c39a-8259-45d6-aa00-4cc41f14e7ee-kube-api-access-gfztq\") pod \"f924c39a-8259-45d6-aa00-4cc41f14e7ee\" (UID: \"f924c39a-8259-45d6-aa00-4cc41f14e7ee\") " Nov 24 13:32:27 crc kubenswrapper[4790]: I1124 13:32:27.078590 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f924c39a-8259-45d6-aa00-4cc41f14e7ee-config\") pod \"f924c39a-8259-45d6-aa00-4cc41f14e7ee\" (UID: \"f924c39a-8259-45d6-aa00-4cc41f14e7ee\") " Nov 24 13:32:27 crc kubenswrapper[4790]: I1124 13:32:27.089243 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f924c39a-8259-45d6-aa00-4cc41f14e7ee-kube-api-access-gfztq" (OuterVolumeSpecName: "kube-api-access-gfztq") pod "f924c39a-8259-45d6-aa00-4cc41f14e7ee" (UID: "f924c39a-8259-45d6-aa00-4cc41f14e7ee"). InnerVolumeSpecName "kube-api-access-gfztq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:32:27 crc kubenswrapper[4790]: I1124 13:32:27.130015 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f924c39a-8259-45d6-aa00-4cc41f14e7ee-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f924c39a-8259-45d6-aa00-4cc41f14e7ee" (UID: "f924c39a-8259-45d6-aa00-4cc41f14e7ee"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:27 crc kubenswrapper[4790]: I1124 13:32:27.149415 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f924c39a-8259-45d6-aa00-4cc41f14e7ee-config" (OuterVolumeSpecName: "config") pod "f924c39a-8259-45d6-aa00-4cc41f14e7ee" (UID: "f924c39a-8259-45d6-aa00-4cc41f14e7ee"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:27 crc kubenswrapper[4790]: I1124 13:32:27.180145 4790 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f924c39a-8259-45d6-aa00-4cc41f14e7ee-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:27 crc kubenswrapper[4790]: I1124 13:32:27.180173 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfztq\" (UniqueName: \"kubernetes.io/projected/f924c39a-8259-45d6-aa00-4cc41f14e7ee-kube-api-access-gfztq\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:27 crc kubenswrapper[4790]: I1124 13:32:27.180183 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f924c39a-8259-45d6-aa00-4cc41f14e7ee-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:27 crc kubenswrapper[4790]: I1124 13:32:27.915977 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-48flt" event={"ID":"87e08198-d2d0-4608-87f3-02d832fe7c55","Type":"ContainerStarted","Data":"2d459199980f3814665ba6c630a8bf310ef6efb41a3f2368c9edbeb004c62897"} Nov 24 13:32:27 crc kubenswrapper[4790]: I1124 13:32:27.918449 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-4d4ck" Nov 24 13:32:27 crc kubenswrapper[4790]: I1124 13:32:27.918475 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a3fa97ac-7189-4828-8b95-664e852dea8e","Type":"ContainerStarted","Data":"889043e225f3f968efc7dfcf48a4fcb3822a1e4d8cf106f1af9b36a1cda57399"} Nov 24 13:32:27 crc kubenswrapper[4790]: I1124 13:32:27.918529 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a3fa97ac-7189-4828-8b95-664e852dea8e","Type":"ContainerStarted","Data":"c71cf899ea7e4a2eff4caa8290abac4a40d0072f283f44d449955d488101621e"} Nov 24 13:32:27 crc kubenswrapper[4790]: I1124 13:32:27.934657 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-48flt" podStartSLOduration=2.597286532 podStartE2EDuration="5.934641841s" podCreationTimestamp="2025-11-24 13:32:22 +0000 UTC" firstStartedPulling="2025-11-24 13:32:23.613897137 +0000 UTC m=+1191.993790799" lastFinishedPulling="2025-11-24 13:32:26.951252426 +0000 UTC m=+1195.331146108" observedRunningTime="2025-11-24 13:32:27.934387274 +0000 UTC m=+1196.314280956" watchObservedRunningTime="2025-11-24 13:32:27.934641841 +0000 UTC m=+1196.314535503" Nov 24 13:32:27 crc kubenswrapper[4790]: I1124 13:32:27.955905 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.550422288 podStartE2EDuration="3.95587632s" podCreationTimestamp="2025-11-24 13:32:24 +0000 UTC" firstStartedPulling="2025-11-24 13:32:25.040978973 +0000 UTC m=+1193.420872635" lastFinishedPulling="2025-11-24 13:32:27.446433005 +0000 UTC m=+1195.826326667" observedRunningTime="2025-11-24 13:32:27.951988522 +0000 UTC m=+1196.331882184" watchObservedRunningTime="2025-11-24 13:32:27.95587632 +0000 UTC m=+1196.335769982" Nov 24 13:32:27 crc kubenswrapper[4790]: I1124 13:32:27.976207 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-4d4ck"] Nov 24 13:32:27 crc kubenswrapper[4790]: I1124 13:32:27.981630 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-4d4ck"] Nov 24 13:32:28 crc kubenswrapper[4790]: I1124 13:32:28.325464 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f924c39a-8259-45d6-aa00-4cc41f14e7ee" path="/var/lib/kubelet/pods/f924c39a-8259-45d6-aa00-4cc41f14e7ee/volumes" Nov 24 13:32:28 crc kubenswrapper[4790]: I1124 13:32:28.925606 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.183344 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6aea-account-create-hfz2v"] Nov 24 13:32:29 crc kubenswrapper[4790]: E1124 13:32:29.183696 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f924c39a-8259-45d6-aa00-4cc41f14e7ee" containerName="dnsmasq-dns" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.183708 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="f924c39a-8259-45d6-aa00-4cc41f14e7ee" containerName="dnsmasq-dns" Nov 24 13:32:29 crc kubenswrapper[4790]: E1124 13:32:29.183722 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc896e16-9261-443a-86d7-c45c2fbef3d8" containerName="dnsmasq-dns" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.183728 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc896e16-9261-443a-86d7-c45c2fbef3d8" containerName="dnsmasq-dns" Nov 24 13:32:29 crc kubenswrapper[4790]: E1124 13:32:29.183748 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f924c39a-8259-45d6-aa00-4cc41f14e7ee" containerName="init" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.183754 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="f924c39a-8259-45d6-aa00-4cc41f14e7ee" containerName="init" Nov 24 13:32:29 crc kubenswrapper[4790]: E1124 13:32:29.183769 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc896e16-9261-443a-86d7-c45c2fbef3d8" containerName="init" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.183775 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc896e16-9261-443a-86d7-c45c2fbef3d8" containerName="init" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.183951 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="f924c39a-8259-45d6-aa00-4cc41f14e7ee" containerName="dnsmasq-dns" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.183969 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc896e16-9261-443a-86d7-c45c2fbef3d8" containerName="dnsmasq-dns" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.184559 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6aea-account-create-hfz2v" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.186346 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.190194 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6aea-account-create-hfz2v"] Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.222112 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-xsphw"] Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.223133 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-xsphw" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.232064 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-xsphw"] Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.313209 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tc2ck\" (UniqueName: \"kubernetes.io/projected/8a30bb00-53b7-4128-954f-d5183fec7675-kube-api-access-tc2ck\") pod \"keystone-6aea-account-create-hfz2v\" (UID: \"8a30bb00-53b7-4128-954f-d5183fec7675\") " pod="openstack/keystone-6aea-account-create-hfz2v" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.313476 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a30bb00-53b7-4128-954f-d5183fec7675-operator-scripts\") pod \"keystone-6aea-account-create-hfz2v\" (UID: \"8a30bb00-53b7-4128-954f-d5183fec7675\") " pod="openstack/keystone-6aea-account-create-hfz2v" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.414737 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tc2ck\" (UniqueName: \"kubernetes.io/projected/8a30bb00-53b7-4128-954f-d5183fec7675-kube-api-access-tc2ck\") pod \"keystone-6aea-account-create-hfz2v\" (UID: \"8a30bb00-53b7-4128-954f-d5183fec7675\") " pod="openstack/keystone-6aea-account-create-hfz2v" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.414872 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a30bb00-53b7-4128-954f-d5183fec7675-operator-scripts\") pod \"keystone-6aea-account-create-hfz2v\" (UID: \"8a30bb00-53b7-4128-954f-d5183fec7675\") " pod="openstack/keystone-6aea-account-create-hfz2v" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.414960 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8zwb\" (UniqueName: \"kubernetes.io/projected/85671e80-3009-49dc-a308-e126b08847a2-kube-api-access-s8zwb\") pod \"keystone-db-create-xsphw\" (UID: \"85671e80-3009-49dc-a308-e126b08847a2\") " pod="openstack/keystone-db-create-xsphw" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.415084 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85671e80-3009-49dc-a308-e126b08847a2-operator-scripts\") pod \"keystone-db-create-xsphw\" (UID: \"85671e80-3009-49dc-a308-e126b08847a2\") " pod="openstack/keystone-db-create-xsphw" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.416796 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a30bb00-53b7-4128-954f-d5183fec7675-operator-scripts\") pod \"keystone-6aea-account-create-hfz2v\" (UID: \"8a30bb00-53b7-4128-954f-d5183fec7675\") " pod="openstack/keystone-6aea-account-create-hfz2v" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.434056 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tc2ck\" (UniqueName: \"kubernetes.io/projected/8a30bb00-53b7-4128-954f-d5183fec7675-kube-api-access-tc2ck\") pod \"keystone-6aea-account-create-hfz2v\" (UID: \"8a30bb00-53b7-4128-954f-d5183fec7675\") " pod="openstack/keystone-6aea-account-create-hfz2v" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.472735 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-67mw7"] Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.473964 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-67mw7" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.501744 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-67mw7"] Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.506603 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6aea-account-create-hfz2v" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.516686 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8zwb\" (UniqueName: \"kubernetes.io/projected/85671e80-3009-49dc-a308-e126b08847a2-kube-api-access-s8zwb\") pod \"keystone-db-create-xsphw\" (UID: \"85671e80-3009-49dc-a308-e126b08847a2\") " pod="openstack/keystone-db-create-xsphw" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.516818 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85671e80-3009-49dc-a308-e126b08847a2-operator-scripts\") pod \"keystone-db-create-xsphw\" (UID: \"85671e80-3009-49dc-a308-e126b08847a2\") " pod="openstack/keystone-db-create-xsphw" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.517911 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85671e80-3009-49dc-a308-e126b08847a2-operator-scripts\") pod \"keystone-db-create-xsphw\" (UID: \"85671e80-3009-49dc-a308-e126b08847a2\") " pod="openstack/keystone-db-create-xsphw" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.544771 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8zwb\" (UniqueName: \"kubernetes.io/projected/85671e80-3009-49dc-a308-e126b08847a2-kube-api-access-s8zwb\") pod \"keystone-db-create-xsphw\" (UID: \"85671e80-3009-49dc-a308-e126b08847a2\") " pod="openstack/keystone-db-create-xsphw" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.549690 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-xsphw" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.561306 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-0fc6-account-create-7zc7c"] Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.563639 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-0fc6-account-create-7zc7c" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.567216 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.583158 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-0fc6-account-create-7zc7c"] Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.618475 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee85107b-89f0-48c3-aecf-05295449f025-operator-scripts\") pod \"placement-db-create-67mw7\" (UID: \"ee85107b-89f0-48c3-aecf-05295449f025\") " pod="openstack/placement-db-create-67mw7" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.618541 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhrln\" (UniqueName: \"kubernetes.io/projected/ee85107b-89f0-48c3-aecf-05295449f025-kube-api-access-qhrln\") pod \"placement-db-create-67mw7\" (UID: \"ee85107b-89f0-48c3-aecf-05295449f025\") " pod="openstack/placement-db-create-67mw7" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.721273 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/967da6fe-b0e0-4559-8db1-eea829746d13-operator-scripts\") pod \"placement-0fc6-account-create-7zc7c\" (UID: \"967da6fe-b0e0-4559-8db1-eea829746d13\") " pod="openstack/placement-0fc6-account-create-7zc7c" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.721387 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee85107b-89f0-48c3-aecf-05295449f025-operator-scripts\") pod \"placement-db-create-67mw7\" (UID: \"ee85107b-89f0-48c3-aecf-05295449f025\") " pod="openstack/placement-db-create-67mw7" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.721454 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sp6rj\" (UniqueName: \"kubernetes.io/projected/967da6fe-b0e0-4559-8db1-eea829746d13-kube-api-access-sp6rj\") pod \"placement-0fc6-account-create-7zc7c\" (UID: \"967da6fe-b0e0-4559-8db1-eea829746d13\") " pod="openstack/placement-0fc6-account-create-7zc7c" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.721489 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhrln\" (UniqueName: \"kubernetes.io/projected/ee85107b-89f0-48c3-aecf-05295449f025-kube-api-access-qhrln\") pod \"placement-db-create-67mw7\" (UID: \"ee85107b-89f0-48c3-aecf-05295449f025\") " pod="openstack/placement-db-create-67mw7" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.722525 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee85107b-89f0-48c3-aecf-05295449f025-operator-scripts\") pod \"placement-db-create-67mw7\" (UID: \"ee85107b-89f0-48c3-aecf-05295449f025\") " pod="openstack/placement-db-create-67mw7" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.763108 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhrln\" (UniqueName: \"kubernetes.io/projected/ee85107b-89f0-48c3-aecf-05295449f025-kube-api-access-qhrln\") pod \"placement-db-create-67mw7\" (UID: \"ee85107b-89f0-48c3-aecf-05295449f025\") " pod="openstack/placement-db-create-67mw7" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.801045 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-ff7zf"] Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.802210 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ff7zf" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.810636 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-67mw7" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.811556 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-ff7zf"] Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.822947 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/967da6fe-b0e0-4559-8db1-eea829746d13-operator-scripts\") pod \"placement-0fc6-account-create-7zc7c\" (UID: \"967da6fe-b0e0-4559-8db1-eea829746d13\") " pod="openstack/placement-0fc6-account-create-7zc7c" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.823025 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sp6rj\" (UniqueName: \"kubernetes.io/projected/967da6fe-b0e0-4559-8db1-eea829746d13-kube-api-access-sp6rj\") pod \"placement-0fc6-account-create-7zc7c\" (UID: \"967da6fe-b0e0-4559-8db1-eea829746d13\") " pod="openstack/placement-0fc6-account-create-7zc7c" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.823849 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/967da6fe-b0e0-4559-8db1-eea829746d13-operator-scripts\") pod \"placement-0fc6-account-create-7zc7c\" (UID: \"967da6fe-b0e0-4559-8db1-eea829746d13\") " pod="openstack/placement-0fc6-account-create-7zc7c" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.841133 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sp6rj\" (UniqueName: \"kubernetes.io/projected/967da6fe-b0e0-4559-8db1-eea829746d13-kube-api-access-sp6rj\") pod \"placement-0fc6-account-create-7zc7c\" (UID: \"967da6fe-b0e0-4559-8db1-eea829746d13\") " pod="openstack/placement-0fc6-account-create-7zc7c" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.915047 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-388c-account-create-p6bcq"] Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.916514 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-388c-account-create-p6bcq" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.920092 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-388c-account-create-p6bcq"] Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.921436 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.924585 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5bee7f9-da20-4e5d-9816-62bb3e9d6bce-operator-scripts\") pod \"glance-db-create-ff7zf\" (UID: \"a5bee7f9-da20-4e5d-9816-62bb3e9d6bce\") " pod="openstack/glance-db-create-ff7zf" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.924652 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkg75\" (UniqueName: \"kubernetes.io/projected/a5bee7f9-da20-4e5d-9816-62bb3e9d6bce-kube-api-access-bkg75\") pod \"glance-db-create-ff7zf\" (UID: \"a5bee7f9-da20-4e5d-9816-62bb3e9d6bce\") " pod="openstack/glance-db-create-ff7zf" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.966376 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-0fc6-account-create-7zc7c" Nov 24 13:32:29 crc kubenswrapper[4790]: I1124 13:32:29.991560 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6aea-account-create-hfz2v"] Nov 24 13:32:30 crc kubenswrapper[4790]: I1124 13:32:30.026010 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5bee7f9-da20-4e5d-9816-62bb3e9d6bce-operator-scripts\") pod \"glance-db-create-ff7zf\" (UID: \"a5bee7f9-da20-4e5d-9816-62bb3e9d6bce\") " pod="openstack/glance-db-create-ff7zf" Nov 24 13:32:30 crc kubenswrapper[4790]: I1124 13:32:30.026069 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkg75\" (UniqueName: \"kubernetes.io/projected/a5bee7f9-da20-4e5d-9816-62bb3e9d6bce-kube-api-access-bkg75\") pod \"glance-db-create-ff7zf\" (UID: \"a5bee7f9-da20-4e5d-9816-62bb3e9d6bce\") " pod="openstack/glance-db-create-ff7zf" Nov 24 13:32:30 crc kubenswrapper[4790]: I1124 13:32:30.026109 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f95eea4d-a9fd-49e3-98d3-3a871e945550-operator-scripts\") pod \"glance-388c-account-create-p6bcq\" (UID: \"f95eea4d-a9fd-49e3-98d3-3a871e945550\") " pod="openstack/glance-388c-account-create-p6bcq" Nov 24 13:32:30 crc kubenswrapper[4790]: I1124 13:32:30.026132 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4h8d6\" (UniqueName: \"kubernetes.io/projected/f95eea4d-a9fd-49e3-98d3-3a871e945550-kube-api-access-4h8d6\") pod \"glance-388c-account-create-p6bcq\" (UID: \"f95eea4d-a9fd-49e3-98d3-3a871e945550\") " pod="openstack/glance-388c-account-create-p6bcq" Nov 24 13:32:30 crc kubenswrapper[4790]: I1124 13:32:30.027089 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5bee7f9-da20-4e5d-9816-62bb3e9d6bce-operator-scripts\") pod \"glance-db-create-ff7zf\" (UID: \"a5bee7f9-da20-4e5d-9816-62bb3e9d6bce\") " pod="openstack/glance-db-create-ff7zf" Nov 24 13:32:30 crc kubenswrapper[4790]: I1124 13:32:30.057466 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkg75\" (UniqueName: \"kubernetes.io/projected/a5bee7f9-da20-4e5d-9816-62bb3e9d6bce-kube-api-access-bkg75\") pod \"glance-db-create-ff7zf\" (UID: \"a5bee7f9-da20-4e5d-9816-62bb3e9d6bce\") " pod="openstack/glance-db-create-ff7zf" Nov 24 13:32:30 crc kubenswrapper[4790]: I1124 13:32:30.122187 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-xsphw"] Nov 24 13:32:30 crc kubenswrapper[4790]: I1124 13:32:30.127758 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f95eea4d-a9fd-49e3-98d3-3a871e945550-operator-scripts\") pod \"glance-388c-account-create-p6bcq\" (UID: \"f95eea4d-a9fd-49e3-98d3-3a871e945550\") " pod="openstack/glance-388c-account-create-p6bcq" Nov 24 13:32:30 crc kubenswrapper[4790]: I1124 13:32:30.127790 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4h8d6\" (UniqueName: \"kubernetes.io/projected/f95eea4d-a9fd-49e3-98d3-3a871e945550-kube-api-access-4h8d6\") pod \"glance-388c-account-create-p6bcq\" (UID: \"f95eea4d-a9fd-49e3-98d3-3a871e945550\") " pod="openstack/glance-388c-account-create-p6bcq" Nov 24 13:32:30 crc kubenswrapper[4790]: I1124 13:32:30.128826 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ff7zf" Nov 24 13:32:30 crc kubenswrapper[4790]: I1124 13:32:30.129114 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f95eea4d-a9fd-49e3-98d3-3a871e945550-operator-scripts\") pod \"glance-388c-account-create-p6bcq\" (UID: \"f95eea4d-a9fd-49e3-98d3-3a871e945550\") " pod="openstack/glance-388c-account-create-p6bcq" Nov 24 13:32:30 crc kubenswrapper[4790]: I1124 13:32:30.152636 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4h8d6\" (UniqueName: \"kubernetes.io/projected/f95eea4d-a9fd-49e3-98d3-3a871e945550-kube-api-access-4h8d6\") pod \"glance-388c-account-create-p6bcq\" (UID: \"f95eea4d-a9fd-49e3-98d3-3a871e945550\") " pod="openstack/glance-388c-account-create-p6bcq" Nov 24 13:32:30 crc kubenswrapper[4790]: I1124 13:32:30.247663 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/63ce5126-cdab-4cdc-968a-acc26c6caf87-etc-swift\") pod \"swift-storage-0\" (UID: \"63ce5126-cdab-4cdc-968a-acc26c6caf87\") " pod="openstack/swift-storage-0" Nov 24 13:32:30 crc kubenswrapper[4790]: E1124 13:32:30.248081 4790 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 13:32:30 crc kubenswrapper[4790]: E1124 13:32:30.248114 4790 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 13:32:30 crc kubenswrapper[4790]: E1124 13:32:30.248175 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/63ce5126-cdab-4cdc-968a-acc26c6caf87-etc-swift podName:63ce5126-cdab-4cdc-968a-acc26c6caf87 nodeName:}" failed. No retries permitted until 2025-11-24 13:32:38.248153484 +0000 UTC m=+1206.628047146 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/63ce5126-cdab-4cdc-968a-acc26c6caf87-etc-swift") pod "swift-storage-0" (UID: "63ce5126-cdab-4cdc-968a-acc26c6caf87") : configmap "swift-ring-files" not found Nov 24 13:32:30 crc kubenswrapper[4790]: I1124 13:32:30.248460 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-388c-account-create-p6bcq" Nov 24 13:32:30 crc kubenswrapper[4790]: I1124 13:32:30.299120 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-67mw7"] Nov 24 13:32:30 crc kubenswrapper[4790]: W1124 13:32:30.305183 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee85107b_89f0_48c3_aecf_05295449f025.slice/crio-8b3876815da2a8c3885b21926cc6cbcc4c5989daad93f5e4facc70f70038077e WatchSource:0}: Error finding container 8b3876815da2a8c3885b21926cc6cbcc4c5989daad93f5e4facc70f70038077e: Status 404 returned error can't find the container with id 8b3876815da2a8c3885b21926cc6cbcc4c5989daad93f5e4facc70f70038077e Nov 24 13:32:30 crc kubenswrapper[4790]: I1124 13:32:30.429870 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-0fc6-account-create-7zc7c"] Nov 24 13:32:30 crc kubenswrapper[4790]: W1124 13:32:30.438163 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod967da6fe_b0e0_4559_8db1_eea829746d13.slice/crio-c088a91c68bbc40427587bcf50025373ebe8f76967f7f3aefb2d693a6664bc55 WatchSource:0}: Error finding container c088a91c68bbc40427587bcf50025373ebe8f76967f7f3aefb2d693a6664bc55: Status 404 returned error can't find the container with id c088a91c68bbc40427587bcf50025373ebe8f76967f7f3aefb2d693a6664bc55 Nov 24 13:32:30 crc kubenswrapper[4790]: I1124 13:32:30.620616 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-ff7zf"] Nov 24 13:32:30 crc kubenswrapper[4790]: W1124 13:32:30.630097 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5bee7f9_da20_4e5d_9816_62bb3e9d6bce.slice/crio-3dbbd4535310694d3745a590a90b75ca2fefccfd2f56981b373479d8e6d55428 WatchSource:0}: Error finding container 3dbbd4535310694d3745a590a90b75ca2fefccfd2f56981b373479d8e6d55428: Status 404 returned error can't find the container with id 3dbbd4535310694d3745a590a90b75ca2fefccfd2f56981b373479d8e6d55428 Nov 24 13:32:30 crc kubenswrapper[4790]: I1124 13:32:30.740130 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-388c-account-create-p6bcq"] Nov 24 13:32:30 crc kubenswrapper[4790]: W1124 13:32:30.757232 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf95eea4d_a9fd_49e3_98d3_3a871e945550.slice/crio-03a0ff28fcb177eef7ace6b6a72b8789e7ae4c80df3ea94b8df08d10b9c871af WatchSource:0}: Error finding container 03a0ff28fcb177eef7ace6b6a72b8789e7ae4c80df3ea94b8df08d10b9c871af: Status 404 returned error can't find the container with id 03a0ff28fcb177eef7ace6b6a72b8789e7ae4c80df3ea94b8df08d10b9c871af Nov 24 13:32:30 crc kubenswrapper[4790]: I1124 13:32:30.963527 4790 generic.go:334] "Generic (PLEG): container finished" podID="967da6fe-b0e0-4559-8db1-eea829746d13" containerID="489363e812cab6ae0ad579bf95670ebf0ce1d05dfc24e582a5b1891a50435049" exitCode=0 Nov 24 13:32:30 crc kubenswrapper[4790]: I1124 13:32:30.963605 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-0fc6-account-create-7zc7c" event={"ID":"967da6fe-b0e0-4559-8db1-eea829746d13","Type":"ContainerDied","Data":"489363e812cab6ae0ad579bf95670ebf0ce1d05dfc24e582a5b1891a50435049"} Nov 24 13:32:30 crc kubenswrapper[4790]: I1124 13:32:30.963633 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-0fc6-account-create-7zc7c" event={"ID":"967da6fe-b0e0-4559-8db1-eea829746d13","Type":"ContainerStarted","Data":"c088a91c68bbc40427587bcf50025373ebe8f76967f7f3aefb2d693a6664bc55"} Nov 24 13:32:30 crc kubenswrapper[4790]: I1124 13:32:30.965563 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ff7zf" event={"ID":"a5bee7f9-da20-4e5d-9816-62bb3e9d6bce","Type":"ContainerStarted","Data":"3dbbd4535310694d3745a590a90b75ca2fefccfd2f56981b373479d8e6d55428"} Nov 24 13:32:30 crc kubenswrapper[4790]: I1124 13:32:30.967505 4790 generic.go:334] "Generic (PLEG): container finished" podID="ee85107b-89f0-48c3-aecf-05295449f025" containerID="169855f1ee71f39432956cfa61f341f14290544d596d6f99664ca020d983892a" exitCode=0 Nov 24 13:32:30 crc kubenswrapper[4790]: I1124 13:32:30.967538 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-67mw7" event={"ID":"ee85107b-89f0-48c3-aecf-05295449f025","Type":"ContainerDied","Data":"169855f1ee71f39432956cfa61f341f14290544d596d6f99664ca020d983892a"} Nov 24 13:32:30 crc kubenswrapper[4790]: I1124 13:32:30.967564 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-67mw7" event={"ID":"ee85107b-89f0-48c3-aecf-05295449f025","Type":"ContainerStarted","Data":"8b3876815da2a8c3885b21926cc6cbcc4c5989daad93f5e4facc70f70038077e"} Nov 24 13:32:30 crc kubenswrapper[4790]: I1124 13:32:30.969213 4790 generic.go:334] "Generic (PLEG): container finished" podID="8a30bb00-53b7-4128-954f-d5183fec7675" containerID="ad9512a19986dcbf34683ed61d8fb2ead36a72f27f516136c67f069e610076cf" exitCode=0 Nov 24 13:32:30 crc kubenswrapper[4790]: I1124 13:32:30.969443 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6aea-account-create-hfz2v" event={"ID":"8a30bb00-53b7-4128-954f-d5183fec7675","Type":"ContainerDied","Data":"ad9512a19986dcbf34683ed61d8fb2ead36a72f27f516136c67f069e610076cf"} Nov 24 13:32:30 crc kubenswrapper[4790]: I1124 13:32:30.969478 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6aea-account-create-hfz2v" event={"ID":"8a30bb00-53b7-4128-954f-d5183fec7675","Type":"ContainerStarted","Data":"be9587f845d043ed4fa3ab27bca7a5c21fafc801be5a7d27d8bf12cd7d469a39"} Nov 24 13:32:30 crc kubenswrapper[4790]: I1124 13:32:30.970606 4790 generic.go:334] "Generic (PLEG): container finished" podID="85671e80-3009-49dc-a308-e126b08847a2" containerID="eab70d8990c01fc8d4d58c7a888bef266d435891d2a2affc19e30cc46f143f8a" exitCode=0 Nov 24 13:32:30 crc kubenswrapper[4790]: I1124 13:32:30.970657 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-xsphw" event={"ID":"85671e80-3009-49dc-a308-e126b08847a2","Type":"ContainerDied","Data":"eab70d8990c01fc8d4d58c7a888bef266d435891d2a2affc19e30cc46f143f8a"} Nov 24 13:32:30 crc kubenswrapper[4790]: I1124 13:32:30.970670 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-xsphw" event={"ID":"85671e80-3009-49dc-a308-e126b08847a2","Type":"ContainerStarted","Data":"6bbdfb1c799994251f1972b39f44c26d72ed0b811d6f2b263ea0a35348bbef6c"} Nov 24 13:32:30 crc kubenswrapper[4790]: I1124 13:32:30.971801 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-388c-account-create-p6bcq" event={"ID":"f95eea4d-a9fd-49e3-98d3-3a871e945550","Type":"ContainerStarted","Data":"03a0ff28fcb177eef7ace6b6a72b8789e7ae4c80df3ea94b8df08d10b9c871af"} Nov 24 13:32:31 crc kubenswrapper[4790]: I1124 13:32:31.987238 4790 generic.go:334] "Generic (PLEG): container finished" podID="f95eea4d-a9fd-49e3-98d3-3a871e945550" containerID="7976704e51ab332073a36acf2fadd9c4a5760783d9fc94a68e424b56646d342e" exitCode=0 Nov 24 13:32:31 crc kubenswrapper[4790]: I1124 13:32:31.987589 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-388c-account-create-p6bcq" event={"ID":"f95eea4d-a9fd-49e3-98d3-3a871e945550","Type":"ContainerDied","Data":"7976704e51ab332073a36acf2fadd9c4a5760783d9fc94a68e424b56646d342e"} Nov 24 13:32:31 crc kubenswrapper[4790]: I1124 13:32:31.990601 4790 generic.go:334] "Generic (PLEG): container finished" podID="a5bee7f9-da20-4e5d-9816-62bb3e9d6bce" containerID="d03a01bc9550ec4a70da4d90c581b8b702ead5abff5bbee34ae2f7e4898a6b05" exitCode=0 Nov 24 13:32:31 crc kubenswrapper[4790]: I1124 13:32:31.991371 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ff7zf" event={"ID":"a5bee7f9-da20-4e5d-9816-62bb3e9d6bce","Type":"ContainerDied","Data":"d03a01bc9550ec4a70da4d90c581b8b702ead5abff5bbee34ae2f7e4898a6b05"} Nov 24 13:32:32 crc kubenswrapper[4790]: I1124 13:32:32.365544 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6aea-account-create-hfz2v" Nov 24 13:32:32 crc kubenswrapper[4790]: E1124 13:32:32.366961 4790 info.go:109] Failed to get network devices: open /sys/class/net/6bbdfb1c7999942/address: no such file or directory Nov 24 13:32:32 crc kubenswrapper[4790]: I1124 13:32:32.398722 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tc2ck\" (UniqueName: \"kubernetes.io/projected/8a30bb00-53b7-4128-954f-d5183fec7675-kube-api-access-tc2ck\") pod \"8a30bb00-53b7-4128-954f-d5183fec7675\" (UID: \"8a30bb00-53b7-4128-954f-d5183fec7675\") " Nov 24 13:32:32 crc kubenswrapper[4790]: I1124 13:32:32.398856 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a30bb00-53b7-4128-954f-d5183fec7675-operator-scripts\") pod \"8a30bb00-53b7-4128-954f-d5183fec7675\" (UID: \"8a30bb00-53b7-4128-954f-d5183fec7675\") " Nov 24 13:32:32 crc kubenswrapper[4790]: I1124 13:32:32.400003 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a30bb00-53b7-4128-954f-d5183fec7675-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8a30bb00-53b7-4128-954f-d5183fec7675" (UID: "8a30bb00-53b7-4128-954f-d5183fec7675"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:32 crc kubenswrapper[4790]: I1124 13:32:32.406542 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a30bb00-53b7-4128-954f-d5183fec7675-kube-api-access-tc2ck" (OuterVolumeSpecName: "kube-api-access-tc2ck") pod "8a30bb00-53b7-4128-954f-d5183fec7675" (UID: "8a30bb00-53b7-4128-954f-d5183fec7675"). InnerVolumeSpecName "kube-api-access-tc2ck". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:32:32 crc kubenswrapper[4790]: I1124 13:32:32.500797 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a30bb00-53b7-4128-954f-d5183fec7675-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:32 crc kubenswrapper[4790]: I1124 13:32:32.500826 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tc2ck\" (UniqueName: \"kubernetes.io/projected/8a30bb00-53b7-4128-954f-d5183fec7675-kube-api-access-tc2ck\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:32 crc kubenswrapper[4790]: I1124 13:32:32.530518 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-xsphw" Nov 24 13:32:32 crc kubenswrapper[4790]: I1124 13:32:32.545127 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-0fc6-account-create-7zc7c" Nov 24 13:32:32 crc kubenswrapper[4790]: I1124 13:32:32.574859 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-67mw7" Nov 24 13:32:32 crc kubenswrapper[4790]: I1124 13:32:32.601697 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8zwb\" (UniqueName: \"kubernetes.io/projected/85671e80-3009-49dc-a308-e126b08847a2-kube-api-access-s8zwb\") pod \"85671e80-3009-49dc-a308-e126b08847a2\" (UID: \"85671e80-3009-49dc-a308-e126b08847a2\") " Nov 24 13:32:32 crc kubenswrapper[4790]: I1124 13:32:32.601859 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/967da6fe-b0e0-4559-8db1-eea829746d13-operator-scripts\") pod \"967da6fe-b0e0-4559-8db1-eea829746d13\" (UID: \"967da6fe-b0e0-4559-8db1-eea829746d13\") " Nov 24 13:32:32 crc kubenswrapper[4790]: I1124 13:32:32.601924 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhrln\" (UniqueName: \"kubernetes.io/projected/ee85107b-89f0-48c3-aecf-05295449f025-kube-api-access-qhrln\") pod \"ee85107b-89f0-48c3-aecf-05295449f025\" (UID: \"ee85107b-89f0-48c3-aecf-05295449f025\") " Nov 24 13:32:32 crc kubenswrapper[4790]: I1124 13:32:32.601948 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sp6rj\" (UniqueName: \"kubernetes.io/projected/967da6fe-b0e0-4559-8db1-eea829746d13-kube-api-access-sp6rj\") pod \"967da6fe-b0e0-4559-8db1-eea829746d13\" (UID: \"967da6fe-b0e0-4559-8db1-eea829746d13\") " Nov 24 13:32:32 crc kubenswrapper[4790]: I1124 13:32:32.601974 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85671e80-3009-49dc-a308-e126b08847a2-operator-scripts\") pod \"85671e80-3009-49dc-a308-e126b08847a2\" (UID: \"85671e80-3009-49dc-a308-e126b08847a2\") " Nov 24 13:32:32 crc kubenswrapper[4790]: I1124 13:32:32.602030 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee85107b-89f0-48c3-aecf-05295449f025-operator-scripts\") pod \"ee85107b-89f0-48c3-aecf-05295449f025\" (UID: \"ee85107b-89f0-48c3-aecf-05295449f025\") " Nov 24 13:32:32 crc kubenswrapper[4790]: I1124 13:32:32.603126 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/967da6fe-b0e0-4559-8db1-eea829746d13-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "967da6fe-b0e0-4559-8db1-eea829746d13" (UID: "967da6fe-b0e0-4559-8db1-eea829746d13"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:32 crc kubenswrapper[4790]: I1124 13:32:32.603212 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85671e80-3009-49dc-a308-e126b08847a2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "85671e80-3009-49dc-a308-e126b08847a2" (UID: "85671e80-3009-49dc-a308-e126b08847a2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:32 crc kubenswrapper[4790]: I1124 13:32:32.603463 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee85107b-89f0-48c3-aecf-05295449f025-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ee85107b-89f0-48c3-aecf-05295449f025" (UID: "ee85107b-89f0-48c3-aecf-05295449f025"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:32 crc kubenswrapper[4790]: I1124 13:32:32.608256 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85671e80-3009-49dc-a308-e126b08847a2-kube-api-access-s8zwb" (OuterVolumeSpecName: "kube-api-access-s8zwb") pod "85671e80-3009-49dc-a308-e126b08847a2" (UID: "85671e80-3009-49dc-a308-e126b08847a2"). InnerVolumeSpecName "kube-api-access-s8zwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:32:32 crc kubenswrapper[4790]: I1124 13:32:32.608612 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee85107b-89f0-48c3-aecf-05295449f025-kube-api-access-qhrln" (OuterVolumeSpecName: "kube-api-access-qhrln") pod "ee85107b-89f0-48c3-aecf-05295449f025" (UID: "ee85107b-89f0-48c3-aecf-05295449f025"). InnerVolumeSpecName "kube-api-access-qhrln". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:32:32 crc kubenswrapper[4790]: I1124 13:32:32.610060 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/967da6fe-b0e0-4559-8db1-eea829746d13-kube-api-access-sp6rj" (OuterVolumeSpecName: "kube-api-access-sp6rj") pod "967da6fe-b0e0-4559-8db1-eea829746d13" (UID: "967da6fe-b0e0-4559-8db1-eea829746d13"). InnerVolumeSpecName "kube-api-access-sp6rj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:32:32 crc kubenswrapper[4790]: I1124 13:32:32.703181 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhrln\" (UniqueName: \"kubernetes.io/projected/ee85107b-89f0-48c3-aecf-05295449f025-kube-api-access-qhrln\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:32 crc kubenswrapper[4790]: I1124 13:32:32.703210 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sp6rj\" (UniqueName: \"kubernetes.io/projected/967da6fe-b0e0-4559-8db1-eea829746d13-kube-api-access-sp6rj\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:32 crc kubenswrapper[4790]: I1124 13:32:32.703219 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85671e80-3009-49dc-a308-e126b08847a2-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:32 crc kubenswrapper[4790]: I1124 13:32:32.703229 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee85107b-89f0-48c3-aecf-05295449f025-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:32 crc kubenswrapper[4790]: I1124 13:32:32.703237 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8zwb\" (UniqueName: \"kubernetes.io/projected/85671e80-3009-49dc-a308-e126b08847a2-kube-api-access-s8zwb\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:32 crc kubenswrapper[4790]: I1124 13:32:32.703245 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/967da6fe-b0e0-4559-8db1-eea829746d13-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:33 crc kubenswrapper[4790]: I1124 13:32:33.002849 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-67mw7" event={"ID":"ee85107b-89f0-48c3-aecf-05295449f025","Type":"ContainerDied","Data":"8b3876815da2a8c3885b21926cc6cbcc4c5989daad93f5e4facc70f70038077e"} Nov 24 13:32:33 crc kubenswrapper[4790]: I1124 13:32:33.002937 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b3876815da2a8c3885b21926cc6cbcc4c5989daad93f5e4facc70f70038077e" Nov 24 13:32:33 crc kubenswrapper[4790]: I1124 13:32:33.003821 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-67mw7" Nov 24 13:32:33 crc kubenswrapper[4790]: I1124 13:32:33.005489 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6aea-account-create-hfz2v" event={"ID":"8a30bb00-53b7-4128-954f-d5183fec7675","Type":"ContainerDied","Data":"be9587f845d043ed4fa3ab27bca7a5c21fafc801be5a7d27d8bf12cd7d469a39"} Nov 24 13:32:33 crc kubenswrapper[4790]: I1124 13:32:33.005532 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be9587f845d043ed4fa3ab27bca7a5c21fafc801be5a7d27d8bf12cd7d469a39" Nov 24 13:32:33 crc kubenswrapper[4790]: I1124 13:32:33.005620 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6aea-account-create-hfz2v" Nov 24 13:32:33 crc kubenswrapper[4790]: I1124 13:32:33.015053 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-xsphw" event={"ID":"85671e80-3009-49dc-a308-e126b08847a2","Type":"ContainerDied","Data":"6bbdfb1c799994251f1972b39f44c26d72ed0b811d6f2b263ea0a35348bbef6c"} Nov 24 13:32:33 crc kubenswrapper[4790]: I1124 13:32:33.015094 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6bbdfb1c799994251f1972b39f44c26d72ed0b811d6f2b263ea0a35348bbef6c" Nov 24 13:32:33 crc kubenswrapper[4790]: I1124 13:32:33.015152 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-xsphw" Nov 24 13:32:33 crc kubenswrapper[4790]: I1124 13:32:33.019165 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-0fc6-account-create-7zc7c" event={"ID":"967da6fe-b0e0-4559-8db1-eea829746d13","Type":"ContainerDied","Data":"c088a91c68bbc40427587bcf50025373ebe8f76967f7f3aefb2d693a6664bc55"} Nov 24 13:32:33 crc kubenswrapper[4790]: I1124 13:32:33.019205 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c088a91c68bbc40427587bcf50025373ebe8f76967f7f3aefb2d693a6664bc55" Nov 24 13:32:33 crc kubenswrapper[4790]: I1124 13:32:33.019342 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-0fc6-account-create-7zc7c" Nov 24 13:32:33 crc kubenswrapper[4790]: I1124 13:32:33.317550 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-388c-account-create-p6bcq" Nov 24 13:32:33 crc kubenswrapper[4790]: I1124 13:32:33.327834 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ff7zf" Nov 24 13:32:33 crc kubenswrapper[4790]: I1124 13:32:33.417125 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5bee7f9-da20-4e5d-9816-62bb3e9d6bce-operator-scripts\") pod \"a5bee7f9-da20-4e5d-9816-62bb3e9d6bce\" (UID: \"a5bee7f9-da20-4e5d-9816-62bb3e9d6bce\") " Nov 24 13:32:33 crc kubenswrapper[4790]: I1124 13:32:33.417212 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4h8d6\" (UniqueName: \"kubernetes.io/projected/f95eea4d-a9fd-49e3-98d3-3a871e945550-kube-api-access-4h8d6\") pod \"f95eea4d-a9fd-49e3-98d3-3a871e945550\" (UID: \"f95eea4d-a9fd-49e3-98d3-3a871e945550\") " Nov 24 13:32:33 crc kubenswrapper[4790]: I1124 13:32:33.417380 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f95eea4d-a9fd-49e3-98d3-3a871e945550-operator-scripts\") pod \"f95eea4d-a9fd-49e3-98d3-3a871e945550\" (UID: \"f95eea4d-a9fd-49e3-98d3-3a871e945550\") " Nov 24 13:32:33 crc kubenswrapper[4790]: I1124 13:32:33.417439 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkg75\" (UniqueName: \"kubernetes.io/projected/a5bee7f9-da20-4e5d-9816-62bb3e9d6bce-kube-api-access-bkg75\") pod \"a5bee7f9-da20-4e5d-9816-62bb3e9d6bce\" (UID: \"a5bee7f9-da20-4e5d-9816-62bb3e9d6bce\") " Nov 24 13:32:33 crc kubenswrapper[4790]: I1124 13:32:33.418189 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f95eea4d-a9fd-49e3-98d3-3a871e945550-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f95eea4d-a9fd-49e3-98d3-3a871e945550" (UID: "f95eea4d-a9fd-49e3-98d3-3a871e945550"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:33 crc kubenswrapper[4790]: I1124 13:32:33.418531 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5bee7f9-da20-4e5d-9816-62bb3e9d6bce-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a5bee7f9-da20-4e5d-9816-62bb3e9d6bce" (UID: "a5bee7f9-da20-4e5d-9816-62bb3e9d6bce"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:33 crc kubenswrapper[4790]: I1124 13:32:33.419748 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f95eea4d-a9fd-49e3-98d3-3a871e945550-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:33 crc kubenswrapper[4790]: I1124 13:32:33.419767 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5bee7f9-da20-4e5d-9816-62bb3e9d6bce-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:33 crc kubenswrapper[4790]: I1124 13:32:33.422345 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5bee7f9-da20-4e5d-9816-62bb3e9d6bce-kube-api-access-bkg75" (OuterVolumeSpecName: "kube-api-access-bkg75") pod "a5bee7f9-da20-4e5d-9816-62bb3e9d6bce" (UID: "a5bee7f9-da20-4e5d-9816-62bb3e9d6bce"). InnerVolumeSpecName "kube-api-access-bkg75". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:32:33 crc kubenswrapper[4790]: I1124 13:32:33.422821 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f95eea4d-a9fd-49e3-98d3-3a871e945550-kube-api-access-4h8d6" (OuterVolumeSpecName: "kube-api-access-4h8d6") pod "f95eea4d-a9fd-49e3-98d3-3a871e945550" (UID: "f95eea4d-a9fd-49e3-98d3-3a871e945550"). InnerVolumeSpecName "kube-api-access-4h8d6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:32:33 crc kubenswrapper[4790]: I1124 13:32:33.464692 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8cc7fc4dc-5d55x" Nov 24 13:32:33 crc kubenswrapper[4790]: I1124 13:32:33.521295 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4h8d6\" (UniqueName: \"kubernetes.io/projected/f95eea4d-a9fd-49e3-98d3-3a871e945550-kube-api-access-4h8d6\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:33 crc kubenswrapper[4790]: I1124 13:32:33.521436 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkg75\" (UniqueName: \"kubernetes.io/projected/a5bee7f9-da20-4e5d-9816-62bb3e9d6bce-kube-api-access-bkg75\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:34 crc kubenswrapper[4790]: I1124 13:32:34.004107 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-b4bt2" Nov 24 13:32:34 crc kubenswrapper[4790]: I1124 13:32:34.027182 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ff7zf" event={"ID":"a5bee7f9-da20-4e5d-9816-62bb3e9d6bce","Type":"ContainerDied","Data":"3dbbd4535310694d3745a590a90b75ca2fefccfd2f56981b373479d8e6d55428"} Nov 24 13:32:34 crc kubenswrapper[4790]: I1124 13:32:34.027218 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3dbbd4535310694d3745a590a90b75ca2fefccfd2f56981b373479d8e6d55428" Nov 24 13:32:34 crc kubenswrapper[4790]: I1124 13:32:34.027236 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ff7zf" Nov 24 13:32:34 crc kubenswrapper[4790]: I1124 13:32:34.029869 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-388c-account-create-p6bcq" event={"ID":"f95eea4d-a9fd-49e3-98d3-3a871e945550","Type":"ContainerDied","Data":"03a0ff28fcb177eef7ace6b6a72b8789e7ae4c80df3ea94b8df08d10b9c871af"} Nov 24 13:32:34 crc kubenswrapper[4790]: I1124 13:32:34.029919 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03a0ff28fcb177eef7ace6b6a72b8789e7ae4c80df3ea94b8df08d10b9c871af" Nov 24 13:32:34 crc kubenswrapper[4790]: I1124 13:32:34.030132 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-388c-account-create-p6bcq" Nov 24 13:32:34 crc kubenswrapper[4790]: I1124 13:32:34.031273 4790 generic.go:334] "Generic (PLEG): container finished" podID="87e08198-d2d0-4608-87f3-02d832fe7c55" containerID="2d459199980f3814665ba6c630a8bf310ef6efb41a3f2368c9edbeb004c62897" exitCode=0 Nov 24 13:32:34 crc kubenswrapper[4790]: I1124 13:32:34.031321 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-48flt" event={"ID":"87e08198-d2d0-4608-87f3-02d832fe7c55","Type":"ContainerDied","Data":"2d459199980f3814665ba6c630a8bf310ef6efb41a3f2368c9edbeb004c62897"} Nov 24 13:32:34 crc kubenswrapper[4790]: I1124 13:32:34.068514 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8cc7fc4dc-5d55x"] Nov 24 13:32:34 crc kubenswrapper[4790]: I1124 13:32:34.068748 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8cc7fc4dc-5d55x" podUID="81f66d48-47b6-4a1c-afd1-12ce651dda1f" containerName="dnsmasq-dns" containerID="cri-o://d1247e35b9ee7e3f73a2cb7b18e8833595b28040ca68121c03b967719171cc2c" gracePeriod=10 Nov 24 13:32:34 crc kubenswrapper[4790]: I1124 13:32:34.475932 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8cc7fc4dc-5d55x" Nov 24 13:32:34 crc kubenswrapper[4790]: I1124 13:32:34.536311 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49vwz\" (UniqueName: \"kubernetes.io/projected/81f66d48-47b6-4a1c-afd1-12ce651dda1f-kube-api-access-49vwz\") pod \"81f66d48-47b6-4a1c-afd1-12ce651dda1f\" (UID: \"81f66d48-47b6-4a1c-afd1-12ce651dda1f\") " Nov 24 13:32:34 crc kubenswrapper[4790]: I1124 13:32:34.536372 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/81f66d48-47b6-4a1c-afd1-12ce651dda1f-ovsdbserver-sb\") pod \"81f66d48-47b6-4a1c-afd1-12ce651dda1f\" (UID: \"81f66d48-47b6-4a1c-afd1-12ce651dda1f\") " Nov 24 13:32:34 crc kubenswrapper[4790]: I1124 13:32:34.536393 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81f66d48-47b6-4a1c-afd1-12ce651dda1f-config\") pod \"81f66d48-47b6-4a1c-afd1-12ce651dda1f\" (UID: \"81f66d48-47b6-4a1c-afd1-12ce651dda1f\") " Nov 24 13:32:34 crc kubenswrapper[4790]: I1124 13:32:34.536418 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81f66d48-47b6-4a1c-afd1-12ce651dda1f-dns-svc\") pod \"81f66d48-47b6-4a1c-afd1-12ce651dda1f\" (UID: \"81f66d48-47b6-4a1c-afd1-12ce651dda1f\") " Nov 24 13:32:34 crc kubenswrapper[4790]: I1124 13:32:34.539935 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81f66d48-47b6-4a1c-afd1-12ce651dda1f-kube-api-access-49vwz" (OuterVolumeSpecName: "kube-api-access-49vwz") pod "81f66d48-47b6-4a1c-afd1-12ce651dda1f" (UID: "81f66d48-47b6-4a1c-afd1-12ce651dda1f"). InnerVolumeSpecName "kube-api-access-49vwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:32:34 crc kubenswrapper[4790]: I1124 13:32:34.571907 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81f66d48-47b6-4a1c-afd1-12ce651dda1f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "81f66d48-47b6-4a1c-afd1-12ce651dda1f" (UID: "81f66d48-47b6-4a1c-afd1-12ce651dda1f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:34 crc kubenswrapper[4790]: E1124 13:32:34.574226 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/81f66d48-47b6-4a1c-afd1-12ce651dda1f-config podName:81f66d48-47b6-4a1c-afd1-12ce651dda1f nodeName:}" failed. No retries permitted until 2025-11-24 13:32:35.074194884 +0000 UTC m=+1203.454088546 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config" (UniqueName: "kubernetes.io/configmap/81f66d48-47b6-4a1c-afd1-12ce651dda1f-config") pod "81f66d48-47b6-4a1c-afd1-12ce651dda1f" (UID: "81f66d48-47b6-4a1c-afd1-12ce651dda1f") : error deleting /var/lib/kubelet/pods/81f66d48-47b6-4a1c-afd1-12ce651dda1f/volume-subpaths: remove /var/lib/kubelet/pods/81f66d48-47b6-4a1c-afd1-12ce651dda1f/volume-subpaths: no such file or directory Nov 24 13:32:34 crc kubenswrapper[4790]: I1124 13:32:34.574598 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81f66d48-47b6-4a1c-afd1-12ce651dda1f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "81f66d48-47b6-4a1c-afd1-12ce651dda1f" (UID: "81f66d48-47b6-4a1c-afd1-12ce651dda1f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:34 crc kubenswrapper[4790]: I1124 13:32:34.638256 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49vwz\" (UniqueName: \"kubernetes.io/projected/81f66d48-47b6-4a1c-afd1-12ce651dda1f-kube-api-access-49vwz\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:34 crc kubenswrapper[4790]: I1124 13:32:34.638288 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/81f66d48-47b6-4a1c-afd1-12ce651dda1f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:34 crc kubenswrapper[4790]: I1124 13:32:34.638296 4790 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81f66d48-47b6-4a1c-afd1-12ce651dda1f-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.072159 4790 generic.go:334] "Generic (PLEG): container finished" podID="81f66d48-47b6-4a1c-afd1-12ce651dda1f" containerID="d1247e35b9ee7e3f73a2cb7b18e8833595b28040ca68121c03b967719171cc2c" exitCode=0 Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.072448 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8cc7fc4dc-5d55x" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.073035 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8cc7fc4dc-5d55x" event={"ID":"81f66d48-47b6-4a1c-afd1-12ce651dda1f","Type":"ContainerDied","Data":"d1247e35b9ee7e3f73a2cb7b18e8833595b28040ca68121c03b967719171cc2c"} Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.073101 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8cc7fc4dc-5d55x" event={"ID":"81f66d48-47b6-4a1c-afd1-12ce651dda1f","Type":"ContainerDied","Data":"edb7a9441d0c953eaf381d64c78881b291536f48e18dc01c346fe734a68b6c1a"} Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.073123 4790 scope.go:117] "RemoveContainer" containerID="d1247e35b9ee7e3f73a2cb7b18e8833595b28040ca68121c03b967719171cc2c" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.119300 4790 scope.go:117] "RemoveContainer" containerID="36419dcb1166d10a4271c8f86e7b0e58aa5b800327848c11a2786cf78cbde7f7" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.145070 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81f66d48-47b6-4a1c-afd1-12ce651dda1f-config\") pod \"81f66d48-47b6-4a1c-afd1-12ce651dda1f\" (UID: \"81f66d48-47b6-4a1c-afd1-12ce651dda1f\") " Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.145169 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-chcfj"] Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.145691 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81f66d48-47b6-4a1c-afd1-12ce651dda1f-config" (OuterVolumeSpecName: "config") pod "81f66d48-47b6-4a1c-afd1-12ce651dda1f" (UID: "81f66d48-47b6-4a1c-afd1-12ce651dda1f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:35 crc kubenswrapper[4790]: E1124 13:32:35.145741 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5bee7f9-da20-4e5d-9816-62bb3e9d6bce" containerName="mariadb-database-create" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.145755 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5bee7f9-da20-4e5d-9816-62bb3e9d6bce" containerName="mariadb-database-create" Nov 24 13:32:35 crc kubenswrapper[4790]: E1124 13:32:35.145766 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81f66d48-47b6-4a1c-afd1-12ce651dda1f" containerName="init" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.145774 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="81f66d48-47b6-4a1c-afd1-12ce651dda1f" containerName="init" Nov 24 13:32:35 crc kubenswrapper[4790]: E1124 13:32:35.145785 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85671e80-3009-49dc-a308-e126b08847a2" containerName="mariadb-database-create" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.145793 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="85671e80-3009-49dc-a308-e126b08847a2" containerName="mariadb-database-create" Nov 24 13:32:35 crc kubenswrapper[4790]: E1124 13:32:35.145807 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81f66d48-47b6-4a1c-afd1-12ce651dda1f" containerName="dnsmasq-dns" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.145816 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="81f66d48-47b6-4a1c-afd1-12ce651dda1f" containerName="dnsmasq-dns" Nov 24 13:32:35 crc kubenswrapper[4790]: E1124 13:32:35.145826 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="967da6fe-b0e0-4559-8db1-eea829746d13" containerName="mariadb-account-create" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.145832 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="967da6fe-b0e0-4559-8db1-eea829746d13" containerName="mariadb-account-create" Nov 24 13:32:35 crc kubenswrapper[4790]: E1124 13:32:35.145841 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a30bb00-53b7-4128-954f-d5183fec7675" containerName="mariadb-account-create" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.145847 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a30bb00-53b7-4128-954f-d5183fec7675" containerName="mariadb-account-create" Nov 24 13:32:35 crc kubenswrapper[4790]: E1124 13:32:35.145859 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee85107b-89f0-48c3-aecf-05295449f025" containerName="mariadb-database-create" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.145867 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee85107b-89f0-48c3-aecf-05295449f025" containerName="mariadb-database-create" Nov 24 13:32:35 crc kubenswrapper[4790]: E1124 13:32:35.145897 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f95eea4d-a9fd-49e3-98d3-3a871e945550" containerName="mariadb-account-create" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.145904 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="f95eea4d-a9fd-49e3-98d3-3a871e945550" containerName="mariadb-account-create" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.146047 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5bee7f9-da20-4e5d-9816-62bb3e9d6bce" containerName="mariadb-database-create" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.146082 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="f95eea4d-a9fd-49e3-98d3-3a871e945550" containerName="mariadb-account-create" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.146102 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a30bb00-53b7-4128-954f-d5183fec7675" containerName="mariadb-account-create" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.146117 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="967da6fe-b0e0-4559-8db1-eea829746d13" containerName="mariadb-account-create" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.146136 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="85671e80-3009-49dc-a308-e126b08847a2" containerName="mariadb-database-create" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.146154 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee85107b-89f0-48c3-aecf-05295449f025" containerName="mariadb-database-create" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.146165 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="81f66d48-47b6-4a1c-afd1-12ce651dda1f" containerName="dnsmasq-dns" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.146720 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-chcfj" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.148605 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-nrdsr" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.159056 4790 scope.go:117] "RemoveContainer" containerID="d1247e35b9ee7e3f73a2cb7b18e8833595b28040ca68121c03b967719171cc2c" Nov 24 13:32:35 crc kubenswrapper[4790]: E1124 13:32:35.159443 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1247e35b9ee7e3f73a2cb7b18e8833595b28040ca68121c03b967719171cc2c\": container with ID starting with d1247e35b9ee7e3f73a2cb7b18e8833595b28040ca68121c03b967719171cc2c not found: ID does not exist" containerID="d1247e35b9ee7e3f73a2cb7b18e8833595b28040ca68121c03b967719171cc2c" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.159496 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1247e35b9ee7e3f73a2cb7b18e8833595b28040ca68121c03b967719171cc2c"} err="failed to get container status \"d1247e35b9ee7e3f73a2cb7b18e8833595b28040ca68121c03b967719171cc2c\": rpc error: code = NotFound desc = could not find container \"d1247e35b9ee7e3f73a2cb7b18e8833595b28040ca68121c03b967719171cc2c\": container with ID starting with d1247e35b9ee7e3f73a2cb7b18e8833595b28040ca68121c03b967719171cc2c not found: ID does not exist" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.159524 4790 scope.go:117] "RemoveContainer" containerID="36419dcb1166d10a4271c8f86e7b0e58aa5b800327848c11a2786cf78cbde7f7" Nov 24 13:32:35 crc kubenswrapper[4790]: E1124 13:32:35.163036 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36419dcb1166d10a4271c8f86e7b0e58aa5b800327848c11a2786cf78cbde7f7\": container with ID starting with 36419dcb1166d10a4271c8f86e7b0e58aa5b800327848c11a2786cf78cbde7f7 not found: ID does not exist" containerID="36419dcb1166d10a4271c8f86e7b0e58aa5b800327848c11a2786cf78cbde7f7" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.163218 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36419dcb1166d10a4271c8f86e7b0e58aa5b800327848c11a2786cf78cbde7f7"} err="failed to get container status \"36419dcb1166d10a4271c8f86e7b0e58aa5b800327848c11a2786cf78cbde7f7\": rpc error: code = NotFound desc = could not find container \"36419dcb1166d10a4271c8f86e7b0e58aa5b800327848c11a2786cf78cbde7f7\": container with ID starting with 36419dcb1166d10a4271c8f86e7b0e58aa5b800327848c11a2786cf78cbde7f7 not found: ID does not exist" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.163771 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-chcfj"] Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.164351 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.248392 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c49c82dc-b231-4f2f-8553-323fcb24f67f-combined-ca-bundle\") pod \"glance-db-sync-chcfj\" (UID: \"c49c82dc-b231-4f2f-8553-323fcb24f67f\") " pod="openstack/glance-db-sync-chcfj" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.248467 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crwkk\" (UniqueName: \"kubernetes.io/projected/c49c82dc-b231-4f2f-8553-323fcb24f67f-kube-api-access-crwkk\") pod \"glance-db-sync-chcfj\" (UID: \"c49c82dc-b231-4f2f-8553-323fcb24f67f\") " pod="openstack/glance-db-sync-chcfj" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.248578 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c49c82dc-b231-4f2f-8553-323fcb24f67f-db-sync-config-data\") pod \"glance-db-sync-chcfj\" (UID: \"c49c82dc-b231-4f2f-8553-323fcb24f67f\") " pod="openstack/glance-db-sync-chcfj" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.248600 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c49c82dc-b231-4f2f-8553-323fcb24f67f-config-data\") pod \"glance-db-sync-chcfj\" (UID: \"c49c82dc-b231-4f2f-8553-323fcb24f67f\") " pod="openstack/glance-db-sync-chcfj" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.248643 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81f66d48-47b6-4a1c-afd1-12ce651dda1f-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.325566 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-48flt" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.349503 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/87e08198-d2d0-4608-87f3-02d832fe7c55-etc-swift\") pod \"87e08198-d2d0-4608-87f3-02d832fe7c55\" (UID: \"87e08198-d2d0-4608-87f3-02d832fe7c55\") " Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.349610 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/87e08198-d2d0-4608-87f3-02d832fe7c55-ring-data-devices\") pod \"87e08198-d2d0-4608-87f3-02d832fe7c55\" (UID: \"87e08198-d2d0-4608-87f3-02d832fe7c55\") " Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.349640 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/87e08198-d2d0-4608-87f3-02d832fe7c55-swiftconf\") pod \"87e08198-d2d0-4608-87f3-02d832fe7c55\" (UID: \"87e08198-d2d0-4608-87f3-02d832fe7c55\") " Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.349689 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87e08198-d2d0-4608-87f3-02d832fe7c55-combined-ca-bundle\") pod \"87e08198-d2d0-4608-87f3-02d832fe7c55\" (UID: \"87e08198-d2d0-4608-87f3-02d832fe7c55\") " Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.349759 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-485t4\" (UniqueName: \"kubernetes.io/projected/87e08198-d2d0-4608-87f3-02d832fe7c55-kube-api-access-485t4\") pod \"87e08198-d2d0-4608-87f3-02d832fe7c55\" (UID: \"87e08198-d2d0-4608-87f3-02d832fe7c55\") " Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.349790 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/87e08198-d2d0-4608-87f3-02d832fe7c55-dispersionconf\") pod \"87e08198-d2d0-4608-87f3-02d832fe7c55\" (UID: \"87e08198-d2d0-4608-87f3-02d832fe7c55\") " Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.349865 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/87e08198-d2d0-4608-87f3-02d832fe7c55-scripts\") pod \"87e08198-d2d0-4608-87f3-02d832fe7c55\" (UID: \"87e08198-d2d0-4608-87f3-02d832fe7c55\") " Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.350136 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crwkk\" (UniqueName: \"kubernetes.io/projected/c49c82dc-b231-4f2f-8553-323fcb24f67f-kube-api-access-crwkk\") pod \"glance-db-sync-chcfj\" (UID: \"c49c82dc-b231-4f2f-8553-323fcb24f67f\") " pod="openstack/glance-db-sync-chcfj" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.350273 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87e08198-d2d0-4608-87f3-02d832fe7c55-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "87e08198-d2d0-4608-87f3-02d832fe7c55" (UID: "87e08198-d2d0-4608-87f3-02d832fe7c55"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.350282 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c49c82dc-b231-4f2f-8553-323fcb24f67f-db-sync-config-data\") pod \"glance-db-sync-chcfj\" (UID: \"c49c82dc-b231-4f2f-8553-323fcb24f67f\") " pod="openstack/glance-db-sync-chcfj" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.350325 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c49c82dc-b231-4f2f-8553-323fcb24f67f-config-data\") pod \"glance-db-sync-chcfj\" (UID: \"c49c82dc-b231-4f2f-8553-323fcb24f67f\") " pod="openstack/glance-db-sync-chcfj" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.350356 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c49c82dc-b231-4f2f-8553-323fcb24f67f-combined-ca-bundle\") pod \"glance-db-sync-chcfj\" (UID: \"c49c82dc-b231-4f2f-8553-323fcb24f67f\") " pod="openstack/glance-db-sync-chcfj" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.350425 4790 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/87e08198-d2d0-4608-87f3-02d832fe7c55-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.351111 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87e08198-d2d0-4608-87f3-02d832fe7c55-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "87e08198-d2d0-4608-87f3-02d832fe7c55" (UID: "87e08198-d2d0-4608-87f3-02d832fe7c55"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.354425 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87e08198-d2d0-4608-87f3-02d832fe7c55-kube-api-access-485t4" (OuterVolumeSpecName: "kube-api-access-485t4") pod "87e08198-d2d0-4608-87f3-02d832fe7c55" (UID: "87e08198-d2d0-4608-87f3-02d832fe7c55"). InnerVolumeSpecName "kube-api-access-485t4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.355283 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c49c82dc-b231-4f2f-8553-323fcb24f67f-config-data\") pod \"glance-db-sync-chcfj\" (UID: \"c49c82dc-b231-4f2f-8553-323fcb24f67f\") " pod="openstack/glance-db-sync-chcfj" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.355332 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c49c82dc-b231-4f2f-8553-323fcb24f67f-combined-ca-bundle\") pod \"glance-db-sync-chcfj\" (UID: \"c49c82dc-b231-4f2f-8553-323fcb24f67f\") " pod="openstack/glance-db-sync-chcfj" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.356386 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c49c82dc-b231-4f2f-8553-323fcb24f67f-db-sync-config-data\") pod \"glance-db-sync-chcfj\" (UID: \"c49c82dc-b231-4f2f-8553-323fcb24f67f\") " pod="openstack/glance-db-sync-chcfj" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.362516 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87e08198-d2d0-4608-87f3-02d832fe7c55-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "87e08198-d2d0-4608-87f3-02d832fe7c55" (UID: "87e08198-d2d0-4608-87f3-02d832fe7c55"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.373134 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87e08198-d2d0-4608-87f3-02d832fe7c55-scripts" (OuterVolumeSpecName: "scripts") pod "87e08198-d2d0-4608-87f3-02d832fe7c55" (UID: "87e08198-d2d0-4608-87f3-02d832fe7c55"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.375817 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crwkk\" (UniqueName: \"kubernetes.io/projected/c49c82dc-b231-4f2f-8553-323fcb24f67f-kube-api-access-crwkk\") pod \"glance-db-sync-chcfj\" (UID: \"c49c82dc-b231-4f2f-8553-323fcb24f67f\") " pod="openstack/glance-db-sync-chcfj" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.376190 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87e08198-d2d0-4608-87f3-02d832fe7c55-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "87e08198-d2d0-4608-87f3-02d832fe7c55" (UID: "87e08198-d2d0-4608-87f3-02d832fe7c55"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.385167 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87e08198-d2d0-4608-87f3-02d832fe7c55-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "87e08198-d2d0-4608-87f3-02d832fe7c55" (UID: "87e08198-d2d0-4608-87f3-02d832fe7c55"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.406949 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8cc7fc4dc-5d55x"] Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.413649 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8cc7fc4dc-5d55x"] Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.452691 4790 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/87e08198-d2d0-4608-87f3-02d832fe7c55-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.452727 4790 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/87e08198-d2d0-4608-87f3-02d832fe7c55-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.452735 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87e08198-d2d0-4608-87f3-02d832fe7c55-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.452744 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-485t4\" (UniqueName: \"kubernetes.io/projected/87e08198-d2d0-4608-87f3-02d832fe7c55-kube-api-access-485t4\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.452756 4790 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/87e08198-d2d0-4608-87f3-02d832fe7c55-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.452764 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/87e08198-d2d0-4608-87f3-02d832fe7c55-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.484162 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-chcfj" Nov 24 13:32:35 crc kubenswrapper[4790]: I1124 13:32:35.999060 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-chcfj"] Nov 24 13:32:36 crc kubenswrapper[4790]: I1124 13:32:36.081947 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-48flt" event={"ID":"87e08198-d2d0-4608-87f3-02d832fe7c55","Type":"ContainerDied","Data":"cf9fcff8296cbb4ced33aad8d54fb43654949ffd0d30ffca394d175e5e4723f0"} Nov 24 13:32:36 crc kubenswrapper[4790]: I1124 13:32:36.081992 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf9fcff8296cbb4ced33aad8d54fb43654949ffd0d30ffca394d175e5e4723f0" Nov 24 13:32:36 crc kubenswrapper[4790]: I1124 13:32:36.082374 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-48flt" Nov 24 13:32:36 crc kubenswrapper[4790]: I1124 13:32:36.084095 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-chcfj" event={"ID":"c49c82dc-b231-4f2f-8553-323fcb24f67f","Type":"ContainerStarted","Data":"8e9e5ce942fbf16d6dba6cf2aeab2b51467613456380497f0bb6e48142077c47"} Nov 24 13:32:36 crc kubenswrapper[4790]: I1124 13:32:36.322314 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81f66d48-47b6-4a1c-afd1-12ce651dda1f" path="/var/lib/kubelet/pods/81f66d48-47b6-4a1c-afd1-12ce651dda1f/volumes" Nov 24 13:32:38 crc kubenswrapper[4790]: I1124 13:32:38.325138 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/63ce5126-cdab-4cdc-968a-acc26c6caf87-etc-swift\") pod \"swift-storage-0\" (UID: \"63ce5126-cdab-4cdc-968a-acc26c6caf87\") " pod="openstack/swift-storage-0" Nov 24 13:32:38 crc kubenswrapper[4790]: I1124 13:32:38.333324 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/63ce5126-cdab-4cdc-968a-acc26c6caf87-etc-swift\") pod \"swift-storage-0\" (UID: \"63ce5126-cdab-4cdc-968a-acc26c6caf87\") " pod="openstack/swift-storage-0" Nov 24 13:32:38 crc kubenswrapper[4790]: I1124 13:32:38.577467 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 24 13:32:39 crc kubenswrapper[4790]: I1124 13:32:39.067670 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 24 13:32:39 crc kubenswrapper[4790]: W1124 13:32:39.073752 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63ce5126_cdab_4cdc_968a_acc26c6caf87.slice/crio-141512db35629d7b3d10ad972903d58fad16a3d6829a08d08bd2eddc6dffaf3d WatchSource:0}: Error finding container 141512db35629d7b3d10ad972903d58fad16a3d6829a08d08bd2eddc6dffaf3d: Status 404 returned error can't find the container with id 141512db35629d7b3d10ad972903d58fad16a3d6829a08d08bd2eddc6dffaf3d Nov 24 13:32:39 crc kubenswrapper[4790]: I1124 13:32:39.106552 4790 generic.go:334] "Generic (PLEG): container finished" podID="85e04b92-ae17-4eb6-a736-395c5d44f563" containerID="569c178276bd7deaa7ff29136e16e02f63f79ee3c9e59aae1ab276151057e59e" exitCode=0 Nov 24 13:32:39 crc kubenswrapper[4790]: I1124 13:32:39.106648 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"85e04b92-ae17-4eb6-a736-395c5d44f563","Type":"ContainerDied","Data":"569c178276bd7deaa7ff29136e16e02f63f79ee3c9e59aae1ab276151057e59e"} Nov 24 13:32:39 crc kubenswrapper[4790]: I1124 13:32:39.108608 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63ce5126-cdab-4cdc-968a-acc26c6caf87","Type":"ContainerStarted","Data":"141512db35629d7b3d10ad972903d58fad16a3d6829a08d08bd2eddc6dffaf3d"} Nov 24 13:32:39 crc kubenswrapper[4790]: I1124 13:32:39.134980 4790 generic.go:334] "Generic (PLEG): container finished" podID="39508987-8655-46ed-861b-e1448652ddf6" containerID="9ed4735bba7ba4c006885cc1b29d1af4db7e93db39ebe812c28f73a898f8fa68" exitCode=0 Nov 24 13:32:39 crc kubenswrapper[4790]: I1124 13:32:39.135039 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"39508987-8655-46ed-861b-e1448652ddf6","Type":"ContainerDied","Data":"9ed4735bba7ba4c006885cc1b29d1af4db7e93db39ebe812c28f73a898f8fa68"} Nov 24 13:32:39 crc kubenswrapper[4790]: I1124 13:32:39.538989 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 24 13:32:40 crc kubenswrapper[4790]: I1124 13:32:40.145502 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"39508987-8655-46ed-861b-e1448652ddf6","Type":"ContainerStarted","Data":"921d14419dc93745c5bfa78ec576180ba71427f41b71bc188365cd076418023e"} Nov 24 13:32:40 crc kubenswrapper[4790]: I1124 13:32:40.145779 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 24 13:32:40 crc kubenswrapper[4790]: I1124 13:32:40.148401 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"85e04b92-ae17-4eb6-a736-395c5d44f563","Type":"ContainerStarted","Data":"8c7b1f8ad061938b73c9c94a4ef4bfe3303a40fe30294f1a92f9a97da7d69a3e"} Nov 24 13:32:40 crc kubenswrapper[4790]: I1124 13:32:40.148619 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:32:40 crc kubenswrapper[4790]: I1124 13:32:40.168763 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.032954537 podStartE2EDuration="56.168744834s" podCreationTimestamp="2025-11-24 13:31:44 +0000 UTC" firstStartedPulling="2025-11-24 13:31:46.426716638 +0000 UTC m=+1154.806610300" lastFinishedPulling="2025-11-24 13:32:05.562506935 +0000 UTC m=+1173.942400597" observedRunningTime="2025-11-24 13:32:40.165901486 +0000 UTC m=+1208.545795158" watchObservedRunningTime="2025-11-24 13:32:40.168744834 +0000 UTC m=+1208.548638496" Nov 24 13:32:40 crc kubenswrapper[4790]: I1124 13:32:40.196089 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.743144598 podStartE2EDuration="55.196072272s" podCreationTimestamp="2025-11-24 13:31:45 +0000 UTC" firstStartedPulling="2025-11-24 13:31:47.238334751 +0000 UTC m=+1155.618228413" lastFinishedPulling="2025-11-24 13:32:05.691262425 +0000 UTC m=+1174.071156087" observedRunningTime="2025-11-24 13:32:40.189717636 +0000 UTC m=+1208.569611308" watchObservedRunningTime="2025-11-24 13:32:40.196072272 +0000 UTC m=+1208.575965934" Nov 24 13:32:41 crc kubenswrapper[4790]: I1124 13:32:41.157605 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63ce5126-cdab-4cdc-968a-acc26c6caf87","Type":"ContainerStarted","Data":"54b741f22ad4d511172f550e71786b130b052e565c75124d9ad50b4066de4deb"} Nov 24 13:32:41 crc kubenswrapper[4790]: I1124 13:32:41.158262 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63ce5126-cdab-4cdc-968a-acc26c6caf87","Type":"ContainerStarted","Data":"23963383dd944af8fb81fedb5a24119440fcae6b836f04def348d26a2568c09f"} Nov 24 13:32:43 crc kubenswrapper[4790]: I1124 13:32:43.600142 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-rkxr6" podUID="c8fa811d-92cf-46a2-a9fd-539130369eb1" containerName="ovn-controller" probeResult="failure" output=< Nov 24 13:32:43 crc kubenswrapper[4790]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 24 13:32:43 crc kubenswrapper[4790]: > Nov 24 13:32:43 crc kubenswrapper[4790]: I1124 13:32:43.939254 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:32:43 crc kubenswrapper[4790]: I1124 13:32:43.939542 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:32:48 crc kubenswrapper[4790]: I1124 13:32:48.605180 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-rkxr6" podUID="c8fa811d-92cf-46a2-a9fd-539130369eb1" containerName="ovn-controller" probeResult="failure" output=< Nov 24 13:32:48 crc kubenswrapper[4790]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 24 13:32:48 crc kubenswrapper[4790]: > Nov 24 13:32:48 crc kubenswrapper[4790]: I1124 13:32:48.665799 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-28cwj" Nov 24 13:32:51 crc kubenswrapper[4790]: E1124 13:32:51.708196 4790 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Nov 24 13:32:51 crc kubenswrapper[4790]: E1124 13:32:51.708676 4790 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-crwkk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-chcfj_openstack(c49c82dc-b231-4f2f-8553-323fcb24f67f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:32:51 crc kubenswrapper[4790]: E1124 13:32:51.709811 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-chcfj" podUID="c49c82dc-b231-4f2f-8553-323fcb24f67f" Nov 24 13:32:52 crc kubenswrapper[4790]: I1124 13:32:52.244158 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63ce5126-cdab-4cdc-968a-acc26c6caf87","Type":"ContainerStarted","Data":"a9637198b1cf74e71f1558bec0b1a2c24a29c81f04cc33568637f3f6a30f9759"} Nov 24 13:32:52 crc kubenswrapper[4790]: I1124 13:32:52.244457 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63ce5126-cdab-4cdc-968a-acc26c6caf87","Type":"ContainerStarted","Data":"b3904f2398af315163d3ca972a5d587c4d9aa8034f444bc2b60c9b44aa285d9d"} Nov 24 13:32:52 crc kubenswrapper[4790]: E1124 13:32:52.245688 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-chcfj" podUID="c49c82dc-b231-4f2f-8553-323fcb24f67f" Nov 24 13:32:53 crc kubenswrapper[4790]: I1124 13:32:53.253684 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63ce5126-cdab-4cdc-968a-acc26c6caf87","Type":"ContainerStarted","Data":"3122135b69c368fd946107ad82977792d44cf6ae8bfae58942ad6057773fa295"} Nov 24 13:32:53 crc kubenswrapper[4790]: I1124 13:32:53.596687 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-rkxr6" podUID="c8fa811d-92cf-46a2-a9fd-539130369eb1" containerName="ovn-controller" probeResult="failure" output=< Nov 24 13:32:53 crc kubenswrapper[4790]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 24 13:32:53 crc kubenswrapper[4790]: > Nov 24 13:32:53 crc kubenswrapper[4790]: I1124 13:32:53.636141 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-28cwj" Nov 24 13:32:53 crc kubenswrapper[4790]: I1124 13:32:53.845341 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-rkxr6-config-hdb6r"] Nov 24 13:32:53 crc kubenswrapper[4790]: E1124 13:32:53.845685 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87e08198-d2d0-4608-87f3-02d832fe7c55" containerName="swift-ring-rebalance" Nov 24 13:32:53 crc kubenswrapper[4790]: I1124 13:32:53.845702 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="87e08198-d2d0-4608-87f3-02d832fe7c55" containerName="swift-ring-rebalance" Nov 24 13:32:53 crc kubenswrapper[4790]: I1124 13:32:53.845848 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="87e08198-d2d0-4608-87f3-02d832fe7c55" containerName="swift-ring-rebalance" Nov 24 13:32:53 crc kubenswrapper[4790]: I1124 13:32:53.846383 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rkxr6-config-hdb6r" Nov 24 13:32:53 crc kubenswrapper[4790]: I1124 13:32:53.848968 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 24 13:32:53 crc kubenswrapper[4790]: I1124 13:32:53.864314 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rkxr6-config-hdb6r"] Nov 24 13:32:53 crc kubenswrapper[4790]: I1124 13:32:53.921288 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9a5d8060-35d1-4fc5-9037-d6eb09762735-var-run-ovn\") pod \"ovn-controller-rkxr6-config-hdb6r\" (UID: \"9a5d8060-35d1-4fc5-9037-d6eb09762735\") " pod="openstack/ovn-controller-rkxr6-config-hdb6r" Nov 24 13:32:53 crc kubenswrapper[4790]: I1124 13:32:53.921343 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9a5d8060-35d1-4fc5-9037-d6eb09762735-scripts\") pod \"ovn-controller-rkxr6-config-hdb6r\" (UID: \"9a5d8060-35d1-4fc5-9037-d6eb09762735\") " pod="openstack/ovn-controller-rkxr6-config-hdb6r" Nov 24 13:32:53 crc kubenswrapper[4790]: I1124 13:32:53.921368 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9a5d8060-35d1-4fc5-9037-d6eb09762735-additional-scripts\") pod \"ovn-controller-rkxr6-config-hdb6r\" (UID: \"9a5d8060-35d1-4fc5-9037-d6eb09762735\") " pod="openstack/ovn-controller-rkxr6-config-hdb6r" Nov 24 13:32:53 crc kubenswrapper[4790]: I1124 13:32:53.921389 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9a5d8060-35d1-4fc5-9037-d6eb09762735-var-log-ovn\") pod \"ovn-controller-rkxr6-config-hdb6r\" (UID: \"9a5d8060-35d1-4fc5-9037-d6eb09762735\") " pod="openstack/ovn-controller-rkxr6-config-hdb6r" Nov 24 13:32:53 crc kubenswrapper[4790]: I1124 13:32:53.921411 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lg86\" (UniqueName: \"kubernetes.io/projected/9a5d8060-35d1-4fc5-9037-d6eb09762735-kube-api-access-2lg86\") pod \"ovn-controller-rkxr6-config-hdb6r\" (UID: \"9a5d8060-35d1-4fc5-9037-d6eb09762735\") " pod="openstack/ovn-controller-rkxr6-config-hdb6r" Nov 24 13:32:53 crc kubenswrapper[4790]: I1124 13:32:53.921471 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9a5d8060-35d1-4fc5-9037-d6eb09762735-var-run\") pod \"ovn-controller-rkxr6-config-hdb6r\" (UID: \"9a5d8060-35d1-4fc5-9037-d6eb09762735\") " pod="openstack/ovn-controller-rkxr6-config-hdb6r" Nov 24 13:32:54 crc kubenswrapper[4790]: I1124 13:32:54.022799 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9a5d8060-35d1-4fc5-9037-d6eb09762735-var-run-ovn\") pod \"ovn-controller-rkxr6-config-hdb6r\" (UID: \"9a5d8060-35d1-4fc5-9037-d6eb09762735\") " pod="openstack/ovn-controller-rkxr6-config-hdb6r" Nov 24 13:32:54 crc kubenswrapper[4790]: I1124 13:32:54.022863 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9a5d8060-35d1-4fc5-9037-d6eb09762735-scripts\") pod \"ovn-controller-rkxr6-config-hdb6r\" (UID: \"9a5d8060-35d1-4fc5-9037-d6eb09762735\") " pod="openstack/ovn-controller-rkxr6-config-hdb6r" Nov 24 13:32:54 crc kubenswrapper[4790]: I1124 13:32:54.022914 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9a5d8060-35d1-4fc5-9037-d6eb09762735-additional-scripts\") pod \"ovn-controller-rkxr6-config-hdb6r\" (UID: \"9a5d8060-35d1-4fc5-9037-d6eb09762735\") " pod="openstack/ovn-controller-rkxr6-config-hdb6r" Nov 24 13:32:54 crc kubenswrapper[4790]: I1124 13:32:54.022940 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9a5d8060-35d1-4fc5-9037-d6eb09762735-var-log-ovn\") pod \"ovn-controller-rkxr6-config-hdb6r\" (UID: \"9a5d8060-35d1-4fc5-9037-d6eb09762735\") " pod="openstack/ovn-controller-rkxr6-config-hdb6r" Nov 24 13:32:54 crc kubenswrapper[4790]: I1124 13:32:54.022968 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lg86\" (UniqueName: \"kubernetes.io/projected/9a5d8060-35d1-4fc5-9037-d6eb09762735-kube-api-access-2lg86\") pod \"ovn-controller-rkxr6-config-hdb6r\" (UID: \"9a5d8060-35d1-4fc5-9037-d6eb09762735\") " pod="openstack/ovn-controller-rkxr6-config-hdb6r" Nov 24 13:32:54 crc kubenswrapper[4790]: I1124 13:32:54.023229 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9a5d8060-35d1-4fc5-9037-d6eb09762735-var-run-ovn\") pod \"ovn-controller-rkxr6-config-hdb6r\" (UID: \"9a5d8060-35d1-4fc5-9037-d6eb09762735\") " pod="openstack/ovn-controller-rkxr6-config-hdb6r" Nov 24 13:32:54 crc kubenswrapper[4790]: I1124 13:32:54.023245 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9a5d8060-35d1-4fc5-9037-d6eb09762735-var-log-ovn\") pod \"ovn-controller-rkxr6-config-hdb6r\" (UID: \"9a5d8060-35d1-4fc5-9037-d6eb09762735\") " pod="openstack/ovn-controller-rkxr6-config-hdb6r" Nov 24 13:32:54 crc kubenswrapper[4790]: I1124 13:32:54.023384 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9a5d8060-35d1-4fc5-9037-d6eb09762735-var-run\") pod \"ovn-controller-rkxr6-config-hdb6r\" (UID: \"9a5d8060-35d1-4fc5-9037-d6eb09762735\") " pod="openstack/ovn-controller-rkxr6-config-hdb6r" Nov 24 13:32:54 crc kubenswrapper[4790]: I1124 13:32:54.023487 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9a5d8060-35d1-4fc5-9037-d6eb09762735-var-run\") pod \"ovn-controller-rkxr6-config-hdb6r\" (UID: \"9a5d8060-35d1-4fc5-9037-d6eb09762735\") " pod="openstack/ovn-controller-rkxr6-config-hdb6r" Nov 24 13:32:54 crc kubenswrapper[4790]: I1124 13:32:54.023726 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9a5d8060-35d1-4fc5-9037-d6eb09762735-additional-scripts\") pod \"ovn-controller-rkxr6-config-hdb6r\" (UID: \"9a5d8060-35d1-4fc5-9037-d6eb09762735\") " pod="openstack/ovn-controller-rkxr6-config-hdb6r" Nov 24 13:32:54 crc kubenswrapper[4790]: I1124 13:32:54.025011 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9a5d8060-35d1-4fc5-9037-d6eb09762735-scripts\") pod \"ovn-controller-rkxr6-config-hdb6r\" (UID: \"9a5d8060-35d1-4fc5-9037-d6eb09762735\") " pod="openstack/ovn-controller-rkxr6-config-hdb6r" Nov 24 13:32:54 crc kubenswrapper[4790]: I1124 13:32:54.041708 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lg86\" (UniqueName: \"kubernetes.io/projected/9a5d8060-35d1-4fc5-9037-d6eb09762735-kube-api-access-2lg86\") pod \"ovn-controller-rkxr6-config-hdb6r\" (UID: \"9a5d8060-35d1-4fc5-9037-d6eb09762735\") " pod="openstack/ovn-controller-rkxr6-config-hdb6r" Nov 24 13:32:54 crc kubenswrapper[4790]: I1124 13:32:54.163596 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rkxr6-config-hdb6r" Nov 24 13:32:54 crc kubenswrapper[4790]: I1124 13:32:54.307350 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63ce5126-cdab-4cdc-968a-acc26c6caf87","Type":"ContainerStarted","Data":"09f43e9aef5856b296a5ef62f8bc35fdd45d245130733c53d7e35377992a954b"} Nov 24 13:32:54 crc kubenswrapper[4790]: I1124 13:32:54.307788 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63ce5126-cdab-4cdc-968a-acc26c6caf87","Type":"ContainerStarted","Data":"c08ee946a46469851d49ebb63d97976087646f7f74bc00d1aba900aeb1b780eb"} Nov 24 13:32:54 crc kubenswrapper[4790]: I1124 13:32:54.307798 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63ce5126-cdab-4cdc-968a-acc26c6caf87","Type":"ContainerStarted","Data":"891362e93b7800953eed0ac6306309542589bcd92b30aa5ef910fdb82394fd3d"} Nov 24 13:32:54 crc kubenswrapper[4790]: I1124 13:32:54.593983 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rkxr6-config-hdb6r"] Nov 24 13:32:55 crc kubenswrapper[4790]: I1124 13:32:55.318191 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rkxr6-config-hdb6r" event={"ID":"9a5d8060-35d1-4fc5-9037-d6eb09762735","Type":"ContainerStarted","Data":"3ba62451b6b4646aa85e946c0688dc0e1bb36dca41e3cd99558373a169b528f8"} Nov 24 13:32:55 crc kubenswrapper[4790]: I1124 13:32:55.318448 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rkxr6-config-hdb6r" event={"ID":"9a5d8060-35d1-4fc5-9037-d6eb09762735","Type":"ContainerStarted","Data":"89c658dc9dad0d83350d72f7a08e25ea73aaf2cfbe16b1b7a16a384f8d4e6fcc"} Nov 24 13:32:55 crc kubenswrapper[4790]: I1124 13:32:55.336736 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-rkxr6-config-hdb6r" podStartSLOduration=2.336711282 podStartE2EDuration="2.336711282s" podCreationTimestamp="2025-11-24 13:32:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:32:55.330938892 +0000 UTC m=+1223.710832554" watchObservedRunningTime="2025-11-24 13:32:55.336711282 +0000 UTC m=+1223.716604954" Nov 24 13:32:55 crc kubenswrapper[4790]: I1124 13:32:55.932065 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.334238 4790 generic.go:334] "Generic (PLEG): container finished" podID="9a5d8060-35d1-4fc5-9037-d6eb09762735" containerID="3ba62451b6b4646aa85e946c0688dc0e1bb36dca41e3cd99558373a169b528f8" exitCode=0 Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.338558 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rkxr6-config-hdb6r" event={"ID":"9a5d8060-35d1-4fc5-9037-d6eb09762735","Type":"ContainerDied","Data":"3ba62451b6b4646aa85e946c0688dc0e1bb36dca41e3cd99558373a169b528f8"} Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.358176 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63ce5126-cdab-4cdc-968a-acc26c6caf87","Type":"ContainerStarted","Data":"b471731f94d67b5b7e6123cfb1a2a5c179f4f9ec0d47b37c19690321409b4711"} Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.358239 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63ce5126-cdab-4cdc-968a-acc26c6caf87","Type":"ContainerStarted","Data":"d23adafa420c3790e972ee180913c8f5a9453ac2fffd584b5675a60e8c5a418a"} Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.358253 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63ce5126-cdab-4cdc-968a-acc26c6caf87","Type":"ContainerStarted","Data":"ca07858ec58622d34fa209b9cd36987ee9d21a975afb78c55e5eebc7646c3b1a"} Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.384260 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-8vcj4"] Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.386254 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-8vcj4" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.436803 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-8vcj4"] Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.467276 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-c394-account-create-6kqhg"] Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.469029 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-c394-account-create-6kqhg" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.472846 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.502959 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-c394-account-create-6kqhg"] Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.518889 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-a018-account-create-q2z5q"] Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.520276 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a018-account-create-q2z5q" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.523795 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.529564 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f14beb2d-e165-43a0-a038-4cb2a3a86491-operator-scripts\") pod \"cinder-db-create-8vcj4\" (UID: \"f14beb2d-e165-43a0-a038-4cb2a3a86491\") " pod="openstack/cinder-db-create-8vcj4" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.529646 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6c6p\" (UniqueName: \"kubernetes.io/projected/f14beb2d-e165-43a0-a038-4cb2a3a86491-kube-api-access-p6c6p\") pod \"cinder-db-create-8vcj4\" (UID: \"f14beb2d-e165-43a0-a038-4cb2a3a86491\") " pod="openstack/cinder-db-create-8vcj4" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.551261 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-mlh9g"] Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.552735 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-mlh9g" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.581606 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-a018-account-create-q2z5q"] Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.594009 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-mlh9g"] Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.631661 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gr4w9\" (UniqueName: \"kubernetes.io/projected/937f8268-ef33-440e-a297-d7c1b1cbefc0-kube-api-access-gr4w9\") pod \"barbican-db-create-mlh9g\" (UID: \"937f8268-ef33-440e-a297-d7c1b1cbefc0\") " pod="openstack/barbican-db-create-mlh9g" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.631733 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8p8n\" (UniqueName: \"kubernetes.io/projected/88b90d01-f375-4158-9dd6-8042fdfbbd74-kube-api-access-z8p8n\") pod \"cinder-c394-account-create-6kqhg\" (UID: \"88b90d01-f375-4158-9dd6-8042fdfbbd74\") " pod="openstack/cinder-c394-account-create-6kqhg" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.631834 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f14beb2d-e165-43a0-a038-4cb2a3a86491-operator-scripts\") pod \"cinder-db-create-8vcj4\" (UID: \"f14beb2d-e165-43a0-a038-4cb2a3a86491\") " pod="openstack/cinder-db-create-8vcj4" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.631910 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6c6p\" (UniqueName: \"kubernetes.io/projected/f14beb2d-e165-43a0-a038-4cb2a3a86491-kube-api-access-p6c6p\") pod \"cinder-db-create-8vcj4\" (UID: \"f14beb2d-e165-43a0-a038-4cb2a3a86491\") " pod="openstack/cinder-db-create-8vcj4" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.632016 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c2d52e6-25f1-4ef0-a50e-6a8564c5302a-operator-scripts\") pod \"barbican-a018-account-create-q2z5q\" (UID: \"4c2d52e6-25f1-4ef0-a50e-6a8564c5302a\") " pod="openstack/barbican-a018-account-create-q2z5q" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.632079 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzphc\" (UniqueName: \"kubernetes.io/projected/4c2d52e6-25f1-4ef0-a50e-6a8564c5302a-kube-api-access-wzphc\") pod \"barbican-a018-account-create-q2z5q\" (UID: \"4c2d52e6-25f1-4ef0-a50e-6a8564c5302a\") " pod="openstack/barbican-a018-account-create-q2z5q" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.632104 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/937f8268-ef33-440e-a297-d7c1b1cbefc0-operator-scripts\") pod \"barbican-db-create-mlh9g\" (UID: \"937f8268-ef33-440e-a297-d7c1b1cbefc0\") " pod="openstack/barbican-db-create-mlh9g" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.632136 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/88b90d01-f375-4158-9dd6-8042fdfbbd74-operator-scripts\") pod \"cinder-c394-account-create-6kqhg\" (UID: \"88b90d01-f375-4158-9dd6-8042fdfbbd74\") " pod="openstack/cinder-c394-account-create-6kqhg" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.633037 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f14beb2d-e165-43a0-a038-4cb2a3a86491-operator-scripts\") pod \"cinder-db-create-8vcj4\" (UID: \"f14beb2d-e165-43a0-a038-4cb2a3a86491\") " pod="openstack/cinder-db-create-8vcj4" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.641106 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.669306 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6c6p\" (UniqueName: \"kubernetes.io/projected/f14beb2d-e165-43a0-a038-4cb2a3a86491-kube-api-access-p6c6p\") pod \"cinder-db-create-8vcj4\" (UID: \"f14beb2d-e165-43a0-a038-4cb2a3a86491\") " pod="openstack/cinder-db-create-8vcj4" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.697664 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-86tcg"] Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.699147 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-86tcg" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.716825 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-86tcg"] Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.734779 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzphc\" (UniqueName: \"kubernetes.io/projected/4c2d52e6-25f1-4ef0-a50e-6a8564c5302a-kube-api-access-wzphc\") pod \"barbican-a018-account-create-q2z5q\" (UID: \"4c2d52e6-25f1-4ef0-a50e-6a8564c5302a\") " pod="openstack/barbican-a018-account-create-q2z5q" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.734829 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/937f8268-ef33-440e-a297-d7c1b1cbefc0-operator-scripts\") pod \"barbican-db-create-mlh9g\" (UID: \"937f8268-ef33-440e-a297-d7c1b1cbefc0\") " pod="openstack/barbican-db-create-mlh9g" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.734858 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/88b90d01-f375-4158-9dd6-8042fdfbbd74-operator-scripts\") pod \"cinder-c394-account-create-6kqhg\" (UID: \"88b90d01-f375-4158-9dd6-8042fdfbbd74\") " pod="openstack/cinder-c394-account-create-6kqhg" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.734874 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gr4w9\" (UniqueName: \"kubernetes.io/projected/937f8268-ef33-440e-a297-d7c1b1cbefc0-kube-api-access-gr4w9\") pod \"barbican-db-create-mlh9g\" (UID: \"937f8268-ef33-440e-a297-d7c1b1cbefc0\") " pod="openstack/barbican-db-create-mlh9g" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.734907 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8p8n\" (UniqueName: \"kubernetes.io/projected/88b90d01-f375-4158-9dd6-8042fdfbbd74-kube-api-access-z8p8n\") pod \"cinder-c394-account-create-6kqhg\" (UID: \"88b90d01-f375-4158-9dd6-8042fdfbbd74\") " pod="openstack/cinder-c394-account-create-6kqhg" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.735068 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c2d52e6-25f1-4ef0-a50e-6a8564c5302a-operator-scripts\") pod \"barbican-a018-account-create-q2z5q\" (UID: \"4c2d52e6-25f1-4ef0-a50e-6a8564c5302a\") " pod="openstack/barbican-a018-account-create-q2z5q" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.735784 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c2d52e6-25f1-4ef0-a50e-6a8564c5302a-operator-scripts\") pod \"barbican-a018-account-create-q2z5q\" (UID: \"4c2d52e6-25f1-4ef0-a50e-6a8564c5302a\") " pod="openstack/barbican-a018-account-create-q2z5q" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.736568 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/88b90d01-f375-4158-9dd6-8042fdfbbd74-operator-scripts\") pod \"cinder-c394-account-create-6kqhg\" (UID: \"88b90d01-f375-4158-9dd6-8042fdfbbd74\") " pod="openstack/cinder-c394-account-create-6kqhg" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.737237 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/937f8268-ef33-440e-a297-d7c1b1cbefc0-operator-scripts\") pod \"barbican-db-create-mlh9g\" (UID: \"937f8268-ef33-440e-a297-d7c1b1cbefc0\") " pod="openstack/barbican-db-create-mlh9g" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.766462 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-8vcj4" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.768048 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-f6b9-account-create-zfn8m"] Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.769129 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f6b9-account-create-zfn8m" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.771990 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzphc\" (UniqueName: \"kubernetes.io/projected/4c2d52e6-25f1-4ef0-a50e-6a8564c5302a-kube-api-access-wzphc\") pod \"barbican-a018-account-create-q2z5q\" (UID: \"4c2d52e6-25f1-4ef0-a50e-6a8564c5302a\") " pod="openstack/barbican-a018-account-create-q2z5q" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.774906 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.782172 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-f6b9-account-create-zfn8m"] Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.782648 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gr4w9\" (UniqueName: \"kubernetes.io/projected/937f8268-ef33-440e-a297-d7c1b1cbefc0-kube-api-access-gr4w9\") pod \"barbican-db-create-mlh9g\" (UID: \"937f8268-ef33-440e-a297-d7c1b1cbefc0\") " pod="openstack/barbican-db-create-mlh9g" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.782831 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8p8n\" (UniqueName: \"kubernetes.io/projected/88b90d01-f375-4158-9dd6-8042fdfbbd74-kube-api-access-z8p8n\") pod \"cinder-c394-account-create-6kqhg\" (UID: \"88b90d01-f375-4158-9dd6-8042fdfbbd74\") " pod="openstack/cinder-c394-account-create-6kqhg" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.815612 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-c394-account-create-6kqhg" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.836690 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad24288a-7c41-4e73-b1f3-c403b8646ff7-operator-scripts\") pod \"neutron-f6b9-account-create-zfn8m\" (UID: \"ad24288a-7c41-4e73-b1f3-c403b8646ff7\") " pod="openstack/neutron-f6b9-account-create-zfn8m" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.836757 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmm7w\" (UniqueName: \"kubernetes.io/projected/294bfffa-4d3e-43ff-b86c-26cc00b81b19-kube-api-access-hmm7w\") pod \"neutron-db-create-86tcg\" (UID: \"294bfffa-4d3e-43ff-b86c-26cc00b81b19\") " pod="openstack/neutron-db-create-86tcg" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.836799 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/294bfffa-4d3e-43ff-b86c-26cc00b81b19-operator-scripts\") pod \"neutron-db-create-86tcg\" (UID: \"294bfffa-4d3e-43ff-b86c-26cc00b81b19\") " pod="openstack/neutron-db-create-86tcg" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.836829 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tc6zb\" (UniqueName: \"kubernetes.io/projected/ad24288a-7c41-4e73-b1f3-c403b8646ff7-kube-api-access-tc6zb\") pod \"neutron-f6b9-account-create-zfn8m\" (UID: \"ad24288a-7c41-4e73-b1f3-c403b8646ff7\") " pod="openstack/neutron-f6b9-account-create-zfn8m" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.858239 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a018-account-create-q2z5q" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.885465 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-mlh9g" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.894508 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-m27jn"] Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.895857 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-m27jn" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.913567 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.918520 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xgjnb" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.918703 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.918805 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.933366 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-m27jn"] Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.939202 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad24288a-7c41-4e73-b1f3-c403b8646ff7-operator-scripts\") pod \"neutron-f6b9-account-create-zfn8m\" (UID: \"ad24288a-7c41-4e73-b1f3-c403b8646ff7\") " pod="openstack/neutron-f6b9-account-create-zfn8m" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.939263 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmm7w\" (UniqueName: \"kubernetes.io/projected/294bfffa-4d3e-43ff-b86c-26cc00b81b19-kube-api-access-hmm7w\") pod \"neutron-db-create-86tcg\" (UID: \"294bfffa-4d3e-43ff-b86c-26cc00b81b19\") " pod="openstack/neutron-db-create-86tcg" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.939291 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/294bfffa-4d3e-43ff-b86c-26cc00b81b19-operator-scripts\") pod \"neutron-db-create-86tcg\" (UID: \"294bfffa-4d3e-43ff-b86c-26cc00b81b19\") " pod="openstack/neutron-db-create-86tcg" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.939314 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tc6zb\" (UniqueName: \"kubernetes.io/projected/ad24288a-7c41-4e73-b1f3-c403b8646ff7-kube-api-access-tc6zb\") pod \"neutron-f6b9-account-create-zfn8m\" (UID: \"ad24288a-7c41-4e73-b1f3-c403b8646ff7\") " pod="openstack/neutron-f6b9-account-create-zfn8m" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.940526 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad24288a-7c41-4e73-b1f3-c403b8646ff7-operator-scripts\") pod \"neutron-f6b9-account-create-zfn8m\" (UID: \"ad24288a-7c41-4e73-b1f3-c403b8646ff7\") " pod="openstack/neutron-f6b9-account-create-zfn8m" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.941027 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/294bfffa-4d3e-43ff-b86c-26cc00b81b19-operator-scripts\") pod \"neutron-db-create-86tcg\" (UID: \"294bfffa-4d3e-43ff-b86c-26cc00b81b19\") " pod="openstack/neutron-db-create-86tcg" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.987403 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tc6zb\" (UniqueName: \"kubernetes.io/projected/ad24288a-7c41-4e73-b1f3-c403b8646ff7-kube-api-access-tc6zb\") pod \"neutron-f6b9-account-create-zfn8m\" (UID: \"ad24288a-7c41-4e73-b1f3-c403b8646ff7\") " pod="openstack/neutron-f6b9-account-create-zfn8m" Nov 24 13:32:56 crc kubenswrapper[4790]: I1124 13:32:56.988442 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmm7w\" (UniqueName: \"kubernetes.io/projected/294bfffa-4d3e-43ff-b86c-26cc00b81b19-kube-api-access-hmm7w\") pod \"neutron-db-create-86tcg\" (UID: \"294bfffa-4d3e-43ff-b86c-26cc00b81b19\") " pod="openstack/neutron-db-create-86tcg" Nov 24 13:32:57 crc kubenswrapper[4790]: I1124 13:32:57.041176 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87slr\" (UniqueName: \"kubernetes.io/projected/7f04f724-1dc9-4e6a-80ab-bd139438d1b8-kube-api-access-87slr\") pod \"keystone-db-sync-m27jn\" (UID: \"7f04f724-1dc9-4e6a-80ab-bd139438d1b8\") " pod="openstack/keystone-db-sync-m27jn" Nov 24 13:32:57 crc kubenswrapper[4790]: I1124 13:32:57.041381 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f04f724-1dc9-4e6a-80ab-bd139438d1b8-combined-ca-bundle\") pod \"keystone-db-sync-m27jn\" (UID: \"7f04f724-1dc9-4e6a-80ab-bd139438d1b8\") " pod="openstack/keystone-db-sync-m27jn" Nov 24 13:32:57 crc kubenswrapper[4790]: I1124 13:32:57.041505 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f04f724-1dc9-4e6a-80ab-bd139438d1b8-config-data\") pod \"keystone-db-sync-m27jn\" (UID: \"7f04f724-1dc9-4e6a-80ab-bd139438d1b8\") " pod="openstack/keystone-db-sync-m27jn" Nov 24 13:32:57 crc kubenswrapper[4790]: I1124 13:32:57.143791 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f04f724-1dc9-4e6a-80ab-bd139438d1b8-combined-ca-bundle\") pod \"keystone-db-sync-m27jn\" (UID: \"7f04f724-1dc9-4e6a-80ab-bd139438d1b8\") " pod="openstack/keystone-db-sync-m27jn" Nov 24 13:32:57 crc kubenswrapper[4790]: I1124 13:32:57.144091 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f04f724-1dc9-4e6a-80ab-bd139438d1b8-config-data\") pod \"keystone-db-sync-m27jn\" (UID: \"7f04f724-1dc9-4e6a-80ab-bd139438d1b8\") " pod="openstack/keystone-db-sync-m27jn" Nov 24 13:32:57 crc kubenswrapper[4790]: I1124 13:32:57.144182 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87slr\" (UniqueName: \"kubernetes.io/projected/7f04f724-1dc9-4e6a-80ab-bd139438d1b8-kube-api-access-87slr\") pod \"keystone-db-sync-m27jn\" (UID: \"7f04f724-1dc9-4e6a-80ab-bd139438d1b8\") " pod="openstack/keystone-db-sync-m27jn" Nov 24 13:32:57 crc kubenswrapper[4790]: I1124 13:32:57.149720 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f04f724-1dc9-4e6a-80ab-bd139438d1b8-combined-ca-bundle\") pod \"keystone-db-sync-m27jn\" (UID: \"7f04f724-1dc9-4e6a-80ab-bd139438d1b8\") " pod="openstack/keystone-db-sync-m27jn" Nov 24 13:32:57 crc kubenswrapper[4790]: I1124 13:32:57.154190 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f04f724-1dc9-4e6a-80ab-bd139438d1b8-config-data\") pod \"keystone-db-sync-m27jn\" (UID: \"7f04f724-1dc9-4e6a-80ab-bd139438d1b8\") " pod="openstack/keystone-db-sync-m27jn" Nov 24 13:32:57 crc kubenswrapper[4790]: I1124 13:32:57.180373 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87slr\" (UniqueName: \"kubernetes.io/projected/7f04f724-1dc9-4e6a-80ab-bd139438d1b8-kube-api-access-87slr\") pod \"keystone-db-sync-m27jn\" (UID: \"7f04f724-1dc9-4e6a-80ab-bd139438d1b8\") " pod="openstack/keystone-db-sync-m27jn" Nov 24 13:32:57 crc kubenswrapper[4790]: I1124 13:32:57.236753 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-86tcg" Nov 24 13:32:57 crc kubenswrapper[4790]: I1124 13:32:57.263281 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f6b9-account-create-zfn8m" Nov 24 13:32:57 crc kubenswrapper[4790]: I1124 13:32:57.293198 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-m27jn" Nov 24 13:32:57 crc kubenswrapper[4790]: I1124 13:32:57.494073 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63ce5126-cdab-4cdc-968a-acc26c6caf87","Type":"ContainerStarted","Data":"7d0ce12b3e15290c2d364a5d9b9eb84a5315993172cd4dded95d480621cad447"} Nov 24 13:32:57 crc kubenswrapper[4790]: I1124 13:32:57.494537 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63ce5126-cdab-4cdc-968a-acc26c6caf87","Type":"ContainerStarted","Data":"c2b3e00524e13f39df13207ebcb11976c8f2a411cb2cf3f7922cf052a42ee301"} Nov 24 13:32:57 crc kubenswrapper[4790]: I1124 13:32:57.494548 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63ce5126-cdab-4cdc-968a-acc26c6caf87","Type":"ContainerStarted","Data":"578285e8c4f233014516fdb85665a982ed0526d96886404924f292cd08a273f2"} Nov 24 13:32:57 crc kubenswrapper[4790]: I1124 13:32:57.529179 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-c394-account-create-6kqhg"] Nov 24 13:32:57 crc kubenswrapper[4790]: I1124 13:32:57.690201 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-mlh9g"] Nov 24 13:32:57 crc kubenswrapper[4790]: I1124 13:32:57.708768 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-8vcj4"] Nov 24 13:32:57 crc kubenswrapper[4790]: W1124 13:32:57.720861 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod937f8268_ef33_440e_a297_d7c1b1cbefc0.slice/crio-30663fa0be6150e35ffd149c0b3ea2b7027bc63aa9eaa7d07d59a6a70e3304d2 WatchSource:0}: Error finding container 30663fa0be6150e35ffd149c0b3ea2b7027bc63aa9eaa7d07d59a6a70e3304d2: Status 404 returned error can't find the container with id 30663fa0be6150e35ffd149c0b3ea2b7027bc63aa9eaa7d07d59a6a70e3304d2 Nov 24 13:32:57 crc kubenswrapper[4790]: I1124 13:32:57.971147 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-f6b9-account-create-zfn8m"] Nov 24 13:32:57 crc kubenswrapper[4790]: I1124 13:32:57.981653 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-a018-account-create-q2z5q"] Nov 24 13:32:58 crc kubenswrapper[4790]: W1124 13:32:58.052308 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c2d52e6_25f1_4ef0_a50e_6a8564c5302a.slice/crio-9ca730be406563db199d9c919b056772e1222e6945baba2d26e1999c0274c0a8 WatchSource:0}: Error finding container 9ca730be406563db199d9c919b056772e1222e6945baba2d26e1999c0274c0a8: Status 404 returned error can't find the container with id 9ca730be406563db199d9c919b056772e1222e6945baba2d26e1999c0274c0a8 Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.061548 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-m27jn"] Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.098536 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-86tcg"] Nov 24 13:32:58 crc kubenswrapper[4790]: W1124 13:32:58.114152 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f04f724_1dc9_4e6a_80ab_bd139438d1b8.slice/crio-890dbde712c96bcd10883b865bd78dcbeb77ee324edd124af00c634f08250adc WatchSource:0}: Error finding container 890dbde712c96bcd10883b865bd78dcbeb77ee324edd124af00c634f08250adc: Status 404 returned error can't find the container with id 890dbde712c96bcd10883b865bd78dcbeb77ee324edd124af00c634f08250adc Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.206936 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rkxr6-config-hdb6r" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.308420 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9a5d8060-35d1-4fc5-9037-d6eb09762735-var-run\") pod \"9a5d8060-35d1-4fc5-9037-d6eb09762735\" (UID: \"9a5d8060-35d1-4fc5-9037-d6eb09762735\") " Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.308507 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9a5d8060-35d1-4fc5-9037-d6eb09762735-var-log-ovn\") pod \"9a5d8060-35d1-4fc5-9037-d6eb09762735\" (UID: \"9a5d8060-35d1-4fc5-9037-d6eb09762735\") " Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.308543 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9a5d8060-35d1-4fc5-9037-d6eb09762735-var-run-ovn\") pod \"9a5d8060-35d1-4fc5-9037-d6eb09762735\" (UID: \"9a5d8060-35d1-4fc5-9037-d6eb09762735\") " Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.308597 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a5d8060-35d1-4fc5-9037-d6eb09762735-var-run" (OuterVolumeSpecName: "var-run") pod "9a5d8060-35d1-4fc5-9037-d6eb09762735" (UID: "9a5d8060-35d1-4fc5-9037-d6eb09762735"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.308642 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a5d8060-35d1-4fc5-9037-d6eb09762735-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "9a5d8060-35d1-4fc5-9037-d6eb09762735" (UID: "9a5d8060-35d1-4fc5-9037-d6eb09762735"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.308656 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lg86\" (UniqueName: \"kubernetes.io/projected/9a5d8060-35d1-4fc5-9037-d6eb09762735-kube-api-access-2lg86\") pod \"9a5d8060-35d1-4fc5-9037-d6eb09762735\" (UID: \"9a5d8060-35d1-4fc5-9037-d6eb09762735\") " Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.308700 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9a5d8060-35d1-4fc5-9037-d6eb09762735-scripts\") pod \"9a5d8060-35d1-4fc5-9037-d6eb09762735\" (UID: \"9a5d8060-35d1-4fc5-9037-d6eb09762735\") " Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.308795 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9a5d8060-35d1-4fc5-9037-d6eb09762735-additional-scripts\") pod \"9a5d8060-35d1-4fc5-9037-d6eb09762735\" (UID: \"9a5d8060-35d1-4fc5-9037-d6eb09762735\") " Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.308660 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a5d8060-35d1-4fc5-9037-d6eb09762735-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "9a5d8060-35d1-4fc5-9037-d6eb09762735" (UID: "9a5d8060-35d1-4fc5-9037-d6eb09762735"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.309327 4790 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9a5d8060-35d1-4fc5-9037-d6eb09762735-var-run\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.309351 4790 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9a5d8060-35d1-4fc5-9037-d6eb09762735-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.309364 4790 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9a5d8060-35d1-4fc5-9037-d6eb09762735-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.309475 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a5d8060-35d1-4fc5-9037-d6eb09762735-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "9a5d8060-35d1-4fc5-9037-d6eb09762735" (UID: "9a5d8060-35d1-4fc5-9037-d6eb09762735"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.309676 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a5d8060-35d1-4fc5-9037-d6eb09762735-scripts" (OuterVolumeSpecName: "scripts") pod "9a5d8060-35d1-4fc5-9037-d6eb09762735" (UID: "9a5d8060-35d1-4fc5-9037-d6eb09762735"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.315031 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a5d8060-35d1-4fc5-9037-d6eb09762735-kube-api-access-2lg86" (OuterVolumeSpecName: "kube-api-access-2lg86") pod "9a5d8060-35d1-4fc5-9037-d6eb09762735" (UID: "9a5d8060-35d1-4fc5-9037-d6eb09762735"). InnerVolumeSpecName "kube-api-access-2lg86". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.411051 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lg86\" (UniqueName: \"kubernetes.io/projected/9a5d8060-35d1-4fc5-9037-d6eb09762735-kube-api-access-2lg86\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.411090 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9a5d8060-35d1-4fc5-9037-d6eb09762735-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.411101 4790 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9a5d8060-35d1-4fc5-9037-d6eb09762735-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.502940 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rkxr6-config-hdb6r" event={"ID":"9a5d8060-35d1-4fc5-9037-d6eb09762735","Type":"ContainerDied","Data":"89c658dc9dad0d83350d72f7a08e25ea73aaf2cfbe16b1b7a16a384f8d4e6fcc"} Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.502983 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="89c658dc9dad0d83350d72f7a08e25ea73aaf2cfbe16b1b7a16a384f8d4e6fcc" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.503034 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rkxr6-config-hdb6r" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.511864 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63ce5126-cdab-4cdc-968a-acc26c6caf87","Type":"ContainerStarted","Data":"b7e87dedf7cf3279e0cd55e6ed5885652aac1f5a97cb5db5d608e23cb07f7dc4"} Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.514897 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-m27jn" event={"ID":"7f04f724-1dc9-4e6a-80ab-bd139438d1b8","Type":"ContainerStarted","Data":"890dbde712c96bcd10883b865bd78dcbeb77ee324edd124af00c634f08250adc"} Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.516376 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-8vcj4" event={"ID":"f14beb2d-e165-43a0-a038-4cb2a3a86491","Type":"ContainerStarted","Data":"215ea9dbc1c1da6e42c9fce4e68a1f3b57fb111ed332cc067f9ccf293122ec60"} Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.516419 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-8vcj4" event={"ID":"f14beb2d-e165-43a0-a038-4cb2a3a86491","Type":"ContainerStarted","Data":"d6f8195a12fbdd03f7d450e3c6c70e53e605ac5e849b9689447a65e36a4bfcb1"} Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.517439 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-a018-account-create-q2z5q" event={"ID":"4c2d52e6-25f1-4ef0-a50e-6a8564c5302a","Type":"ContainerStarted","Data":"9ca730be406563db199d9c919b056772e1222e6945baba2d26e1999c0274c0a8"} Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.519301 4790 generic.go:334] "Generic (PLEG): container finished" podID="88b90d01-f375-4158-9dd6-8042fdfbbd74" containerID="5da42ceece6abe14f86b2b9b233db8ef83224e750e7ddae6f4086373f0ef1127" exitCode=0 Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.519355 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-c394-account-create-6kqhg" event={"ID":"88b90d01-f375-4158-9dd6-8042fdfbbd74","Type":"ContainerDied","Data":"5da42ceece6abe14f86b2b9b233db8ef83224e750e7ddae6f4086373f0ef1127"} Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.519375 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-c394-account-create-6kqhg" event={"ID":"88b90d01-f375-4158-9dd6-8042fdfbbd74","Type":"ContainerStarted","Data":"c4055770b5243bb8f7a6d1524becdf159ede30fa0bc0add1be6bba1f0dd5774c"} Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.520338 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f6b9-account-create-zfn8m" event={"ID":"ad24288a-7c41-4e73-b1f3-c403b8646ff7","Type":"ContainerStarted","Data":"f4c81e0fe10dfe47ccbe94c83ab5e606859df620569ecbae0e935942f6362b6b"} Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.522008 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-mlh9g" event={"ID":"937f8268-ef33-440e-a297-d7c1b1cbefc0","Type":"ContainerStarted","Data":"4e4860f11978bcccff47dec8491807d068e5a1c38c0c9e47770b0e3c265e76f7"} Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.522304 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-mlh9g" event={"ID":"937f8268-ef33-440e-a297-d7c1b1cbefc0","Type":"ContainerStarted","Data":"30663fa0be6150e35ffd149c0b3ea2b7027bc63aa9eaa7d07d59a6a70e3304d2"} Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.523173 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-86tcg" event={"ID":"294bfffa-4d3e-43ff-b86c-26cc00b81b19","Type":"ContainerStarted","Data":"d7bdf799b645ae7a914b16c2892fc2f10f2ecb930234bb1c59cbb6d747e9c2f8"} Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.543634 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-rkxr6-config-hdb6r"] Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.553445 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-rkxr6-config-hdb6r"] Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.592298 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=21.110051328 podStartE2EDuration="37.592278494s" podCreationTimestamp="2025-11-24 13:32:21 +0000 UTC" firstStartedPulling="2025-11-24 13:32:39.079458664 +0000 UTC m=+1207.459352326" lastFinishedPulling="2025-11-24 13:32:55.56168583 +0000 UTC m=+1223.941579492" observedRunningTime="2025-11-24 13:32:58.5777115 +0000 UTC m=+1226.957605162" watchObservedRunningTime="2025-11-24 13:32:58.592278494 +0000 UTC m=+1226.972172156" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.641033 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-8vcj4" podStartSLOduration=2.641008485 podStartE2EDuration="2.641008485s" podCreationTimestamp="2025-11-24 13:32:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:32:58.637063596 +0000 UTC m=+1227.016957258" watchObservedRunningTime="2025-11-24 13:32:58.641008485 +0000 UTC m=+1227.020902147" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.660393 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-mlh9g" podStartSLOduration=2.660376202 podStartE2EDuration="2.660376202s" podCreationTimestamp="2025-11-24 13:32:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:32:58.655372984 +0000 UTC m=+1227.035266636" watchObservedRunningTime="2025-11-24 13:32:58.660376202 +0000 UTC m=+1227.040269864" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.677595 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-rkxr6" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.723140 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-rkxr6-config-dgdsl"] Nov 24 13:32:58 crc kubenswrapper[4790]: E1124 13:32:58.723572 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a5d8060-35d1-4fc5-9037-d6eb09762735" containerName="ovn-config" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.723595 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a5d8060-35d1-4fc5-9037-d6eb09762735" containerName="ovn-config" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.723854 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a5d8060-35d1-4fc5-9037-d6eb09762735" containerName="ovn-config" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.724535 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rkxr6-config-dgdsl" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.727677 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.738784 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rkxr6-config-dgdsl"] Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.817992 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5398c0fb-71a5-4a3b-8750-0974556acd9b-scripts\") pod \"ovn-controller-rkxr6-config-dgdsl\" (UID: \"5398c0fb-71a5-4a3b-8750-0974556acd9b\") " pod="openstack/ovn-controller-rkxr6-config-dgdsl" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.818036 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzwd5\" (UniqueName: \"kubernetes.io/projected/5398c0fb-71a5-4a3b-8750-0974556acd9b-kube-api-access-pzwd5\") pod \"ovn-controller-rkxr6-config-dgdsl\" (UID: \"5398c0fb-71a5-4a3b-8750-0974556acd9b\") " pod="openstack/ovn-controller-rkxr6-config-dgdsl" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.818111 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5398c0fb-71a5-4a3b-8750-0974556acd9b-var-run\") pod \"ovn-controller-rkxr6-config-dgdsl\" (UID: \"5398c0fb-71a5-4a3b-8750-0974556acd9b\") " pod="openstack/ovn-controller-rkxr6-config-dgdsl" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.818189 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5398c0fb-71a5-4a3b-8750-0974556acd9b-var-log-ovn\") pod \"ovn-controller-rkxr6-config-dgdsl\" (UID: \"5398c0fb-71a5-4a3b-8750-0974556acd9b\") " pod="openstack/ovn-controller-rkxr6-config-dgdsl" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.818231 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5398c0fb-71a5-4a3b-8750-0974556acd9b-additional-scripts\") pod \"ovn-controller-rkxr6-config-dgdsl\" (UID: \"5398c0fb-71a5-4a3b-8750-0974556acd9b\") " pod="openstack/ovn-controller-rkxr6-config-dgdsl" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.818276 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5398c0fb-71a5-4a3b-8750-0974556acd9b-var-run-ovn\") pod \"ovn-controller-rkxr6-config-dgdsl\" (UID: \"5398c0fb-71a5-4a3b-8750-0974556acd9b\") " pod="openstack/ovn-controller-rkxr6-config-dgdsl" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.920353 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5398c0fb-71a5-4a3b-8750-0974556acd9b-scripts\") pod \"ovn-controller-rkxr6-config-dgdsl\" (UID: \"5398c0fb-71a5-4a3b-8750-0974556acd9b\") " pod="openstack/ovn-controller-rkxr6-config-dgdsl" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.920736 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzwd5\" (UniqueName: \"kubernetes.io/projected/5398c0fb-71a5-4a3b-8750-0974556acd9b-kube-api-access-pzwd5\") pod \"ovn-controller-rkxr6-config-dgdsl\" (UID: \"5398c0fb-71a5-4a3b-8750-0974556acd9b\") " pod="openstack/ovn-controller-rkxr6-config-dgdsl" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.920861 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5398c0fb-71a5-4a3b-8750-0974556acd9b-var-run\") pod \"ovn-controller-rkxr6-config-dgdsl\" (UID: \"5398c0fb-71a5-4a3b-8750-0974556acd9b\") " pod="openstack/ovn-controller-rkxr6-config-dgdsl" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.921001 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5398c0fb-71a5-4a3b-8750-0974556acd9b-var-log-ovn\") pod \"ovn-controller-rkxr6-config-dgdsl\" (UID: \"5398c0fb-71a5-4a3b-8750-0974556acd9b\") " pod="openstack/ovn-controller-rkxr6-config-dgdsl" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.921099 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5398c0fb-71a5-4a3b-8750-0974556acd9b-additional-scripts\") pod \"ovn-controller-rkxr6-config-dgdsl\" (UID: \"5398c0fb-71a5-4a3b-8750-0974556acd9b\") " pod="openstack/ovn-controller-rkxr6-config-dgdsl" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.921218 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5398c0fb-71a5-4a3b-8750-0974556acd9b-var-run-ovn\") pod \"ovn-controller-rkxr6-config-dgdsl\" (UID: \"5398c0fb-71a5-4a3b-8750-0974556acd9b\") " pod="openstack/ovn-controller-rkxr6-config-dgdsl" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.921417 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5398c0fb-71a5-4a3b-8750-0974556acd9b-var-run\") pod \"ovn-controller-rkxr6-config-dgdsl\" (UID: \"5398c0fb-71a5-4a3b-8750-0974556acd9b\") " pod="openstack/ovn-controller-rkxr6-config-dgdsl" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.921434 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5398c0fb-71a5-4a3b-8750-0974556acd9b-var-run-ovn\") pod \"ovn-controller-rkxr6-config-dgdsl\" (UID: \"5398c0fb-71a5-4a3b-8750-0974556acd9b\") " pod="openstack/ovn-controller-rkxr6-config-dgdsl" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.921573 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5398c0fb-71a5-4a3b-8750-0974556acd9b-var-log-ovn\") pod \"ovn-controller-rkxr6-config-dgdsl\" (UID: \"5398c0fb-71a5-4a3b-8750-0974556acd9b\") " pod="openstack/ovn-controller-rkxr6-config-dgdsl" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.922191 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5398c0fb-71a5-4a3b-8750-0974556acd9b-additional-scripts\") pod \"ovn-controller-rkxr6-config-dgdsl\" (UID: \"5398c0fb-71a5-4a3b-8750-0974556acd9b\") " pod="openstack/ovn-controller-rkxr6-config-dgdsl" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.922501 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5398c0fb-71a5-4a3b-8750-0974556acd9b-scripts\") pod \"ovn-controller-rkxr6-config-dgdsl\" (UID: \"5398c0fb-71a5-4a3b-8750-0974556acd9b\") " pod="openstack/ovn-controller-rkxr6-config-dgdsl" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.953629 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-q9jn5"] Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.955535 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-q9jn5" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.956231 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzwd5\" (UniqueName: \"kubernetes.io/projected/5398c0fb-71a5-4a3b-8750-0974556acd9b-kube-api-access-pzwd5\") pod \"ovn-controller-rkxr6-config-dgdsl\" (UID: \"5398c0fb-71a5-4a3b-8750-0974556acd9b\") " pod="openstack/ovn-controller-rkxr6-config-dgdsl" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.958105 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Nov 24 13:32:58 crc kubenswrapper[4790]: I1124 13:32:58.974814 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-q9jn5"] Nov 24 13:32:59 crc kubenswrapper[4790]: I1124 13:32:59.124512 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d291f904-9c55-4213-9cde-69d6f71b67df-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-q9jn5\" (UID: \"d291f904-9c55-4213-9cde-69d6f71b67df\") " pod="openstack/dnsmasq-dns-5c79d794d7-q9jn5" Nov 24 13:32:59 crc kubenswrapper[4790]: I1124 13:32:59.124669 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d291f904-9c55-4213-9cde-69d6f71b67df-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-q9jn5\" (UID: \"d291f904-9c55-4213-9cde-69d6f71b67df\") " pod="openstack/dnsmasq-dns-5c79d794d7-q9jn5" Nov 24 13:32:59 crc kubenswrapper[4790]: I1124 13:32:59.124747 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7n9n5\" (UniqueName: \"kubernetes.io/projected/d291f904-9c55-4213-9cde-69d6f71b67df-kube-api-access-7n9n5\") pod \"dnsmasq-dns-5c79d794d7-q9jn5\" (UID: \"d291f904-9c55-4213-9cde-69d6f71b67df\") " pod="openstack/dnsmasq-dns-5c79d794d7-q9jn5" Nov 24 13:32:59 crc kubenswrapper[4790]: I1124 13:32:59.124825 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d291f904-9c55-4213-9cde-69d6f71b67df-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-q9jn5\" (UID: \"d291f904-9c55-4213-9cde-69d6f71b67df\") " pod="openstack/dnsmasq-dns-5c79d794d7-q9jn5" Nov 24 13:32:59 crc kubenswrapper[4790]: I1124 13:32:59.124920 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d291f904-9c55-4213-9cde-69d6f71b67df-config\") pod \"dnsmasq-dns-5c79d794d7-q9jn5\" (UID: \"d291f904-9c55-4213-9cde-69d6f71b67df\") " pod="openstack/dnsmasq-dns-5c79d794d7-q9jn5" Nov 24 13:32:59 crc kubenswrapper[4790]: I1124 13:32:59.125000 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d291f904-9c55-4213-9cde-69d6f71b67df-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-q9jn5\" (UID: \"d291f904-9c55-4213-9cde-69d6f71b67df\") " pod="openstack/dnsmasq-dns-5c79d794d7-q9jn5" Nov 24 13:32:59 crc kubenswrapper[4790]: I1124 13:32:59.144965 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rkxr6-config-dgdsl" Nov 24 13:32:59 crc kubenswrapper[4790]: I1124 13:32:59.226074 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d291f904-9c55-4213-9cde-69d6f71b67df-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-q9jn5\" (UID: \"d291f904-9c55-4213-9cde-69d6f71b67df\") " pod="openstack/dnsmasq-dns-5c79d794d7-q9jn5" Nov 24 13:32:59 crc kubenswrapper[4790]: I1124 13:32:59.226441 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d291f904-9c55-4213-9cde-69d6f71b67df-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-q9jn5\" (UID: \"d291f904-9c55-4213-9cde-69d6f71b67df\") " pod="openstack/dnsmasq-dns-5c79d794d7-q9jn5" Nov 24 13:32:59 crc kubenswrapper[4790]: I1124 13:32:59.226505 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7n9n5\" (UniqueName: \"kubernetes.io/projected/d291f904-9c55-4213-9cde-69d6f71b67df-kube-api-access-7n9n5\") pod \"dnsmasq-dns-5c79d794d7-q9jn5\" (UID: \"d291f904-9c55-4213-9cde-69d6f71b67df\") " pod="openstack/dnsmasq-dns-5c79d794d7-q9jn5" Nov 24 13:32:59 crc kubenswrapper[4790]: I1124 13:32:59.226578 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d291f904-9c55-4213-9cde-69d6f71b67df-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-q9jn5\" (UID: \"d291f904-9c55-4213-9cde-69d6f71b67df\") " pod="openstack/dnsmasq-dns-5c79d794d7-q9jn5" Nov 24 13:32:59 crc kubenswrapper[4790]: I1124 13:32:59.226606 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d291f904-9c55-4213-9cde-69d6f71b67df-config\") pod \"dnsmasq-dns-5c79d794d7-q9jn5\" (UID: \"d291f904-9c55-4213-9cde-69d6f71b67df\") " pod="openstack/dnsmasq-dns-5c79d794d7-q9jn5" Nov 24 13:32:59 crc kubenswrapper[4790]: I1124 13:32:59.226647 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d291f904-9c55-4213-9cde-69d6f71b67df-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-q9jn5\" (UID: \"d291f904-9c55-4213-9cde-69d6f71b67df\") " pod="openstack/dnsmasq-dns-5c79d794d7-q9jn5" Nov 24 13:32:59 crc kubenswrapper[4790]: I1124 13:32:59.227484 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d291f904-9c55-4213-9cde-69d6f71b67df-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-q9jn5\" (UID: \"d291f904-9c55-4213-9cde-69d6f71b67df\") " pod="openstack/dnsmasq-dns-5c79d794d7-q9jn5" Nov 24 13:32:59 crc kubenswrapper[4790]: I1124 13:32:59.227908 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d291f904-9c55-4213-9cde-69d6f71b67df-config\") pod \"dnsmasq-dns-5c79d794d7-q9jn5\" (UID: \"d291f904-9c55-4213-9cde-69d6f71b67df\") " pod="openstack/dnsmasq-dns-5c79d794d7-q9jn5" Nov 24 13:32:59 crc kubenswrapper[4790]: I1124 13:32:59.228160 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d291f904-9c55-4213-9cde-69d6f71b67df-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-q9jn5\" (UID: \"d291f904-9c55-4213-9cde-69d6f71b67df\") " pod="openstack/dnsmasq-dns-5c79d794d7-q9jn5" Nov 24 13:32:59 crc kubenswrapper[4790]: I1124 13:32:59.228358 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d291f904-9c55-4213-9cde-69d6f71b67df-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-q9jn5\" (UID: \"d291f904-9c55-4213-9cde-69d6f71b67df\") " pod="openstack/dnsmasq-dns-5c79d794d7-q9jn5" Nov 24 13:32:59 crc kubenswrapper[4790]: I1124 13:32:59.228603 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d291f904-9c55-4213-9cde-69d6f71b67df-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-q9jn5\" (UID: \"d291f904-9c55-4213-9cde-69d6f71b67df\") " pod="openstack/dnsmasq-dns-5c79d794d7-q9jn5" Nov 24 13:32:59 crc kubenswrapper[4790]: I1124 13:32:59.271246 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7n9n5\" (UniqueName: \"kubernetes.io/projected/d291f904-9c55-4213-9cde-69d6f71b67df-kube-api-access-7n9n5\") pod \"dnsmasq-dns-5c79d794d7-q9jn5\" (UID: \"d291f904-9c55-4213-9cde-69d6f71b67df\") " pod="openstack/dnsmasq-dns-5c79d794d7-q9jn5" Nov 24 13:32:59 crc kubenswrapper[4790]: I1124 13:32:59.292284 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-q9jn5" Nov 24 13:32:59 crc kubenswrapper[4790]: I1124 13:32:59.554172 4790 generic.go:334] "Generic (PLEG): container finished" podID="937f8268-ef33-440e-a297-d7c1b1cbefc0" containerID="4e4860f11978bcccff47dec8491807d068e5a1c38c0c9e47770b0e3c265e76f7" exitCode=0 Nov 24 13:32:59 crc kubenswrapper[4790]: I1124 13:32:59.554499 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-mlh9g" event={"ID":"937f8268-ef33-440e-a297-d7c1b1cbefc0","Type":"ContainerDied","Data":"4e4860f11978bcccff47dec8491807d068e5a1c38c0c9e47770b0e3c265e76f7"} Nov 24 13:32:59 crc kubenswrapper[4790]: I1124 13:32:59.560449 4790 generic.go:334] "Generic (PLEG): container finished" podID="294bfffa-4d3e-43ff-b86c-26cc00b81b19" containerID="8e0b47428124851c84a07cc1375f10d5a8146d6c54766f0c2f6d6b97e1386e0f" exitCode=0 Nov 24 13:32:59 crc kubenswrapper[4790]: I1124 13:32:59.560560 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-86tcg" event={"ID":"294bfffa-4d3e-43ff-b86c-26cc00b81b19","Type":"ContainerDied","Data":"8e0b47428124851c84a07cc1375f10d5a8146d6c54766f0c2f6d6b97e1386e0f"} Nov 24 13:32:59 crc kubenswrapper[4790]: I1124 13:32:59.562747 4790 generic.go:334] "Generic (PLEG): container finished" podID="f14beb2d-e165-43a0-a038-4cb2a3a86491" containerID="215ea9dbc1c1da6e42c9fce4e68a1f3b57fb111ed332cc067f9ccf293122ec60" exitCode=0 Nov 24 13:32:59 crc kubenswrapper[4790]: I1124 13:32:59.562866 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-8vcj4" event={"ID":"f14beb2d-e165-43a0-a038-4cb2a3a86491","Type":"ContainerDied","Data":"215ea9dbc1c1da6e42c9fce4e68a1f3b57fb111ed332cc067f9ccf293122ec60"} Nov 24 13:32:59 crc kubenswrapper[4790]: I1124 13:32:59.564280 4790 generic.go:334] "Generic (PLEG): container finished" podID="4c2d52e6-25f1-4ef0-a50e-6a8564c5302a" containerID="cc10010eead4062b594ab18421178a8aa7d04ff50aeac7bc2ef5d887df82e45c" exitCode=0 Nov 24 13:32:59 crc kubenswrapper[4790]: I1124 13:32:59.564339 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-a018-account-create-q2z5q" event={"ID":"4c2d52e6-25f1-4ef0-a50e-6a8564c5302a","Type":"ContainerDied","Data":"cc10010eead4062b594ab18421178a8aa7d04ff50aeac7bc2ef5d887df82e45c"} Nov 24 13:32:59 crc kubenswrapper[4790]: I1124 13:32:59.572818 4790 generic.go:334] "Generic (PLEG): container finished" podID="ad24288a-7c41-4e73-b1f3-c403b8646ff7" containerID="d7420f1383875329ff22e1b2127be77d6b3b85841829420c95d5e97849a31ae4" exitCode=0 Nov 24 13:32:59 crc kubenswrapper[4790]: I1124 13:32:59.573008 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f6b9-account-create-zfn8m" event={"ID":"ad24288a-7c41-4e73-b1f3-c403b8646ff7","Type":"ContainerDied","Data":"d7420f1383875329ff22e1b2127be77d6b3b85841829420c95d5e97849a31ae4"} Nov 24 13:32:59 crc kubenswrapper[4790]: I1124 13:32:59.646271 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rkxr6-config-dgdsl"] Nov 24 13:32:59 crc kubenswrapper[4790]: I1124 13:32:59.835677 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-q9jn5"] Nov 24 13:33:00 crc kubenswrapper[4790]: I1124 13:33:00.025258 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-c394-account-create-6kqhg" Nov 24 13:33:00 crc kubenswrapper[4790]: I1124 13:33:00.150442 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8p8n\" (UniqueName: \"kubernetes.io/projected/88b90d01-f375-4158-9dd6-8042fdfbbd74-kube-api-access-z8p8n\") pod \"88b90d01-f375-4158-9dd6-8042fdfbbd74\" (UID: \"88b90d01-f375-4158-9dd6-8042fdfbbd74\") " Nov 24 13:33:00 crc kubenswrapper[4790]: I1124 13:33:00.150490 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/88b90d01-f375-4158-9dd6-8042fdfbbd74-operator-scripts\") pod \"88b90d01-f375-4158-9dd6-8042fdfbbd74\" (UID: \"88b90d01-f375-4158-9dd6-8042fdfbbd74\") " Nov 24 13:33:00 crc kubenswrapper[4790]: I1124 13:33:00.151450 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88b90d01-f375-4158-9dd6-8042fdfbbd74-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "88b90d01-f375-4158-9dd6-8042fdfbbd74" (UID: "88b90d01-f375-4158-9dd6-8042fdfbbd74"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:00 crc kubenswrapper[4790]: I1124 13:33:00.154578 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88b90d01-f375-4158-9dd6-8042fdfbbd74-kube-api-access-z8p8n" (OuterVolumeSpecName: "kube-api-access-z8p8n") pod "88b90d01-f375-4158-9dd6-8042fdfbbd74" (UID: "88b90d01-f375-4158-9dd6-8042fdfbbd74"). InnerVolumeSpecName "kube-api-access-z8p8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:00 crc kubenswrapper[4790]: I1124 13:33:00.252663 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8p8n\" (UniqueName: \"kubernetes.io/projected/88b90d01-f375-4158-9dd6-8042fdfbbd74-kube-api-access-z8p8n\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:00 crc kubenswrapper[4790]: I1124 13:33:00.252973 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/88b90d01-f375-4158-9dd6-8042fdfbbd74-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:00 crc kubenswrapper[4790]: I1124 13:33:00.335332 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a5d8060-35d1-4fc5-9037-d6eb09762735" path="/var/lib/kubelet/pods/9a5d8060-35d1-4fc5-9037-d6eb09762735/volumes" Nov 24 13:33:00 crc kubenswrapper[4790]: I1124 13:33:00.590798 4790 generic.go:334] "Generic (PLEG): container finished" podID="5398c0fb-71a5-4a3b-8750-0974556acd9b" containerID="f4493288d1d8aa4448843cf480499bfdbc6588351887c3da3c497a1384b9f401" exitCode=0 Nov 24 13:33:00 crc kubenswrapper[4790]: I1124 13:33:00.590901 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rkxr6-config-dgdsl" event={"ID":"5398c0fb-71a5-4a3b-8750-0974556acd9b","Type":"ContainerDied","Data":"f4493288d1d8aa4448843cf480499bfdbc6588351887c3da3c497a1384b9f401"} Nov 24 13:33:00 crc kubenswrapper[4790]: I1124 13:33:00.590940 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rkxr6-config-dgdsl" event={"ID":"5398c0fb-71a5-4a3b-8750-0974556acd9b","Type":"ContainerStarted","Data":"0158f417dd5916c9a47ae5cf350176392d53d6aa162362011580d0ac5dadad2e"} Nov 24 13:33:00 crc kubenswrapper[4790]: I1124 13:33:00.594233 4790 generic.go:334] "Generic (PLEG): container finished" podID="d291f904-9c55-4213-9cde-69d6f71b67df" containerID="af9af167ccc6e18abbb926da632a2ecaa83cb236485e364edf0866450b15d5c0" exitCode=0 Nov 24 13:33:00 crc kubenswrapper[4790]: I1124 13:33:00.594313 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-q9jn5" event={"ID":"d291f904-9c55-4213-9cde-69d6f71b67df","Type":"ContainerDied","Data":"af9af167ccc6e18abbb926da632a2ecaa83cb236485e364edf0866450b15d5c0"} Nov 24 13:33:00 crc kubenswrapper[4790]: I1124 13:33:00.594351 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-q9jn5" event={"ID":"d291f904-9c55-4213-9cde-69d6f71b67df","Type":"ContainerStarted","Data":"40cfeb31462c0912bc5fcbc5d3a82655bb12b78c2e36a398a999e9ea2589fc8c"} Nov 24 13:33:00 crc kubenswrapper[4790]: I1124 13:33:00.598261 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-c394-account-create-6kqhg" event={"ID":"88b90d01-f375-4158-9dd6-8042fdfbbd74","Type":"ContainerDied","Data":"c4055770b5243bb8f7a6d1524becdf159ede30fa0bc0add1be6bba1f0dd5774c"} Nov 24 13:33:00 crc kubenswrapper[4790]: I1124 13:33:00.598304 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c4055770b5243bb8f7a6d1524becdf159ede30fa0bc0add1be6bba1f0dd5774c" Nov 24 13:33:00 crc kubenswrapper[4790]: I1124 13:33:00.598343 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-c394-account-create-6kqhg" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.004464 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f6b9-account-create-zfn8m" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.179483 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tc6zb\" (UniqueName: \"kubernetes.io/projected/ad24288a-7c41-4e73-b1f3-c403b8646ff7-kube-api-access-tc6zb\") pod \"ad24288a-7c41-4e73-b1f3-c403b8646ff7\" (UID: \"ad24288a-7c41-4e73-b1f3-c403b8646ff7\") " Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.179819 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad24288a-7c41-4e73-b1f3-c403b8646ff7-operator-scripts\") pod \"ad24288a-7c41-4e73-b1f3-c403b8646ff7\" (UID: \"ad24288a-7c41-4e73-b1f3-c403b8646ff7\") " Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.182088 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad24288a-7c41-4e73-b1f3-c403b8646ff7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ad24288a-7c41-4e73-b1f3-c403b8646ff7" (UID: "ad24288a-7c41-4e73-b1f3-c403b8646ff7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.189243 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad24288a-7c41-4e73-b1f3-c403b8646ff7-kube-api-access-tc6zb" (OuterVolumeSpecName: "kube-api-access-tc6zb") pod "ad24288a-7c41-4e73-b1f3-c403b8646ff7" (UID: "ad24288a-7c41-4e73-b1f3-c403b8646ff7"). InnerVolumeSpecName "kube-api-access-tc6zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.257547 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-8vcj4" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.286683 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-86tcg" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.287536 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tc6zb\" (UniqueName: \"kubernetes.io/projected/ad24288a-7c41-4e73-b1f3-c403b8646ff7-kube-api-access-tc6zb\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.287566 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad24288a-7c41-4e73-b1f3-c403b8646ff7-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.293672 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-mlh9g" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.311669 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a018-account-create-q2z5q" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.388688 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6c6p\" (UniqueName: \"kubernetes.io/projected/f14beb2d-e165-43a0-a038-4cb2a3a86491-kube-api-access-p6c6p\") pod \"f14beb2d-e165-43a0-a038-4cb2a3a86491\" (UID: \"f14beb2d-e165-43a0-a038-4cb2a3a86491\") " Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.388746 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/294bfffa-4d3e-43ff-b86c-26cc00b81b19-operator-scripts\") pod \"294bfffa-4d3e-43ff-b86c-26cc00b81b19\" (UID: \"294bfffa-4d3e-43ff-b86c-26cc00b81b19\") " Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.388811 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmm7w\" (UniqueName: \"kubernetes.io/projected/294bfffa-4d3e-43ff-b86c-26cc00b81b19-kube-api-access-hmm7w\") pod \"294bfffa-4d3e-43ff-b86c-26cc00b81b19\" (UID: \"294bfffa-4d3e-43ff-b86c-26cc00b81b19\") " Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.388860 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f14beb2d-e165-43a0-a038-4cb2a3a86491-operator-scripts\") pod \"f14beb2d-e165-43a0-a038-4cb2a3a86491\" (UID: \"f14beb2d-e165-43a0-a038-4cb2a3a86491\") " Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.389010 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/937f8268-ef33-440e-a297-d7c1b1cbefc0-operator-scripts\") pod \"937f8268-ef33-440e-a297-d7c1b1cbefc0\" (UID: \"937f8268-ef33-440e-a297-d7c1b1cbefc0\") " Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.389037 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gr4w9\" (UniqueName: \"kubernetes.io/projected/937f8268-ef33-440e-a297-d7c1b1cbefc0-kube-api-access-gr4w9\") pod \"937f8268-ef33-440e-a297-d7c1b1cbefc0\" (UID: \"937f8268-ef33-440e-a297-d7c1b1cbefc0\") " Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.389587 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f14beb2d-e165-43a0-a038-4cb2a3a86491-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f14beb2d-e165-43a0-a038-4cb2a3a86491" (UID: "f14beb2d-e165-43a0-a038-4cb2a3a86491"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.389996 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/937f8268-ef33-440e-a297-d7c1b1cbefc0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "937f8268-ef33-440e-a297-d7c1b1cbefc0" (UID: "937f8268-ef33-440e-a297-d7c1b1cbefc0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.391195 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/294bfffa-4d3e-43ff-b86c-26cc00b81b19-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "294bfffa-4d3e-43ff-b86c-26cc00b81b19" (UID: "294bfffa-4d3e-43ff-b86c-26cc00b81b19"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.393146 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/937f8268-ef33-440e-a297-d7c1b1cbefc0-kube-api-access-gr4w9" (OuterVolumeSpecName: "kube-api-access-gr4w9") pod "937f8268-ef33-440e-a297-d7c1b1cbefc0" (UID: "937f8268-ef33-440e-a297-d7c1b1cbefc0"). InnerVolumeSpecName "kube-api-access-gr4w9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.393357 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f14beb2d-e165-43a0-a038-4cb2a3a86491-kube-api-access-p6c6p" (OuterVolumeSpecName: "kube-api-access-p6c6p") pod "f14beb2d-e165-43a0-a038-4cb2a3a86491" (UID: "f14beb2d-e165-43a0-a038-4cb2a3a86491"). InnerVolumeSpecName "kube-api-access-p6c6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.393055 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/294bfffa-4d3e-43ff-b86c-26cc00b81b19-kube-api-access-hmm7w" (OuterVolumeSpecName: "kube-api-access-hmm7w") pod "294bfffa-4d3e-43ff-b86c-26cc00b81b19" (UID: "294bfffa-4d3e-43ff-b86c-26cc00b81b19"). InnerVolumeSpecName "kube-api-access-hmm7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.490163 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c2d52e6-25f1-4ef0-a50e-6a8564c5302a-operator-scripts\") pod \"4c2d52e6-25f1-4ef0-a50e-6a8564c5302a\" (UID: \"4c2d52e6-25f1-4ef0-a50e-6a8564c5302a\") " Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.490212 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wzphc\" (UniqueName: \"kubernetes.io/projected/4c2d52e6-25f1-4ef0-a50e-6a8564c5302a-kube-api-access-wzphc\") pod \"4c2d52e6-25f1-4ef0-a50e-6a8564c5302a\" (UID: \"4c2d52e6-25f1-4ef0-a50e-6a8564c5302a\") " Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.490635 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6c6p\" (UniqueName: \"kubernetes.io/projected/f14beb2d-e165-43a0-a038-4cb2a3a86491-kube-api-access-p6c6p\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.490646 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/294bfffa-4d3e-43ff-b86c-26cc00b81b19-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.490656 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmm7w\" (UniqueName: \"kubernetes.io/projected/294bfffa-4d3e-43ff-b86c-26cc00b81b19-kube-api-access-hmm7w\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.490664 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f14beb2d-e165-43a0-a038-4cb2a3a86491-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.490672 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/937f8268-ef33-440e-a297-d7c1b1cbefc0-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.490681 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gr4w9\" (UniqueName: \"kubernetes.io/projected/937f8268-ef33-440e-a297-d7c1b1cbefc0-kube-api-access-gr4w9\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.491429 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c2d52e6-25f1-4ef0-a50e-6a8564c5302a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4c2d52e6-25f1-4ef0-a50e-6a8564c5302a" (UID: "4c2d52e6-25f1-4ef0-a50e-6a8564c5302a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.494148 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c2d52e6-25f1-4ef0-a50e-6a8564c5302a-kube-api-access-wzphc" (OuterVolumeSpecName: "kube-api-access-wzphc") pod "4c2d52e6-25f1-4ef0-a50e-6a8564c5302a" (UID: "4c2d52e6-25f1-4ef0-a50e-6a8564c5302a"). InnerVolumeSpecName "kube-api-access-wzphc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.593053 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wzphc\" (UniqueName: \"kubernetes.io/projected/4c2d52e6-25f1-4ef0-a50e-6a8564c5302a-kube-api-access-wzphc\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.593095 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c2d52e6-25f1-4ef0-a50e-6a8564c5302a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.607689 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-8vcj4" event={"ID":"f14beb2d-e165-43a0-a038-4cb2a3a86491","Type":"ContainerDied","Data":"d6f8195a12fbdd03f7d450e3c6c70e53e605ac5e849b9689447a65e36a4bfcb1"} Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.607738 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6f8195a12fbdd03f7d450e3c6c70e53e605ac5e849b9689447a65e36a4bfcb1" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.607805 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-8vcj4" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.610457 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-q9jn5" event={"ID":"d291f904-9c55-4213-9cde-69d6f71b67df","Type":"ContainerStarted","Data":"083bb5dfa444f0e05367f4a3e328194b3c06970f40e5a951abc781f5875f0852"} Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.610535 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c79d794d7-q9jn5" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.612690 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-a018-account-create-q2z5q" event={"ID":"4c2d52e6-25f1-4ef0-a50e-6a8564c5302a","Type":"ContainerDied","Data":"9ca730be406563db199d9c919b056772e1222e6945baba2d26e1999c0274c0a8"} Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.612744 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a018-account-create-q2z5q" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.612751 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ca730be406563db199d9c919b056772e1222e6945baba2d26e1999c0274c0a8" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.614219 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f6b9-account-create-zfn8m" event={"ID":"ad24288a-7c41-4e73-b1f3-c403b8646ff7","Type":"ContainerDied","Data":"f4c81e0fe10dfe47ccbe94c83ab5e606859df620569ecbae0e935942f6362b6b"} Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.614242 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4c81e0fe10dfe47ccbe94c83ab5e606859df620569ecbae0e935942f6362b6b" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.614251 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f6b9-account-create-zfn8m" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.621697 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-mlh9g" event={"ID":"937f8268-ef33-440e-a297-d7c1b1cbefc0","Type":"ContainerDied","Data":"30663fa0be6150e35ffd149c0b3ea2b7027bc63aa9eaa7d07d59a6a70e3304d2"} Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.621747 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30663fa0be6150e35ffd149c0b3ea2b7027bc63aa9eaa7d07d59a6a70e3304d2" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.621815 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-mlh9g" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.628087 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-86tcg" Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.636755 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-86tcg" event={"ID":"294bfffa-4d3e-43ff-b86c-26cc00b81b19","Type":"ContainerDied","Data":"d7bdf799b645ae7a914b16c2892fc2f10f2ecb930234bb1c59cbb6d747e9c2f8"} Nov 24 13:33:01 crc kubenswrapper[4790]: I1124 13:33:01.636792 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7bdf799b645ae7a914b16c2892fc2f10f2ecb930234bb1c59cbb6d747e9c2f8" Nov 24 13:33:02 crc kubenswrapper[4790]: I1124 13:33:02.036257 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c79d794d7-q9jn5" podStartSLOduration=4.036235179 podStartE2EDuration="4.036235179s" podCreationTimestamp="2025-11-24 13:32:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:33:01.63808781 +0000 UTC m=+1230.017981482" watchObservedRunningTime="2025-11-24 13:33:02.036235179 +0000 UTC m=+1230.416128841" Nov 24 13:33:04 crc kubenswrapper[4790]: I1124 13:33:04.467482 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rkxr6-config-dgdsl" Nov 24 13:33:04 crc kubenswrapper[4790]: I1124 13:33:04.638540 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5398c0fb-71a5-4a3b-8750-0974556acd9b-var-run\") pod \"5398c0fb-71a5-4a3b-8750-0974556acd9b\" (UID: \"5398c0fb-71a5-4a3b-8750-0974556acd9b\") " Nov 24 13:33:04 crc kubenswrapper[4790]: I1124 13:33:04.638704 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5398c0fb-71a5-4a3b-8750-0974556acd9b-var-log-ovn\") pod \"5398c0fb-71a5-4a3b-8750-0974556acd9b\" (UID: \"5398c0fb-71a5-4a3b-8750-0974556acd9b\") " Nov 24 13:33:04 crc kubenswrapper[4790]: I1124 13:33:04.638749 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5398c0fb-71a5-4a3b-8750-0974556acd9b-var-run-ovn\") pod \"5398c0fb-71a5-4a3b-8750-0974556acd9b\" (UID: \"5398c0fb-71a5-4a3b-8750-0974556acd9b\") " Nov 24 13:33:04 crc kubenswrapper[4790]: I1124 13:33:04.638855 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5398c0fb-71a5-4a3b-8750-0974556acd9b-scripts\") pod \"5398c0fb-71a5-4a3b-8750-0974556acd9b\" (UID: \"5398c0fb-71a5-4a3b-8750-0974556acd9b\") " Nov 24 13:33:04 crc kubenswrapper[4790]: I1124 13:33:04.638927 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5398c0fb-71a5-4a3b-8750-0974556acd9b-additional-scripts\") pod \"5398c0fb-71a5-4a3b-8750-0974556acd9b\" (UID: \"5398c0fb-71a5-4a3b-8750-0974556acd9b\") " Nov 24 13:33:04 crc kubenswrapper[4790]: I1124 13:33:04.638926 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5398c0fb-71a5-4a3b-8750-0974556acd9b-var-run" (OuterVolumeSpecName: "var-run") pod "5398c0fb-71a5-4a3b-8750-0974556acd9b" (UID: "5398c0fb-71a5-4a3b-8750-0974556acd9b"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:33:04 crc kubenswrapper[4790]: I1124 13:33:04.638972 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzwd5\" (UniqueName: \"kubernetes.io/projected/5398c0fb-71a5-4a3b-8750-0974556acd9b-kube-api-access-pzwd5\") pod \"5398c0fb-71a5-4a3b-8750-0974556acd9b\" (UID: \"5398c0fb-71a5-4a3b-8750-0974556acd9b\") " Nov 24 13:33:04 crc kubenswrapper[4790]: I1124 13:33:04.638988 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5398c0fb-71a5-4a3b-8750-0974556acd9b-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "5398c0fb-71a5-4a3b-8750-0974556acd9b" (UID: "5398c0fb-71a5-4a3b-8750-0974556acd9b"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:33:04 crc kubenswrapper[4790]: I1124 13:33:04.639012 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5398c0fb-71a5-4a3b-8750-0974556acd9b-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "5398c0fb-71a5-4a3b-8750-0974556acd9b" (UID: "5398c0fb-71a5-4a3b-8750-0974556acd9b"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:33:04 crc kubenswrapper[4790]: I1124 13:33:04.639416 4790 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5398c0fb-71a5-4a3b-8750-0974556acd9b-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:04 crc kubenswrapper[4790]: I1124 13:33:04.639436 4790 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5398c0fb-71a5-4a3b-8750-0974556acd9b-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:04 crc kubenswrapper[4790]: I1124 13:33:04.639449 4790 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5398c0fb-71a5-4a3b-8750-0974556acd9b-var-run\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:04 crc kubenswrapper[4790]: I1124 13:33:04.640181 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5398c0fb-71a5-4a3b-8750-0974556acd9b-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "5398c0fb-71a5-4a3b-8750-0974556acd9b" (UID: "5398c0fb-71a5-4a3b-8750-0974556acd9b"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:04 crc kubenswrapper[4790]: I1124 13:33:04.640842 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5398c0fb-71a5-4a3b-8750-0974556acd9b-scripts" (OuterVolumeSpecName: "scripts") pod "5398c0fb-71a5-4a3b-8750-0974556acd9b" (UID: "5398c0fb-71a5-4a3b-8750-0974556acd9b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:04 crc kubenswrapper[4790]: I1124 13:33:04.643862 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5398c0fb-71a5-4a3b-8750-0974556acd9b-kube-api-access-pzwd5" (OuterVolumeSpecName: "kube-api-access-pzwd5") pod "5398c0fb-71a5-4a3b-8750-0974556acd9b" (UID: "5398c0fb-71a5-4a3b-8750-0974556acd9b"). InnerVolumeSpecName "kube-api-access-pzwd5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:04 crc kubenswrapper[4790]: I1124 13:33:04.657173 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rkxr6-config-dgdsl" event={"ID":"5398c0fb-71a5-4a3b-8750-0974556acd9b","Type":"ContainerDied","Data":"0158f417dd5916c9a47ae5cf350176392d53d6aa162362011580d0ac5dadad2e"} Nov 24 13:33:04 crc kubenswrapper[4790]: I1124 13:33:04.657229 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0158f417dd5916c9a47ae5cf350176392d53d6aa162362011580d0ac5dadad2e" Nov 24 13:33:04 crc kubenswrapper[4790]: I1124 13:33:04.657259 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rkxr6-config-dgdsl" Nov 24 13:33:04 crc kubenswrapper[4790]: I1124 13:33:04.741980 4790 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5398c0fb-71a5-4a3b-8750-0974556acd9b-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:04 crc kubenswrapper[4790]: I1124 13:33:04.742276 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzwd5\" (UniqueName: \"kubernetes.io/projected/5398c0fb-71a5-4a3b-8750-0974556acd9b-kube-api-access-pzwd5\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:04 crc kubenswrapper[4790]: I1124 13:33:04.742286 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5398c0fb-71a5-4a3b-8750-0974556acd9b-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:05 crc kubenswrapper[4790]: I1124 13:33:05.540354 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-rkxr6-config-dgdsl"] Nov 24 13:33:05 crc kubenswrapper[4790]: I1124 13:33:05.548778 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-rkxr6-config-dgdsl"] Nov 24 13:33:05 crc kubenswrapper[4790]: I1124 13:33:05.667151 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-chcfj" event={"ID":"c49c82dc-b231-4f2f-8553-323fcb24f67f","Type":"ContainerStarted","Data":"2c6c3b4bcd309e9bbfe6304002cfac512cde3ab86acfec0e937b2af1a62d7ef2"} Nov 24 13:33:05 crc kubenswrapper[4790]: I1124 13:33:05.668460 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-m27jn" event={"ID":"7f04f724-1dc9-4e6a-80ab-bd139438d1b8","Type":"ContainerStarted","Data":"a92ec3cde75257483e0a8e30ef0c90a883c89a010f5b7ce77f5f75db1ebfab68"} Nov 24 13:33:05 crc kubenswrapper[4790]: I1124 13:33:05.684738 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-chcfj" podStartSLOduration=2.183169019 podStartE2EDuration="30.684719225s" podCreationTimestamp="2025-11-24 13:32:35 +0000 UTC" firstStartedPulling="2025-11-24 13:32:36.002253507 +0000 UTC m=+1204.382147169" lastFinishedPulling="2025-11-24 13:33:04.503803713 +0000 UTC m=+1232.883697375" observedRunningTime="2025-11-24 13:33:05.682753111 +0000 UTC m=+1234.062646773" watchObservedRunningTime="2025-11-24 13:33:05.684719225 +0000 UTC m=+1234.064612887" Nov 24 13:33:05 crc kubenswrapper[4790]: I1124 13:33:05.704854 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-m27jn" podStartSLOduration=3.340698035 podStartE2EDuration="9.704833343s" podCreationTimestamp="2025-11-24 13:32:56 +0000 UTC" firstStartedPulling="2025-11-24 13:32:58.154458595 +0000 UTC m=+1226.534352267" lastFinishedPulling="2025-11-24 13:33:04.518593913 +0000 UTC m=+1232.898487575" observedRunningTime="2025-11-24 13:33:05.702255741 +0000 UTC m=+1234.082149403" watchObservedRunningTime="2025-11-24 13:33:05.704833343 +0000 UTC m=+1234.084727005" Nov 24 13:33:06 crc kubenswrapper[4790]: I1124 13:33:06.324440 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5398c0fb-71a5-4a3b-8750-0974556acd9b" path="/var/lib/kubelet/pods/5398c0fb-71a5-4a3b-8750-0974556acd9b/volumes" Nov 24 13:33:09 crc kubenswrapper[4790]: I1124 13:33:09.294009 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c79d794d7-q9jn5" Nov 24 13:33:09 crc kubenswrapper[4790]: I1124 13:33:09.346336 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-b4bt2"] Nov 24 13:33:09 crc kubenswrapper[4790]: I1124 13:33:09.346620 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-b4bt2" podUID="5bde9a3b-eafa-454f-b6f2-af094d594752" containerName="dnsmasq-dns" containerID="cri-o://f33e9fc8e484ad15ca70bba7e44986f290e2e29ff65185a8d0930c4eb8085c4a" gracePeriod=10 Nov 24 13:33:09 crc kubenswrapper[4790]: I1124 13:33:09.706524 4790 generic.go:334] "Generic (PLEG): container finished" podID="5bde9a3b-eafa-454f-b6f2-af094d594752" containerID="f33e9fc8e484ad15ca70bba7e44986f290e2e29ff65185a8d0930c4eb8085c4a" exitCode=0 Nov 24 13:33:09 crc kubenswrapper[4790]: I1124 13:33:09.706567 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-b4bt2" event={"ID":"5bde9a3b-eafa-454f-b6f2-af094d594752","Type":"ContainerDied","Data":"f33e9fc8e484ad15ca70bba7e44986f290e2e29ff65185a8d0930c4eb8085c4a"} Nov 24 13:33:09 crc kubenswrapper[4790]: I1124 13:33:09.893903 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-b4bt2" Nov 24 13:33:10 crc kubenswrapper[4790]: I1124 13:33:10.025387 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5bde9a3b-eafa-454f-b6f2-af094d594752-ovsdbserver-nb\") pod \"5bde9a3b-eafa-454f-b6f2-af094d594752\" (UID: \"5bde9a3b-eafa-454f-b6f2-af094d594752\") " Nov 24 13:33:10 crc kubenswrapper[4790]: I1124 13:33:10.025564 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wm5qx\" (UniqueName: \"kubernetes.io/projected/5bde9a3b-eafa-454f-b6f2-af094d594752-kube-api-access-wm5qx\") pod \"5bde9a3b-eafa-454f-b6f2-af094d594752\" (UID: \"5bde9a3b-eafa-454f-b6f2-af094d594752\") " Nov 24 13:33:10 crc kubenswrapper[4790]: I1124 13:33:10.025605 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5bde9a3b-eafa-454f-b6f2-af094d594752-dns-svc\") pod \"5bde9a3b-eafa-454f-b6f2-af094d594752\" (UID: \"5bde9a3b-eafa-454f-b6f2-af094d594752\") " Nov 24 13:33:10 crc kubenswrapper[4790]: I1124 13:33:10.025644 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bde9a3b-eafa-454f-b6f2-af094d594752-config\") pod \"5bde9a3b-eafa-454f-b6f2-af094d594752\" (UID: \"5bde9a3b-eafa-454f-b6f2-af094d594752\") " Nov 24 13:33:10 crc kubenswrapper[4790]: I1124 13:33:10.025706 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5bde9a3b-eafa-454f-b6f2-af094d594752-ovsdbserver-sb\") pod \"5bde9a3b-eafa-454f-b6f2-af094d594752\" (UID: \"5bde9a3b-eafa-454f-b6f2-af094d594752\") " Nov 24 13:33:10 crc kubenswrapper[4790]: I1124 13:33:10.047135 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bde9a3b-eafa-454f-b6f2-af094d594752-kube-api-access-wm5qx" (OuterVolumeSpecName: "kube-api-access-wm5qx") pod "5bde9a3b-eafa-454f-b6f2-af094d594752" (UID: "5bde9a3b-eafa-454f-b6f2-af094d594752"). InnerVolumeSpecName "kube-api-access-wm5qx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:10 crc kubenswrapper[4790]: I1124 13:33:10.071093 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bde9a3b-eafa-454f-b6f2-af094d594752-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5bde9a3b-eafa-454f-b6f2-af094d594752" (UID: "5bde9a3b-eafa-454f-b6f2-af094d594752"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:10 crc kubenswrapper[4790]: I1124 13:33:10.077966 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bde9a3b-eafa-454f-b6f2-af094d594752-config" (OuterVolumeSpecName: "config") pod "5bde9a3b-eafa-454f-b6f2-af094d594752" (UID: "5bde9a3b-eafa-454f-b6f2-af094d594752"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:10 crc kubenswrapper[4790]: I1124 13:33:10.090440 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bde9a3b-eafa-454f-b6f2-af094d594752-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5bde9a3b-eafa-454f-b6f2-af094d594752" (UID: "5bde9a3b-eafa-454f-b6f2-af094d594752"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:10 crc kubenswrapper[4790]: I1124 13:33:10.091951 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bde9a3b-eafa-454f-b6f2-af094d594752-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5bde9a3b-eafa-454f-b6f2-af094d594752" (UID: "5bde9a3b-eafa-454f-b6f2-af094d594752"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:10 crc kubenswrapper[4790]: I1124 13:33:10.127261 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wm5qx\" (UniqueName: \"kubernetes.io/projected/5bde9a3b-eafa-454f-b6f2-af094d594752-kube-api-access-wm5qx\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:10 crc kubenswrapper[4790]: I1124 13:33:10.127295 4790 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5bde9a3b-eafa-454f-b6f2-af094d594752-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:10 crc kubenswrapper[4790]: I1124 13:33:10.127303 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bde9a3b-eafa-454f-b6f2-af094d594752-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:10 crc kubenswrapper[4790]: I1124 13:33:10.127313 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5bde9a3b-eafa-454f-b6f2-af094d594752-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:10 crc kubenswrapper[4790]: I1124 13:33:10.127322 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5bde9a3b-eafa-454f-b6f2-af094d594752-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:10 crc kubenswrapper[4790]: I1124 13:33:10.715031 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-b4bt2" event={"ID":"5bde9a3b-eafa-454f-b6f2-af094d594752","Type":"ContainerDied","Data":"565093f3d6e5aae5cfe76c48ac351d7692a9092157b87fc00681a0de3cccd850"} Nov 24 13:33:10 crc kubenswrapper[4790]: I1124 13:33:10.715091 4790 scope.go:117] "RemoveContainer" containerID="f33e9fc8e484ad15ca70bba7e44986f290e2e29ff65185a8d0930c4eb8085c4a" Nov 24 13:33:10 crc kubenswrapper[4790]: I1124 13:33:10.715103 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-b4bt2" Nov 24 13:33:10 crc kubenswrapper[4790]: I1124 13:33:10.771668 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-b4bt2"] Nov 24 13:33:10 crc kubenswrapper[4790]: I1124 13:33:10.776582 4790 scope.go:117] "RemoveContainer" containerID="78514cecd06fc8a736fc2455d2533f997ed9388d52795cf38e89cb7ee03442f8" Nov 24 13:33:10 crc kubenswrapper[4790]: I1124 13:33:10.779268 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-b4bt2"] Nov 24 13:33:12 crc kubenswrapper[4790]: I1124 13:33:12.326931 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bde9a3b-eafa-454f-b6f2-af094d594752" path="/var/lib/kubelet/pods/5bde9a3b-eafa-454f-b6f2-af094d594752/volumes" Nov 24 13:33:13 crc kubenswrapper[4790]: I1124 13:33:13.741906 4790 generic.go:334] "Generic (PLEG): container finished" podID="7f04f724-1dc9-4e6a-80ab-bd139438d1b8" containerID="a92ec3cde75257483e0a8e30ef0c90a883c89a010f5b7ce77f5f75db1ebfab68" exitCode=0 Nov 24 13:33:13 crc kubenswrapper[4790]: I1124 13:33:13.741946 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-m27jn" event={"ID":"7f04f724-1dc9-4e6a-80ab-bd139438d1b8","Type":"ContainerDied","Data":"a92ec3cde75257483e0a8e30ef0c90a883c89a010f5b7ce77f5f75db1ebfab68"} Nov 24 13:33:13 crc kubenswrapper[4790]: I1124 13:33:13.938434 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:33:13 crc kubenswrapper[4790]: I1124 13:33:13.938501 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:33:15 crc kubenswrapper[4790]: I1124 13:33:15.017548 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-m27jn" Nov 24 13:33:15 crc kubenswrapper[4790]: I1124 13:33:15.129194 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87slr\" (UniqueName: \"kubernetes.io/projected/7f04f724-1dc9-4e6a-80ab-bd139438d1b8-kube-api-access-87slr\") pod \"7f04f724-1dc9-4e6a-80ab-bd139438d1b8\" (UID: \"7f04f724-1dc9-4e6a-80ab-bd139438d1b8\") " Nov 24 13:33:15 crc kubenswrapper[4790]: I1124 13:33:15.129295 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f04f724-1dc9-4e6a-80ab-bd139438d1b8-config-data\") pod \"7f04f724-1dc9-4e6a-80ab-bd139438d1b8\" (UID: \"7f04f724-1dc9-4e6a-80ab-bd139438d1b8\") " Nov 24 13:33:15 crc kubenswrapper[4790]: I1124 13:33:15.129383 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f04f724-1dc9-4e6a-80ab-bd139438d1b8-combined-ca-bundle\") pod \"7f04f724-1dc9-4e6a-80ab-bd139438d1b8\" (UID: \"7f04f724-1dc9-4e6a-80ab-bd139438d1b8\") " Nov 24 13:33:15 crc kubenswrapper[4790]: I1124 13:33:15.156103 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f04f724-1dc9-4e6a-80ab-bd139438d1b8-kube-api-access-87slr" (OuterVolumeSpecName: "kube-api-access-87slr") pod "7f04f724-1dc9-4e6a-80ab-bd139438d1b8" (UID: "7f04f724-1dc9-4e6a-80ab-bd139438d1b8"). InnerVolumeSpecName "kube-api-access-87slr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:15 crc kubenswrapper[4790]: I1124 13:33:15.189164 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f04f724-1dc9-4e6a-80ab-bd139438d1b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7f04f724-1dc9-4e6a-80ab-bd139438d1b8" (UID: "7f04f724-1dc9-4e6a-80ab-bd139438d1b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:15 crc kubenswrapper[4790]: I1124 13:33:15.228348 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f04f724-1dc9-4e6a-80ab-bd139438d1b8-config-data" (OuterVolumeSpecName: "config-data") pod "7f04f724-1dc9-4e6a-80ab-bd139438d1b8" (UID: "7f04f724-1dc9-4e6a-80ab-bd139438d1b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:15 crc kubenswrapper[4790]: I1124 13:33:15.231287 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f04f724-1dc9-4e6a-80ab-bd139438d1b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:15 crc kubenswrapper[4790]: I1124 13:33:15.231320 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87slr\" (UniqueName: \"kubernetes.io/projected/7f04f724-1dc9-4e6a-80ab-bd139438d1b8-kube-api-access-87slr\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:15 crc kubenswrapper[4790]: I1124 13:33:15.231331 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f04f724-1dc9-4e6a-80ab-bd139438d1b8-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:15 crc kubenswrapper[4790]: I1124 13:33:15.759310 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-m27jn" event={"ID":"7f04f724-1dc9-4e6a-80ab-bd139438d1b8","Type":"ContainerDied","Data":"890dbde712c96bcd10883b865bd78dcbeb77ee324edd124af00c634f08250adc"} Nov 24 13:33:15 crc kubenswrapper[4790]: I1124 13:33:15.759524 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="890dbde712c96bcd10883b865bd78dcbeb77ee324edd124af00c634f08250adc" Nov 24 13:33:15 crc kubenswrapper[4790]: I1124 13:33:15.759381 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-m27jn" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.041628 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b868669f-8nj6l"] Nov 24 13:33:16 crc kubenswrapper[4790]: E1124 13:33:16.042362 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88b90d01-f375-4158-9dd6-8042fdfbbd74" containerName="mariadb-account-create" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.042379 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="88b90d01-f375-4158-9dd6-8042fdfbbd74" containerName="mariadb-account-create" Nov 24 13:33:16 crc kubenswrapper[4790]: E1124 13:33:16.042407 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bde9a3b-eafa-454f-b6f2-af094d594752" containerName="dnsmasq-dns" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.042415 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bde9a3b-eafa-454f-b6f2-af094d594752" containerName="dnsmasq-dns" Nov 24 13:33:16 crc kubenswrapper[4790]: E1124 13:33:16.042428 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5398c0fb-71a5-4a3b-8750-0974556acd9b" containerName="ovn-config" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.042435 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="5398c0fb-71a5-4a3b-8750-0974556acd9b" containerName="ovn-config" Nov 24 13:33:16 crc kubenswrapper[4790]: E1124 13:33:16.042447 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bde9a3b-eafa-454f-b6f2-af094d594752" containerName="init" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.042456 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bde9a3b-eafa-454f-b6f2-af094d594752" containerName="init" Nov 24 13:33:16 crc kubenswrapper[4790]: E1124 13:33:16.042470 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="937f8268-ef33-440e-a297-d7c1b1cbefc0" containerName="mariadb-database-create" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.042477 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="937f8268-ef33-440e-a297-d7c1b1cbefc0" containerName="mariadb-database-create" Nov 24 13:33:16 crc kubenswrapper[4790]: E1124 13:33:16.042491 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="294bfffa-4d3e-43ff-b86c-26cc00b81b19" containerName="mariadb-database-create" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.042499 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="294bfffa-4d3e-43ff-b86c-26cc00b81b19" containerName="mariadb-database-create" Nov 24 13:33:16 crc kubenswrapper[4790]: E1124 13:33:16.042515 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c2d52e6-25f1-4ef0-a50e-6a8564c5302a" containerName="mariadb-account-create" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.042523 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c2d52e6-25f1-4ef0-a50e-6a8564c5302a" containerName="mariadb-account-create" Nov 24 13:33:16 crc kubenswrapper[4790]: E1124 13:33:16.042538 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f04f724-1dc9-4e6a-80ab-bd139438d1b8" containerName="keystone-db-sync" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.042545 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f04f724-1dc9-4e6a-80ab-bd139438d1b8" containerName="keystone-db-sync" Nov 24 13:33:16 crc kubenswrapper[4790]: E1124 13:33:16.042557 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad24288a-7c41-4e73-b1f3-c403b8646ff7" containerName="mariadb-account-create" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.042564 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad24288a-7c41-4e73-b1f3-c403b8646ff7" containerName="mariadb-account-create" Nov 24 13:33:16 crc kubenswrapper[4790]: E1124 13:33:16.042582 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f14beb2d-e165-43a0-a038-4cb2a3a86491" containerName="mariadb-database-create" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.042590 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="f14beb2d-e165-43a0-a038-4cb2a3a86491" containerName="mariadb-database-create" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.043823 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f04f724-1dc9-4e6a-80ab-bd139438d1b8" containerName="keystone-db-sync" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.043863 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="f14beb2d-e165-43a0-a038-4cb2a3a86491" containerName="mariadb-database-create" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.043895 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="294bfffa-4d3e-43ff-b86c-26cc00b81b19" containerName="mariadb-database-create" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.043908 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="88b90d01-f375-4158-9dd6-8042fdfbbd74" containerName="mariadb-account-create" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.043927 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="937f8268-ef33-440e-a297-d7c1b1cbefc0" containerName="mariadb-database-create" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.043939 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="5398c0fb-71a5-4a3b-8750-0974556acd9b" containerName="ovn-config" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.043953 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bde9a3b-eafa-454f-b6f2-af094d594752" containerName="dnsmasq-dns" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.043964 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c2d52e6-25f1-4ef0-a50e-6a8564c5302a" containerName="mariadb-account-create" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.043975 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad24288a-7c41-4e73-b1f3-c403b8646ff7" containerName="mariadb-account-create" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.045257 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-8nj6l" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.060638 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-8nj6l"] Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.100365 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-b77mf"] Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.107440 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-b77mf" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.111441 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.111665 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.111818 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xgjnb" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.112123 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.112451 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.127753 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-b77mf"] Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.146019 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-dns-svc\") pod \"dnsmasq-dns-5b868669f-8nj6l\" (UID: \"b99bce66-1dc0-40e9-8cbc-6e0eb09ba126\") " pod="openstack/dnsmasq-dns-5b868669f-8nj6l" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.146061 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-config\") pod \"dnsmasq-dns-5b868669f-8nj6l\" (UID: \"b99bce66-1dc0-40e9-8cbc-6e0eb09ba126\") " pod="openstack/dnsmasq-dns-5b868669f-8nj6l" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.146094 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7m96v\" (UniqueName: \"kubernetes.io/projected/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-kube-api-access-7m96v\") pod \"dnsmasq-dns-5b868669f-8nj6l\" (UID: \"b99bce66-1dc0-40e9-8cbc-6e0eb09ba126\") " pod="openstack/dnsmasq-dns-5b868669f-8nj6l" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.146222 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-ovsdbserver-nb\") pod \"dnsmasq-dns-5b868669f-8nj6l\" (UID: \"b99bce66-1dc0-40e9-8cbc-6e0eb09ba126\") " pod="openstack/dnsmasq-dns-5b868669f-8nj6l" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.146252 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-ovsdbserver-sb\") pod \"dnsmasq-dns-5b868669f-8nj6l\" (UID: \"b99bce66-1dc0-40e9-8cbc-6e0eb09ba126\") " pod="openstack/dnsmasq-dns-5b868669f-8nj6l" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.146370 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-dns-swift-storage-0\") pod \"dnsmasq-dns-5b868669f-8nj6l\" (UID: \"b99bce66-1dc0-40e9-8cbc-6e0eb09ba126\") " pod="openstack/dnsmasq-dns-5b868669f-8nj6l" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.243048 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-8dvn5"] Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.244218 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8dvn5" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.250070 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf19b969-f00b-40f9-9ef0-e7556903f76e-scripts\") pod \"keystone-bootstrap-b77mf\" (UID: \"bf19b969-f00b-40f9-9ef0-e7556903f76e\") " pod="openstack/keystone-bootstrap-b77mf" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.250137 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6jz4\" (UniqueName: \"kubernetes.io/projected/bf19b969-f00b-40f9-9ef0-e7556903f76e-kube-api-access-m6jz4\") pod \"keystone-bootstrap-b77mf\" (UID: \"bf19b969-f00b-40f9-9ef0-e7556903f76e\") " pod="openstack/keystone-bootstrap-b77mf" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.250169 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf19b969-f00b-40f9-9ef0-e7556903f76e-config-data\") pod \"keystone-bootstrap-b77mf\" (UID: \"bf19b969-f00b-40f9-9ef0-e7556903f76e\") " pod="openstack/keystone-bootstrap-b77mf" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.250183 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf19b969-f00b-40f9-9ef0-e7556903f76e-combined-ca-bundle\") pod \"keystone-bootstrap-b77mf\" (UID: \"bf19b969-f00b-40f9-9ef0-e7556903f76e\") " pod="openstack/keystone-bootstrap-b77mf" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.250219 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-dns-svc\") pod \"dnsmasq-dns-5b868669f-8nj6l\" (UID: \"b99bce66-1dc0-40e9-8cbc-6e0eb09ba126\") " pod="openstack/dnsmasq-dns-5b868669f-8nj6l" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.250237 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-config\") pod \"dnsmasq-dns-5b868669f-8nj6l\" (UID: \"b99bce66-1dc0-40e9-8cbc-6e0eb09ba126\") " pod="openstack/dnsmasq-dns-5b868669f-8nj6l" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.250255 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7m96v\" (UniqueName: \"kubernetes.io/projected/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-kube-api-access-7m96v\") pod \"dnsmasq-dns-5b868669f-8nj6l\" (UID: \"b99bce66-1dc0-40e9-8cbc-6e0eb09ba126\") " pod="openstack/dnsmasq-dns-5b868669f-8nj6l" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.250285 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-ovsdbserver-nb\") pod \"dnsmasq-dns-5b868669f-8nj6l\" (UID: \"b99bce66-1dc0-40e9-8cbc-6e0eb09ba126\") " pod="openstack/dnsmasq-dns-5b868669f-8nj6l" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.250303 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-ovsdbserver-sb\") pod \"dnsmasq-dns-5b868669f-8nj6l\" (UID: \"b99bce66-1dc0-40e9-8cbc-6e0eb09ba126\") " pod="openstack/dnsmasq-dns-5b868669f-8nj6l" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.250326 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/bf19b969-f00b-40f9-9ef0-e7556903f76e-credential-keys\") pod \"keystone-bootstrap-b77mf\" (UID: \"bf19b969-f00b-40f9-9ef0-e7556903f76e\") " pod="openstack/keystone-bootstrap-b77mf" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.250347 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bf19b969-f00b-40f9-9ef0-e7556903f76e-fernet-keys\") pod \"keystone-bootstrap-b77mf\" (UID: \"bf19b969-f00b-40f9-9ef0-e7556903f76e\") " pod="openstack/keystone-bootstrap-b77mf" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.250388 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-dns-swift-storage-0\") pod \"dnsmasq-dns-5b868669f-8nj6l\" (UID: \"b99bce66-1dc0-40e9-8cbc-6e0eb09ba126\") " pod="openstack/dnsmasq-dns-5b868669f-8nj6l" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.251378 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-dns-swift-storage-0\") pod \"dnsmasq-dns-5b868669f-8nj6l\" (UID: \"b99bce66-1dc0-40e9-8cbc-6e0eb09ba126\") " pod="openstack/dnsmasq-dns-5b868669f-8nj6l" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.251684 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-p5msf"] Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.252026 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-bsm2t" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.252154 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.252288 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.252747 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-config\") pod \"dnsmasq-dns-5b868669f-8nj6l\" (UID: \"b99bce66-1dc0-40e9-8cbc-6e0eb09ba126\") " pod="openstack/dnsmasq-dns-5b868669f-8nj6l" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.252930 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-ovsdbserver-sb\") pod \"dnsmasq-dns-5b868669f-8nj6l\" (UID: \"b99bce66-1dc0-40e9-8cbc-6e0eb09ba126\") " pod="openstack/dnsmasq-dns-5b868669f-8nj6l" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.253040 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-dns-svc\") pod \"dnsmasq-dns-5b868669f-8nj6l\" (UID: \"b99bce66-1dc0-40e9-8cbc-6e0eb09ba126\") " pod="openstack/dnsmasq-dns-5b868669f-8nj6l" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.253474 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-ovsdbserver-nb\") pod \"dnsmasq-dns-5b868669f-8nj6l\" (UID: \"b99bce66-1dc0-40e9-8cbc-6e0eb09ba126\") " pod="openstack/dnsmasq-dns-5b868669f-8nj6l" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.262204 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-8dvn5"] Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.262303 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-p5msf" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.271372 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.271394 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-pws5r" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.271586 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.296489 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-p5msf"] Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.299099 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7m96v\" (UniqueName: \"kubernetes.io/projected/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-kube-api-access-7m96v\") pod \"dnsmasq-dns-5b868669f-8nj6l\" (UID: \"b99bce66-1dc0-40e9-8cbc-6e0eb09ba126\") " pod="openstack/dnsmasq-dns-5b868669f-8nj6l" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.357155 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-combined-ca-bundle\") pod \"cinder-db-sync-8dvn5\" (UID: \"3e5246df-ca01-43e6-bcd9-c8d9b6b78279\") " pod="openstack/cinder-db-sync-8dvn5" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.357222 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-config-data\") pod \"cinder-db-sync-8dvn5\" (UID: \"3e5246df-ca01-43e6-bcd9-c8d9b6b78279\") " pod="openstack/cinder-db-sync-8dvn5" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.357275 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf19b969-f00b-40f9-9ef0-e7556903f76e-scripts\") pod \"keystone-bootstrap-b77mf\" (UID: \"bf19b969-f00b-40f9-9ef0-e7556903f76e\") " pod="openstack/keystone-bootstrap-b77mf" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.357322 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6jz4\" (UniqueName: \"kubernetes.io/projected/bf19b969-f00b-40f9-9ef0-e7556903f76e-kube-api-access-m6jz4\") pod \"keystone-bootstrap-b77mf\" (UID: \"bf19b969-f00b-40f9-9ef0-e7556903f76e\") " pod="openstack/keystone-bootstrap-b77mf" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.357358 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-422jt\" (UniqueName: \"kubernetes.io/projected/84a50ea0-c597-42e2-b4cd-7e628a517ddf-kube-api-access-422jt\") pod \"neutron-db-sync-p5msf\" (UID: \"84a50ea0-c597-42e2-b4cd-7e628a517ddf\") " pod="openstack/neutron-db-sync-p5msf" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.357411 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-scripts\") pod \"cinder-db-sync-8dvn5\" (UID: \"3e5246df-ca01-43e6-bcd9-c8d9b6b78279\") " pod="openstack/cinder-db-sync-8dvn5" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.357434 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/84a50ea0-c597-42e2-b4cd-7e628a517ddf-config\") pod \"neutron-db-sync-p5msf\" (UID: \"84a50ea0-c597-42e2-b4cd-7e628a517ddf\") " pod="openstack/neutron-db-sync-p5msf" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.357455 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf19b969-f00b-40f9-9ef0-e7556903f76e-config-data\") pod \"keystone-bootstrap-b77mf\" (UID: \"bf19b969-f00b-40f9-9ef0-e7556903f76e\") " pod="openstack/keystone-bootstrap-b77mf" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.357477 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf19b969-f00b-40f9-9ef0-e7556903f76e-combined-ca-bundle\") pod \"keystone-bootstrap-b77mf\" (UID: \"bf19b969-f00b-40f9-9ef0-e7556903f76e\") " pod="openstack/keystone-bootstrap-b77mf" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.357545 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2w4sf\" (UniqueName: \"kubernetes.io/projected/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-kube-api-access-2w4sf\") pod \"cinder-db-sync-8dvn5\" (UID: \"3e5246df-ca01-43e6-bcd9-c8d9b6b78279\") " pod="openstack/cinder-db-sync-8dvn5" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.357573 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/bf19b969-f00b-40f9-9ef0-e7556903f76e-credential-keys\") pod \"keystone-bootstrap-b77mf\" (UID: \"bf19b969-f00b-40f9-9ef0-e7556903f76e\") " pod="openstack/keystone-bootstrap-b77mf" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.357604 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-etc-machine-id\") pod \"cinder-db-sync-8dvn5\" (UID: \"3e5246df-ca01-43e6-bcd9-c8d9b6b78279\") " pod="openstack/cinder-db-sync-8dvn5" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.357629 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bf19b969-f00b-40f9-9ef0-e7556903f76e-fernet-keys\") pod \"keystone-bootstrap-b77mf\" (UID: \"bf19b969-f00b-40f9-9ef0-e7556903f76e\") " pod="openstack/keystone-bootstrap-b77mf" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.357655 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-db-sync-config-data\") pod \"cinder-db-sync-8dvn5\" (UID: \"3e5246df-ca01-43e6-bcd9-c8d9b6b78279\") " pod="openstack/cinder-db-sync-8dvn5" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.357696 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84a50ea0-c597-42e2-b4cd-7e628a517ddf-combined-ca-bundle\") pod \"neutron-db-sync-p5msf\" (UID: \"84a50ea0-c597-42e2-b4cd-7e628a517ddf\") " pod="openstack/neutron-db-sync-p5msf" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.381435 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/bf19b969-f00b-40f9-9ef0-e7556903f76e-credential-keys\") pod \"keystone-bootstrap-b77mf\" (UID: \"bf19b969-f00b-40f9-9ef0-e7556903f76e\") " pod="openstack/keystone-bootstrap-b77mf" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.381758 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf19b969-f00b-40f9-9ef0-e7556903f76e-scripts\") pod \"keystone-bootstrap-b77mf\" (UID: \"bf19b969-f00b-40f9-9ef0-e7556903f76e\") " pod="openstack/keystone-bootstrap-b77mf" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.391448 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf19b969-f00b-40f9-9ef0-e7556903f76e-combined-ca-bundle\") pod \"keystone-bootstrap-b77mf\" (UID: \"bf19b969-f00b-40f9-9ef0-e7556903f76e\") " pod="openstack/keystone-bootstrap-b77mf" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.410927 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.426020 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bf19b969-f00b-40f9-9ef0-e7556903f76e-fernet-keys\") pod \"keystone-bootstrap-b77mf\" (UID: \"bf19b969-f00b-40f9-9ef0-e7556903f76e\") " pod="openstack/keystone-bootstrap-b77mf" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.427906 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-8nj6l" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.437917 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.438055 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.441065 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6jz4\" (UniqueName: \"kubernetes.io/projected/bf19b969-f00b-40f9-9ef0-e7556903f76e-kube-api-access-m6jz4\") pod \"keystone-bootstrap-b77mf\" (UID: \"bf19b969-f00b-40f9-9ef0-e7556903f76e\") " pod="openstack/keystone-bootstrap-b77mf" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.449516 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.449768 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.451859 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf19b969-f00b-40f9-9ef0-e7556903f76e-config-data\") pod \"keystone-bootstrap-b77mf\" (UID: \"bf19b969-f00b-40f9-9ef0-e7556903f76e\") " pod="openstack/keystone-bootstrap-b77mf" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.461846 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2w4sf\" (UniqueName: \"kubernetes.io/projected/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-kube-api-access-2w4sf\") pod \"cinder-db-sync-8dvn5\" (UID: \"3e5246df-ca01-43e6-bcd9-c8d9b6b78279\") " pod="openstack/cinder-db-sync-8dvn5" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.461936 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-etc-machine-id\") pod \"cinder-db-sync-8dvn5\" (UID: \"3e5246df-ca01-43e6-bcd9-c8d9b6b78279\") " pod="openstack/cinder-db-sync-8dvn5" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.462048 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-etc-machine-id\") pod \"cinder-db-sync-8dvn5\" (UID: \"3e5246df-ca01-43e6-bcd9-c8d9b6b78279\") " pod="openstack/cinder-db-sync-8dvn5" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.462145 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-db-sync-config-data\") pod \"cinder-db-sync-8dvn5\" (UID: \"3e5246df-ca01-43e6-bcd9-c8d9b6b78279\") " pod="openstack/cinder-db-sync-8dvn5" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.462226 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84a50ea0-c597-42e2-b4cd-7e628a517ddf-combined-ca-bundle\") pod \"neutron-db-sync-p5msf\" (UID: \"84a50ea0-c597-42e2-b4cd-7e628a517ddf\") " pod="openstack/neutron-db-sync-p5msf" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.462327 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-combined-ca-bundle\") pod \"cinder-db-sync-8dvn5\" (UID: \"3e5246df-ca01-43e6-bcd9-c8d9b6b78279\") " pod="openstack/cinder-db-sync-8dvn5" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.462356 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-config-data\") pod \"cinder-db-sync-8dvn5\" (UID: \"3e5246df-ca01-43e6-bcd9-c8d9b6b78279\") " pod="openstack/cinder-db-sync-8dvn5" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.462466 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-422jt\" (UniqueName: \"kubernetes.io/projected/84a50ea0-c597-42e2-b4cd-7e628a517ddf-kube-api-access-422jt\") pod \"neutron-db-sync-p5msf\" (UID: \"84a50ea0-c597-42e2-b4cd-7e628a517ddf\") " pod="openstack/neutron-db-sync-p5msf" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.462507 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-scripts\") pod \"cinder-db-sync-8dvn5\" (UID: \"3e5246df-ca01-43e6-bcd9-c8d9b6b78279\") " pod="openstack/cinder-db-sync-8dvn5" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.462573 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/84a50ea0-c597-42e2-b4cd-7e628a517ddf-config\") pod \"neutron-db-sync-p5msf\" (UID: \"84a50ea0-c597-42e2-b4cd-7e628a517ddf\") " pod="openstack/neutron-db-sync-p5msf" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.471182 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/84a50ea0-c597-42e2-b4cd-7e628a517ddf-config\") pod \"neutron-db-sync-p5msf\" (UID: \"84a50ea0-c597-42e2-b4cd-7e628a517ddf\") " pod="openstack/neutron-db-sync-p5msf" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.471899 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-scripts\") pod \"cinder-db-sync-8dvn5\" (UID: \"3e5246df-ca01-43e6-bcd9-c8d9b6b78279\") " pod="openstack/cinder-db-sync-8dvn5" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.472182 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-config-data\") pod \"cinder-db-sync-8dvn5\" (UID: \"3e5246df-ca01-43e6-bcd9-c8d9b6b78279\") " pod="openstack/cinder-db-sync-8dvn5" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.476012 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-combined-ca-bundle\") pod \"cinder-db-sync-8dvn5\" (UID: \"3e5246df-ca01-43e6-bcd9-c8d9b6b78279\") " pod="openstack/cinder-db-sync-8dvn5" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.477098 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84a50ea0-c597-42e2-b4cd-7e628a517ddf-combined-ca-bundle\") pod \"neutron-db-sync-p5msf\" (UID: \"84a50ea0-c597-42e2-b4cd-7e628a517ddf\") " pod="openstack/neutron-db-sync-p5msf" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.479300 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-mbm86"] Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.482196 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-mbm86" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.488348 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-db-sync-config-data\") pod \"cinder-db-sync-8dvn5\" (UID: \"3e5246df-ca01-43e6-bcd9-c8d9b6b78279\") " pod="openstack/cinder-db-sync-8dvn5" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.492080 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.492669 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-d9ljp" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.532475 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-422jt\" (UniqueName: \"kubernetes.io/projected/84a50ea0-c597-42e2-b4cd-7e628a517ddf-kube-api-access-422jt\") pod \"neutron-db-sync-p5msf\" (UID: \"84a50ea0-c597-42e2-b4cd-7e628a517ddf\") " pod="openstack/neutron-db-sync-p5msf" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.542077 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2w4sf\" (UniqueName: \"kubernetes.io/projected/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-kube-api-access-2w4sf\") pod \"cinder-db-sync-8dvn5\" (UID: \"3e5246df-ca01-43e6-bcd9-c8d9b6b78279\") " pod="openstack/cinder-db-sync-8dvn5" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.543356 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-mbm86"] Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.559494 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-8nj6l"] Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.566761 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-q4sb2"] Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.567012 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca484cf0-4ab0-4581-bb65-b1701db63df3-combined-ca-bundle\") pod \"barbican-db-sync-mbm86\" (UID: \"ca484cf0-4ab0-4581-bb65-b1701db63df3\") " pod="openstack/barbican-db-sync-mbm86" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.567095 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\") " pod="openstack/ceilometer-0" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.567145 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-config-data\") pod \"ceilometer-0\" (UID: \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\") " pod="openstack/ceilometer-0" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.567187 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ca484cf0-4ab0-4581-bb65-b1701db63df3-db-sync-config-data\") pod \"barbican-db-sync-mbm86\" (UID: \"ca484cf0-4ab0-4581-bb65-b1701db63df3\") " pod="openstack/barbican-db-sync-mbm86" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.567216 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\") " pod="openstack/ceilometer-0" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.567264 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-scripts\") pod \"ceilometer-0\" (UID: \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\") " pod="openstack/ceilometer-0" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.567300 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8m6zh\" (UniqueName: \"kubernetes.io/projected/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-kube-api-access-8m6zh\") pod \"ceilometer-0\" (UID: \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\") " pod="openstack/ceilometer-0" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.567322 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-log-httpd\") pod \"ceilometer-0\" (UID: \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\") " pod="openstack/ceilometer-0" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.567349 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-run-httpd\") pod \"ceilometer-0\" (UID: \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\") " pod="openstack/ceilometer-0" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.567394 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fcc7\" (UniqueName: \"kubernetes.io/projected/ca484cf0-4ab0-4581-bb65-b1701db63df3-kube-api-access-2fcc7\") pod \"barbican-db-sync-mbm86\" (UID: \"ca484cf0-4ab0-4581-bb65-b1701db63df3\") " pod="openstack/barbican-db-sync-mbm86" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.568220 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-q4sb2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.569478 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8dvn5" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.576471 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-q4sb2"] Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.576728 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.577016 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-ljf7m" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.577198 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.578169 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-p5msf" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.600869 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-qqzt2"] Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.605961 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-qqzt2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.618113 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-qqzt2"] Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.673062 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca484cf0-4ab0-4581-bb65-b1701db63df3-combined-ca-bundle\") pod \"barbican-db-sync-mbm86\" (UID: \"ca484cf0-4ab0-4581-bb65-b1701db63df3\") " pod="openstack/barbican-db-sync-mbm86" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.673110 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-qqzt2\" (UID: \"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2\") " pod="openstack/dnsmasq-dns-cf78879c9-qqzt2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.673140 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-config\") pod \"dnsmasq-dns-cf78879c9-qqzt2\" (UID: \"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2\") " pod="openstack/dnsmasq-dns-cf78879c9-qqzt2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.673174 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\") " pod="openstack/ceilometer-0" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.673198 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-qqzt2\" (UID: \"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2\") " pod="openstack/dnsmasq-dns-cf78879c9-qqzt2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.673225 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-config-data\") pod \"ceilometer-0\" (UID: \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\") " pod="openstack/ceilometer-0" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.673242 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f5d721b-e52b-490d-89f3-4a5edaf403ec-scripts\") pod \"placement-db-sync-q4sb2\" (UID: \"6f5d721b-e52b-490d-89f3-4a5edaf403ec\") " pod="openstack/placement-db-sync-q4sb2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.673268 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ca484cf0-4ab0-4581-bb65-b1701db63df3-db-sync-config-data\") pod \"barbican-db-sync-mbm86\" (UID: \"ca484cf0-4ab0-4581-bb65-b1701db63df3\") " pod="openstack/barbican-db-sync-mbm86" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.673287 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\") " pod="openstack/ceilometer-0" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.673315 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f5d721b-e52b-490d-89f3-4a5edaf403ec-combined-ca-bundle\") pod \"placement-db-sync-q4sb2\" (UID: \"6f5d721b-e52b-490d-89f3-4a5edaf403ec\") " pod="openstack/placement-db-sync-q4sb2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.673338 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-qqzt2\" (UID: \"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2\") " pod="openstack/dnsmasq-dns-cf78879c9-qqzt2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.673365 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-scripts\") pod \"ceilometer-0\" (UID: \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\") " pod="openstack/ceilometer-0" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.673385 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfqw2\" (UniqueName: \"kubernetes.io/projected/6f5d721b-e52b-490d-89f3-4a5edaf403ec-kube-api-access-kfqw2\") pod \"placement-db-sync-q4sb2\" (UID: \"6f5d721b-e52b-490d-89f3-4a5edaf403ec\") " pod="openstack/placement-db-sync-q4sb2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.673405 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8m6zh\" (UniqueName: \"kubernetes.io/projected/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-kube-api-access-8m6zh\") pod \"ceilometer-0\" (UID: \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\") " pod="openstack/ceilometer-0" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.673422 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-log-httpd\") pod \"ceilometer-0\" (UID: \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\") " pod="openstack/ceilometer-0" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.673439 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-dns-svc\") pod \"dnsmasq-dns-cf78879c9-qqzt2\" (UID: \"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2\") " pod="openstack/dnsmasq-dns-cf78879c9-qqzt2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.673454 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f5d721b-e52b-490d-89f3-4a5edaf403ec-config-data\") pod \"placement-db-sync-q4sb2\" (UID: \"6f5d721b-e52b-490d-89f3-4a5edaf403ec\") " pod="openstack/placement-db-sync-q4sb2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.673473 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-run-httpd\") pod \"ceilometer-0\" (UID: \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\") " pod="openstack/ceilometer-0" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.673498 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f5d721b-e52b-490d-89f3-4a5edaf403ec-logs\") pod \"placement-db-sync-q4sb2\" (UID: \"6f5d721b-e52b-490d-89f3-4a5edaf403ec\") " pod="openstack/placement-db-sync-q4sb2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.673523 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pw598\" (UniqueName: \"kubernetes.io/projected/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-kube-api-access-pw598\") pod \"dnsmasq-dns-cf78879c9-qqzt2\" (UID: \"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2\") " pod="openstack/dnsmasq-dns-cf78879c9-qqzt2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.673551 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fcc7\" (UniqueName: \"kubernetes.io/projected/ca484cf0-4ab0-4581-bb65-b1701db63df3-kube-api-access-2fcc7\") pod \"barbican-db-sync-mbm86\" (UID: \"ca484cf0-4ab0-4581-bb65-b1701db63df3\") " pod="openstack/barbican-db-sync-mbm86" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.678024 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-run-httpd\") pod \"ceilometer-0\" (UID: \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\") " pod="openstack/ceilometer-0" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.678445 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-log-httpd\") pod \"ceilometer-0\" (UID: \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\") " pod="openstack/ceilometer-0" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.682357 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca484cf0-4ab0-4581-bb65-b1701db63df3-combined-ca-bundle\") pod \"barbican-db-sync-mbm86\" (UID: \"ca484cf0-4ab0-4581-bb65-b1701db63df3\") " pod="openstack/barbican-db-sync-mbm86" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.682692 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-scripts\") pod \"ceilometer-0\" (UID: \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\") " pod="openstack/ceilometer-0" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.687993 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\") " pod="openstack/ceilometer-0" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.689080 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ca484cf0-4ab0-4581-bb65-b1701db63df3-db-sync-config-data\") pod \"barbican-db-sync-mbm86\" (UID: \"ca484cf0-4ab0-4581-bb65-b1701db63df3\") " pod="openstack/barbican-db-sync-mbm86" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.689817 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\") " pod="openstack/ceilometer-0" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.690961 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-config-data\") pod \"ceilometer-0\" (UID: \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\") " pod="openstack/ceilometer-0" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.695920 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fcc7\" (UniqueName: \"kubernetes.io/projected/ca484cf0-4ab0-4581-bb65-b1701db63df3-kube-api-access-2fcc7\") pod \"barbican-db-sync-mbm86\" (UID: \"ca484cf0-4ab0-4581-bb65-b1701db63df3\") " pod="openstack/barbican-db-sync-mbm86" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.700867 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8m6zh\" (UniqueName: \"kubernetes.io/projected/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-kube-api-access-8m6zh\") pod \"ceilometer-0\" (UID: \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\") " pod="openstack/ceilometer-0" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.743996 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-b77mf" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.775097 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-qqzt2\" (UID: \"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2\") " pod="openstack/dnsmasq-dns-cf78879c9-qqzt2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.775442 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-config\") pod \"dnsmasq-dns-cf78879c9-qqzt2\" (UID: \"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2\") " pod="openstack/dnsmasq-dns-cf78879c9-qqzt2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.775494 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-qqzt2\" (UID: \"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2\") " pod="openstack/dnsmasq-dns-cf78879c9-qqzt2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.775541 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f5d721b-e52b-490d-89f3-4a5edaf403ec-scripts\") pod \"placement-db-sync-q4sb2\" (UID: \"6f5d721b-e52b-490d-89f3-4a5edaf403ec\") " pod="openstack/placement-db-sync-q4sb2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.775604 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f5d721b-e52b-490d-89f3-4a5edaf403ec-combined-ca-bundle\") pod \"placement-db-sync-q4sb2\" (UID: \"6f5d721b-e52b-490d-89f3-4a5edaf403ec\") " pod="openstack/placement-db-sync-q4sb2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.775645 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-qqzt2\" (UID: \"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2\") " pod="openstack/dnsmasq-dns-cf78879c9-qqzt2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.775681 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfqw2\" (UniqueName: \"kubernetes.io/projected/6f5d721b-e52b-490d-89f3-4a5edaf403ec-kube-api-access-kfqw2\") pod \"placement-db-sync-q4sb2\" (UID: \"6f5d721b-e52b-490d-89f3-4a5edaf403ec\") " pod="openstack/placement-db-sync-q4sb2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.775717 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-dns-svc\") pod \"dnsmasq-dns-cf78879c9-qqzt2\" (UID: \"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2\") " pod="openstack/dnsmasq-dns-cf78879c9-qqzt2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.775757 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f5d721b-e52b-490d-89f3-4a5edaf403ec-config-data\") pod \"placement-db-sync-q4sb2\" (UID: \"6f5d721b-e52b-490d-89f3-4a5edaf403ec\") " pod="openstack/placement-db-sync-q4sb2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.775806 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f5d721b-e52b-490d-89f3-4a5edaf403ec-logs\") pod \"placement-db-sync-q4sb2\" (UID: \"6f5d721b-e52b-490d-89f3-4a5edaf403ec\") " pod="openstack/placement-db-sync-q4sb2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.775833 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pw598\" (UniqueName: \"kubernetes.io/projected/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-kube-api-access-pw598\") pod \"dnsmasq-dns-cf78879c9-qqzt2\" (UID: \"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2\") " pod="openstack/dnsmasq-dns-cf78879c9-qqzt2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.777118 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-config\") pod \"dnsmasq-dns-cf78879c9-qqzt2\" (UID: \"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2\") " pod="openstack/dnsmasq-dns-cf78879c9-qqzt2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.778023 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-qqzt2\" (UID: \"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2\") " pod="openstack/dnsmasq-dns-cf78879c9-qqzt2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.778760 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-dns-svc\") pod \"dnsmasq-dns-cf78879c9-qqzt2\" (UID: \"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2\") " pod="openstack/dnsmasq-dns-cf78879c9-qqzt2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.782181 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f5d721b-e52b-490d-89f3-4a5edaf403ec-logs\") pod \"placement-db-sync-q4sb2\" (UID: \"6f5d721b-e52b-490d-89f3-4a5edaf403ec\") " pod="openstack/placement-db-sync-q4sb2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.782258 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-qqzt2\" (UID: \"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2\") " pod="openstack/dnsmasq-dns-cf78879c9-qqzt2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.782936 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-qqzt2\" (UID: \"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2\") " pod="openstack/dnsmasq-dns-cf78879c9-qqzt2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.783117 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f5d721b-e52b-490d-89f3-4a5edaf403ec-combined-ca-bundle\") pod \"placement-db-sync-q4sb2\" (UID: \"6f5d721b-e52b-490d-89f3-4a5edaf403ec\") " pod="openstack/placement-db-sync-q4sb2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.788311 4790 generic.go:334] "Generic (PLEG): container finished" podID="c49c82dc-b231-4f2f-8553-323fcb24f67f" containerID="2c6c3b4bcd309e9bbfe6304002cfac512cde3ab86acfec0e937b2af1a62d7ef2" exitCode=0 Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.788367 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-chcfj" event={"ID":"c49c82dc-b231-4f2f-8553-323fcb24f67f","Type":"ContainerDied","Data":"2c6c3b4bcd309e9bbfe6304002cfac512cde3ab86acfec0e937b2af1a62d7ef2"} Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.788706 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f5d721b-e52b-490d-89f3-4a5edaf403ec-config-data\") pod \"placement-db-sync-q4sb2\" (UID: \"6f5d721b-e52b-490d-89f3-4a5edaf403ec\") " pod="openstack/placement-db-sync-q4sb2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.792221 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f5d721b-e52b-490d-89f3-4a5edaf403ec-scripts\") pod \"placement-db-sync-q4sb2\" (UID: \"6f5d721b-e52b-490d-89f3-4a5edaf403ec\") " pod="openstack/placement-db-sync-q4sb2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.795365 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfqw2\" (UniqueName: \"kubernetes.io/projected/6f5d721b-e52b-490d-89f3-4a5edaf403ec-kube-api-access-kfqw2\") pod \"placement-db-sync-q4sb2\" (UID: \"6f5d721b-e52b-490d-89f3-4a5edaf403ec\") " pod="openstack/placement-db-sync-q4sb2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.803341 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pw598\" (UniqueName: \"kubernetes.io/projected/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-kube-api-access-pw598\") pod \"dnsmasq-dns-cf78879c9-qqzt2\" (UID: \"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2\") " pod="openstack/dnsmasq-dns-cf78879c9-qqzt2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.844380 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.859515 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-mbm86" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.912283 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-q4sb2" Nov 24 13:33:16 crc kubenswrapper[4790]: I1124 13:33:16.935047 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-qqzt2" Nov 24 13:33:17 crc kubenswrapper[4790]: I1124 13:33:17.075948 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-8nj6l"] Nov 24 13:33:17 crc kubenswrapper[4790]: W1124 13:33:17.128678 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb99bce66_1dc0_40e9_8cbc_6e0eb09ba126.slice/crio-4a879fd8ea87706b1e3738a0a410572c37e895cfacfa79ce509d411a13abc1e0 WatchSource:0}: Error finding container 4a879fd8ea87706b1e3738a0a410572c37e895cfacfa79ce509d411a13abc1e0: Status 404 returned error can't find the container with id 4a879fd8ea87706b1e3738a0a410572c37e895cfacfa79ce509d411a13abc1e0 Nov 24 13:33:17 crc kubenswrapper[4790]: I1124 13:33:17.203253 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-p5msf"] Nov 24 13:33:17 crc kubenswrapper[4790]: W1124 13:33:17.205417 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e5246df_ca01_43e6_bcd9_c8d9b6b78279.slice/crio-32a5e77128d6ef141be8ff5534229731c523fc643794e1407b8166941a35e49a WatchSource:0}: Error finding container 32a5e77128d6ef141be8ff5534229731c523fc643794e1407b8166941a35e49a: Status 404 returned error can't find the container with id 32a5e77128d6ef141be8ff5534229731c523fc643794e1407b8166941a35e49a Nov 24 13:33:17 crc kubenswrapper[4790]: I1124 13:33:17.230267 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-8dvn5"] Nov 24 13:33:17 crc kubenswrapper[4790]: I1124 13:33:17.419390 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-b77mf"] Nov 24 13:33:17 crc kubenswrapper[4790]: I1124 13:33:17.568367 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-q4sb2"] Nov 24 13:33:17 crc kubenswrapper[4790]: W1124 13:33:17.582072 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f5d721b_e52b_490d_89f3_4a5edaf403ec.slice/crio-0fdb94317be2e50fbd9a8f04489ec9a294110f094caaadb913457c5f4a6db87a WatchSource:0}: Error finding container 0fdb94317be2e50fbd9a8f04489ec9a294110f094caaadb913457c5f4a6db87a: Status 404 returned error can't find the container with id 0fdb94317be2e50fbd9a8f04489ec9a294110f094caaadb913457c5f4a6db87a Nov 24 13:33:17 crc kubenswrapper[4790]: W1124 13:33:17.591830 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca484cf0_4ab0_4581_bb65_b1701db63df3.slice/crio-d8d66e587d7088f54f5678772108443371f6ea405b639c1292f06c359440c6fe WatchSource:0}: Error finding container d8d66e587d7088f54f5678772108443371f6ea405b639c1292f06c359440c6fe: Status 404 returned error can't find the container with id d8d66e587d7088f54f5678772108443371f6ea405b639c1292f06c359440c6fe Nov 24 13:33:17 crc kubenswrapper[4790]: I1124 13:33:17.594452 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:33:17 crc kubenswrapper[4790]: I1124 13:33:17.601750 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-qqzt2"] Nov 24 13:33:17 crc kubenswrapper[4790]: I1124 13:33:17.608839 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-mbm86"] Nov 24 13:33:17 crc kubenswrapper[4790]: I1124 13:33:17.800198 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f","Type":"ContainerStarted","Data":"35ba9ca9fd843161f03f5d502d310845016dd199433abf5abbe6cc64039bcf25"} Nov 24 13:33:17 crc kubenswrapper[4790]: I1124 13:33:17.801713 4790 generic.go:334] "Generic (PLEG): container finished" podID="b99bce66-1dc0-40e9-8cbc-6e0eb09ba126" containerID="5e7e9a3fa40e4670cc44060febcd9c248befc8d1064223ad0a80d64eca91e260" exitCode=0 Nov 24 13:33:17 crc kubenswrapper[4790]: I1124 13:33:17.801918 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b868669f-8nj6l" event={"ID":"b99bce66-1dc0-40e9-8cbc-6e0eb09ba126","Type":"ContainerDied","Data":"5e7e9a3fa40e4670cc44060febcd9c248befc8d1064223ad0a80d64eca91e260"} Nov 24 13:33:17 crc kubenswrapper[4790]: I1124 13:33:17.801957 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b868669f-8nj6l" event={"ID":"b99bce66-1dc0-40e9-8cbc-6e0eb09ba126","Type":"ContainerStarted","Data":"4a879fd8ea87706b1e3738a0a410572c37e895cfacfa79ce509d411a13abc1e0"} Nov 24 13:33:17 crc kubenswrapper[4790]: I1124 13:33:17.811472 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-p5msf" event={"ID":"84a50ea0-c597-42e2-b4cd-7e628a517ddf","Type":"ContainerStarted","Data":"cf5790b66698273019b8f16b970dbdbffc943209fdca11bb7061f46cb7418370"} Nov 24 13:33:17 crc kubenswrapper[4790]: I1124 13:33:17.811524 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-p5msf" event={"ID":"84a50ea0-c597-42e2-b4cd-7e628a517ddf","Type":"ContainerStarted","Data":"374c89fcbac95affb0b9cf6110fb31cf2a5aceeebec36e8d8a100419b679018e"} Nov 24 13:33:17 crc kubenswrapper[4790]: I1124 13:33:17.818541 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-b77mf" event={"ID":"bf19b969-f00b-40f9-9ef0-e7556903f76e","Type":"ContainerStarted","Data":"1cdea1e01c6022d8752115c1dbd9b9cd87f839392ad6920de4f0d664706cf7e2"} Nov 24 13:33:17 crc kubenswrapper[4790]: I1124 13:33:17.818593 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-b77mf" event={"ID":"bf19b969-f00b-40f9-9ef0-e7556903f76e","Type":"ContainerStarted","Data":"07293e6ee1c4372b9655c1aac99208516ba7d229d59bbd0ae7b4942acef7d7dc"} Nov 24 13:33:17 crc kubenswrapper[4790]: I1124 13:33:17.823641 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8dvn5" event={"ID":"3e5246df-ca01-43e6-bcd9-c8d9b6b78279","Type":"ContainerStarted","Data":"32a5e77128d6ef141be8ff5534229731c523fc643794e1407b8166941a35e49a"} Nov 24 13:33:17 crc kubenswrapper[4790]: I1124 13:33:17.829869 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-qqzt2" event={"ID":"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2","Type":"ContainerStarted","Data":"5b52e40c19049d488eb9bca51afc3064c16e94264443da3acadd6e63b4102e3b"} Nov 24 13:33:17 crc kubenswrapper[4790]: I1124 13:33:17.829949 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-qqzt2" event={"ID":"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2","Type":"ContainerStarted","Data":"0156680f4e851ebeb52d1e61fac1dfcba1ac5b5223e107b8ee6c2a0fc65a7353"} Nov 24 13:33:17 crc kubenswrapper[4790]: I1124 13:33:17.841622 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-mbm86" event={"ID":"ca484cf0-4ab0-4581-bb65-b1701db63df3","Type":"ContainerStarted","Data":"d8d66e587d7088f54f5678772108443371f6ea405b639c1292f06c359440c6fe"} Nov 24 13:33:17 crc kubenswrapper[4790]: I1124 13:33:17.843576 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-q4sb2" event={"ID":"6f5d721b-e52b-490d-89f3-4a5edaf403ec","Type":"ContainerStarted","Data":"0fdb94317be2e50fbd9a8f04489ec9a294110f094caaadb913457c5f4a6db87a"} Nov 24 13:33:17 crc kubenswrapper[4790]: I1124 13:33:17.866282 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-b77mf" podStartSLOduration=1.866248461 podStartE2EDuration="1.866248461s" podCreationTimestamp="2025-11-24 13:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:33:17.854001932 +0000 UTC m=+1246.233895614" watchObservedRunningTime="2025-11-24 13:33:17.866248461 +0000 UTC m=+1246.246142123" Nov 24 13:33:17 crc kubenswrapper[4790]: I1124 13:33:17.881419 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-p5msf" podStartSLOduration=1.881313089 podStartE2EDuration="1.881313089s" podCreationTimestamp="2025-11-24 13:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:33:17.876245288 +0000 UTC m=+1246.256138950" watchObservedRunningTime="2025-11-24 13:33:17.881313089 +0000 UTC m=+1246.261206751" Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.224195 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-8nj6l" Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.326248 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-ovsdbserver-sb\") pod \"b99bce66-1dc0-40e9-8cbc-6e0eb09ba126\" (UID: \"b99bce66-1dc0-40e9-8cbc-6e0eb09ba126\") " Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.326561 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-dns-swift-storage-0\") pod \"b99bce66-1dc0-40e9-8cbc-6e0eb09ba126\" (UID: \"b99bce66-1dc0-40e9-8cbc-6e0eb09ba126\") " Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.326647 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7m96v\" (UniqueName: \"kubernetes.io/projected/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-kube-api-access-7m96v\") pod \"b99bce66-1dc0-40e9-8cbc-6e0eb09ba126\" (UID: \"b99bce66-1dc0-40e9-8cbc-6e0eb09ba126\") " Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.326739 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-config\") pod \"b99bce66-1dc0-40e9-8cbc-6e0eb09ba126\" (UID: \"b99bce66-1dc0-40e9-8cbc-6e0eb09ba126\") " Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.326807 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-dns-svc\") pod \"b99bce66-1dc0-40e9-8cbc-6e0eb09ba126\" (UID: \"b99bce66-1dc0-40e9-8cbc-6e0eb09ba126\") " Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.326898 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-ovsdbserver-nb\") pod \"b99bce66-1dc0-40e9-8cbc-6e0eb09ba126\" (UID: \"b99bce66-1dc0-40e9-8cbc-6e0eb09ba126\") " Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.340412 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-kube-api-access-7m96v" (OuterVolumeSpecName: "kube-api-access-7m96v") pod "b99bce66-1dc0-40e9-8cbc-6e0eb09ba126" (UID: "b99bce66-1dc0-40e9-8cbc-6e0eb09ba126"). InnerVolumeSpecName "kube-api-access-7m96v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.377758 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b99bce66-1dc0-40e9-8cbc-6e0eb09ba126" (UID: "b99bce66-1dc0-40e9-8cbc-6e0eb09ba126"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.396486 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b99bce66-1dc0-40e9-8cbc-6e0eb09ba126" (UID: "b99bce66-1dc0-40e9-8cbc-6e0eb09ba126"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.396568 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-config" (OuterVolumeSpecName: "config") pod "b99bce66-1dc0-40e9-8cbc-6e0eb09ba126" (UID: "b99bce66-1dc0-40e9-8cbc-6e0eb09ba126"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.397284 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b99bce66-1dc0-40e9-8cbc-6e0eb09ba126" (UID: "b99bce66-1dc0-40e9-8cbc-6e0eb09ba126"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.402159 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b99bce66-1dc0-40e9-8cbc-6e0eb09ba126" (UID: "b99bce66-1dc0-40e9-8cbc-6e0eb09ba126"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.441299 4790 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.441330 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7m96v\" (UniqueName: \"kubernetes.io/projected/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-kube-api-access-7m96v\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.441342 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.441350 4790 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.441358 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.441365 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.465995 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-chcfj" Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.551464 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crwkk\" (UniqueName: \"kubernetes.io/projected/c49c82dc-b231-4f2f-8553-323fcb24f67f-kube-api-access-crwkk\") pod \"c49c82dc-b231-4f2f-8553-323fcb24f67f\" (UID: \"c49c82dc-b231-4f2f-8553-323fcb24f67f\") " Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.551553 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c49c82dc-b231-4f2f-8553-323fcb24f67f-db-sync-config-data\") pod \"c49c82dc-b231-4f2f-8553-323fcb24f67f\" (UID: \"c49c82dc-b231-4f2f-8553-323fcb24f67f\") " Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.551771 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c49c82dc-b231-4f2f-8553-323fcb24f67f-combined-ca-bundle\") pod \"c49c82dc-b231-4f2f-8553-323fcb24f67f\" (UID: \"c49c82dc-b231-4f2f-8553-323fcb24f67f\") " Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.551801 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c49c82dc-b231-4f2f-8553-323fcb24f67f-config-data\") pod \"c49c82dc-b231-4f2f-8553-323fcb24f67f\" (UID: \"c49c82dc-b231-4f2f-8553-323fcb24f67f\") " Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.555022 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c49c82dc-b231-4f2f-8553-323fcb24f67f-kube-api-access-crwkk" (OuterVolumeSpecName: "kube-api-access-crwkk") pod "c49c82dc-b231-4f2f-8553-323fcb24f67f" (UID: "c49c82dc-b231-4f2f-8553-323fcb24f67f"). InnerVolumeSpecName "kube-api-access-crwkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.559018 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c49c82dc-b231-4f2f-8553-323fcb24f67f-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "c49c82dc-b231-4f2f-8553-323fcb24f67f" (UID: "c49c82dc-b231-4f2f-8553-323fcb24f67f"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.602338 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c49c82dc-b231-4f2f-8553-323fcb24f67f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c49c82dc-b231-4f2f-8553-323fcb24f67f" (UID: "c49c82dc-b231-4f2f-8553-323fcb24f67f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.638192 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c49c82dc-b231-4f2f-8553-323fcb24f67f-config-data" (OuterVolumeSpecName: "config-data") pod "c49c82dc-b231-4f2f-8553-323fcb24f67f" (UID: "c49c82dc-b231-4f2f-8553-323fcb24f67f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.656292 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crwkk\" (UniqueName: \"kubernetes.io/projected/c49c82dc-b231-4f2f-8553-323fcb24f67f-kube-api-access-crwkk\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.656333 4790 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c49c82dc-b231-4f2f-8553-323fcb24f67f-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.656347 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c49c82dc-b231-4f2f-8553-323fcb24f67f-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.656359 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c49c82dc-b231-4f2f-8553-323fcb24f67f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.897486 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b868669f-8nj6l" event={"ID":"b99bce66-1dc0-40e9-8cbc-6e0eb09ba126","Type":"ContainerDied","Data":"4a879fd8ea87706b1e3738a0a410572c37e895cfacfa79ce509d411a13abc1e0"} Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.897542 4790 scope.go:117] "RemoveContainer" containerID="5e7e9a3fa40e4670cc44060febcd9c248befc8d1064223ad0a80d64eca91e260" Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.897697 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-8nj6l" Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.910570 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.912413 4790 generic.go:334] "Generic (PLEG): container finished" podID="7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2" containerID="5b52e40c19049d488eb9bca51afc3064c16e94264443da3acadd6e63b4102e3b" exitCode=0 Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.913252 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-qqzt2" event={"ID":"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2","Type":"ContainerDied","Data":"5b52e40c19049d488eb9bca51afc3064c16e94264443da3acadd6e63b4102e3b"} Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.930704 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-chcfj" event={"ID":"c49c82dc-b231-4f2f-8553-323fcb24f67f","Type":"ContainerDied","Data":"8e9e5ce942fbf16d6dba6cf2aeab2b51467613456380497f0bb6e48142077c47"} Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.930788 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e9e5ce942fbf16d6dba6cf2aeab2b51467613456380497f0bb6e48142077c47" Nov 24 13:33:18 crc kubenswrapper[4790]: I1124 13:33:18.930902 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-chcfj" Nov 24 13:33:19 crc kubenswrapper[4790]: I1124 13:33:19.014945 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-8nj6l"] Nov 24 13:33:19 crc kubenswrapper[4790]: I1124 13:33:19.021559 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-8nj6l"] Nov 24 13:33:19 crc kubenswrapper[4790]: I1124 13:33:19.200006 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-qqzt2"] Nov 24 13:33:19 crc kubenswrapper[4790]: I1124 13:33:19.215492 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-c2jtn"] Nov 24 13:33:19 crc kubenswrapper[4790]: E1124 13:33:19.215859 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c49c82dc-b231-4f2f-8553-323fcb24f67f" containerName="glance-db-sync" Nov 24 13:33:19 crc kubenswrapper[4790]: I1124 13:33:19.215894 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="c49c82dc-b231-4f2f-8553-323fcb24f67f" containerName="glance-db-sync" Nov 24 13:33:19 crc kubenswrapper[4790]: E1124 13:33:19.215917 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b99bce66-1dc0-40e9-8cbc-6e0eb09ba126" containerName="init" Nov 24 13:33:19 crc kubenswrapper[4790]: I1124 13:33:19.215923 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="b99bce66-1dc0-40e9-8cbc-6e0eb09ba126" containerName="init" Nov 24 13:33:19 crc kubenswrapper[4790]: I1124 13:33:19.216093 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="b99bce66-1dc0-40e9-8cbc-6e0eb09ba126" containerName="init" Nov 24 13:33:19 crc kubenswrapper[4790]: I1124 13:33:19.216108 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="c49c82dc-b231-4f2f-8553-323fcb24f67f" containerName="glance-db-sync" Nov 24 13:33:19 crc kubenswrapper[4790]: I1124 13:33:19.216985 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-c2jtn" Nov 24 13:33:19 crc kubenswrapper[4790]: I1124 13:33:19.238182 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-c2jtn"] Nov 24 13:33:19 crc kubenswrapper[4790]: I1124 13:33:19.280310 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-config\") pod \"dnsmasq-dns-56df8fb6b7-c2jtn\" (UID: \"871c7d3b-ecde-472b-a0ff-61c1a7a24f30\") " pod="openstack/dnsmasq-dns-56df8fb6b7-c2jtn" Nov 24 13:33:19 crc kubenswrapper[4790]: I1124 13:33:19.280358 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-c2jtn\" (UID: \"871c7d3b-ecde-472b-a0ff-61c1a7a24f30\") " pod="openstack/dnsmasq-dns-56df8fb6b7-c2jtn" Nov 24 13:33:19 crc kubenswrapper[4790]: I1124 13:33:19.280383 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-c2jtn\" (UID: \"871c7d3b-ecde-472b-a0ff-61c1a7a24f30\") " pod="openstack/dnsmasq-dns-56df8fb6b7-c2jtn" Nov 24 13:33:19 crc kubenswrapper[4790]: I1124 13:33:19.280434 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-c2jtn\" (UID: \"871c7d3b-ecde-472b-a0ff-61c1a7a24f30\") " pod="openstack/dnsmasq-dns-56df8fb6b7-c2jtn" Nov 24 13:33:19 crc kubenswrapper[4790]: I1124 13:33:19.280480 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-c2jtn\" (UID: \"871c7d3b-ecde-472b-a0ff-61c1a7a24f30\") " pod="openstack/dnsmasq-dns-56df8fb6b7-c2jtn" Nov 24 13:33:19 crc kubenswrapper[4790]: I1124 13:33:19.280528 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5c6p7\" (UniqueName: \"kubernetes.io/projected/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-kube-api-access-5c6p7\") pod \"dnsmasq-dns-56df8fb6b7-c2jtn\" (UID: \"871c7d3b-ecde-472b-a0ff-61c1a7a24f30\") " pod="openstack/dnsmasq-dns-56df8fb6b7-c2jtn" Nov 24 13:33:19 crc kubenswrapper[4790]: I1124 13:33:19.383241 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-config\") pod \"dnsmasq-dns-56df8fb6b7-c2jtn\" (UID: \"871c7d3b-ecde-472b-a0ff-61c1a7a24f30\") " pod="openstack/dnsmasq-dns-56df8fb6b7-c2jtn" Nov 24 13:33:19 crc kubenswrapper[4790]: I1124 13:33:19.383325 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-c2jtn\" (UID: \"871c7d3b-ecde-472b-a0ff-61c1a7a24f30\") " pod="openstack/dnsmasq-dns-56df8fb6b7-c2jtn" Nov 24 13:33:19 crc kubenswrapper[4790]: I1124 13:33:19.383390 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-c2jtn\" (UID: \"871c7d3b-ecde-472b-a0ff-61c1a7a24f30\") " pod="openstack/dnsmasq-dns-56df8fb6b7-c2jtn" Nov 24 13:33:19 crc kubenswrapper[4790]: I1124 13:33:19.383440 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-c2jtn\" (UID: \"871c7d3b-ecde-472b-a0ff-61c1a7a24f30\") " pod="openstack/dnsmasq-dns-56df8fb6b7-c2jtn" Nov 24 13:33:19 crc kubenswrapper[4790]: I1124 13:33:19.383479 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-c2jtn\" (UID: \"871c7d3b-ecde-472b-a0ff-61c1a7a24f30\") " pod="openstack/dnsmasq-dns-56df8fb6b7-c2jtn" Nov 24 13:33:19 crc kubenswrapper[4790]: I1124 13:33:19.383542 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5c6p7\" (UniqueName: \"kubernetes.io/projected/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-kube-api-access-5c6p7\") pod \"dnsmasq-dns-56df8fb6b7-c2jtn\" (UID: \"871c7d3b-ecde-472b-a0ff-61c1a7a24f30\") " pod="openstack/dnsmasq-dns-56df8fb6b7-c2jtn" Nov 24 13:33:19 crc kubenswrapper[4790]: I1124 13:33:19.386285 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-config\") pod \"dnsmasq-dns-56df8fb6b7-c2jtn\" (UID: \"871c7d3b-ecde-472b-a0ff-61c1a7a24f30\") " pod="openstack/dnsmasq-dns-56df8fb6b7-c2jtn" Nov 24 13:33:19 crc kubenswrapper[4790]: I1124 13:33:19.387504 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-c2jtn\" (UID: \"871c7d3b-ecde-472b-a0ff-61c1a7a24f30\") " pod="openstack/dnsmasq-dns-56df8fb6b7-c2jtn" Nov 24 13:33:19 crc kubenswrapper[4790]: I1124 13:33:19.389090 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-c2jtn\" (UID: \"871c7d3b-ecde-472b-a0ff-61c1a7a24f30\") " pod="openstack/dnsmasq-dns-56df8fb6b7-c2jtn" Nov 24 13:33:19 crc kubenswrapper[4790]: I1124 13:33:19.390671 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-c2jtn\" (UID: \"871c7d3b-ecde-472b-a0ff-61c1a7a24f30\") " pod="openstack/dnsmasq-dns-56df8fb6b7-c2jtn" Nov 24 13:33:19 crc kubenswrapper[4790]: I1124 13:33:19.393185 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-c2jtn\" (UID: \"871c7d3b-ecde-472b-a0ff-61c1a7a24f30\") " pod="openstack/dnsmasq-dns-56df8fb6b7-c2jtn" Nov 24 13:33:19 crc kubenswrapper[4790]: I1124 13:33:19.422751 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5c6p7\" (UniqueName: \"kubernetes.io/projected/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-kube-api-access-5c6p7\") pod \"dnsmasq-dns-56df8fb6b7-c2jtn\" (UID: \"871c7d3b-ecde-472b-a0ff-61c1a7a24f30\") " pod="openstack/dnsmasq-dns-56df8fb6b7-c2jtn" Nov 24 13:33:19 crc kubenswrapper[4790]: I1124 13:33:19.553036 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-c2jtn" Nov 24 13:33:19 crc kubenswrapper[4790]: I1124 13:33:19.940773 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-qqzt2" event={"ID":"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2","Type":"ContainerStarted","Data":"66f5ea547ee89aef5100d0c34e763ebff615070dd32d43842f8ab4c00325722c"} Nov 24 13:33:19 crc kubenswrapper[4790]: I1124 13:33:19.941180 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cf78879c9-qqzt2" podUID="7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2" containerName="dnsmasq-dns" containerID="cri-o://66f5ea547ee89aef5100d0c34e763ebff615070dd32d43842f8ab4c00325722c" gracePeriod=10 Nov 24 13:33:19 crc kubenswrapper[4790]: I1124 13:33:19.941440 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cf78879c9-qqzt2" Nov 24 13:33:19 crc kubenswrapper[4790]: I1124 13:33:19.960906 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cf78879c9-qqzt2" podStartSLOduration=3.960890246 podStartE2EDuration="3.960890246s" podCreationTimestamp="2025-11-24 13:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:33:19.956178205 +0000 UTC m=+1248.336071877" watchObservedRunningTime="2025-11-24 13:33:19.960890246 +0000 UTC m=+1248.340783908" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.107951 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.113804 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.116048 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-nrdsr" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.116404 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.120301 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.123405 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.173994 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-c2jtn"] Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.206833 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af74b160-8348-44b1-ad80-92bb0a1c34ad-config-data\") pod \"glance-default-external-api-0\" (UID: \"af74b160-8348-44b1-ad80-92bb0a1c34ad\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.206971 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af74b160-8348-44b1-ad80-92bb0a1c34ad-logs\") pod \"glance-default-external-api-0\" (UID: \"af74b160-8348-44b1-ad80-92bb0a1c34ad\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.207078 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"af74b160-8348-44b1-ad80-92bb0a1c34ad\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.207197 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af74b160-8348-44b1-ad80-92bb0a1c34ad-scripts\") pod \"glance-default-external-api-0\" (UID: \"af74b160-8348-44b1-ad80-92bb0a1c34ad\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.207253 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af74b160-8348-44b1-ad80-92bb0a1c34ad-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"af74b160-8348-44b1-ad80-92bb0a1c34ad\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.207331 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/af74b160-8348-44b1-ad80-92bb0a1c34ad-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"af74b160-8348-44b1-ad80-92bb0a1c34ad\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.207475 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnj6g\" (UniqueName: \"kubernetes.io/projected/af74b160-8348-44b1-ad80-92bb0a1c34ad-kube-api-access-pnj6g\") pod \"glance-default-external-api-0\" (UID: \"af74b160-8348-44b1-ad80-92bb0a1c34ad\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.308793 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"af74b160-8348-44b1-ad80-92bb0a1c34ad\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.308927 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af74b160-8348-44b1-ad80-92bb0a1c34ad-scripts\") pod \"glance-default-external-api-0\" (UID: \"af74b160-8348-44b1-ad80-92bb0a1c34ad\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.308978 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af74b160-8348-44b1-ad80-92bb0a1c34ad-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"af74b160-8348-44b1-ad80-92bb0a1c34ad\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.309069 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/af74b160-8348-44b1-ad80-92bb0a1c34ad-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"af74b160-8348-44b1-ad80-92bb0a1c34ad\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.309092 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnj6g\" (UniqueName: \"kubernetes.io/projected/af74b160-8348-44b1-ad80-92bb0a1c34ad-kube-api-access-pnj6g\") pod \"glance-default-external-api-0\" (UID: \"af74b160-8348-44b1-ad80-92bb0a1c34ad\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.309115 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af74b160-8348-44b1-ad80-92bb0a1c34ad-config-data\") pod \"glance-default-external-api-0\" (UID: \"af74b160-8348-44b1-ad80-92bb0a1c34ad\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.309144 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af74b160-8348-44b1-ad80-92bb0a1c34ad-logs\") pod \"glance-default-external-api-0\" (UID: \"af74b160-8348-44b1-ad80-92bb0a1c34ad\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.311811 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af74b160-8348-44b1-ad80-92bb0a1c34ad-logs\") pod \"glance-default-external-api-0\" (UID: \"af74b160-8348-44b1-ad80-92bb0a1c34ad\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.312198 4790 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"af74b160-8348-44b1-ad80-92bb0a1c34ad\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.319255 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/af74b160-8348-44b1-ad80-92bb0a1c34ad-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"af74b160-8348-44b1-ad80-92bb0a1c34ad\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.323077 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af74b160-8348-44b1-ad80-92bb0a1c34ad-scripts\") pod \"glance-default-external-api-0\" (UID: \"af74b160-8348-44b1-ad80-92bb0a1c34ad\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.324756 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af74b160-8348-44b1-ad80-92bb0a1c34ad-config-data\") pod \"glance-default-external-api-0\" (UID: \"af74b160-8348-44b1-ad80-92bb0a1c34ad\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.332156 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af74b160-8348-44b1-ad80-92bb0a1c34ad-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"af74b160-8348-44b1-ad80-92bb0a1c34ad\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.347957 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnj6g\" (UniqueName: \"kubernetes.io/projected/af74b160-8348-44b1-ad80-92bb0a1c34ad-kube-api-access-pnj6g\") pod \"glance-default-external-api-0\" (UID: \"af74b160-8348-44b1-ad80-92bb0a1c34ad\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.349407 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b99bce66-1dc0-40e9-8cbc-6e0eb09ba126" path="/var/lib/kubelet/pods/b99bce66-1dc0-40e9-8cbc-6e0eb09ba126/volumes" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.363065 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"af74b160-8348-44b1-ad80-92bb0a1c34ad\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.363256 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.364872 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.376072 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.392211 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.416090 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be3f173b-f907-4f35-bb20-3afe1e874583-config-data\") pod \"glance-default-internal-api-0\" (UID: \"be3f173b-f907-4f35-bb20-3afe1e874583\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.416548 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-248q5\" (UniqueName: \"kubernetes.io/projected/be3f173b-f907-4f35-bb20-3afe1e874583-kube-api-access-248q5\") pod \"glance-default-internal-api-0\" (UID: \"be3f173b-f907-4f35-bb20-3afe1e874583\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.416576 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"be3f173b-f907-4f35-bb20-3afe1e874583\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.416609 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be3f173b-f907-4f35-bb20-3afe1e874583-logs\") pod \"glance-default-internal-api-0\" (UID: \"be3f173b-f907-4f35-bb20-3afe1e874583\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.416845 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be3f173b-f907-4f35-bb20-3afe1e874583-scripts\") pod \"glance-default-internal-api-0\" (UID: \"be3f173b-f907-4f35-bb20-3afe1e874583\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.417215 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be3f173b-f907-4f35-bb20-3afe1e874583-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"be3f173b-f907-4f35-bb20-3afe1e874583\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.417524 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be3f173b-f907-4f35-bb20-3afe1e874583-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"be3f173b-f907-4f35-bb20-3afe1e874583\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.448105 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.522097 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-248q5\" (UniqueName: \"kubernetes.io/projected/be3f173b-f907-4f35-bb20-3afe1e874583-kube-api-access-248q5\") pod \"glance-default-internal-api-0\" (UID: \"be3f173b-f907-4f35-bb20-3afe1e874583\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.522334 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"be3f173b-f907-4f35-bb20-3afe1e874583\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.522369 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be3f173b-f907-4f35-bb20-3afe1e874583-logs\") pod \"glance-default-internal-api-0\" (UID: \"be3f173b-f907-4f35-bb20-3afe1e874583\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.522437 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be3f173b-f907-4f35-bb20-3afe1e874583-scripts\") pod \"glance-default-internal-api-0\" (UID: \"be3f173b-f907-4f35-bb20-3afe1e874583\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.522477 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be3f173b-f907-4f35-bb20-3afe1e874583-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"be3f173b-f907-4f35-bb20-3afe1e874583\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.522497 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be3f173b-f907-4f35-bb20-3afe1e874583-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"be3f173b-f907-4f35-bb20-3afe1e874583\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.522530 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be3f173b-f907-4f35-bb20-3afe1e874583-config-data\") pod \"glance-default-internal-api-0\" (UID: \"be3f173b-f907-4f35-bb20-3afe1e874583\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.522630 4790 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"be3f173b-f907-4f35-bb20-3afe1e874583\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.523634 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be3f173b-f907-4f35-bb20-3afe1e874583-logs\") pod \"glance-default-internal-api-0\" (UID: \"be3f173b-f907-4f35-bb20-3afe1e874583\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.524538 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be3f173b-f907-4f35-bb20-3afe1e874583-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"be3f173b-f907-4f35-bb20-3afe1e874583\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.530133 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be3f173b-f907-4f35-bb20-3afe1e874583-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"be3f173b-f907-4f35-bb20-3afe1e874583\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.534598 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be3f173b-f907-4f35-bb20-3afe1e874583-scripts\") pod \"glance-default-internal-api-0\" (UID: \"be3f173b-f907-4f35-bb20-3afe1e874583\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.535484 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be3f173b-f907-4f35-bb20-3afe1e874583-config-data\") pod \"glance-default-internal-api-0\" (UID: \"be3f173b-f907-4f35-bb20-3afe1e874583\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.544857 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-248q5\" (UniqueName: \"kubernetes.io/projected/be3f173b-f907-4f35-bb20-3afe1e874583-kube-api-access-248q5\") pod \"glance-default-internal-api-0\" (UID: \"be3f173b-f907-4f35-bb20-3afe1e874583\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.569425 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"be3f173b-f907-4f35-bb20-3afe1e874583\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.725932 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.963017 4790 generic.go:334] "Generic (PLEG): container finished" podID="7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2" containerID="66f5ea547ee89aef5100d0c34e763ebff615070dd32d43842f8ab4c00325722c" exitCode=0 Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.963271 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-qqzt2" event={"ID":"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2","Type":"ContainerDied","Data":"66f5ea547ee89aef5100d0c34e763ebff615070dd32d43842f8ab4c00325722c"} Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.963297 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-qqzt2" event={"ID":"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2","Type":"ContainerDied","Data":"0156680f4e851ebeb52d1e61fac1dfcba1ac5b5223e107b8ee6c2a0fc65a7353"} Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.963307 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0156680f4e851ebeb52d1e61fac1dfcba1ac5b5223e107b8ee6c2a0fc65a7353" Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.965666 4790 generic.go:334] "Generic (PLEG): container finished" podID="871c7d3b-ecde-472b-a0ff-61c1a7a24f30" containerID="d59fc437b66b47fcc10e49fb481cf5d86df8e0a79f0c8fbf5ac234655c8d9a5d" exitCode=0 Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.965709 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-c2jtn" event={"ID":"871c7d3b-ecde-472b-a0ff-61c1a7a24f30","Type":"ContainerDied","Data":"d59fc437b66b47fcc10e49fb481cf5d86df8e0a79f0c8fbf5ac234655c8d9a5d"} Nov 24 13:33:20 crc kubenswrapper[4790]: I1124 13:33:20.965735 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-c2jtn" event={"ID":"871c7d3b-ecde-472b-a0ff-61c1a7a24f30","Type":"ContainerStarted","Data":"43c992ed439be035efc7997e34b933d881c57f94b9e3162d65353c55b85c90cb"} Nov 24 13:33:21 crc kubenswrapper[4790]: I1124 13:33:21.137708 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 13:33:21 crc kubenswrapper[4790]: W1124 13:33:21.150363 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf74b160_8348_44b1_ad80_92bb0a1c34ad.slice/crio-eedf60aefb2065fa246c8ccb6d94be23de7d8cce902f9ec10140fc36e0f8223b WatchSource:0}: Error finding container eedf60aefb2065fa246c8ccb6d94be23de7d8cce902f9ec10140fc36e0f8223b: Status 404 returned error can't find the container with id eedf60aefb2065fa246c8ccb6d94be23de7d8cce902f9ec10140fc36e0f8223b Nov 24 13:33:21 crc kubenswrapper[4790]: I1124 13:33:21.228588 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-qqzt2" Nov 24 13:33:21 crc kubenswrapper[4790]: I1124 13:33:21.268164 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:33:21 crc kubenswrapper[4790]: I1124 13:33:21.336484 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-dns-svc\") pod \"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2\" (UID: \"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2\") " Nov 24 13:33:21 crc kubenswrapper[4790]: I1124 13:33:21.336604 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pw598\" (UniqueName: \"kubernetes.io/projected/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-kube-api-access-pw598\") pod \"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2\" (UID: \"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2\") " Nov 24 13:33:21 crc kubenswrapper[4790]: I1124 13:33:21.336695 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-dns-swift-storage-0\") pod \"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2\" (UID: \"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2\") " Nov 24 13:33:21 crc kubenswrapper[4790]: I1124 13:33:21.336821 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-ovsdbserver-nb\") pod \"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2\" (UID: \"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2\") " Nov 24 13:33:21 crc kubenswrapper[4790]: I1124 13:33:21.336846 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-config\") pod \"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2\" (UID: \"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2\") " Nov 24 13:33:21 crc kubenswrapper[4790]: I1124 13:33:21.336915 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-ovsdbserver-sb\") pod \"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2\" (UID: \"7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2\") " Nov 24 13:33:21 crc kubenswrapper[4790]: W1124 13:33:21.338820 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbe3f173b_f907_4f35_bb20_3afe1e874583.slice/crio-2d85f3b00fe43e02223548e90a3a92d081c60c9e702d046f75c41e7bc9eda4ae WatchSource:0}: Error finding container 2d85f3b00fe43e02223548e90a3a92d081c60c9e702d046f75c41e7bc9eda4ae: Status 404 returned error can't find the container with id 2d85f3b00fe43e02223548e90a3a92d081c60c9e702d046f75c41e7bc9eda4ae Nov 24 13:33:21 crc kubenswrapper[4790]: I1124 13:33:21.347154 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-kube-api-access-pw598" (OuterVolumeSpecName: "kube-api-access-pw598") pod "7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2" (UID: "7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2"). InnerVolumeSpecName "kube-api-access-pw598". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:21 crc kubenswrapper[4790]: I1124 13:33:21.391030 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2" (UID: "7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:21 crc kubenswrapper[4790]: I1124 13:33:21.394250 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2" (UID: "7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:21 crc kubenswrapper[4790]: I1124 13:33:21.397871 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2" (UID: "7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:21 crc kubenswrapper[4790]: I1124 13:33:21.398020 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-config" (OuterVolumeSpecName: "config") pod "7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2" (UID: "7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:21 crc kubenswrapper[4790]: I1124 13:33:21.418579 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2" (UID: "7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:21 crc kubenswrapper[4790]: I1124 13:33:21.439154 4790 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:21 crc kubenswrapper[4790]: I1124 13:33:21.439204 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:21 crc kubenswrapper[4790]: I1124 13:33:21.439214 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:21 crc kubenswrapper[4790]: I1124 13:33:21.439224 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:21 crc kubenswrapper[4790]: I1124 13:33:21.439233 4790 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:21 crc kubenswrapper[4790]: I1124 13:33:21.439241 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pw598\" (UniqueName: \"kubernetes.io/projected/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2-kube-api-access-pw598\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:21 crc kubenswrapper[4790]: I1124 13:33:21.980306 4790 generic.go:334] "Generic (PLEG): container finished" podID="bf19b969-f00b-40f9-9ef0-e7556903f76e" containerID="1cdea1e01c6022d8752115c1dbd9b9cd87f839392ad6920de4f0d664706cf7e2" exitCode=0 Nov 24 13:33:21 crc kubenswrapper[4790]: I1124 13:33:21.980337 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-b77mf" event={"ID":"bf19b969-f00b-40f9-9ef0-e7556903f76e","Type":"ContainerDied","Data":"1cdea1e01c6022d8752115c1dbd9b9cd87f839392ad6920de4f0d664706cf7e2"} Nov 24 13:33:21 crc kubenswrapper[4790]: I1124 13:33:21.983143 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"be3f173b-f907-4f35-bb20-3afe1e874583","Type":"ContainerStarted","Data":"2d85f3b00fe43e02223548e90a3a92d081c60c9e702d046f75c41e7bc9eda4ae"} Nov 24 13:33:21 crc kubenswrapper[4790]: I1124 13:33:21.990397 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-c2jtn" event={"ID":"871c7d3b-ecde-472b-a0ff-61c1a7a24f30","Type":"ContainerStarted","Data":"14a2345853543961bb0465b42e0dd8b869d2c3096eaf9c9311f17fad8ea912de"} Nov 24 13:33:21 crc kubenswrapper[4790]: I1124 13:33:21.991762 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56df8fb6b7-c2jtn" Nov 24 13:33:21 crc kubenswrapper[4790]: I1124 13:33:21.994781 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-qqzt2" Nov 24 13:33:21 crc kubenswrapper[4790]: I1124 13:33:21.995091 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"af74b160-8348-44b1-ad80-92bb0a1c34ad","Type":"ContainerStarted","Data":"b7f86aeef5273169f1a971a910b999e251459aa52cccb8d5b05853980caf75da"} Nov 24 13:33:21 crc kubenswrapper[4790]: I1124 13:33:21.995138 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"af74b160-8348-44b1-ad80-92bb0a1c34ad","Type":"ContainerStarted","Data":"eedf60aefb2065fa246c8ccb6d94be23de7d8cce902f9ec10140fc36e0f8223b"} Nov 24 13:33:22 crc kubenswrapper[4790]: I1124 13:33:22.037775 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56df8fb6b7-c2jtn" podStartSLOduration=3.037757758 podStartE2EDuration="3.037757758s" podCreationTimestamp="2025-11-24 13:33:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:33:22.016653893 +0000 UTC m=+1250.396547555" watchObservedRunningTime="2025-11-24 13:33:22.037757758 +0000 UTC m=+1250.417651410" Nov 24 13:33:22 crc kubenswrapper[4790]: I1124 13:33:22.068174 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-qqzt2"] Nov 24 13:33:22 crc kubenswrapper[4790]: I1124 13:33:22.076064 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-qqzt2"] Nov 24 13:33:22 crc kubenswrapper[4790]: I1124 13:33:22.331112 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2" path="/var/lib/kubelet/pods/7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2/volumes" Nov 24 13:33:23 crc kubenswrapper[4790]: I1124 13:33:23.008480 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"be3f173b-f907-4f35-bb20-3afe1e874583","Type":"ContainerStarted","Data":"e4ce8052121e4bc044cba22d485867c5d86471221f0babfc7af4072083c45a26"} Nov 24 13:33:26 crc kubenswrapper[4790]: I1124 13:33:26.367665 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 13:33:26 crc kubenswrapper[4790]: I1124 13:33:26.437058 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:33:27 crc kubenswrapper[4790]: I1124 13:33:27.235508 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-b77mf" Nov 24 13:33:27 crc kubenswrapper[4790]: I1124 13:33:27.360367 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf19b969-f00b-40f9-9ef0-e7556903f76e-config-data\") pod \"bf19b969-f00b-40f9-9ef0-e7556903f76e\" (UID: \"bf19b969-f00b-40f9-9ef0-e7556903f76e\") " Nov 24 13:33:27 crc kubenswrapper[4790]: I1124 13:33:27.361036 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bf19b969-f00b-40f9-9ef0-e7556903f76e-fernet-keys\") pod \"bf19b969-f00b-40f9-9ef0-e7556903f76e\" (UID: \"bf19b969-f00b-40f9-9ef0-e7556903f76e\") " Nov 24 13:33:27 crc kubenswrapper[4790]: I1124 13:33:27.361118 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf19b969-f00b-40f9-9ef0-e7556903f76e-scripts\") pod \"bf19b969-f00b-40f9-9ef0-e7556903f76e\" (UID: \"bf19b969-f00b-40f9-9ef0-e7556903f76e\") " Nov 24 13:33:27 crc kubenswrapper[4790]: I1124 13:33:27.361426 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/bf19b969-f00b-40f9-9ef0-e7556903f76e-credential-keys\") pod \"bf19b969-f00b-40f9-9ef0-e7556903f76e\" (UID: \"bf19b969-f00b-40f9-9ef0-e7556903f76e\") " Nov 24 13:33:27 crc kubenswrapper[4790]: I1124 13:33:27.361495 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6jz4\" (UniqueName: \"kubernetes.io/projected/bf19b969-f00b-40f9-9ef0-e7556903f76e-kube-api-access-m6jz4\") pod \"bf19b969-f00b-40f9-9ef0-e7556903f76e\" (UID: \"bf19b969-f00b-40f9-9ef0-e7556903f76e\") " Nov 24 13:33:27 crc kubenswrapper[4790]: I1124 13:33:27.361526 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf19b969-f00b-40f9-9ef0-e7556903f76e-combined-ca-bundle\") pod \"bf19b969-f00b-40f9-9ef0-e7556903f76e\" (UID: \"bf19b969-f00b-40f9-9ef0-e7556903f76e\") " Nov 24 13:33:27 crc kubenswrapper[4790]: I1124 13:33:27.366111 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf19b969-f00b-40f9-9ef0-e7556903f76e-scripts" (OuterVolumeSpecName: "scripts") pod "bf19b969-f00b-40f9-9ef0-e7556903f76e" (UID: "bf19b969-f00b-40f9-9ef0-e7556903f76e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:27 crc kubenswrapper[4790]: I1124 13:33:27.366698 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf19b969-f00b-40f9-9ef0-e7556903f76e-kube-api-access-m6jz4" (OuterVolumeSpecName: "kube-api-access-m6jz4") pod "bf19b969-f00b-40f9-9ef0-e7556903f76e" (UID: "bf19b969-f00b-40f9-9ef0-e7556903f76e"). InnerVolumeSpecName "kube-api-access-m6jz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:27 crc kubenswrapper[4790]: I1124 13:33:27.370321 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf19b969-f00b-40f9-9ef0-e7556903f76e-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "bf19b969-f00b-40f9-9ef0-e7556903f76e" (UID: "bf19b969-f00b-40f9-9ef0-e7556903f76e"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:27 crc kubenswrapper[4790]: I1124 13:33:27.370758 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf19b969-f00b-40f9-9ef0-e7556903f76e-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "bf19b969-f00b-40f9-9ef0-e7556903f76e" (UID: "bf19b969-f00b-40f9-9ef0-e7556903f76e"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:27 crc kubenswrapper[4790]: I1124 13:33:27.391293 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf19b969-f00b-40f9-9ef0-e7556903f76e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bf19b969-f00b-40f9-9ef0-e7556903f76e" (UID: "bf19b969-f00b-40f9-9ef0-e7556903f76e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:27 crc kubenswrapper[4790]: I1124 13:33:27.391432 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf19b969-f00b-40f9-9ef0-e7556903f76e-config-data" (OuterVolumeSpecName: "config-data") pod "bf19b969-f00b-40f9-9ef0-e7556903f76e" (UID: "bf19b969-f00b-40f9-9ef0-e7556903f76e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:27 crc kubenswrapper[4790]: I1124 13:33:27.463396 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf19b969-f00b-40f9-9ef0-e7556903f76e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:27 crc kubenswrapper[4790]: I1124 13:33:27.463430 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf19b969-f00b-40f9-9ef0-e7556903f76e-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:27 crc kubenswrapper[4790]: I1124 13:33:27.463440 4790 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bf19b969-f00b-40f9-9ef0-e7556903f76e-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:27 crc kubenswrapper[4790]: I1124 13:33:27.463451 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf19b969-f00b-40f9-9ef0-e7556903f76e-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:27 crc kubenswrapper[4790]: I1124 13:33:27.463462 4790 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/bf19b969-f00b-40f9-9ef0-e7556903f76e-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:27 crc kubenswrapper[4790]: I1124 13:33:27.463473 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6jz4\" (UniqueName: \"kubernetes.io/projected/bf19b969-f00b-40f9-9ef0-e7556903f76e-kube-api-access-m6jz4\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.063478 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-b77mf" event={"ID":"bf19b969-f00b-40f9-9ef0-e7556903f76e","Type":"ContainerDied","Data":"07293e6ee1c4372b9655c1aac99208516ba7d229d59bbd0ae7b4942acef7d7dc"} Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.063801 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07293e6ee1c4372b9655c1aac99208516ba7d229d59bbd0ae7b4942acef7d7dc" Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.063518 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-b77mf" Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.342736 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-b77mf"] Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.350118 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-b77mf"] Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.443330 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-zqb2l"] Nov 24 13:33:28 crc kubenswrapper[4790]: E1124 13:33:28.443760 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf19b969-f00b-40f9-9ef0-e7556903f76e" containerName="keystone-bootstrap" Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.443798 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf19b969-f00b-40f9-9ef0-e7556903f76e" containerName="keystone-bootstrap" Nov 24 13:33:28 crc kubenswrapper[4790]: E1124 13:33:28.443822 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2" containerName="dnsmasq-dns" Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.443829 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2" containerName="dnsmasq-dns" Nov 24 13:33:28 crc kubenswrapper[4790]: E1124 13:33:28.443852 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2" containerName="init" Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.443864 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2" containerName="init" Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.444095 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="7af1ea8a-a063-419e-9b6f-f3dc4ac4aff2" containerName="dnsmasq-dns" Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.444130 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf19b969-f00b-40f9-9ef0-e7556903f76e" containerName="keystone-bootstrap" Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.444906 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-zqb2l" Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.446999 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.447338 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.448123 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xgjnb" Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.448501 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.448750 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.474856 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-zqb2l"] Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.583742 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrx9v\" (UniqueName: \"kubernetes.io/projected/482e2d7c-282b-490c-b476-955ddc498707-kube-api-access-jrx9v\") pod \"keystone-bootstrap-zqb2l\" (UID: \"482e2d7c-282b-490c-b476-955ddc498707\") " pod="openstack/keystone-bootstrap-zqb2l" Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.583796 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/482e2d7c-282b-490c-b476-955ddc498707-combined-ca-bundle\") pod \"keystone-bootstrap-zqb2l\" (UID: \"482e2d7c-282b-490c-b476-955ddc498707\") " pod="openstack/keystone-bootstrap-zqb2l" Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.583831 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/482e2d7c-282b-490c-b476-955ddc498707-credential-keys\") pod \"keystone-bootstrap-zqb2l\" (UID: \"482e2d7c-282b-490c-b476-955ddc498707\") " pod="openstack/keystone-bootstrap-zqb2l" Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.583911 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/482e2d7c-282b-490c-b476-955ddc498707-config-data\") pod \"keystone-bootstrap-zqb2l\" (UID: \"482e2d7c-282b-490c-b476-955ddc498707\") " pod="openstack/keystone-bootstrap-zqb2l" Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.584015 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/482e2d7c-282b-490c-b476-955ddc498707-scripts\") pod \"keystone-bootstrap-zqb2l\" (UID: \"482e2d7c-282b-490c-b476-955ddc498707\") " pod="openstack/keystone-bootstrap-zqb2l" Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.584035 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/482e2d7c-282b-490c-b476-955ddc498707-fernet-keys\") pod \"keystone-bootstrap-zqb2l\" (UID: \"482e2d7c-282b-490c-b476-955ddc498707\") " pod="openstack/keystone-bootstrap-zqb2l" Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.685845 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/482e2d7c-282b-490c-b476-955ddc498707-scripts\") pod \"keystone-bootstrap-zqb2l\" (UID: \"482e2d7c-282b-490c-b476-955ddc498707\") " pod="openstack/keystone-bootstrap-zqb2l" Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.685914 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/482e2d7c-282b-490c-b476-955ddc498707-fernet-keys\") pod \"keystone-bootstrap-zqb2l\" (UID: \"482e2d7c-282b-490c-b476-955ddc498707\") " pod="openstack/keystone-bootstrap-zqb2l" Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.685999 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrx9v\" (UniqueName: \"kubernetes.io/projected/482e2d7c-282b-490c-b476-955ddc498707-kube-api-access-jrx9v\") pod \"keystone-bootstrap-zqb2l\" (UID: \"482e2d7c-282b-490c-b476-955ddc498707\") " pod="openstack/keystone-bootstrap-zqb2l" Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.686028 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/482e2d7c-282b-490c-b476-955ddc498707-combined-ca-bundle\") pod \"keystone-bootstrap-zqb2l\" (UID: \"482e2d7c-282b-490c-b476-955ddc498707\") " pod="openstack/keystone-bootstrap-zqb2l" Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.686051 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/482e2d7c-282b-490c-b476-955ddc498707-credential-keys\") pod \"keystone-bootstrap-zqb2l\" (UID: \"482e2d7c-282b-490c-b476-955ddc498707\") " pod="openstack/keystone-bootstrap-zqb2l" Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.686088 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/482e2d7c-282b-490c-b476-955ddc498707-config-data\") pod \"keystone-bootstrap-zqb2l\" (UID: \"482e2d7c-282b-490c-b476-955ddc498707\") " pod="openstack/keystone-bootstrap-zqb2l" Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.775999 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/482e2d7c-282b-490c-b476-955ddc498707-credential-keys\") pod \"keystone-bootstrap-zqb2l\" (UID: \"482e2d7c-282b-490c-b476-955ddc498707\") " pod="openstack/keystone-bootstrap-zqb2l" Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.776095 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrx9v\" (UniqueName: \"kubernetes.io/projected/482e2d7c-282b-490c-b476-955ddc498707-kube-api-access-jrx9v\") pod \"keystone-bootstrap-zqb2l\" (UID: \"482e2d7c-282b-490c-b476-955ddc498707\") " pod="openstack/keystone-bootstrap-zqb2l" Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.777411 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/482e2d7c-282b-490c-b476-955ddc498707-fernet-keys\") pod \"keystone-bootstrap-zqb2l\" (UID: \"482e2d7c-282b-490c-b476-955ddc498707\") " pod="openstack/keystone-bootstrap-zqb2l" Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.777501 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/482e2d7c-282b-490c-b476-955ddc498707-scripts\") pod \"keystone-bootstrap-zqb2l\" (UID: \"482e2d7c-282b-490c-b476-955ddc498707\") " pod="openstack/keystone-bootstrap-zqb2l" Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.777773 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/482e2d7c-282b-490c-b476-955ddc498707-config-data\") pod \"keystone-bootstrap-zqb2l\" (UID: \"482e2d7c-282b-490c-b476-955ddc498707\") " pod="openstack/keystone-bootstrap-zqb2l" Nov 24 13:33:28 crc kubenswrapper[4790]: I1124 13:33:28.777798 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/482e2d7c-282b-490c-b476-955ddc498707-combined-ca-bundle\") pod \"keystone-bootstrap-zqb2l\" (UID: \"482e2d7c-282b-490c-b476-955ddc498707\") " pod="openstack/keystone-bootstrap-zqb2l" Nov 24 13:33:29 crc kubenswrapper[4790]: I1124 13:33:29.067689 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-zqb2l" Nov 24 13:33:29 crc kubenswrapper[4790]: I1124 13:33:29.555733 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56df8fb6b7-c2jtn" Nov 24 13:33:29 crc kubenswrapper[4790]: I1124 13:33:29.634185 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-q9jn5"] Nov 24 13:33:29 crc kubenswrapper[4790]: I1124 13:33:29.634528 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c79d794d7-q9jn5" podUID="d291f904-9c55-4213-9cde-69d6f71b67df" containerName="dnsmasq-dns" containerID="cri-o://083bb5dfa444f0e05367f4a3e328194b3c06970f40e5a951abc781f5875f0852" gracePeriod=10 Nov 24 13:33:30 crc kubenswrapper[4790]: I1124 13:33:30.080535 4790 generic.go:334] "Generic (PLEG): container finished" podID="d291f904-9c55-4213-9cde-69d6f71b67df" containerID="083bb5dfa444f0e05367f4a3e328194b3c06970f40e5a951abc781f5875f0852" exitCode=0 Nov 24 13:33:30 crc kubenswrapper[4790]: I1124 13:33:30.080577 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-q9jn5" event={"ID":"d291f904-9c55-4213-9cde-69d6f71b67df","Type":"ContainerDied","Data":"083bb5dfa444f0e05367f4a3e328194b3c06970f40e5a951abc781f5875f0852"} Nov 24 13:33:30 crc kubenswrapper[4790]: I1124 13:33:30.328975 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf19b969-f00b-40f9-9ef0-e7556903f76e" path="/var/lib/kubelet/pods/bf19b969-f00b-40f9-9ef0-e7556903f76e/volumes" Nov 24 13:33:34 crc kubenswrapper[4790]: I1124 13:33:34.294282 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c79d794d7-q9jn5" podUID="d291f904-9c55-4213-9cde-69d6f71b67df" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: connect: connection refused" Nov 24 13:33:39 crc kubenswrapper[4790]: E1124 13:33:39.133569 4790 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Nov 24 13:33:39 crc kubenswrapper[4790]: E1124 13:33:39.134529 4790 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n76h5cbh67bh8dh84h67fhc6h697hc5h58bh5dhchcbh645h568h668h9h598h699h559h67ch647hb8h5d7h5cdhb4h59fh655h599h65dhdbh5d4q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8m6zh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(4d574c9c-ecb5-4b8c-b0b3-a058c141c09f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:33:42 crc kubenswrapper[4790]: E1124 13:33:42.223720 4790 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Nov 24 13:33:42 crc kubenswrapper[4790]: E1124 13:33:42.224297 4790 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2fcc7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-mbm86_openstack(ca484cf0-4ab0-4581-bb65-b1701db63df3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:33:42 crc kubenswrapper[4790]: E1124 13:33:42.225563 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-mbm86" podUID="ca484cf0-4ab0-4581-bb65-b1701db63df3" Nov 24 13:33:42 crc kubenswrapper[4790]: I1124 13:33:42.380531 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-q9jn5" Nov 24 13:33:42 crc kubenswrapper[4790]: E1124 13:33:42.381685 4790 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Nov 24 13:33:42 crc kubenswrapper[4790]: E1124 13:33:42.381863 4790 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2w4sf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-8dvn5_openstack(3e5246df-ca01-43e6-bcd9-c8d9b6b78279): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:33:42 crc kubenswrapper[4790]: E1124 13:33:42.383136 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-8dvn5" podUID="3e5246df-ca01-43e6-bcd9-c8d9b6b78279" Nov 24 13:33:42 crc kubenswrapper[4790]: I1124 13:33:42.467214 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d291f904-9c55-4213-9cde-69d6f71b67df-ovsdbserver-nb\") pod \"d291f904-9c55-4213-9cde-69d6f71b67df\" (UID: \"d291f904-9c55-4213-9cde-69d6f71b67df\") " Nov 24 13:33:42 crc kubenswrapper[4790]: I1124 13:33:42.467262 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d291f904-9c55-4213-9cde-69d6f71b67df-dns-swift-storage-0\") pod \"d291f904-9c55-4213-9cde-69d6f71b67df\" (UID: \"d291f904-9c55-4213-9cde-69d6f71b67df\") " Nov 24 13:33:42 crc kubenswrapper[4790]: I1124 13:33:42.467295 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d291f904-9c55-4213-9cde-69d6f71b67df-ovsdbserver-sb\") pod \"d291f904-9c55-4213-9cde-69d6f71b67df\" (UID: \"d291f904-9c55-4213-9cde-69d6f71b67df\") " Nov 24 13:33:42 crc kubenswrapper[4790]: I1124 13:33:42.467417 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7n9n5\" (UniqueName: \"kubernetes.io/projected/d291f904-9c55-4213-9cde-69d6f71b67df-kube-api-access-7n9n5\") pod \"d291f904-9c55-4213-9cde-69d6f71b67df\" (UID: \"d291f904-9c55-4213-9cde-69d6f71b67df\") " Nov 24 13:33:42 crc kubenswrapper[4790]: I1124 13:33:42.467476 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d291f904-9c55-4213-9cde-69d6f71b67df-dns-svc\") pod \"d291f904-9c55-4213-9cde-69d6f71b67df\" (UID: \"d291f904-9c55-4213-9cde-69d6f71b67df\") " Nov 24 13:33:42 crc kubenswrapper[4790]: I1124 13:33:42.467537 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d291f904-9c55-4213-9cde-69d6f71b67df-config\") pod \"d291f904-9c55-4213-9cde-69d6f71b67df\" (UID: \"d291f904-9c55-4213-9cde-69d6f71b67df\") " Nov 24 13:33:42 crc kubenswrapper[4790]: I1124 13:33:42.472914 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d291f904-9c55-4213-9cde-69d6f71b67df-kube-api-access-7n9n5" (OuterVolumeSpecName: "kube-api-access-7n9n5") pod "d291f904-9c55-4213-9cde-69d6f71b67df" (UID: "d291f904-9c55-4213-9cde-69d6f71b67df"). InnerVolumeSpecName "kube-api-access-7n9n5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:42 crc kubenswrapper[4790]: I1124 13:33:42.522343 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d291f904-9c55-4213-9cde-69d6f71b67df-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d291f904-9c55-4213-9cde-69d6f71b67df" (UID: "d291f904-9c55-4213-9cde-69d6f71b67df"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:42 crc kubenswrapper[4790]: I1124 13:33:42.523708 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d291f904-9c55-4213-9cde-69d6f71b67df-config" (OuterVolumeSpecName: "config") pod "d291f904-9c55-4213-9cde-69d6f71b67df" (UID: "d291f904-9c55-4213-9cde-69d6f71b67df"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:42 crc kubenswrapper[4790]: I1124 13:33:42.533296 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d291f904-9c55-4213-9cde-69d6f71b67df-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d291f904-9c55-4213-9cde-69d6f71b67df" (UID: "d291f904-9c55-4213-9cde-69d6f71b67df"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:42 crc kubenswrapper[4790]: I1124 13:33:42.533394 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d291f904-9c55-4213-9cde-69d6f71b67df-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d291f904-9c55-4213-9cde-69d6f71b67df" (UID: "d291f904-9c55-4213-9cde-69d6f71b67df"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:42 crc kubenswrapper[4790]: I1124 13:33:42.551520 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d291f904-9c55-4213-9cde-69d6f71b67df-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d291f904-9c55-4213-9cde-69d6f71b67df" (UID: "d291f904-9c55-4213-9cde-69d6f71b67df"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:42 crc kubenswrapper[4790]: I1124 13:33:42.574099 4790 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d291f904-9c55-4213-9cde-69d6f71b67df-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:42 crc kubenswrapper[4790]: I1124 13:33:42.574445 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d291f904-9c55-4213-9cde-69d6f71b67df-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:42 crc kubenswrapper[4790]: I1124 13:33:42.574460 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d291f904-9c55-4213-9cde-69d6f71b67df-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:42 crc kubenswrapper[4790]: I1124 13:33:42.574471 4790 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d291f904-9c55-4213-9cde-69d6f71b67df-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:42 crc kubenswrapper[4790]: I1124 13:33:42.574484 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d291f904-9c55-4213-9cde-69d6f71b67df-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:42 crc kubenswrapper[4790]: I1124 13:33:42.574495 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7n9n5\" (UniqueName: \"kubernetes.io/projected/d291f904-9c55-4213-9cde-69d6f71b67df-kube-api-access-7n9n5\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:42 crc kubenswrapper[4790]: I1124 13:33:42.773767 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-zqb2l"] Nov 24 13:33:42 crc kubenswrapper[4790]: W1124 13:33:42.782726 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod482e2d7c_282b_490c_b476_955ddc498707.slice/crio-3b4b9483e2d9a84c545a5b970e2550c2498d8887c62b854d55c0b393409877cd WatchSource:0}: Error finding container 3b4b9483e2d9a84c545a5b970e2550c2498d8887c62b854d55c0b393409877cd: Status 404 returned error can't find the container with id 3b4b9483e2d9a84c545a5b970e2550c2498d8887c62b854d55c0b393409877cd Nov 24 13:33:42 crc kubenswrapper[4790]: I1124 13:33:42.788108 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 13:33:43 crc kubenswrapper[4790]: I1124 13:33:43.208235 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-zqb2l" event={"ID":"482e2d7c-282b-490c-b476-955ddc498707","Type":"ContainerStarted","Data":"147300b703a9e3feb3b2b670b1b8b7a5e8e9817668e01ca6b34aaceb64126d0d"} Nov 24 13:33:43 crc kubenswrapper[4790]: I1124 13:33:43.208615 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-zqb2l" event={"ID":"482e2d7c-282b-490c-b476-955ddc498707","Type":"ContainerStarted","Data":"3b4b9483e2d9a84c545a5b970e2550c2498d8887c62b854d55c0b393409877cd"} Nov 24 13:33:43 crc kubenswrapper[4790]: I1124 13:33:43.216350 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"be3f173b-f907-4f35-bb20-3afe1e874583","Type":"ContainerStarted","Data":"28c80113b6a0beb50ce8b88b824e7d6e482ff1de8d5f6bbe43b263537bcd974b"} Nov 24 13:33:43 crc kubenswrapper[4790]: I1124 13:33:43.216388 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="be3f173b-f907-4f35-bb20-3afe1e874583" containerName="glance-log" containerID="cri-o://e4ce8052121e4bc044cba22d485867c5d86471221f0babfc7af4072083c45a26" gracePeriod=30 Nov 24 13:33:43 crc kubenswrapper[4790]: I1124 13:33:43.216476 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="be3f173b-f907-4f35-bb20-3afe1e874583" containerName="glance-httpd" containerID="cri-o://28c80113b6a0beb50ce8b88b824e7d6e482ff1de8d5f6bbe43b263537bcd974b" gracePeriod=30 Nov 24 13:33:43 crc kubenswrapper[4790]: I1124 13:33:43.254747 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-zqb2l" podStartSLOduration=15.254731133 podStartE2EDuration="15.254731133s" podCreationTimestamp="2025-11-24 13:33:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:33:43.226996994 +0000 UTC m=+1271.606890656" watchObservedRunningTime="2025-11-24 13:33:43.254731133 +0000 UTC m=+1271.634624795" Nov 24 13:33:43 crc kubenswrapper[4790]: I1124 13:33:43.256096 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-q9jn5" event={"ID":"d291f904-9c55-4213-9cde-69d6f71b67df","Type":"ContainerDied","Data":"40cfeb31462c0912bc5fcbc5d3a82655bb12b78c2e36a398a999e9ea2589fc8c"} Nov 24 13:33:43 crc kubenswrapper[4790]: I1124 13:33:43.256159 4790 scope.go:117] "RemoveContainer" containerID="083bb5dfa444f0e05367f4a3e328194b3c06970f40e5a951abc781f5875f0852" Nov 24 13:33:43 crc kubenswrapper[4790]: I1124 13:33:43.256315 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-q9jn5" Nov 24 13:33:43 crc kubenswrapper[4790]: I1124 13:33:43.262925 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-q4sb2" event={"ID":"6f5d721b-e52b-490d-89f3-4a5edaf403ec","Type":"ContainerStarted","Data":"1080017b2c864b1e6ae073dbeb6f0f37bbb80e9d0e0f77e3c13cf7a6fbb41f70"} Nov 24 13:33:43 crc kubenswrapper[4790]: I1124 13:33:43.269361 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="af74b160-8348-44b1-ad80-92bb0a1c34ad" containerName="glance-log" containerID="cri-o://b7f86aeef5273169f1a971a910b999e251459aa52cccb8d5b05853980caf75da" gracePeriod=30 Nov 24 13:33:43 crc kubenswrapper[4790]: I1124 13:33:43.269531 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="af74b160-8348-44b1-ad80-92bb0a1c34ad" containerName="glance-httpd" containerID="cri-o://2fc9cec7284506d3b37f47291b8ca79f29733436cf720e2a8d5dbd81e315744f" gracePeriod=30 Nov 24 13:33:43 crc kubenswrapper[4790]: I1124 13:33:43.269792 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"af74b160-8348-44b1-ad80-92bb0a1c34ad","Type":"ContainerStarted","Data":"2fc9cec7284506d3b37f47291b8ca79f29733436cf720e2a8d5dbd81e315744f"} Nov 24 13:33:43 crc kubenswrapper[4790]: E1124 13:33:43.271790 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-mbm86" podUID="ca484cf0-4ab0-4581-bb65-b1701db63df3" Nov 24 13:33:43 crc kubenswrapper[4790]: E1124 13:33:43.272298 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-8dvn5" podUID="3e5246df-ca01-43e6-bcd9-c8d9b6b78279" Nov 24 13:33:43 crc kubenswrapper[4790]: I1124 13:33:43.281112 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=24.281088612 podStartE2EDuration="24.281088612s" podCreationTimestamp="2025-11-24 13:33:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:33:43.275818342 +0000 UTC m=+1271.655712034" watchObservedRunningTime="2025-11-24 13:33:43.281088612 +0000 UTC m=+1271.660982264" Nov 24 13:33:43 crc kubenswrapper[4790]: I1124 13:33:43.315309 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-q4sb2" podStartSLOduration=2.620409453 podStartE2EDuration="27.315291215s" podCreationTimestamp="2025-11-24 13:33:16 +0000 UTC" firstStartedPulling="2025-11-24 13:33:17.584365716 +0000 UTC m=+1245.964259378" lastFinishedPulling="2025-11-24 13:33:42.279247438 +0000 UTC m=+1270.659141140" observedRunningTime="2025-11-24 13:33:43.308057749 +0000 UTC m=+1271.687951411" watchObservedRunningTime="2025-11-24 13:33:43.315291215 +0000 UTC m=+1271.695184877" Nov 24 13:33:43 crc kubenswrapper[4790]: I1124 13:33:43.339280 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=24.339260107 podStartE2EDuration="24.339260107s" podCreationTimestamp="2025-11-24 13:33:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:33:43.329783747 +0000 UTC m=+1271.709677409" watchObservedRunningTime="2025-11-24 13:33:43.339260107 +0000 UTC m=+1271.719153769" Nov 24 13:33:43 crc kubenswrapper[4790]: I1124 13:33:43.371955 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-q9jn5"] Nov 24 13:33:43 crc kubenswrapper[4790]: I1124 13:33:43.377794 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-q9jn5"] Nov 24 13:33:43 crc kubenswrapper[4790]: I1124 13:33:43.630713 4790 scope.go:117] "RemoveContainer" containerID="af9af167ccc6e18abbb926da632a2ecaa83cb236485e364edf0866450b15d5c0" Nov 24 13:33:43 crc kubenswrapper[4790]: I1124 13:33:43.938987 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:33:43 crc kubenswrapper[4790]: I1124 13:33:43.939054 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:33:43 crc kubenswrapper[4790]: I1124 13:33:43.939097 4790 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 13:33:43 crc kubenswrapper[4790]: I1124 13:33:43.939736 4790 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"56e28eeaa0ab287a217413e69cba64d88c1ede93557f55581b2a1331d38d609a"} pod="openshift-machine-config-operator/machine-config-daemon-xz49t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 13:33:43 crc kubenswrapper[4790]: I1124 13:33:43.939792 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" containerID="cri-o://56e28eeaa0ab287a217413e69cba64d88c1ede93557f55581b2a1331d38d609a" gracePeriod=600 Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.133046 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.205618 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"be3f173b-f907-4f35-bb20-3afe1e874583\" (UID: \"be3f173b-f907-4f35-bb20-3afe1e874583\") " Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.205713 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be3f173b-f907-4f35-bb20-3afe1e874583-config-data\") pod \"be3f173b-f907-4f35-bb20-3afe1e874583\" (UID: \"be3f173b-f907-4f35-bb20-3afe1e874583\") " Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.205750 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be3f173b-f907-4f35-bb20-3afe1e874583-logs\") pod \"be3f173b-f907-4f35-bb20-3afe1e874583\" (UID: \"be3f173b-f907-4f35-bb20-3afe1e874583\") " Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.205810 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be3f173b-f907-4f35-bb20-3afe1e874583-combined-ca-bundle\") pod \"be3f173b-f907-4f35-bb20-3afe1e874583\" (UID: \"be3f173b-f907-4f35-bb20-3afe1e874583\") " Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.205848 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be3f173b-f907-4f35-bb20-3afe1e874583-httpd-run\") pod \"be3f173b-f907-4f35-bb20-3afe1e874583\" (UID: \"be3f173b-f907-4f35-bb20-3afe1e874583\") " Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.205914 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-248q5\" (UniqueName: \"kubernetes.io/projected/be3f173b-f907-4f35-bb20-3afe1e874583-kube-api-access-248q5\") pod \"be3f173b-f907-4f35-bb20-3afe1e874583\" (UID: \"be3f173b-f907-4f35-bb20-3afe1e874583\") " Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.205940 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be3f173b-f907-4f35-bb20-3afe1e874583-scripts\") pod \"be3f173b-f907-4f35-bb20-3afe1e874583\" (UID: \"be3f173b-f907-4f35-bb20-3afe1e874583\") " Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.206184 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be3f173b-f907-4f35-bb20-3afe1e874583-logs" (OuterVolumeSpecName: "logs") pod "be3f173b-f907-4f35-bb20-3afe1e874583" (UID: "be3f173b-f907-4f35-bb20-3afe1e874583"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.206445 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be3f173b-f907-4f35-bb20-3afe1e874583-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.210232 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be3f173b-f907-4f35-bb20-3afe1e874583-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "be3f173b-f907-4f35-bb20-3afe1e874583" (UID: "be3f173b-f907-4f35-bb20-3afe1e874583"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.212681 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "be3f173b-f907-4f35-bb20-3afe1e874583" (UID: "be3f173b-f907-4f35-bb20-3afe1e874583"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.216859 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be3f173b-f907-4f35-bb20-3afe1e874583-kube-api-access-248q5" (OuterVolumeSpecName: "kube-api-access-248q5") pod "be3f173b-f907-4f35-bb20-3afe1e874583" (UID: "be3f173b-f907-4f35-bb20-3afe1e874583"). InnerVolumeSpecName "kube-api-access-248q5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.218552 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be3f173b-f907-4f35-bb20-3afe1e874583-scripts" (OuterVolumeSpecName: "scripts") pod "be3f173b-f907-4f35-bb20-3afe1e874583" (UID: "be3f173b-f907-4f35-bb20-3afe1e874583"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.241818 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be3f173b-f907-4f35-bb20-3afe1e874583-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be3f173b-f907-4f35-bb20-3afe1e874583" (UID: "be3f173b-f907-4f35-bb20-3afe1e874583"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.261246 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be3f173b-f907-4f35-bb20-3afe1e874583-config-data" (OuterVolumeSpecName: "config-data") pod "be3f173b-f907-4f35-bb20-3afe1e874583" (UID: "be3f173b-f907-4f35-bb20-3afe1e874583"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.262141 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.293991 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c79d794d7-q9jn5" podUID="d291f904-9c55-4213-9cde-69d6f71b67df" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: i/o timeout" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.304202 4790 generic.go:334] "Generic (PLEG): container finished" podID="0d73b133-48f0-455f-8f6a-742e633f631a" containerID="56e28eeaa0ab287a217413e69cba64d88c1ede93557f55581b2a1331d38d609a" exitCode=0 Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.304275 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerDied","Data":"56e28eeaa0ab287a217413e69cba64d88c1ede93557f55581b2a1331d38d609a"} Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.304473 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"8544370c5579779a1c61d01762c06ad3cb058bc1368085e6b5750e300ddbf449"} Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.304640 4790 scope.go:117] "RemoveContainer" containerID="e635b1d9a632257918b8c7ec939010f4c4491014f07d31a3ef3de514129b90d0" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.308890 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af74b160-8348-44b1-ad80-92bb0a1c34ad-combined-ca-bundle\") pod \"af74b160-8348-44b1-ad80-92bb0a1c34ad\" (UID: \"af74b160-8348-44b1-ad80-92bb0a1c34ad\") " Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.308960 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af74b160-8348-44b1-ad80-92bb0a1c34ad-logs\") pod \"af74b160-8348-44b1-ad80-92bb0a1c34ad\" (UID: \"af74b160-8348-44b1-ad80-92bb0a1c34ad\") " Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.308980 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af74b160-8348-44b1-ad80-92bb0a1c34ad-config-data\") pod \"af74b160-8348-44b1-ad80-92bb0a1c34ad\" (UID: \"af74b160-8348-44b1-ad80-92bb0a1c34ad\") " Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.308995 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af74b160-8348-44b1-ad80-92bb0a1c34ad-scripts\") pod \"af74b160-8348-44b1-ad80-92bb0a1c34ad\" (UID: \"af74b160-8348-44b1-ad80-92bb0a1c34ad\") " Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.309057 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/af74b160-8348-44b1-ad80-92bb0a1c34ad-httpd-run\") pod \"af74b160-8348-44b1-ad80-92bb0a1c34ad\" (UID: \"af74b160-8348-44b1-ad80-92bb0a1c34ad\") " Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.309079 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"af74b160-8348-44b1-ad80-92bb0a1c34ad\" (UID: \"af74b160-8348-44b1-ad80-92bb0a1c34ad\") " Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.309122 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnj6g\" (UniqueName: \"kubernetes.io/projected/af74b160-8348-44b1-ad80-92bb0a1c34ad-kube-api-access-pnj6g\") pod \"af74b160-8348-44b1-ad80-92bb0a1c34ad\" (UID: \"af74b160-8348-44b1-ad80-92bb0a1c34ad\") " Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.309723 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-248q5\" (UniqueName: \"kubernetes.io/projected/be3f173b-f907-4f35-bb20-3afe1e874583-kube-api-access-248q5\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.309756 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be3f173b-f907-4f35-bb20-3afe1e874583-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.309779 4790 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.309789 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be3f173b-f907-4f35-bb20-3afe1e874583-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.309798 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be3f173b-f907-4f35-bb20-3afe1e874583-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.309806 4790 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be3f173b-f907-4f35-bb20-3afe1e874583-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.311424 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af74b160-8348-44b1-ad80-92bb0a1c34ad-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "af74b160-8348-44b1-ad80-92bb0a1c34ad" (UID: "af74b160-8348-44b1-ad80-92bb0a1c34ad"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.311679 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af74b160-8348-44b1-ad80-92bb0a1c34ad-logs" (OuterVolumeSpecName: "logs") pod "af74b160-8348-44b1-ad80-92bb0a1c34ad" (UID: "af74b160-8348-44b1-ad80-92bb0a1c34ad"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.312924 4790 generic.go:334] "Generic (PLEG): container finished" podID="af74b160-8348-44b1-ad80-92bb0a1c34ad" containerID="2fc9cec7284506d3b37f47291b8ca79f29733436cf720e2a8d5dbd81e315744f" exitCode=0 Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.312992 4790 generic.go:334] "Generic (PLEG): container finished" podID="af74b160-8348-44b1-ad80-92bb0a1c34ad" containerID="b7f86aeef5273169f1a971a910b999e251459aa52cccb8d5b05853980caf75da" exitCode=143 Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.313084 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"af74b160-8348-44b1-ad80-92bb0a1c34ad","Type":"ContainerDied","Data":"2fc9cec7284506d3b37f47291b8ca79f29733436cf720e2a8d5dbd81e315744f"} Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.313123 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"af74b160-8348-44b1-ad80-92bb0a1c34ad","Type":"ContainerDied","Data":"b7f86aeef5273169f1a971a910b999e251459aa52cccb8d5b05853980caf75da"} Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.313163 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"af74b160-8348-44b1-ad80-92bb0a1c34ad","Type":"ContainerDied","Data":"eedf60aefb2065fa246c8ccb6d94be23de7d8cce902f9ec10140fc36e0f8223b"} Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.313264 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.322934 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "af74b160-8348-44b1-ad80-92bb0a1c34ad" (UID: "af74b160-8348-44b1-ad80-92bb0a1c34ad"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.324362 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af74b160-8348-44b1-ad80-92bb0a1c34ad-scripts" (OuterVolumeSpecName: "scripts") pod "af74b160-8348-44b1-ad80-92bb0a1c34ad" (UID: "af74b160-8348-44b1-ad80-92bb0a1c34ad"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.330088 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af74b160-8348-44b1-ad80-92bb0a1c34ad-kube-api-access-pnj6g" (OuterVolumeSpecName: "kube-api-access-pnj6g") pod "af74b160-8348-44b1-ad80-92bb0a1c34ad" (UID: "af74b160-8348-44b1-ad80-92bb0a1c34ad"). InnerVolumeSpecName "kube-api-access-pnj6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.330648 4790 generic.go:334] "Generic (PLEG): container finished" podID="be3f173b-f907-4f35-bb20-3afe1e874583" containerID="28c80113b6a0beb50ce8b88b824e7d6e482ff1de8d5f6bbe43b263537bcd974b" exitCode=0 Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.330712 4790 generic.go:334] "Generic (PLEG): container finished" podID="be3f173b-f907-4f35-bb20-3afe1e874583" containerID="e4ce8052121e4bc044cba22d485867c5d86471221f0babfc7af4072083c45a26" exitCode=143 Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.330866 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.346515 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d291f904-9c55-4213-9cde-69d6f71b67df" path="/var/lib/kubelet/pods/d291f904-9c55-4213-9cde-69d6f71b67df/volumes" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.347933 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af74b160-8348-44b1-ad80-92bb0a1c34ad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "af74b160-8348-44b1-ad80-92bb0a1c34ad" (UID: "af74b160-8348-44b1-ad80-92bb0a1c34ad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.366108 4790 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.379913 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af74b160-8348-44b1-ad80-92bb0a1c34ad-config-data" (OuterVolumeSpecName: "config-data") pod "af74b160-8348-44b1-ad80-92bb0a1c34ad" (UID: "af74b160-8348-44b1-ad80-92bb0a1c34ad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.411101 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af74b160-8348-44b1-ad80-92bb0a1c34ad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.411132 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af74b160-8348-44b1-ad80-92bb0a1c34ad-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.411144 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af74b160-8348-44b1-ad80-92bb0a1c34ad-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.411152 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af74b160-8348-44b1-ad80-92bb0a1c34ad-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.411189 4790 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/af74b160-8348-44b1-ad80-92bb0a1c34ad-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.411211 4790 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.411220 4790 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.411229 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnj6g\" (UniqueName: \"kubernetes.io/projected/af74b160-8348-44b1-ad80-92bb0a1c34ad-kube-api-access-pnj6g\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.427827 4790 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.434964 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f","Type":"ContainerStarted","Data":"6a117358d00dc47eb45e9087790aa499d845a99631f8d556a818c3dbb36a7a89"} Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.435021 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"be3f173b-f907-4f35-bb20-3afe1e874583","Type":"ContainerDied","Data":"28c80113b6a0beb50ce8b88b824e7d6e482ff1de8d5f6bbe43b263537bcd974b"} Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.435039 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"be3f173b-f907-4f35-bb20-3afe1e874583","Type":"ContainerDied","Data":"e4ce8052121e4bc044cba22d485867c5d86471221f0babfc7af4072083c45a26"} Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.435052 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"be3f173b-f907-4f35-bb20-3afe1e874583","Type":"ContainerDied","Data":"2d85f3b00fe43e02223548e90a3a92d081c60c9e702d046f75c41e7bc9eda4ae"} Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.455964 4790 scope.go:117] "RemoveContainer" containerID="2fc9cec7284506d3b37f47291b8ca79f29733436cf720e2a8d5dbd81e315744f" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.490862 4790 scope.go:117] "RemoveContainer" containerID="b7f86aeef5273169f1a971a910b999e251459aa52cccb8d5b05853980caf75da" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.514064 4790 scope.go:117] "RemoveContainer" containerID="2fc9cec7284506d3b37f47291b8ca79f29733436cf720e2a8d5dbd81e315744f" Nov 24 13:33:44 crc kubenswrapper[4790]: E1124 13:33:44.514527 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fc9cec7284506d3b37f47291b8ca79f29733436cf720e2a8d5dbd81e315744f\": container with ID starting with 2fc9cec7284506d3b37f47291b8ca79f29733436cf720e2a8d5dbd81e315744f not found: ID does not exist" containerID="2fc9cec7284506d3b37f47291b8ca79f29733436cf720e2a8d5dbd81e315744f" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.514572 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fc9cec7284506d3b37f47291b8ca79f29733436cf720e2a8d5dbd81e315744f"} err="failed to get container status \"2fc9cec7284506d3b37f47291b8ca79f29733436cf720e2a8d5dbd81e315744f\": rpc error: code = NotFound desc = could not find container \"2fc9cec7284506d3b37f47291b8ca79f29733436cf720e2a8d5dbd81e315744f\": container with ID starting with 2fc9cec7284506d3b37f47291b8ca79f29733436cf720e2a8d5dbd81e315744f not found: ID does not exist" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.514599 4790 scope.go:117] "RemoveContainer" containerID="b7f86aeef5273169f1a971a910b999e251459aa52cccb8d5b05853980caf75da" Nov 24 13:33:44 crc kubenswrapper[4790]: E1124 13:33:44.514866 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7f86aeef5273169f1a971a910b999e251459aa52cccb8d5b05853980caf75da\": container with ID starting with b7f86aeef5273169f1a971a910b999e251459aa52cccb8d5b05853980caf75da not found: ID does not exist" containerID="b7f86aeef5273169f1a971a910b999e251459aa52cccb8d5b05853980caf75da" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.514883 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7f86aeef5273169f1a971a910b999e251459aa52cccb8d5b05853980caf75da"} err="failed to get container status \"b7f86aeef5273169f1a971a910b999e251459aa52cccb8d5b05853980caf75da\": rpc error: code = NotFound desc = could not find container \"b7f86aeef5273169f1a971a910b999e251459aa52cccb8d5b05853980caf75da\": container with ID starting with b7f86aeef5273169f1a971a910b999e251459aa52cccb8d5b05853980caf75da not found: ID does not exist" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.514908 4790 scope.go:117] "RemoveContainer" containerID="2fc9cec7284506d3b37f47291b8ca79f29733436cf720e2a8d5dbd81e315744f" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.515056 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fc9cec7284506d3b37f47291b8ca79f29733436cf720e2a8d5dbd81e315744f"} err="failed to get container status \"2fc9cec7284506d3b37f47291b8ca79f29733436cf720e2a8d5dbd81e315744f\": rpc error: code = NotFound desc = could not find container \"2fc9cec7284506d3b37f47291b8ca79f29733436cf720e2a8d5dbd81e315744f\": container with ID starting with 2fc9cec7284506d3b37f47291b8ca79f29733436cf720e2a8d5dbd81e315744f not found: ID does not exist" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.515069 4790 scope.go:117] "RemoveContainer" containerID="b7f86aeef5273169f1a971a910b999e251459aa52cccb8d5b05853980caf75da" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.515087 4790 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.515249 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7f86aeef5273169f1a971a910b999e251459aa52cccb8d5b05853980caf75da"} err="failed to get container status \"b7f86aeef5273169f1a971a910b999e251459aa52cccb8d5b05853980caf75da\": rpc error: code = NotFound desc = could not find container \"b7f86aeef5273169f1a971a910b999e251459aa52cccb8d5b05853980caf75da\": container with ID starting with b7f86aeef5273169f1a971a910b999e251459aa52cccb8d5b05853980caf75da not found: ID does not exist" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.515263 4790 scope.go:117] "RemoveContainer" containerID="28c80113b6a0beb50ce8b88b824e7d6e482ff1de8d5f6bbe43b263537bcd974b" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.539126 4790 scope.go:117] "RemoveContainer" containerID="e4ce8052121e4bc044cba22d485867c5d86471221f0babfc7af4072083c45a26" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.559321 4790 scope.go:117] "RemoveContainer" containerID="28c80113b6a0beb50ce8b88b824e7d6e482ff1de8d5f6bbe43b263537bcd974b" Nov 24 13:33:44 crc kubenswrapper[4790]: E1124 13:33:44.559755 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28c80113b6a0beb50ce8b88b824e7d6e482ff1de8d5f6bbe43b263537bcd974b\": container with ID starting with 28c80113b6a0beb50ce8b88b824e7d6e482ff1de8d5f6bbe43b263537bcd974b not found: ID does not exist" containerID="28c80113b6a0beb50ce8b88b824e7d6e482ff1de8d5f6bbe43b263537bcd974b" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.559789 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28c80113b6a0beb50ce8b88b824e7d6e482ff1de8d5f6bbe43b263537bcd974b"} err="failed to get container status \"28c80113b6a0beb50ce8b88b824e7d6e482ff1de8d5f6bbe43b263537bcd974b\": rpc error: code = NotFound desc = could not find container \"28c80113b6a0beb50ce8b88b824e7d6e482ff1de8d5f6bbe43b263537bcd974b\": container with ID starting with 28c80113b6a0beb50ce8b88b824e7d6e482ff1de8d5f6bbe43b263537bcd974b not found: ID does not exist" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.559818 4790 scope.go:117] "RemoveContainer" containerID="e4ce8052121e4bc044cba22d485867c5d86471221f0babfc7af4072083c45a26" Nov 24 13:33:44 crc kubenswrapper[4790]: E1124 13:33:44.560302 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4ce8052121e4bc044cba22d485867c5d86471221f0babfc7af4072083c45a26\": container with ID starting with e4ce8052121e4bc044cba22d485867c5d86471221f0babfc7af4072083c45a26 not found: ID does not exist" containerID="e4ce8052121e4bc044cba22d485867c5d86471221f0babfc7af4072083c45a26" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.560353 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4ce8052121e4bc044cba22d485867c5d86471221f0babfc7af4072083c45a26"} err="failed to get container status \"e4ce8052121e4bc044cba22d485867c5d86471221f0babfc7af4072083c45a26\": rpc error: code = NotFound desc = could not find container \"e4ce8052121e4bc044cba22d485867c5d86471221f0babfc7af4072083c45a26\": container with ID starting with e4ce8052121e4bc044cba22d485867c5d86471221f0babfc7af4072083c45a26 not found: ID does not exist" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.560385 4790 scope.go:117] "RemoveContainer" containerID="28c80113b6a0beb50ce8b88b824e7d6e482ff1de8d5f6bbe43b263537bcd974b" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.560649 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28c80113b6a0beb50ce8b88b824e7d6e482ff1de8d5f6bbe43b263537bcd974b"} err="failed to get container status \"28c80113b6a0beb50ce8b88b824e7d6e482ff1de8d5f6bbe43b263537bcd974b\": rpc error: code = NotFound desc = could not find container \"28c80113b6a0beb50ce8b88b824e7d6e482ff1de8d5f6bbe43b263537bcd974b\": container with ID starting with 28c80113b6a0beb50ce8b88b824e7d6e482ff1de8d5f6bbe43b263537bcd974b not found: ID does not exist" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.560671 4790 scope.go:117] "RemoveContainer" containerID="e4ce8052121e4bc044cba22d485867c5d86471221f0babfc7af4072083c45a26" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.560966 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4ce8052121e4bc044cba22d485867c5d86471221f0babfc7af4072083c45a26"} err="failed to get container status \"e4ce8052121e4bc044cba22d485867c5d86471221f0babfc7af4072083c45a26\": rpc error: code = NotFound desc = could not find container \"e4ce8052121e4bc044cba22d485867c5d86471221f0babfc7af4072083c45a26\": container with ID starting with e4ce8052121e4bc044cba22d485867c5d86471221f0babfc7af4072083c45a26 not found: ID does not exist" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.646936 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.654727 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.673792 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 13:33:44 crc kubenswrapper[4790]: E1124 13:33:44.674279 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af74b160-8348-44b1-ad80-92bb0a1c34ad" containerName="glance-httpd" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.674301 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="af74b160-8348-44b1-ad80-92bb0a1c34ad" containerName="glance-httpd" Nov 24 13:33:44 crc kubenswrapper[4790]: E1124 13:33:44.674325 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be3f173b-f907-4f35-bb20-3afe1e874583" containerName="glance-httpd" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.674337 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="be3f173b-f907-4f35-bb20-3afe1e874583" containerName="glance-httpd" Nov 24 13:33:44 crc kubenswrapper[4790]: E1124 13:33:44.674350 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d291f904-9c55-4213-9cde-69d6f71b67df" containerName="init" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.674359 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="d291f904-9c55-4213-9cde-69d6f71b67df" containerName="init" Nov 24 13:33:44 crc kubenswrapper[4790]: E1124 13:33:44.674374 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d291f904-9c55-4213-9cde-69d6f71b67df" containerName="dnsmasq-dns" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.674381 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="d291f904-9c55-4213-9cde-69d6f71b67df" containerName="dnsmasq-dns" Nov 24 13:33:44 crc kubenswrapper[4790]: E1124 13:33:44.674400 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af74b160-8348-44b1-ad80-92bb0a1c34ad" containerName="glance-log" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.674407 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="af74b160-8348-44b1-ad80-92bb0a1c34ad" containerName="glance-log" Nov 24 13:33:44 crc kubenswrapper[4790]: E1124 13:33:44.674419 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be3f173b-f907-4f35-bb20-3afe1e874583" containerName="glance-log" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.674429 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="be3f173b-f907-4f35-bb20-3afe1e874583" containerName="glance-log" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.674621 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="be3f173b-f907-4f35-bb20-3afe1e874583" containerName="glance-log" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.674642 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="be3f173b-f907-4f35-bb20-3afe1e874583" containerName="glance-httpd" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.674663 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="af74b160-8348-44b1-ad80-92bb0a1c34ad" containerName="glance-log" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.674674 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="d291f904-9c55-4213-9cde-69d6f71b67df" containerName="dnsmasq-dns" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.674693 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="af74b160-8348-44b1-ad80-92bb0a1c34ad" containerName="glance-httpd" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.675779 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.681504 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.681903 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.681930 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-nrdsr" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.682007 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.688805 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.821710 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69617248-5d4b-4027-a2a9-e972278e9d5e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"69617248-5d4b-4027-a2a9-e972278e9d5e\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.821750 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69617248-5d4b-4027-a2a9-e972278e9d5e-scripts\") pod \"glance-default-external-api-0\" (UID: \"69617248-5d4b-4027-a2a9-e972278e9d5e\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.821810 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thlrp\" (UniqueName: \"kubernetes.io/projected/69617248-5d4b-4027-a2a9-e972278e9d5e-kube-api-access-thlrp\") pod \"glance-default-external-api-0\" (UID: \"69617248-5d4b-4027-a2a9-e972278e9d5e\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.821831 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"69617248-5d4b-4027-a2a9-e972278e9d5e\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.821848 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69617248-5d4b-4027-a2a9-e972278e9d5e-config-data\") pod \"glance-default-external-api-0\" (UID: \"69617248-5d4b-4027-a2a9-e972278e9d5e\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.821867 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/69617248-5d4b-4027-a2a9-e972278e9d5e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"69617248-5d4b-4027-a2a9-e972278e9d5e\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.821913 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/69617248-5d4b-4027-a2a9-e972278e9d5e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"69617248-5d4b-4027-a2a9-e972278e9d5e\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.821968 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69617248-5d4b-4027-a2a9-e972278e9d5e-logs\") pod \"glance-default-external-api-0\" (UID: \"69617248-5d4b-4027-a2a9-e972278e9d5e\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.923749 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69617248-5d4b-4027-a2a9-e972278e9d5e-logs\") pod \"glance-default-external-api-0\" (UID: \"69617248-5d4b-4027-a2a9-e972278e9d5e\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.923817 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69617248-5d4b-4027-a2a9-e972278e9d5e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"69617248-5d4b-4027-a2a9-e972278e9d5e\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.923838 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69617248-5d4b-4027-a2a9-e972278e9d5e-scripts\") pod \"glance-default-external-api-0\" (UID: \"69617248-5d4b-4027-a2a9-e972278e9d5e\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.923915 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thlrp\" (UniqueName: \"kubernetes.io/projected/69617248-5d4b-4027-a2a9-e972278e9d5e-kube-api-access-thlrp\") pod \"glance-default-external-api-0\" (UID: \"69617248-5d4b-4027-a2a9-e972278e9d5e\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.923944 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"69617248-5d4b-4027-a2a9-e972278e9d5e\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.923972 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69617248-5d4b-4027-a2a9-e972278e9d5e-config-data\") pod \"glance-default-external-api-0\" (UID: \"69617248-5d4b-4027-a2a9-e972278e9d5e\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.923987 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/69617248-5d4b-4027-a2a9-e972278e9d5e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"69617248-5d4b-4027-a2a9-e972278e9d5e\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.924015 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/69617248-5d4b-4027-a2a9-e972278e9d5e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"69617248-5d4b-4027-a2a9-e972278e9d5e\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.925086 4790 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"69617248-5d4b-4027-a2a9-e972278e9d5e\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.925400 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69617248-5d4b-4027-a2a9-e972278e9d5e-logs\") pod \"glance-default-external-api-0\" (UID: \"69617248-5d4b-4027-a2a9-e972278e9d5e\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.925557 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/69617248-5d4b-4027-a2a9-e972278e9d5e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"69617248-5d4b-4027-a2a9-e972278e9d5e\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.930446 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69617248-5d4b-4027-a2a9-e972278e9d5e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"69617248-5d4b-4027-a2a9-e972278e9d5e\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.943727 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69617248-5d4b-4027-a2a9-e972278e9d5e-config-data\") pod \"glance-default-external-api-0\" (UID: \"69617248-5d4b-4027-a2a9-e972278e9d5e\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.945302 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/69617248-5d4b-4027-a2a9-e972278e9d5e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"69617248-5d4b-4027-a2a9-e972278e9d5e\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.948215 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thlrp\" (UniqueName: \"kubernetes.io/projected/69617248-5d4b-4027-a2a9-e972278e9d5e-kube-api-access-thlrp\") pod \"glance-default-external-api-0\" (UID: \"69617248-5d4b-4027-a2a9-e972278e9d5e\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.954436 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69617248-5d4b-4027-a2a9-e972278e9d5e-scripts\") pod \"glance-default-external-api-0\" (UID: \"69617248-5d4b-4027-a2a9-e972278e9d5e\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:44 crc kubenswrapper[4790]: I1124 13:33:44.988806 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"69617248-5d4b-4027-a2a9-e972278e9d5e\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:45 crc kubenswrapper[4790]: I1124 13:33:45.028087 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 13:33:45 crc kubenswrapper[4790]: I1124 13:33:45.558596 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 13:33:45 crc kubenswrapper[4790]: W1124 13:33:45.594475 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69617248_5d4b_4027_a2a9_e972278e9d5e.slice/crio-16744171db16176fa3a2049b1e44d1307eb50cc23f3c2b08a0fd5cff09d24569 WatchSource:0}: Error finding container 16744171db16176fa3a2049b1e44d1307eb50cc23f3c2b08a0fd5cff09d24569: Status 404 returned error can't find the container with id 16744171db16176fa3a2049b1e44d1307eb50cc23f3c2b08a0fd5cff09d24569 Nov 24 13:33:46 crc kubenswrapper[4790]: I1124 13:33:46.325111 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af74b160-8348-44b1-ad80-92bb0a1c34ad" path="/var/lib/kubelet/pods/af74b160-8348-44b1-ad80-92bb0a1c34ad/volumes" Nov 24 13:33:46 crc kubenswrapper[4790]: I1124 13:33:46.362747 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"69617248-5d4b-4027-a2a9-e972278e9d5e","Type":"ContainerStarted","Data":"590c7623f08a06bb7521b7040e8300b1caee6cdbd6b60c1d3823cf11d4cb0fbb"} Nov 24 13:33:46 crc kubenswrapper[4790]: I1124 13:33:46.362787 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"69617248-5d4b-4027-a2a9-e972278e9d5e","Type":"ContainerStarted","Data":"16744171db16176fa3a2049b1e44d1307eb50cc23f3c2b08a0fd5cff09d24569"} Nov 24 13:33:47 crc kubenswrapper[4790]: I1124 13:33:47.376251 4790 generic.go:334] "Generic (PLEG): container finished" podID="482e2d7c-282b-490c-b476-955ddc498707" containerID="147300b703a9e3feb3b2b670b1b8b7a5e8e9817668e01ca6b34aaceb64126d0d" exitCode=0 Nov 24 13:33:47 crc kubenswrapper[4790]: I1124 13:33:47.376366 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-zqb2l" event={"ID":"482e2d7c-282b-490c-b476-955ddc498707","Type":"ContainerDied","Data":"147300b703a9e3feb3b2b670b1b8b7a5e8e9817668e01ca6b34aaceb64126d0d"} Nov 24 13:33:48 crc kubenswrapper[4790]: I1124 13:33:48.387562 4790 generic.go:334] "Generic (PLEG): container finished" podID="6f5d721b-e52b-490d-89f3-4a5edaf403ec" containerID="1080017b2c864b1e6ae073dbeb6f0f37bbb80e9d0e0f77e3c13cf7a6fbb41f70" exitCode=0 Nov 24 13:33:48 crc kubenswrapper[4790]: I1124 13:33:48.387602 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-q4sb2" event={"ID":"6f5d721b-e52b-490d-89f3-4a5edaf403ec","Type":"ContainerDied","Data":"1080017b2c864b1e6ae073dbeb6f0f37bbb80e9d0e0f77e3c13cf7a6fbb41f70"} Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.342418 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-zqb2l" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.431001 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-zqb2l" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.431135 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-zqb2l" event={"ID":"482e2d7c-282b-490c-b476-955ddc498707","Type":"ContainerDied","Data":"3b4b9483e2d9a84c545a5b970e2550c2498d8887c62b854d55c0b393409877cd"} Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.431186 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b4b9483e2d9a84c545a5b970e2550c2498d8887c62b854d55c0b393409877cd" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.433096 4790 generic.go:334] "Generic (PLEG): container finished" podID="84a50ea0-c597-42e2-b4cd-7e628a517ddf" containerID="cf5790b66698273019b8f16b970dbdbffc943209fdca11bb7061f46cb7418370" exitCode=0 Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.433209 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-p5msf" event={"ID":"84a50ea0-c597-42e2-b4cd-7e628a517ddf","Type":"ContainerDied","Data":"cf5790b66698273019b8f16b970dbdbffc943209fdca11bb7061f46cb7418370"} Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.497343 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7ddd868cf8-lc6qt"] Nov 24 13:33:49 crc kubenswrapper[4790]: E1124 13:33:49.497741 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="482e2d7c-282b-490c-b476-955ddc498707" containerName="keystone-bootstrap" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.497755 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="482e2d7c-282b-490c-b476-955ddc498707" containerName="keystone-bootstrap" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.497998 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="482e2d7c-282b-490c-b476-955ddc498707" containerName="keystone-bootstrap" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.498606 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7ddd868cf8-lc6qt" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.505232 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.507434 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.509789 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7ddd868cf8-lc6qt"] Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.520387 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/482e2d7c-282b-490c-b476-955ddc498707-fernet-keys\") pod \"482e2d7c-282b-490c-b476-955ddc498707\" (UID: \"482e2d7c-282b-490c-b476-955ddc498707\") " Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.520451 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/482e2d7c-282b-490c-b476-955ddc498707-scripts\") pod \"482e2d7c-282b-490c-b476-955ddc498707\" (UID: \"482e2d7c-282b-490c-b476-955ddc498707\") " Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.520482 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/482e2d7c-282b-490c-b476-955ddc498707-credential-keys\") pod \"482e2d7c-282b-490c-b476-955ddc498707\" (UID: \"482e2d7c-282b-490c-b476-955ddc498707\") " Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.520525 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/482e2d7c-282b-490c-b476-955ddc498707-combined-ca-bundle\") pod \"482e2d7c-282b-490c-b476-955ddc498707\" (UID: \"482e2d7c-282b-490c-b476-955ddc498707\") " Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.520553 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrx9v\" (UniqueName: \"kubernetes.io/projected/482e2d7c-282b-490c-b476-955ddc498707-kube-api-access-jrx9v\") pod \"482e2d7c-282b-490c-b476-955ddc498707\" (UID: \"482e2d7c-282b-490c-b476-955ddc498707\") " Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.520651 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/482e2d7c-282b-490c-b476-955ddc498707-config-data\") pod \"482e2d7c-282b-490c-b476-955ddc498707\" (UID: \"482e2d7c-282b-490c-b476-955ddc498707\") " Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.533700 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/482e2d7c-282b-490c-b476-955ddc498707-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "482e2d7c-282b-490c-b476-955ddc498707" (UID: "482e2d7c-282b-490c-b476-955ddc498707"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.537098 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/482e2d7c-282b-490c-b476-955ddc498707-kube-api-access-jrx9v" (OuterVolumeSpecName: "kube-api-access-jrx9v") pod "482e2d7c-282b-490c-b476-955ddc498707" (UID: "482e2d7c-282b-490c-b476-955ddc498707"). InnerVolumeSpecName "kube-api-access-jrx9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.542156 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/482e2d7c-282b-490c-b476-955ddc498707-scripts" (OuterVolumeSpecName: "scripts") pod "482e2d7c-282b-490c-b476-955ddc498707" (UID: "482e2d7c-282b-490c-b476-955ddc498707"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.542419 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/482e2d7c-282b-490c-b476-955ddc498707-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "482e2d7c-282b-490c-b476-955ddc498707" (UID: "482e2d7c-282b-490c-b476-955ddc498707"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.576022 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/482e2d7c-282b-490c-b476-955ddc498707-config-data" (OuterVolumeSpecName: "config-data") pod "482e2d7c-282b-490c-b476-955ddc498707" (UID: "482e2d7c-282b-490c-b476-955ddc498707"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.596083 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/482e2d7c-282b-490c-b476-955ddc498707-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "482e2d7c-282b-490c-b476-955ddc498707" (UID: "482e2d7c-282b-490c-b476-955ddc498707"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.622663 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-internal-tls-certs\") pod \"keystone-7ddd868cf8-lc6qt\" (UID: \"336619dc-13bf-4168-a58b-b5c724759890\") " pod="openstack/keystone-7ddd868cf8-lc6qt" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.622751 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-public-tls-certs\") pod \"keystone-7ddd868cf8-lc6qt\" (UID: \"336619dc-13bf-4168-a58b-b5c724759890\") " pod="openstack/keystone-7ddd868cf8-lc6qt" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.622959 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-config-data\") pod \"keystone-7ddd868cf8-lc6qt\" (UID: \"336619dc-13bf-4168-a58b-b5c724759890\") " pod="openstack/keystone-7ddd868cf8-lc6qt" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.623014 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dv4jm\" (UniqueName: \"kubernetes.io/projected/336619dc-13bf-4168-a58b-b5c724759890-kube-api-access-dv4jm\") pod \"keystone-7ddd868cf8-lc6qt\" (UID: \"336619dc-13bf-4168-a58b-b5c724759890\") " pod="openstack/keystone-7ddd868cf8-lc6qt" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.623076 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-credential-keys\") pod \"keystone-7ddd868cf8-lc6qt\" (UID: \"336619dc-13bf-4168-a58b-b5c724759890\") " pod="openstack/keystone-7ddd868cf8-lc6qt" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.623276 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-combined-ca-bundle\") pod \"keystone-7ddd868cf8-lc6qt\" (UID: \"336619dc-13bf-4168-a58b-b5c724759890\") " pod="openstack/keystone-7ddd868cf8-lc6qt" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.623314 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-fernet-keys\") pod \"keystone-7ddd868cf8-lc6qt\" (UID: \"336619dc-13bf-4168-a58b-b5c724759890\") " pod="openstack/keystone-7ddd868cf8-lc6qt" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.623360 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-scripts\") pod \"keystone-7ddd868cf8-lc6qt\" (UID: \"336619dc-13bf-4168-a58b-b5c724759890\") " pod="openstack/keystone-7ddd868cf8-lc6qt" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.623467 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/482e2d7c-282b-490c-b476-955ddc498707-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.623483 4790 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/482e2d7c-282b-490c-b476-955ddc498707-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.623497 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrx9v\" (UniqueName: \"kubernetes.io/projected/482e2d7c-282b-490c-b476-955ddc498707-kube-api-access-jrx9v\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.623510 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/482e2d7c-282b-490c-b476-955ddc498707-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.623522 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/482e2d7c-282b-490c-b476-955ddc498707-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.623534 4790 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/482e2d7c-282b-490c-b476-955ddc498707-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.724996 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-config-data\") pod \"keystone-7ddd868cf8-lc6qt\" (UID: \"336619dc-13bf-4168-a58b-b5c724759890\") " pod="openstack/keystone-7ddd868cf8-lc6qt" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.725067 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dv4jm\" (UniqueName: \"kubernetes.io/projected/336619dc-13bf-4168-a58b-b5c724759890-kube-api-access-dv4jm\") pod \"keystone-7ddd868cf8-lc6qt\" (UID: \"336619dc-13bf-4168-a58b-b5c724759890\") " pod="openstack/keystone-7ddd868cf8-lc6qt" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.725232 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-credential-keys\") pod \"keystone-7ddd868cf8-lc6qt\" (UID: \"336619dc-13bf-4168-a58b-b5c724759890\") " pod="openstack/keystone-7ddd868cf8-lc6qt" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.725287 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-combined-ca-bundle\") pod \"keystone-7ddd868cf8-lc6qt\" (UID: \"336619dc-13bf-4168-a58b-b5c724759890\") " pod="openstack/keystone-7ddd868cf8-lc6qt" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.725304 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-fernet-keys\") pod \"keystone-7ddd868cf8-lc6qt\" (UID: \"336619dc-13bf-4168-a58b-b5c724759890\") " pod="openstack/keystone-7ddd868cf8-lc6qt" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.725324 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-scripts\") pod \"keystone-7ddd868cf8-lc6qt\" (UID: \"336619dc-13bf-4168-a58b-b5c724759890\") " pod="openstack/keystone-7ddd868cf8-lc6qt" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.725351 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-internal-tls-certs\") pod \"keystone-7ddd868cf8-lc6qt\" (UID: \"336619dc-13bf-4168-a58b-b5c724759890\") " pod="openstack/keystone-7ddd868cf8-lc6qt" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.725394 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-public-tls-certs\") pod \"keystone-7ddd868cf8-lc6qt\" (UID: \"336619dc-13bf-4168-a58b-b5c724759890\") " pod="openstack/keystone-7ddd868cf8-lc6qt" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.729682 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-combined-ca-bundle\") pod \"keystone-7ddd868cf8-lc6qt\" (UID: \"336619dc-13bf-4168-a58b-b5c724759890\") " pod="openstack/keystone-7ddd868cf8-lc6qt" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.730023 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-public-tls-certs\") pod \"keystone-7ddd868cf8-lc6qt\" (UID: \"336619dc-13bf-4168-a58b-b5c724759890\") " pod="openstack/keystone-7ddd868cf8-lc6qt" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.730521 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-fernet-keys\") pod \"keystone-7ddd868cf8-lc6qt\" (UID: \"336619dc-13bf-4168-a58b-b5c724759890\") " pod="openstack/keystone-7ddd868cf8-lc6qt" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.730869 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-config-data\") pod \"keystone-7ddd868cf8-lc6qt\" (UID: \"336619dc-13bf-4168-a58b-b5c724759890\") " pod="openstack/keystone-7ddd868cf8-lc6qt" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.737915 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-scripts\") pod \"keystone-7ddd868cf8-lc6qt\" (UID: \"336619dc-13bf-4168-a58b-b5c724759890\") " pod="openstack/keystone-7ddd868cf8-lc6qt" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.737999 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-credential-keys\") pod \"keystone-7ddd868cf8-lc6qt\" (UID: \"336619dc-13bf-4168-a58b-b5c724759890\") " pod="openstack/keystone-7ddd868cf8-lc6qt" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.739132 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-q4sb2" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.743757 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dv4jm\" (UniqueName: \"kubernetes.io/projected/336619dc-13bf-4168-a58b-b5c724759890-kube-api-access-dv4jm\") pod \"keystone-7ddd868cf8-lc6qt\" (UID: \"336619dc-13bf-4168-a58b-b5c724759890\") " pod="openstack/keystone-7ddd868cf8-lc6qt" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.744499 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-internal-tls-certs\") pod \"keystone-7ddd868cf8-lc6qt\" (UID: \"336619dc-13bf-4168-a58b-b5c724759890\") " pod="openstack/keystone-7ddd868cf8-lc6qt" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.832924 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7ddd868cf8-lc6qt" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.930206 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f5d721b-e52b-490d-89f3-4a5edaf403ec-logs\") pod \"6f5d721b-e52b-490d-89f3-4a5edaf403ec\" (UID: \"6f5d721b-e52b-490d-89f3-4a5edaf403ec\") " Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.930503 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f5d721b-e52b-490d-89f3-4a5edaf403ec-combined-ca-bundle\") pod \"6f5d721b-e52b-490d-89f3-4a5edaf403ec\" (UID: \"6f5d721b-e52b-490d-89f3-4a5edaf403ec\") " Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.930586 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfqw2\" (UniqueName: \"kubernetes.io/projected/6f5d721b-e52b-490d-89f3-4a5edaf403ec-kube-api-access-kfqw2\") pod \"6f5d721b-e52b-490d-89f3-4a5edaf403ec\" (UID: \"6f5d721b-e52b-490d-89f3-4a5edaf403ec\") " Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.930660 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f5d721b-e52b-490d-89f3-4a5edaf403ec-scripts\") pod \"6f5d721b-e52b-490d-89f3-4a5edaf403ec\" (UID: \"6f5d721b-e52b-490d-89f3-4a5edaf403ec\") " Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.930686 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f5d721b-e52b-490d-89f3-4a5edaf403ec-config-data\") pod \"6f5d721b-e52b-490d-89f3-4a5edaf403ec\" (UID: \"6f5d721b-e52b-490d-89f3-4a5edaf403ec\") " Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.932259 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f5d721b-e52b-490d-89f3-4a5edaf403ec-logs" (OuterVolumeSpecName: "logs") pod "6f5d721b-e52b-490d-89f3-4a5edaf403ec" (UID: "6f5d721b-e52b-490d-89f3-4a5edaf403ec"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.934614 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f5d721b-e52b-490d-89f3-4a5edaf403ec-kube-api-access-kfqw2" (OuterVolumeSpecName: "kube-api-access-kfqw2") pod "6f5d721b-e52b-490d-89f3-4a5edaf403ec" (UID: "6f5d721b-e52b-490d-89f3-4a5edaf403ec"). InnerVolumeSpecName "kube-api-access-kfqw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.935304 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f5d721b-e52b-490d-89f3-4a5edaf403ec-scripts" (OuterVolumeSpecName: "scripts") pod "6f5d721b-e52b-490d-89f3-4a5edaf403ec" (UID: "6f5d721b-e52b-490d-89f3-4a5edaf403ec"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.959078 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f5d721b-e52b-490d-89f3-4a5edaf403ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6f5d721b-e52b-490d-89f3-4a5edaf403ec" (UID: "6f5d721b-e52b-490d-89f3-4a5edaf403ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:49 crc kubenswrapper[4790]: I1124 13:33:49.960864 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f5d721b-e52b-490d-89f3-4a5edaf403ec-config-data" (OuterVolumeSpecName: "config-data") pod "6f5d721b-e52b-490d-89f3-4a5edaf403ec" (UID: "6f5d721b-e52b-490d-89f3-4a5edaf403ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.034244 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfqw2\" (UniqueName: \"kubernetes.io/projected/6f5d721b-e52b-490d-89f3-4a5edaf403ec-kube-api-access-kfqw2\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.034273 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f5d721b-e52b-490d-89f3-4a5edaf403ec-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.034281 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f5d721b-e52b-490d-89f3-4a5edaf403ec-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.034291 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f5d721b-e52b-490d-89f3-4a5edaf403ec-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.034300 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f5d721b-e52b-490d-89f3-4a5edaf403ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:50 crc kubenswrapper[4790]: W1124 13:33:50.260192 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod336619dc_13bf_4168_a58b_b5c724759890.slice/crio-1a431a97315bb5dee273ee36c96be5d5d31b8ac7dddb0b81940b402108384302 WatchSource:0}: Error finding container 1a431a97315bb5dee273ee36c96be5d5d31b8ac7dddb0b81940b402108384302: Status 404 returned error can't find the container with id 1a431a97315bb5dee273ee36c96be5d5d31b8ac7dddb0b81940b402108384302 Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.265573 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7ddd868cf8-lc6qt"] Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.446147 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7ddd868cf8-lc6qt" event={"ID":"336619dc-13bf-4168-a58b-b5c724759890","Type":"ContainerStarted","Data":"1a431a97315bb5dee273ee36c96be5d5d31b8ac7dddb0b81940b402108384302"} Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.447875 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"69617248-5d4b-4027-a2a9-e972278e9d5e","Type":"ContainerStarted","Data":"1783160f79b92abc9fcc7d3d325d99cf4fd66d34a919d93503b4b9eb6f006447"} Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.451169 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-q4sb2" event={"ID":"6f5d721b-e52b-490d-89f3-4a5edaf403ec","Type":"ContainerDied","Data":"0fdb94317be2e50fbd9a8f04489ec9a294110f094caaadb913457c5f4a6db87a"} Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.451194 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0fdb94317be2e50fbd9a8f04489ec9a294110f094caaadb913457c5f4a6db87a" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.451203 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-q4sb2" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.462650 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f","Type":"ContainerStarted","Data":"dc393809fb6808c6307e7e2fc40a54ca96fae2a33dd3a809b34d8dc7c75ce59b"} Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.473546 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.47352856 podStartE2EDuration="6.47352856s" podCreationTimestamp="2025-11-24 13:33:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:33:50.473492429 +0000 UTC m=+1278.853386101" watchObservedRunningTime="2025-11-24 13:33:50.47352856 +0000 UTC m=+1278.853422222" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.576754 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7957b6846-f8n7f"] Nov 24 13:33:50 crc kubenswrapper[4790]: E1124 13:33:50.577726 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f5d721b-e52b-490d-89f3-4a5edaf403ec" containerName="placement-db-sync" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.577752 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f5d721b-e52b-490d-89f3-4a5edaf403ec" containerName="placement-db-sync" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.578002 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f5d721b-e52b-490d-89f3-4a5edaf403ec" containerName="placement-db-sync" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.579858 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7957b6846-f8n7f" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.588291 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.588533 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.588583 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-ljf7m" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.588684 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.588723 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.605025 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7957b6846-f8n7f"] Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.696995 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-p5msf" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.747306 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b507c50-0776-4d50-ad41-10f26f25fd5f-logs\") pod \"placement-7957b6846-f8n7f\" (UID: \"1b507c50-0776-4d50-ad41-10f26f25fd5f\") " pod="openstack/placement-7957b6846-f8n7f" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.747384 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b507c50-0776-4d50-ad41-10f26f25fd5f-internal-tls-certs\") pod \"placement-7957b6846-f8n7f\" (UID: \"1b507c50-0776-4d50-ad41-10f26f25fd5f\") " pod="openstack/placement-7957b6846-f8n7f" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.747436 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b507c50-0776-4d50-ad41-10f26f25fd5f-config-data\") pod \"placement-7957b6846-f8n7f\" (UID: \"1b507c50-0776-4d50-ad41-10f26f25fd5f\") " pod="openstack/placement-7957b6846-f8n7f" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.747461 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b507c50-0776-4d50-ad41-10f26f25fd5f-scripts\") pod \"placement-7957b6846-f8n7f\" (UID: \"1b507c50-0776-4d50-ad41-10f26f25fd5f\") " pod="openstack/placement-7957b6846-f8n7f" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.747508 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b507c50-0776-4d50-ad41-10f26f25fd5f-combined-ca-bundle\") pod \"placement-7957b6846-f8n7f\" (UID: \"1b507c50-0776-4d50-ad41-10f26f25fd5f\") " pod="openstack/placement-7957b6846-f8n7f" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.747528 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b507c50-0776-4d50-ad41-10f26f25fd5f-public-tls-certs\") pod \"placement-7957b6846-f8n7f\" (UID: \"1b507c50-0776-4d50-ad41-10f26f25fd5f\") " pod="openstack/placement-7957b6846-f8n7f" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.747556 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lr2bq\" (UniqueName: \"kubernetes.io/projected/1b507c50-0776-4d50-ad41-10f26f25fd5f-kube-api-access-lr2bq\") pod \"placement-7957b6846-f8n7f\" (UID: \"1b507c50-0776-4d50-ad41-10f26f25fd5f\") " pod="openstack/placement-7957b6846-f8n7f" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.849133 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/84a50ea0-c597-42e2-b4cd-7e628a517ddf-config\") pod \"84a50ea0-c597-42e2-b4cd-7e628a517ddf\" (UID: \"84a50ea0-c597-42e2-b4cd-7e628a517ddf\") " Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.849228 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-422jt\" (UniqueName: \"kubernetes.io/projected/84a50ea0-c597-42e2-b4cd-7e628a517ddf-kube-api-access-422jt\") pod \"84a50ea0-c597-42e2-b4cd-7e628a517ddf\" (UID: \"84a50ea0-c597-42e2-b4cd-7e628a517ddf\") " Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.849277 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84a50ea0-c597-42e2-b4cd-7e628a517ddf-combined-ca-bundle\") pod \"84a50ea0-c597-42e2-b4cd-7e628a517ddf\" (UID: \"84a50ea0-c597-42e2-b4cd-7e628a517ddf\") " Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.849552 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b507c50-0776-4d50-ad41-10f26f25fd5f-combined-ca-bundle\") pod \"placement-7957b6846-f8n7f\" (UID: \"1b507c50-0776-4d50-ad41-10f26f25fd5f\") " pod="openstack/placement-7957b6846-f8n7f" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.849586 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b507c50-0776-4d50-ad41-10f26f25fd5f-public-tls-certs\") pod \"placement-7957b6846-f8n7f\" (UID: \"1b507c50-0776-4d50-ad41-10f26f25fd5f\") " pod="openstack/placement-7957b6846-f8n7f" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.849622 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lr2bq\" (UniqueName: \"kubernetes.io/projected/1b507c50-0776-4d50-ad41-10f26f25fd5f-kube-api-access-lr2bq\") pod \"placement-7957b6846-f8n7f\" (UID: \"1b507c50-0776-4d50-ad41-10f26f25fd5f\") " pod="openstack/placement-7957b6846-f8n7f" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.849695 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b507c50-0776-4d50-ad41-10f26f25fd5f-logs\") pod \"placement-7957b6846-f8n7f\" (UID: \"1b507c50-0776-4d50-ad41-10f26f25fd5f\") " pod="openstack/placement-7957b6846-f8n7f" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.849760 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b507c50-0776-4d50-ad41-10f26f25fd5f-internal-tls-certs\") pod \"placement-7957b6846-f8n7f\" (UID: \"1b507c50-0776-4d50-ad41-10f26f25fd5f\") " pod="openstack/placement-7957b6846-f8n7f" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.849801 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b507c50-0776-4d50-ad41-10f26f25fd5f-config-data\") pod \"placement-7957b6846-f8n7f\" (UID: \"1b507c50-0776-4d50-ad41-10f26f25fd5f\") " pod="openstack/placement-7957b6846-f8n7f" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.849829 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b507c50-0776-4d50-ad41-10f26f25fd5f-scripts\") pod \"placement-7957b6846-f8n7f\" (UID: \"1b507c50-0776-4d50-ad41-10f26f25fd5f\") " pod="openstack/placement-7957b6846-f8n7f" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.850447 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b507c50-0776-4d50-ad41-10f26f25fd5f-logs\") pod \"placement-7957b6846-f8n7f\" (UID: \"1b507c50-0776-4d50-ad41-10f26f25fd5f\") " pod="openstack/placement-7957b6846-f8n7f" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.854031 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b507c50-0776-4d50-ad41-10f26f25fd5f-internal-tls-certs\") pod \"placement-7957b6846-f8n7f\" (UID: \"1b507c50-0776-4d50-ad41-10f26f25fd5f\") " pod="openstack/placement-7957b6846-f8n7f" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.854396 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b507c50-0776-4d50-ad41-10f26f25fd5f-config-data\") pod \"placement-7957b6846-f8n7f\" (UID: \"1b507c50-0776-4d50-ad41-10f26f25fd5f\") " pod="openstack/placement-7957b6846-f8n7f" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.854517 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b507c50-0776-4d50-ad41-10f26f25fd5f-combined-ca-bundle\") pod \"placement-7957b6846-f8n7f\" (UID: \"1b507c50-0776-4d50-ad41-10f26f25fd5f\") " pod="openstack/placement-7957b6846-f8n7f" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.855808 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84a50ea0-c597-42e2-b4cd-7e628a517ddf-kube-api-access-422jt" (OuterVolumeSpecName: "kube-api-access-422jt") pod "84a50ea0-c597-42e2-b4cd-7e628a517ddf" (UID: "84a50ea0-c597-42e2-b4cd-7e628a517ddf"). InnerVolumeSpecName "kube-api-access-422jt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.861586 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b507c50-0776-4d50-ad41-10f26f25fd5f-scripts\") pod \"placement-7957b6846-f8n7f\" (UID: \"1b507c50-0776-4d50-ad41-10f26f25fd5f\") " pod="openstack/placement-7957b6846-f8n7f" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.866353 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b507c50-0776-4d50-ad41-10f26f25fd5f-public-tls-certs\") pod \"placement-7957b6846-f8n7f\" (UID: \"1b507c50-0776-4d50-ad41-10f26f25fd5f\") " pod="openstack/placement-7957b6846-f8n7f" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.869366 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lr2bq\" (UniqueName: \"kubernetes.io/projected/1b507c50-0776-4d50-ad41-10f26f25fd5f-kube-api-access-lr2bq\") pod \"placement-7957b6846-f8n7f\" (UID: \"1b507c50-0776-4d50-ad41-10f26f25fd5f\") " pod="openstack/placement-7957b6846-f8n7f" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.875321 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84a50ea0-c597-42e2-b4cd-7e628a517ddf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "84a50ea0-c597-42e2-b4cd-7e628a517ddf" (UID: "84a50ea0-c597-42e2-b4cd-7e628a517ddf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.878863 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84a50ea0-c597-42e2-b4cd-7e628a517ddf-config" (OuterVolumeSpecName: "config") pod "84a50ea0-c597-42e2-b4cd-7e628a517ddf" (UID: "84a50ea0-c597-42e2-b4cd-7e628a517ddf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.900866 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7957b6846-f8n7f" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.950984 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-422jt\" (UniqueName: \"kubernetes.io/projected/84a50ea0-c597-42e2-b4cd-7e628a517ddf-kube-api-access-422jt\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.951022 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84a50ea0-c597-42e2-b4cd-7e628a517ddf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:50 crc kubenswrapper[4790]: I1124 13:33:50.951032 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/84a50ea0-c597-42e2-b4cd-7e628a517ddf-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.146703 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7957b6846-f8n7f"] Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.494977 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7ddd868cf8-lc6qt" event={"ID":"336619dc-13bf-4168-a58b-b5c724759890","Type":"ContainerStarted","Data":"07ac1c04efe6c3d8ada271427bad943fb80399185792ecba62e72b13d34fc58d"} Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.495404 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-7ddd868cf8-lc6qt" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.497656 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7957b6846-f8n7f" event={"ID":"1b507c50-0776-4d50-ad41-10f26f25fd5f","Type":"ContainerStarted","Data":"96fb30b3c69ce48bcf01e52ed956c73a35d60273c9db96fac4575e7dc8dd1004"} Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.497686 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7957b6846-f8n7f" event={"ID":"1b507c50-0776-4d50-ad41-10f26f25fd5f","Type":"ContainerStarted","Data":"d67870725bbb86c25d9c408bb9a7732e56318396fd04d3647257357971eef254"} Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.501646 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-p5msf" event={"ID":"84a50ea0-c597-42e2-b4cd-7e628a517ddf","Type":"ContainerDied","Data":"374c89fcbac95affb0b9cf6110fb31cf2a5aceeebec36e8d8a100419b679018e"} Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.501747 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="374c89fcbac95affb0b9cf6110fb31cf2a5aceeebec36e8d8a100419b679018e" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.501790 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-p5msf" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.521904 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7ddd868cf8-lc6qt" podStartSLOduration=2.521867416 podStartE2EDuration="2.521867416s" podCreationTimestamp="2025-11-24 13:33:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:33:51.514024803 +0000 UTC m=+1279.893918495" watchObservedRunningTime="2025-11-24 13:33:51.521867416 +0000 UTC m=+1279.901761078" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.659178 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-z76jp"] Nov 24 13:33:51 crc kubenswrapper[4790]: E1124 13:33:51.661992 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84a50ea0-c597-42e2-b4cd-7e628a517ddf" containerName="neutron-db-sync" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.662022 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="84a50ea0-c597-42e2-b4cd-7e628a517ddf" containerName="neutron-db-sync" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.662232 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="84a50ea0-c597-42e2-b4cd-7e628a517ddf" containerName="neutron-db-sync" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.663318 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-z76jp" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.691262 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-z76jp"] Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.765154 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7294e855-6ce3-4807-9871-2fd2efbcfd0f-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-z76jp\" (UID: \"7294e855-6ce3-4807-9871-2fd2efbcfd0f\") " pod="openstack/dnsmasq-dns-6b7b667979-z76jp" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.765208 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7294e855-6ce3-4807-9871-2fd2efbcfd0f-dns-svc\") pod \"dnsmasq-dns-6b7b667979-z76jp\" (UID: \"7294e855-6ce3-4807-9871-2fd2efbcfd0f\") " pod="openstack/dnsmasq-dns-6b7b667979-z76jp" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.765238 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7294e855-6ce3-4807-9871-2fd2efbcfd0f-config\") pod \"dnsmasq-dns-6b7b667979-z76jp\" (UID: \"7294e855-6ce3-4807-9871-2fd2efbcfd0f\") " pod="openstack/dnsmasq-dns-6b7b667979-z76jp" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.765299 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z58nr\" (UniqueName: \"kubernetes.io/projected/7294e855-6ce3-4807-9871-2fd2efbcfd0f-kube-api-access-z58nr\") pod \"dnsmasq-dns-6b7b667979-z76jp\" (UID: \"7294e855-6ce3-4807-9871-2fd2efbcfd0f\") " pod="openstack/dnsmasq-dns-6b7b667979-z76jp" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.765327 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7294e855-6ce3-4807-9871-2fd2efbcfd0f-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-z76jp\" (UID: \"7294e855-6ce3-4807-9871-2fd2efbcfd0f\") " pod="openstack/dnsmasq-dns-6b7b667979-z76jp" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.765355 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7294e855-6ce3-4807-9871-2fd2efbcfd0f-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-z76jp\" (UID: \"7294e855-6ce3-4807-9871-2fd2efbcfd0f\") " pod="openstack/dnsmasq-dns-6b7b667979-z76jp" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.805547 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-69f849cd44-8v55p"] Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.807224 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69f849cd44-8v55p" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.809603 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-pws5r" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.810109 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.811781 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.811983 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.814587 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-69f849cd44-8v55p"] Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.868559 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7294e855-6ce3-4807-9871-2fd2efbcfd0f-config\") pod \"dnsmasq-dns-6b7b667979-z76jp\" (UID: \"7294e855-6ce3-4807-9871-2fd2efbcfd0f\") " pod="openstack/dnsmasq-dns-6b7b667979-z76jp" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.868604 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/001f4ed0-f6e6-440e-a878-2bca6982ea47-config\") pod \"neutron-69f849cd44-8v55p\" (UID: \"001f4ed0-f6e6-440e-a878-2bca6982ea47\") " pod="openstack/neutron-69f849cd44-8v55p" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.868626 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/001f4ed0-f6e6-440e-a878-2bca6982ea47-ovndb-tls-certs\") pod \"neutron-69f849cd44-8v55p\" (UID: \"001f4ed0-f6e6-440e-a878-2bca6982ea47\") " pod="openstack/neutron-69f849cd44-8v55p" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.869037 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/001f4ed0-f6e6-440e-a878-2bca6982ea47-combined-ca-bundle\") pod \"neutron-69f849cd44-8v55p\" (UID: \"001f4ed0-f6e6-440e-a878-2bca6982ea47\") " pod="openstack/neutron-69f849cd44-8v55p" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.869245 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z58nr\" (UniqueName: \"kubernetes.io/projected/7294e855-6ce3-4807-9871-2fd2efbcfd0f-kube-api-access-z58nr\") pod \"dnsmasq-dns-6b7b667979-z76jp\" (UID: \"7294e855-6ce3-4807-9871-2fd2efbcfd0f\") " pod="openstack/dnsmasq-dns-6b7b667979-z76jp" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.869642 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7294e855-6ce3-4807-9871-2fd2efbcfd0f-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-z76jp\" (UID: \"7294e855-6ce3-4807-9871-2fd2efbcfd0f\") " pod="openstack/dnsmasq-dns-6b7b667979-z76jp" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.869820 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7294e855-6ce3-4807-9871-2fd2efbcfd0f-config\") pod \"dnsmasq-dns-6b7b667979-z76jp\" (UID: \"7294e855-6ce3-4807-9871-2fd2efbcfd0f\") " pod="openstack/dnsmasq-dns-6b7b667979-z76jp" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.870415 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7294e855-6ce3-4807-9871-2fd2efbcfd0f-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-z76jp\" (UID: \"7294e855-6ce3-4807-9871-2fd2efbcfd0f\") " pod="openstack/dnsmasq-dns-6b7b667979-z76jp" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.870873 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7294e855-6ce3-4807-9871-2fd2efbcfd0f-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-z76jp\" (UID: \"7294e855-6ce3-4807-9871-2fd2efbcfd0f\") " pod="openstack/dnsmasq-dns-6b7b667979-z76jp" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.871561 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7294e855-6ce3-4807-9871-2fd2efbcfd0f-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-z76jp\" (UID: \"7294e855-6ce3-4807-9871-2fd2efbcfd0f\") " pod="openstack/dnsmasq-dns-6b7b667979-z76jp" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.871757 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/001f4ed0-f6e6-440e-a878-2bca6982ea47-httpd-config\") pod \"neutron-69f849cd44-8v55p\" (UID: \"001f4ed0-f6e6-440e-a878-2bca6982ea47\") " pod="openstack/neutron-69f849cd44-8v55p" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.871914 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7294e855-6ce3-4807-9871-2fd2efbcfd0f-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-z76jp\" (UID: \"7294e855-6ce3-4807-9871-2fd2efbcfd0f\") " pod="openstack/dnsmasq-dns-6b7b667979-z76jp" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.872750 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjlcb\" (UniqueName: \"kubernetes.io/projected/001f4ed0-f6e6-440e-a878-2bca6982ea47-kube-api-access-kjlcb\") pod \"neutron-69f849cd44-8v55p\" (UID: \"001f4ed0-f6e6-440e-a878-2bca6982ea47\") " pod="openstack/neutron-69f849cd44-8v55p" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.872699 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7294e855-6ce3-4807-9871-2fd2efbcfd0f-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-z76jp\" (UID: \"7294e855-6ce3-4807-9871-2fd2efbcfd0f\") " pod="openstack/dnsmasq-dns-6b7b667979-z76jp" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.873003 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7294e855-6ce3-4807-9871-2fd2efbcfd0f-dns-svc\") pod \"dnsmasq-dns-6b7b667979-z76jp\" (UID: \"7294e855-6ce3-4807-9871-2fd2efbcfd0f\") " pod="openstack/dnsmasq-dns-6b7b667979-z76jp" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.873598 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7294e855-6ce3-4807-9871-2fd2efbcfd0f-dns-svc\") pod \"dnsmasq-dns-6b7b667979-z76jp\" (UID: \"7294e855-6ce3-4807-9871-2fd2efbcfd0f\") " pod="openstack/dnsmasq-dns-6b7b667979-z76jp" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.889094 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z58nr\" (UniqueName: \"kubernetes.io/projected/7294e855-6ce3-4807-9871-2fd2efbcfd0f-kube-api-access-z58nr\") pod \"dnsmasq-dns-6b7b667979-z76jp\" (UID: \"7294e855-6ce3-4807-9871-2fd2efbcfd0f\") " pod="openstack/dnsmasq-dns-6b7b667979-z76jp" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.975246 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/001f4ed0-f6e6-440e-a878-2bca6982ea47-httpd-config\") pod \"neutron-69f849cd44-8v55p\" (UID: \"001f4ed0-f6e6-440e-a878-2bca6982ea47\") " pod="openstack/neutron-69f849cd44-8v55p" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.975318 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjlcb\" (UniqueName: \"kubernetes.io/projected/001f4ed0-f6e6-440e-a878-2bca6982ea47-kube-api-access-kjlcb\") pod \"neutron-69f849cd44-8v55p\" (UID: \"001f4ed0-f6e6-440e-a878-2bca6982ea47\") " pod="openstack/neutron-69f849cd44-8v55p" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.975359 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/001f4ed0-f6e6-440e-a878-2bca6982ea47-config\") pod \"neutron-69f849cd44-8v55p\" (UID: \"001f4ed0-f6e6-440e-a878-2bca6982ea47\") " pod="openstack/neutron-69f849cd44-8v55p" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.975393 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/001f4ed0-f6e6-440e-a878-2bca6982ea47-ovndb-tls-certs\") pod \"neutron-69f849cd44-8v55p\" (UID: \"001f4ed0-f6e6-440e-a878-2bca6982ea47\") " pod="openstack/neutron-69f849cd44-8v55p" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.975425 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/001f4ed0-f6e6-440e-a878-2bca6982ea47-combined-ca-bundle\") pod \"neutron-69f849cd44-8v55p\" (UID: \"001f4ed0-f6e6-440e-a878-2bca6982ea47\") " pod="openstack/neutron-69f849cd44-8v55p" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.985453 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/001f4ed0-f6e6-440e-a878-2bca6982ea47-ovndb-tls-certs\") pod \"neutron-69f849cd44-8v55p\" (UID: \"001f4ed0-f6e6-440e-a878-2bca6982ea47\") " pod="openstack/neutron-69f849cd44-8v55p" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.986189 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/001f4ed0-f6e6-440e-a878-2bca6982ea47-httpd-config\") pod \"neutron-69f849cd44-8v55p\" (UID: \"001f4ed0-f6e6-440e-a878-2bca6982ea47\") " pod="openstack/neutron-69f849cd44-8v55p" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.986762 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/001f4ed0-f6e6-440e-a878-2bca6982ea47-combined-ca-bundle\") pod \"neutron-69f849cd44-8v55p\" (UID: \"001f4ed0-f6e6-440e-a878-2bca6982ea47\") " pod="openstack/neutron-69f849cd44-8v55p" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.987730 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/001f4ed0-f6e6-440e-a878-2bca6982ea47-config\") pod \"neutron-69f849cd44-8v55p\" (UID: \"001f4ed0-f6e6-440e-a878-2bca6982ea47\") " pod="openstack/neutron-69f849cd44-8v55p" Nov 24 13:33:51 crc kubenswrapper[4790]: I1124 13:33:51.993399 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjlcb\" (UniqueName: \"kubernetes.io/projected/001f4ed0-f6e6-440e-a878-2bca6982ea47-kube-api-access-kjlcb\") pod \"neutron-69f849cd44-8v55p\" (UID: \"001f4ed0-f6e6-440e-a878-2bca6982ea47\") " pod="openstack/neutron-69f849cd44-8v55p" Nov 24 13:33:52 crc kubenswrapper[4790]: I1124 13:33:52.004838 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-z76jp" Nov 24 13:33:52 crc kubenswrapper[4790]: I1124 13:33:52.131160 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69f849cd44-8v55p" Nov 24 13:33:52 crc kubenswrapper[4790]: I1124 13:33:52.292311 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-z76jp"] Nov 24 13:33:52 crc kubenswrapper[4790]: I1124 13:33:52.510736 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-69f849cd44-8v55p"] Nov 24 13:33:52 crc kubenswrapper[4790]: I1124 13:33:52.514767 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7957b6846-f8n7f" event={"ID":"1b507c50-0776-4d50-ad41-10f26f25fd5f","Type":"ContainerStarted","Data":"2f816c8e5f36c0c446e29e140b1f74aeb5df0880612029daa8ef51dfaee75bee"} Nov 24 13:33:52 crc kubenswrapper[4790]: I1124 13:33:52.514900 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7957b6846-f8n7f" Nov 24 13:33:52 crc kubenswrapper[4790]: I1124 13:33:52.516036 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-z76jp" event={"ID":"7294e855-6ce3-4807-9871-2fd2efbcfd0f","Type":"ContainerStarted","Data":"382b2a2e0f19b90b7f32fc8330f9439f63f759dc23f5e33b3f177f5df34060bc"} Nov 24 13:33:52 crc kubenswrapper[4790]: I1124 13:33:52.563094 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-7957b6846-f8n7f" podStartSLOduration=2.56307217 podStartE2EDuration="2.56307217s" podCreationTimestamp="2025-11-24 13:33:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:33:52.543007329 +0000 UTC m=+1280.922901011" watchObservedRunningTime="2025-11-24 13:33:52.56307217 +0000 UTC m=+1280.942965832" Nov 24 13:33:53 crc kubenswrapper[4790]: I1124 13:33:53.529355 4790 generic.go:334] "Generic (PLEG): container finished" podID="7294e855-6ce3-4807-9871-2fd2efbcfd0f" containerID="70fd1853a5a4f323a22d1dbf5921aaad40f430d3febd19818b17b8ed831a66a4" exitCode=0 Nov 24 13:33:53 crc kubenswrapper[4790]: I1124 13:33:53.529605 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-z76jp" event={"ID":"7294e855-6ce3-4807-9871-2fd2efbcfd0f","Type":"ContainerDied","Data":"70fd1853a5a4f323a22d1dbf5921aaad40f430d3febd19818b17b8ed831a66a4"} Nov 24 13:33:53 crc kubenswrapper[4790]: I1124 13:33:53.534127 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69f849cd44-8v55p" event={"ID":"001f4ed0-f6e6-440e-a878-2bca6982ea47","Type":"ContainerStarted","Data":"f09ab5eeebda745a1632f857ef38ee2cf3134e76bbb87bc46078ee01b4771c6f"} Nov 24 13:33:53 crc kubenswrapper[4790]: I1124 13:33:53.534189 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7957b6846-f8n7f" Nov 24 13:33:53 crc kubenswrapper[4790]: I1124 13:33:53.864742 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-78cd478f4f-bqm94"] Nov 24 13:33:53 crc kubenswrapper[4790]: I1124 13:33:53.867321 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-78cd478f4f-bqm94" Nov 24 13:33:53 crc kubenswrapper[4790]: I1124 13:33:53.869945 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 24 13:33:53 crc kubenswrapper[4790]: I1124 13:33:53.870191 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 24 13:33:53 crc kubenswrapper[4790]: I1124 13:33:53.901281 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-78cd478f4f-bqm94"] Nov 24 13:33:54 crc kubenswrapper[4790]: I1124 13:33:54.019332 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-public-tls-certs\") pod \"neutron-78cd478f4f-bqm94\" (UID: \"4d4e45de-9768-40e9-9c75-6abba8b38559\") " pod="openstack/neutron-78cd478f4f-bqm94" Nov 24 13:33:54 crc kubenswrapper[4790]: I1124 13:33:54.019397 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-httpd-config\") pod \"neutron-78cd478f4f-bqm94\" (UID: \"4d4e45de-9768-40e9-9c75-6abba8b38559\") " pod="openstack/neutron-78cd478f4f-bqm94" Nov 24 13:33:54 crc kubenswrapper[4790]: I1124 13:33:54.019424 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-ovndb-tls-certs\") pod \"neutron-78cd478f4f-bqm94\" (UID: \"4d4e45de-9768-40e9-9c75-6abba8b38559\") " pod="openstack/neutron-78cd478f4f-bqm94" Nov 24 13:33:54 crc kubenswrapper[4790]: I1124 13:33:54.019466 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzgcq\" (UniqueName: \"kubernetes.io/projected/4d4e45de-9768-40e9-9c75-6abba8b38559-kube-api-access-zzgcq\") pod \"neutron-78cd478f4f-bqm94\" (UID: \"4d4e45de-9768-40e9-9c75-6abba8b38559\") " pod="openstack/neutron-78cd478f4f-bqm94" Nov 24 13:33:54 crc kubenswrapper[4790]: I1124 13:33:54.019693 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-config\") pod \"neutron-78cd478f4f-bqm94\" (UID: \"4d4e45de-9768-40e9-9c75-6abba8b38559\") " pod="openstack/neutron-78cd478f4f-bqm94" Nov 24 13:33:54 crc kubenswrapper[4790]: I1124 13:33:54.019933 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-internal-tls-certs\") pod \"neutron-78cd478f4f-bqm94\" (UID: \"4d4e45de-9768-40e9-9c75-6abba8b38559\") " pod="openstack/neutron-78cd478f4f-bqm94" Nov 24 13:33:54 crc kubenswrapper[4790]: I1124 13:33:54.020013 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-combined-ca-bundle\") pod \"neutron-78cd478f4f-bqm94\" (UID: \"4d4e45de-9768-40e9-9c75-6abba8b38559\") " pod="openstack/neutron-78cd478f4f-bqm94" Nov 24 13:33:54 crc kubenswrapper[4790]: I1124 13:33:54.121858 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-internal-tls-certs\") pod \"neutron-78cd478f4f-bqm94\" (UID: \"4d4e45de-9768-40e9-9c75-6abba8b38559\") " pod="openstack/neutron-78cd478f4f-bqm94" Nov 24 13:33:54 crc kubenswrapper[4790]: I1124 13:33:54.122225 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-combined-ca-bundle\") pod \"neutron-78cd478f4f-bqm94\" (UID: \"4d4e45de-9768-40e9-9c75-6abba8b38559\") " pod="openstack/neutron-78cd478f4f-bqm94" Nov 24 13:33:54 crc kubenswrapper[4790]: I1124 13:33:54.122290 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-public-tls-certs\") pod \"neutron-78cd478f4f-bqm94\" (UID: \"4d4e45de-9768-40e9-9c75-6abba8b38559\") " pod="openstack/neutron-78cd478f4f-bqm94" Nov 24 13:33:54 crc kubenswrapper[4790]: I1124 13:33:54.122329 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-httpd-config\") pod \"neutron-78cd478f4f-bqm94\" (UID: \"4d4e45de-9768-40e9-9c75-6abba8b38559\") " pod="openstack/neutron-78cd478f4f-bqm94" Nov 24 13:33:54 crc kubenswrapper[4790]: I1124 13:33:54.122360 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-ovndb-tls-certs\") pod \"neutron-78cd478f4f-bqm94\" (UID: \"4d4e45de-9768-40e9-9c75-6abba8b38559\") " pod="openstack/neutron-78cd478f4f-bqm94" Nov 24 13:33:54 crc kubenswrapper[4790]: I1124 13:33:54.122413 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzgcq\" (UniqueName: \"kubernetes.io/projected/4d4e45de-9768-40e9-9c75-6abba8b38559-kube-api-access-zzgcq\") pod \"neutron-78cd478f4f-bqm94\" (UID: \"4d4e45de-9768-40e9-9c75-6abba8b38559\") " pod="openstack/neutron-78cd478f4f-bqm94" Nov 24 13:33:54 crc kubenswrapper[4790]: I1124 13:33:54.122467 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-config\") pod \"neutron-78cd478f4f-bqm94\" (UID: \"4d4e45de-9768-40e9-9c75-6abba8b38559\") " pod="openstack/neutron-78cd478f4f-bqm94" Nov 24 13:33:54 crc kubenswrapper[4790]: I1124 13:33:54.130593 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-public-tls-certs\") pod \"neutron-78cd478f4f-bqm94\" (UID: \"4d4e45de-9768-40e9-9c75-6abba8b38559\") " pod="openstack/neutron-78cd478f4f-bqm94" Nov 24 13:33:54 crc kubenswrapper[4790]: I1124 13:33:54.130748 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-ovndb-tls-certs\") pod \"neutron-78cd478f4f-bqm94\" (UID: \"4d4e45de-9768-40e9-9c75-6abba8b38559\") " pod="openstack/neutron-78cd478f4f-bqm94" Nov 24 13:33:54 crc kubenswrapper[4790]: I1124 13:33:54.131655 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-internal-tls-certs\") pod \"neutron-78cd478f4f-bqm94\" (UID: \"4d4e45de-9768-40e9-9c75-6abba8b38559\") " pod="openstack/neutron-78cd478f4f-bqm94" Nov 24 13:33:54 crc kubenswrapper[4790]: I1124 13:33:54.131759 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-httpd-config\") pod \"neutron-78cd478f4f-bqm94\" (UID: \"4d4e45de-9768-40e9-9c75-6abba8b38559\") " pod="openstack/neutron-78cd478f4f-bqm94" Nov 24 13:33:54 crc kubenswrapper[4790]: I1124 13:33:54.132421 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-config\") pod \"neutron-78cd478f4f-bqm94\" (UID: \"4d4e45de-9768-40e9-9c75-6abba8b38559\") " pod="openstack/neutron-78cd478f4f-bqm94" Nov 24 13:33:54 crc kubenswrapper[4790]: I1124 13:33:54.136660 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-combined-ca-bundle\") pod \"neutron-78cd478f4f-bqm94\" (UID: \"4d4e45de-9768-40e9-9c75-6abba8b38559\") " pod="openstack/neutron-78cd478f4f-bqm94" Nov 24 13:33:54 crc kubenswrapper[4790]: I1124 13:33:54.142637 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzgcq\" (UniqueName: \"kubernetes.io/projected/4d4e45de-9768-40e9-9c75-6abba8b38559-kube-api-access-zzgcq\") pod \"neutron-78cd478f4f-bqm94\" (UID: \"4d4e45de-9768-40e9-9c75-6abba8b38559\") " pod="openstack/neutron-78cd478f4f-bqm94" Nov 24 13:33:54 crc kubenswrapper[4790]: I1124 13:33:54.200272 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-78cd478f4f-bqm94" Nov 24 13:33:54 crc kubenswrapper[4790]: I1124 13:33:54.559728 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-z76jp" event={"ID":"7294e855-6ce3-4807-9871-2fd2efbcfd0f","Type":"ContainerStarted","Data":"a66746afa43133bf3b2cc2c02574b9bde384aeefd099c3f8982e8a6cf882206d"} Nov 24 13:33:54 crc kubenswrapper[4790]: I1124 13:33:54.559796 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7b667979-z76jp" Nov 24 13:33:54 crc kubenswrapper[4790]: I1124 13:33:54.569141 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69f849cd44-8v55p" event={"ID":"001f4ed0-f6e6-440e-a878-2bca6982ea47","Type":"ContainerStarted","Data":"db4596a34402f208f174677329719e056fd084851d4e5a37a88604237aa9cfd2"} Nov 24 13:33:54 crc kubenswrapper[4790]: I1124 13:33:54.592986 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7b667979-z76jp" podStartSLOduration=3.592923643 podStartE2EDuration="3.592923643s" podCreationTimestamp="2025-11-24 13:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:33:54.590236616 +0000 UTC m=+1282.970130278" watchObservedRunningTime="2025-11-24 13:33:54.592923643 +0000 UTC m=+1282.972817305" Nov 24 13:33:55 crc kubenswrapper[4790]: I1124 13:33:55.029065 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 13:33:55 crc kubenswrapper[4790]: I1124 13:33:55.029132 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 13:33:55 crc kubenswrapper[4790]: I1124 13:33:55.072175 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 13:33:55 crc kubenswrapper[4790]: I1124 13:33:55.077611 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 13:33:55 crc kubenswrapper[4790]: I1124 13:33:55.578660 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 13:33:55 crc kubenswrapper[4790]: I1124 13:33:55.578698 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 13:33:57 crc kubenswrapper[4790]: I1124 13:33:57.565175 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 13:33:57 crc kubenswrapper[4790]: I1124 13:33:57.567058 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 13:33:59 crc kubenswrapper[4790]: E1124 13:33:59.766076 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="4d574c9c-ecb5-4b8c-b0b3-a058c141c09f" Nov 24 13:33:59 crc kubenswrapper[4790]: I1124 13:33:59.968163 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-78cd478f4f-bqm94"] Nov 24 13:34:00 crc kubenswrapper[4790]: I1124 13:34:00.636733 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4d574c9c-ecb5-4b8c-b0b3-a058c141c09f" containerName="ceilometer-notification-agent" containerID="cri-o://6a117358d00dc47eb45e9087790aa499d845a99631f8d556a818c3dbb36a7a89" gracePeriod=30 Nov 24 13:34:00 crc kubenswrapper[4790]: I1124 13:34:00.636800 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4d574c9c-ecb5-4b8c-b0b3-a058c141c09f" containerName="proxy-httpd" containerID="cri-o://fdb5a55d080b1307b7fd39a3e10de2010b66b33a3318bca4aff4b9b6e6d39845" gracePeriod=30 Nov 24 13:34:00 crc kubenswrapper[4790]: I1124 13:34:00.636867 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4d574c9c-ecb5-4b8c-b0b3-a058c141c09f" containerName="sg-core" containerID="cri-o://dc393809fb6808c6307e7e2fc40a54ca96fae2a33dd3a809b34d8dc7c75ce59b" gracePeriod=30 Nov 24 13:34:00 crc kubenswrapper[4790]: I1124 13:34:00.637127 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f","Type":"ContainerStarted","Data":"fdb5a55d080b1307b7fd39a3e10de2010b66b33a3318bca4aff4b9b6e6d39845"} Nov 24 13:34:00 crc kubenswrapper[4790]: I1124 13:34:00.637170 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 13:34:00 crc kubenswrapper[4790]: I1124 13:34:00.649240 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69f849cd44-8v55p" event={"ID":"001f4ed0-f6e6-440e-a878-2bca6982ea47","Type":"ContainerStarted","Data":"6537ec4e93e7d7e332d59fb429a20d95718baf70f342e2c89ccf3f495ddb3c80"} Nov 24 13:34:00 crc kubenswrapper[4790]: I1124 13:34:00.649455 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-69f849cd44-8v55p" Nov 24 13:34:00 crc kubenswrapper[4790]: I1124 13:34:00.655553 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8dvn5" event={"ID":"3e5246df-ca01-43e6-bcd9-c8d9b6b78279","Type":"ContainerStarted","Data":"d54ef9df4953d96ae51b1400969f9bf8a83e542d0262e0cb2e78aa602b92ae44"} Nov 24 13:34:00 crc kubenswrapper[4790]: I1124 13:34:00.674059 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-78cd478f4f-bqm94" event={"ID":"4d4e45de-9768-40e9-9c75-6abba8b38559","Type":"ContainerStarted","Data":"cef1cb69d1e5ecd83dedc3637e51913646334d098991071163c9a6574695cd3e"} Nov 24 13:34:00 crc kubenswrapper[4790]: I1124 13:34:00.674107 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-78cd478f4f-bqm94" event={"ID":"4d4e45de-9768-40e9-9c75-6abba8b38559","Type":"ContainerStarted","Data":"deee7a82aed113ecc740dcca448cfcb53642b959cd494942c6732db6590cb066"} Nov 24 13:34:00 crc kubenswrapper[4790]: I1124 13:34:00.674117 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-78cd478f4f-bqm94" event={"ID":"4d4e45de-9768-40e9-9c75-6abba8b38559","Type":"ContainerStarted","Data":"fcfab0bd76919bb2d0cb9a4d6fe3601824fa84b37b36fb0578f01279ecef5776"} Nov 24 13:34:00 crc kubenswrapper[4790]: I1124 13:34:00.674272 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-78cd478f4f-bqm94" Nov 24 13:34:00 crc kubenswrapper[4790]: I1124 13:34:00.681926 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-mbm86" event={"ID":"ca484cf0-4ab0-4581-bb65-b1701db63df3","Type":"ContainerStarted","Data":"d45798957a393567d315abf87c95bac024e5a20af18679a63ea606b0ce35e55e"} Nov 24 13:34:00 crc kubenswrapper[4790]: I1124 13:34:00.682869 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-8dvn5" podStartSLOduration=2.408364322 podStartE2EDuration="44.682853923s" podCreationTimestamp="2025-11-24 13:33:16 +0000 UTC" firstStartedPulling="2025-11-24 13:33:17.207211889 +0000 UTC m=+1245.587105551" lastFinishedPulling="2025-11-24 13:33:59.48170149 +0000 UTC m=+1287.861595152" observedRunningTime="2025-11-24 13:34:00.678422447 +0000 UTC m=+1289.058316109" watchObservedRunningTime="2025-11-24 13:34:00.682853923 +0000 UTC m=+1289.062747585" Nov 24 13:34:00 crc kubenswrapper[4790]: I1124 13:34:00.705287 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-69f849cd44-8v55p" podStartSLOduration=9.705265071 podStartE2EDuration="9.705265071s" podCreationTimestamp="2025-11-24 13:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:34:00.702112791 +0000 UTC m=+1289.082006473" watchObservedRunningTime="2025-11-24 13:34:00.705265071 +0000 UTC m=+1289.085158733" Nov 24 13:34:00 crc kubenswrapper[4790]: I1124 13:34:00.734222 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-78cd478f4f-bqm94" podStartSLOduration=7.734199504 podStartE2EDuration="7.734199504s" podCreationTimestamp="2025-11-24 13:33:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:34:00.718740264 +0000 UTC m=+1289.098633926" watchObservedRunningTime="2025-11-24 13:34:00.734199504 +0000 UTC m=+1289.114093166" Nov 24 13:34:00 crc kubenswrapper[4790]: I1124 13:34:00.740390 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-mbm86" podStartSLOduration=2.853967578 podStartE2EDuration="44.740376889s" podCreationTimestamp="2025-11-24 13:33:16 +0000 UTC" firstStartedPulling="2025-11-24 13:33:17.595289049 +0000 UTC m=+1245.975182711" lastFinishedPulling="2025-11-24 13:33:59.48169837 +0000 UTC m=+1287.861592022" observedRunningTime="2025-11-24 13:34:00.733564516 +0000 UTC m=+1289.113458178" watchObservedRunningTime="2025-11-24 13:34:00.740376889 +0000 UTC m=+1289.120270551" Nov 24 13:34:01 crc kubenswrapper[4790]: I1124 13:34:01.693438 4790 generic.go:334] "Generic (PLEG): container finished" podID="4d574c9c-ecb5-4b8c-b0b3-a058c141c09f" containerID="fdb5a55d080b1307b7fd39a3e10de2010b66b33a3318bca4aff4b9b6e6d39845" exitCode=0 Nov 24 13:34:01 crc kubenswrapper[4790]: I1124 13:34:01.693835 4790 generic.go:334] "Generic (PLEG): container finished" podID="4d574c9c-ecb5-4b8c-b0b3-a058c141c09f" containerID="dc393809fb6808c6307e7e2fc40a54ca96fae2a33dd3a809b34d8dc7c75ce59b" exitCode=2 Nov 24 13:34:01 crc kubenswrapper[4790]: I1124 13:34:01.693506 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f","Type":"ContainerDied","Data":"fdb5a55d080b1307b7fd39a3e10de2010b66b33a3318bca4aff4b9b6e6d39845"} Nov 24 13:34:01 crc kubenswrapper[4790]: I1124 13:34:01.694021 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f","Type":"ContainerDied","Data":"dc393809fb6808c6307e7e2fc40a54ca96fae2a33dd3a809b34d8dc7c75ce59b"} Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.007203 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b7b667979-z76jp" Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.080473 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-c2jtn"] Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.080704 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56df8fb6b7-c2jtn" podUID="871c7d3b-ecde-472b-a0ff-61c1a7a24f30" containerName="dnsmasq-dns" containerID="cri-o://14a2345853543961bb0465b42e0dd8b869d2c3096eaf9c9311f17fad8ea912de" gracePeriod=10 Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.526812 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-c2jtn" Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.587568 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-ovsdbserver-nb\") pod \"871c7d3b-ecde-472b-a0ff-61c1a7a24f30\" (UID: \"871c7d3b-ecde-472b-a0ff-61c1a7a24f30\") " Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.588114 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-dns-svc\") pod \"871c7d3b-ecde-472b-a0ff-61c1a7a24f30\" (UID: \"871c7d3b-ecde-472b-a0ff-61c1a7a24f30\") " Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.588201 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-config\") pod \"871c7d3b-ecde-472b-a0ff-61c1a7a24f30\" (UID: \"871c7d3b-ecde-472b-a0ff-61c1a7a24f30\") " Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.588347 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-dns-swift-storage-0\") pod \"871c7d3b-ecde-472b-a0ff-61c1a7a24f30\" (UID: \"871c7d3b-ecde-472b-a0ff-61c1a7a24f30\") " Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.588458 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-ovsdbserver-sb\") pod \"871c7d3b-ecde-472b-a0ff-61c1a7a24f30\" (UID: \"871c7d3b-ecde-472b-a0ff-61c1a7a24f30\") " Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.588532 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5c6p7\" (UniqueName: \"kubernetes.io/projected/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-kube-api-access-5c6p7\") pod \"871c7d3b-ecde-472b-a0ff-61c1a7a24f30\" (UID: \"871c7d3b-ecde-472b-a0ff-61c1a7a24f30\") " Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.603183 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-kube-api-access-5c6p7" (OuterVolumeSpecName: "kube-api-access-5c6p7") pod "871c7d3b-ecde-472b-a0ff-61c1a7a24f30" (UID: "871c7d3b-ecde-472b-a0ff-61c1a7a24f30"). InnerVolumeSpecName "kube-api-access-5c6p7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.636674 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "871c7d3b-ecde-472b-a0ff-61c1a7a24f30" (UID: "871c7d3b-ecde-472b-a0ff-61c1a7a24f30"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.644340 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "871c7d3b-ecde-472b-a0ff-61c1a7a24f30" (UID: "871c7d3b-ecde-472b-a0ff-61c1a7a24f30"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.645334 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "871c7d3b-ecde-472b-a0ff-61c1a7a24f30" (UID: "871c7d3b-ecde-472b-a0ff-61c1a7a24f30"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.658089 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-config" (OuterVolumeSpecName: "config") pod "871c7d3b-ecde-472b-a0ff-61c1a7a24f30" (UID: "871c7d3b-ecde-472b-a0ff-61c1a7a24f30"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.658325 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "871c7d3b-ecde-472b-a0ff-61c1a7a24f30" (UID: "871c7d3b-ecde-472b-a0ff-61c1a7a24f30"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.691150 4790 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.691199 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.691210 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5c6p7\" (UniqueName: \"kubernetes.io/projected/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-kube-api-access-5c6p7\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.691221 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.691230 4790 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.691239 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/871c7d3b-ecde-472b-a0ff-61c1a7a24f30-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.706034 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-mbm86" event={"ID":"ca484cf0-4ab0-4581-bb65-b1701db63df3","Type":"ContainerDied","Data":"d45798957a393567d315abf87c95bac024e5a20af18679a63ea606b0ce35e55e"} Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.705940 4790 generic.go:334] "Generic (PLEG): container finished" podID="ca484cf0-4ab0-4581-bb65-b1701db63df3" containerID="d45798957a393567d315abf87c95bac024e5a20af18679a63ea606b0ce35e55e" exitCode=0 Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.710171 4790 generic.go:334] "Generic (PLEG): container finished" podID="871c7d3b-ecde-472b-a0ff-61c1a7a24f30" containerID="14a2345853543961bb0465b42e0dd8b869d2c3096eaf9c9311f17fad8ea912de" exitCode=0 Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.710207 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-c2jtn" event={"ID":"871c7d3b-ecde-472b-a0ff-61c1a7a24f30","Type":"ContainerDied","Data":"14a2345853543961bb0465b42e0dd8b869d2c3096eaf9c9311f17fad8ea912de"} Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.710228 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-c2jtn" Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.710287 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-c2jtn" event={"ID":"871c7d3b-ecde-472b-a0ff-61c1a7a24f30","Type":"ContainerDied","Data":"43c992ed439be035efc7997e34b933d881c57f94b9e3162d65353c55b85c90cb"} Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.710321 4790 scope.go:117] "RemoveContainer" containerID="14a2345853543961bb0465b42e0dd8b869d2c3096eaf9c9311f17fad8ea912de" Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.733914 4790 scope.go:117] "RemoveContainer" containerID="d59fc437b66b47fcc10e49fb481cf5d86df8e0a79f0c8fbf5ac234655c8d9a5d" Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.745371 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-c2jtn"] Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.753196 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-c2jtn"] Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.765677 4790 scope.go:117] "RemoveContainer" containerID="14a2345853543961bb0465b42e0dd8b869d2c3096eaf9c9311f17fad8ea912de" Nov 24 13:34:02 crc kubenswrapper[4790]: E1124 13:34:02.766326 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14a2345853543961bb0465b42e0dd8b869d2c3096eaf9c9311f17fad8ea912de\": container with ID starting with 14a2345853543961bb0465b42e0dd8b869d2c3096eaf9c9311f17fad8ea912de not found: ID does not exist" containerID="14a2345853543961bb0465b42e0dd8b869d2c3096eaf9c9311f17fad8ea912de" Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.766386 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14a2345853543961bb0465b42e0dd8b869d2c3096eaf9c9311f17fad8ea912de"} err="failed to get container status \"14a2345853543961bb0465b42e0dd8b869d2c3096eaf9c9311f17fad8ea912de\": rpc error: code = NotFound desc = could not find container \"14a2345853543961bb0465b42e0dd8b869d2c3096eaf9c9311f17fad8ea912de\": container with ID starting with 14a2345853543961bb0465b42e0dd8b869d2c3096eaf9c9311f17fad8ea912de not found: ID does not exist" Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.766417 4790 scope.go:117] "RemoveContainer" containerID="d59fc437b66b47fcc10e49fb481cf5d86df8e0a79f0c8fbf5ac234655c8d9a5d" Nov 24 13:34:02 crc kubenswrapper[4790]: E1124 13:34:02.766918 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d59fc437b66b47fcc10e49fb481cf5d86df8e0a79f0c8fbf5ac234655c8d9a5d\": container with ID starting with d59fc437b66b47fcc10e49fb481cf5d86df8e0a79f0c8fbf5ac234655c8d9a5d not found: ID does not exist" containerID="d59fc437b66b47fcc10e49fb481cf5d86df8e0a79f0c8fbf5ac234655c8d9a5d" Nov 24 13:34:02 crc kubenswrapper[4790]: I1124 13:34:02.766958 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d59fc437b66b47fcc10e49fb481cf5d86df8e0a79f0c8fbf5ac234655c8d9a5d"} err="failed to get container status \"d59fc437b66b47fcc10e49fb481cf5d86df8e0a79f0c8fbf5ac234655c8d9a5d\": rpc error: code = NotFound desc = could not find container \"d59fc437b66b47fcc10e49fb481cf5d86df8e0a79f0c8fbf5ac234655c8d9a5d\": container with ID starting with d59fc437b66b47fcc10e49fb481cf5d86df8e0a79f0c8fbf5ac234655c8d9a5d not found: ID does not exist" Nov 24 13:34:04 crc kubenswrapper[4790]: I1124 13:34:04.043661 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-mbm86" Nov 24 13:34:04 crc kubenswrapper[4790]: I1124 13:34:04.114279 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fcc7\" (UniqueName: \"kubernetes.io/projected/ca484cf0-4ab0-4581-bb65-b1701db63df3-kube-api-access-2fcc7\") pod \"ca484cf0-4ab0-4581-bb65-b1701db63df3\" (UID: \"ca484cf0-4ab0-4581-bb65-b1701db63df3\") " Nov 24 13:34:04 crc kubenswrapper[4790]: I1124 13:34:04.114329 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ca484cf0-4ab0-4581-bb65-b1701db63df3-db-sync-config-data\") pod \"ca484cf0-4ab0-4581-bb65-b1701db63df3\" (UID: \"ca484cf0-4ab0-4581-bb65-b1701db63df3\") " Nov 24 13:34:04 crc kubenswrapper[4790]: I1124 13:34:04.114464 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca484cf0-4ab0-4581-bb65-b1701db63df3-combined-ca-bundle\") pod \"ca484cf0-4ab0-4581-bb65-b1701db63df3\" (UID: \"ca484cf0-4ab0-4581-bb65-b1701db63df3\") " Nov 24 13:34:04 crc kubenswrapper[4790]: I1124 13:34:04.120150 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca484cf0-4ab0-4581-bb65-b1701db63df3-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ca484cf0-4ab0-4581-bb65-b1701db63df3" (UID: "ca484cf0-4ab0-4581-bb65-b1701db63df3"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:04 crc kubenswrapper[4790]: I1124 13:34:04.120300 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca484cf0-4ab0-4581-bb65-b1701db63df3-kube-api-access-2fcc7" (OuterVolumeSpecName: "kube-api-access-2fcc7") pod "ca484cf0-4ab0-4581-bb65-b1701db63df3" (UID: "ca484cf0-4ab0-4581-bb65-b1701db63df3"). InnerVolumeSpecName "kube-api-access-2fcc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:04 crc kubenswrapper[4790]: I1124 13:34:04.143019 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca484cf0-4ab0-4581-bb65-b1701db63df3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ca484cf0-4ab0-4581-bb65-b1701db63df3" (UID: "ca484cf0-4ab0-4581-bb65-b1701db63df3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:04 crc kubenswrapper[4790]: I1124 13:34:04.216128 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca484cf0-4ab0-4581-bb65-b1701db63df3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:04 crc kubenswrapper[4790]: I1124 13:34:04.216168 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fcc7\" (UniqueName: \"kubernetes.io/projected/ca484cf0-4ab0-4581-bb65-b1701db63df3-kube-api-access-2fcc7\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:04 crc kubenswrapper[4790]: I1124 13:34:04.216184 4790 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ca484cf0-4ab0-4581-bb65-b1701db63df3-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:04 crc kubenswrapper[4790]: I1124 13:34:04.324681 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="871c7d3b-ecde-472b-a0ff-61c1a7a24f30" path="/var/lib/kubelet/pods/871c7d3b-ecde-472b-a0ff-61c1a7a24f30/volumes" Nov 24 13:34:04 crc kubenswrapper[4790]: I1124 13:34:04.737431 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-mbm86" Nov 24 13:34:04 crc kubenswrapper[4790]: I1124 13:34:04.738453 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-mbm86" event={"ID":"ca484cf0-4ab0-4581-bb65-b1701db63df3","Type":"ContainerDied","Data":"d8d66e587d7088f54f5678772108443371f6ea405b639c1292f06c359440c6fe"} Nov 24 13:34:04 crc kubenswrapper[4790]: I1124 13:34:04.738487 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8d66e587d7088f54f5678772108443371f6ea405b639c1292f06c359440c6fe" Nov 24 13:34:04 crc kubenswrapper[4790]: I1124 13:34:04.741792 4790 generic.go:334] "Generic (PLEG): container finished" podID="3e5246df-ca01-43e6-bcd9-c8d9b6b78279" containerID="d54ef9df4953d96ae51b1400969f9bf8a83e542d0262e0cb2e78aa602b92ae44" exitCode=0 Nov 24 13:34:04 crc kubenswrapper[4790]: I1124 13:34:04.741833 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8dvn5" event={"ID":"3e5246df-ca01-43e6-bcd9-c8d9b6b78279","Type":"ContainerDied","Data":"d54ef9df4953d96ae51b1400969f9bf8a83e542d0262e0cb2e78aa602b92ae44"} Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.293645 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-58b5bcbcb-b6kbh"] Nov 24 13:34:05 crc kubenswrapper[4790]: E1124 13:34:05.294145 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca484cf0-4ab0-4581-bb65-b1701db63df3" containerName="barbican-db-sync" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.294161 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca484cf0-4ab0-4581-bb65-b1701db63df3" containerName="barbican-db-sync" Nov 24 13:34:05 crc kubenswrapper[4790]: E1124 13:34:05.294186 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="871c7d3b-ecde-472b-a0ff-61c1a7a24f30" containerName="init" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.294193 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="871c7d3b-ecde-472b-a0ff-61c1a7a24f30" containerName="init" Nov 24 13:34:05 crc kubenswrapper[4790]: E1124 13:34:05.294209 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="871c7d3b-ecde-472b-a0ff-61c1a7a24f30" containerName="dnsmasq-dns" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.294216 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="871c7d3b-ecde-472b-a0ff-61c1a7a24f30" containerName="dnsmasq-dns" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.294417 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca484cf0-4ab0-4581-bb65-b1701db63df3" containerName="barbican-db-sync" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.294442 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="871c7d3b-ecde-472b-a0ff-61c1a7a24f30" containerName="dnsmasq-dns" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.295602 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-58b5bcbcb-b6kbh" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.297791 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.302627 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-d9ljp" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.305818 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.305819 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-845f9d57f-2s48z"] Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.307786 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-845f9d57f-2s48z" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.326123 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.338033 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-58b5bcbcb-b6kbh"] Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.361557 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-845f9d57f-2s48z"] Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.399405 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-8hxrv"] Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.405294 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-8hxrv" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.428190 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-8hxrv"] Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.433572 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkrrs\" (UniqueName: \"kubernetes.io/projected/3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c-kube-api-access-fkrrs\") pod \"barbican-worker-845f9d57f-2s48z\" (UID: \"3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c\") " pod="openstack/barbican-worker-845f9d57f-2s48z" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.433616 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2ee8af7-26fb-4652-be37-594db62f1146-config-data\") pod \"barbican-keystone-listener-58b5bcbcb-b6kbh\" (UID: \"a2ee8af7-26fb-4652-be37-594db62f1146\") " pod="openstack/barbican-keystone-listener-58b5bcbcb-b6kbh" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.433641 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqgn6\" (UniqueName: \"kubernetes.io/projected/a2ee8af7-26fb-4652-be37-594db62f1146-kube-api-access-hqgn6\") pod \"barbican-keystone-listener-58b5bcbcb-b6kbh\" (UID: \"a2ee8af7-26fb-4652-be37-594db62f1146\") " pod="openstack/barbican-keystone-listener-58b5bcbcb-b6kbh" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.433676 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c-config-data\") pod \"barbican-worker-845f9d57f-2s48z\" (UID: \"3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c\") " pod="openstack/barbican-worker-845f9d57f-2s48z" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.433694 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c-config-data-custom\") pod \"barbican-worker-845f9d57f-2s48z\" (UID: \"3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c\") " pod="openstack/barbican-worker-845f9d57f-2s48z" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.433747 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2ee8af7-26fb-4652-be37-594db62f1146-logs\") pod \"barbican-keystone-listener-58b5bcbcb-b6kbh\" (UID: \"a2ee8af7-26fb-4652-be37-594db62f1146\") " pod="openstack/barbican-keystone-listener-58b5bcbcb-b6kbh" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.433763 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c-logs\") pod \"barbican-worker-845f9d57f-2s48z\" (UID: \"3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c\") " pod="openstack/barbican-worker-845f9d57f-2s48z" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.433789 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c-combined-ca-bundle\") pod \"barbican-worker-845f9d57f-2s48z\" (UID: \"3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c\") " pod="openstack/barbican-worker-845f9d57f-2s48z" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.434217 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2ee8af7-26fb-4652-be37-594db62f1146-combined-ca-bundle\") pod \"barbican-keystone-listener-58b5bcbcb-b6kbh\" (UID: \"a2ee8af7-26fb-4652-be37-594db62f1146\") " pod="openstack/barbican-keystone-listener-58b5bcbcb-b6kbh" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.434259 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a2ee8af7-26fb-4652-be37-594db62f1146-config-data-custom\") pod \"barbican-keystone-listener-58b5bcbcb-b6kbh\" (UID: \"a2ee8af7-26fb-4652-be37-594db62f1146\") " pod="openstack/barbican-keystone-listener-58b5bcbcb-b6kbh" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.536340 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2ee8af7-26fb-4652-be37-594db62f1146-combined-ca-bundle\") pod \"barbican-keystone-listener-58b5bcbcb-b6kbh\" (UID: \"a2ee8af7-26fb-4652-be37-594db62f1146\") " pod="openstack/barbican-keystone-listener-58b5bcbcb-b6kbh" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.536438 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a2ee8af7-26fb-4652-be37-594db62f1146-config-data-custom\") pod \"barbican-keystone-listener-58b5bcbcb-b6kbh\" (UID: \"a2ee8af7-26fb-4652-be37-594db62f1146\") " pod="openstack/barbican-keystone-listener-58b5bcbcb-b6kbh" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.537868 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkrrs\" (UniqueName: \"kubernetes.io/projected/3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c-kube-api-access-fkrrs\") pod \"barbican-worker-845f9d57f-2s48z\" (UID: \"3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c\") " pod="openstack/barbican-worker-845f9d57f-2s48z" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.537922 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2ee8af7-26fb-4652-be37-594db62f1146-config-data\") pod \"barbican-keystone-listener-58b5bcbcb-b6kbh\" (UID: \"a2ee8af7-26fb-4652-be37-594db62f1146\") " pod="openstack/barbican-keystone-listener-58b5bcbcb-b6kbh" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.537960 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-8hxrv\" (UID: \"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265\") " pod="openstack/dnsmasq-dns-848cf88cfc-8hxrv" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.537987 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqgn6\" (UniqueName: \"kubernetes.io/projected/a2ee8af7-26fb-4652-be37-594db62f1146-kube-api-access-hqgn6\") pod \"barbican-keystone-listener-58b5bcbcb-b6kbh\" (UID: \"a2ee8af7-26fb-4652-be37-594db62f1146\") " pod="openstack/barbican-keystone-listener-58b5bcbcb-b6kbh" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.538057 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfn8l\" (UniqueName: \"kubernetes.io/projected/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-kube-api-access-tfn8l\") pod \"dnsmasq-dns-848cf88cfc-8hxrv\" (UID: \"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265\") " pod="openstack/dnsmasq-dns-848cf88cfc-8hxrv" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.538081 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-config\") pod \"dnsmasq-dns-848cf88cfc-8hxrv\" (UID: \"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265\") " pod="openstack/dnsmasq-dns-848cf88cfc-8hxrv" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.538118 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c-config-data\") pod \"barbican-worker-845f9d57f-2s48z\" (UID: \"3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c\") " pod="openstack/barbican-worker-845f9d57f-2s48z" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.538148 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c-config-data-custom\") pod \"barbican-worker-845f9d57f-2s48z\" (UID: \"3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c\") " pod="openstack/barbican-worker-845f9d57f-2s48z" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.538255 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2ee8af7-26fb-4652-be37-594db62f1146-logs\") pod \"barbican-keystone-listener-58b5bcbcb-b6kbh\" (UID: \"a2ee8af7-26fb-4652-be37-594db62f1146\") " pod="openstack/barbican-keystone-listener-58b5bcbcb-b6kbh" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.538286 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c-logs\") pod \"barbican-worker-845f9d57f-2s48z\" (UID: \"3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c\") " pod="openstack/barbican-worker-845f9d57f-2s48z" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.538332 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-8hxrv\" (UID: \"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265\") " pod="openstack/dnsmasq-dns-848cf88cfc-8hxrv" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.538361 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-8hxrv\" (UID: \"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265\") " pod="openstack/dnsmasq-dns-848cf88cfc-8hxrv" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.538388 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c-combined-ca-bundle\") pod \"barbican-worker-845f9d57f-2s48z\" (UID: \"3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c\") " pod="openstack/barbican-worker-845f9d57f-2s48z" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.538430 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-8hxrv\" (UID: \"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265\") " pod="openstack/dnsmasq-dns-848cf88cfc-8hxrv" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.538572 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2ee8af7-26fb-4652-be37-594db62f1146-logs\") pod \"barbican-keystone-listener-58b5bcbcb-b6kbh\" (UID: \"a2ee8af7-26fb-4652-be37-594db62f1146\") " pod="openstack/barbican-keystone-listener-58b5bcbcb-b6kbh" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.538848 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c-logs\") pod \"barbican-worker-845f9d57f-2s48z\" (UID: \"3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c\") " pod="openstack/barbican-worker-845f9d57f-2s48z" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.557762 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2ee8af7-26fb-4652-be37-594db62f1146-combined-ca-bundle\") pod \"barbican-keystone-listener-58b5bcbcb-b6kbh\" (UID: \"a2ee8af7-26fb-4652-be37-594db62f1146\") " pod="openstack/barbican-keystone-listener-58b5bcbcb-b6kbh" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.557763 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a2ee8af7-26fb-4652-be37-594db62f1146-config-data-custom\") pod \"barbican-keystone-listener-58b5bcbcb-b6kbh\" (UID: \"a2ee8af7-26fb-4652-be37-594db62f1146\") " pod="openstack/barbican-keystone-listener-58b5bcbcb-b6kbh" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.558196 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c-config-data\") pod \"barbican-worker-845f9d57f-2s48z\" (UID: \"3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c\") " pod="openstack/barbican-worker-845f9d57f-2s48z" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.558863 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2ee8af7-26fb-4652-be37-594db62f1146-config-data\") pod \"barbican-keystone-listener-58b5bcbcb-b6kbh\" (UID: \"a2ee8af7-26fb-4652-be37-594db62f1146\") " pod="openstack/barbican-keystone-listener-58b5bcbcb-b6kbh" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.559797 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c-config-data-custom\") pod \"barbican-worker-845f9d57f-2s48z\" (UID: \"3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c\") " pod="openstack/barbican-worker-845f9d57f-2s48z" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.563000 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqgn6\" (UniqueName: \"kubernetes.io/projected/a2ee8af7-26fb-4652-be37-594db62f1146-kube-api-access-hqgn6\") pod \"barbican-keystone-listener-58b5bcbcb-b6kbh\" (UID: \"a2ee8af7-26fb-4652-be37-594db62f1146\") " pod="openstack/barbican-keystone-listener-58b5bcbcb-b6kbh" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.563622 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkrrs\" (UniqueName: \"kubernetes.io/projected/3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c-kube-api-access-fkrrs\") pod \"barbican-worker-845f9d57f-2s48z\" (UID: \"3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c\") " pod="openstack/barbican-worker-845f9d57f-2s48z" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.569806 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c-combined-ca-bundle\") pod \"barbican-worker-845f9d57f-2s48z\" (UID: \"3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c\") " pod="openstack/barbican-worker-845f9d57f-2s48z" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.596180 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-dd8b84bc8-9g27k"] Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.599299 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-dd8b84bc8-9g27k" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.605190 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.609802 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-dd8b84bc8-9g27k"] Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.620777 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-58b5bcbcb-b6kbh" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.640402 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca62cb72-d397-4ee8-a7c4-15d831f87776-config-data\") pod \"barbican-api-dd8b84bc8-9g27k\" (UID: \"ca62cb72-d397-4ee8-a7c4-15d831f87776\") " pod="openstack/barbican-api-dd8b84bc8-9g27k" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.640481 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca62cb72-d397-4ee8-a7c4-15d831f87776-combined-ca-bundle\") pod \"barbican-api-dd8b84bc8-9g27k\" (UID: \"ca62cb72-d397-4ee8-a7c4-15d831f87776\") " pod="openstack/barbican-api-dd8b84bc8-9g27k" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.640531 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-8hxrv\" (UID: \"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265\") " pod="openstack/dnsmasq-dns-848cf88cfc-8hxrv" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.640564 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-8hxrv\" (UID: \"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265\") " pod="openstack/dnsmasq-dns-848cf88cfc-8hxrv" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.640604 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-8hxrv\" (UID: \"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265\") " pod="openstack/dnsmasq-dns-848cf88cfc-8hxrv" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.640643 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca62cb72-d397-4ee8-a7c4-15d831f87776-config-data-custom\") pod \"barbican-api-dd8b84bc8-9g27k\" (UID: \"ca62cb72-d397-4ee8-a7c4-15d831f87776\") " pod="openstack/barbican-api-dd8b84bc8-9g27k" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.640708 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca62cb72-d397-4ee8-a7c4-15d831f87776-logs\") pod \"barbican-api-dd8b84bc8-9g27k\" (UID: \"ca62cb72-d397-4ee8-a7c4-15d831f87776\") " pod="openstack/barbican-api-dd8b84bc8-9g27k" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.640745 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8ksw\" (UniqueName: \"kubernetes.io/projected/ca62cb72-d397-4ee8-a7c4-15d831f87776-kube-api-access-k8ksw\") pod \"barbican-api-dd8b84bc8-9g27k\" (UID: \"ca62cb72-d397-4ee8-a7c4-15d831f87776\") " pod="openstack/barbican-api-dd8b84bc8-9g27k" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.643818 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-8hxrv\" (UID: \"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265\") " pod="openstack/dnsmasq-dns-848cf88cfc-8hxrv" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.643869 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfn8l\" (UniqueName: \"kubernetes.io/projected/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-kube-api-access-tfn8l\") pod \"dnsmasq-dns-848cf88cfc-8hxrv\" (UID: \"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265\") " pod="openstack/dnsmasq-dns-848cf88cfc-8hxrv" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.643928 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-config\") pod \"dnsmasq-dns-848cf88cfc-8hxrv\" (UID: \"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265\") " pod="openstack/dnsmasq-dns-848cf88cfc-8hxrv" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.643974 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-8hxrv\" (UID: \"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265\") " pod="openstack/dnsmasq-dns-848cf88cfc-8hxrv" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.642730 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-8hxrv\" (UID: \"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265\") " pod="openstack/dnsmasq-dns-848cf88cfc-8hxrv" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.643292 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-8hxrv\" (UID: \"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265\") " pod="openstack/dnsmasq-dns-848cf88cfc-8hxrv" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.645103 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-8hxrv\" (UID: \"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265\") " pod="openstack/dnsmasq-dns-848cf88cfc-8hxrv" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.645154 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-config\") pod \"dnsmasq-dns-848cf88cfc-8hxrv\" (UID: \"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265\") " pod="openstack/dnsmasq-dns-848cf88cfc-8hxrv" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.646353 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-845f9d57f-2s48z" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.663026 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfn8l\" (UniqueName: \"kubernetes.io/projected/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-kube-api-access-tfn8l\") pod \"dnsmasq-dns-848cf88cfc-8hxrv\" (UID: \"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265\") " pod="openstack/dnsmasq-dns-848cf88cfc-8hxrv" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.742314 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-8hxrv" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.745927 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca62cb72-d397-4ee8-a7c4-15d831f87776-combined-ca-bundle\") pod \"barbican-api-dd8b84bc8-9g27k\" (UID: \"ca62cb72-d397-4ee8-a7c4-15d831f87776\") " pod="openstack/barbican-api-dd8b84bc8-9g27k" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.746022 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca62cb72-d397-4ee8-a7c4-15d831f87776-config-data-custom\") pod \"barbican-api-dd8b84bc8-9g27k\" (UID: \"ca62cb72-d397-4ee8-a7c4-15d831f87776\") " pod="openstack/barbican-api-dd8b84bc8-9g27k" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.746090 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca62cb72-d397-4ee8-a7c4-15d831f87776-logs\") pod \"barbican-api-dd8b84bc8-9g27k\" (UID: \"ca62cb72-d397-4ee8-a7c4-15d831f87776\") " pod="openstack/barbican-api-dd8b84bc8-9g27k" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.746125 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8ksw\" (UniqueName: \"kubernetes.io/projected/ca62cb72-d397-4ee8-a7c4-15d831f87776-kube-api-access-k8ksw\") pod \"barbican-api-dd8b84bc8-9g27k\" (UID: \"ca62cb72-d397-4ee8-a7c4-15d831f87776\") " pod="openstack/barbican-api-dd8b84bc8-9g27k" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.746194 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca62cb72-d397-4ee8-a7c4-15d831f87776-config-data\") pod \"barbican-api-dd8b84bc8-9g27k\" (UID: \"ca62cb72-d397-4ee8-a7c4-15d831f87776\") " pod="openstack/barbican-api-dd8b84bc8-9g27k" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.749929 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca62cb72-d397-4ee8-a7c4-15d831f87776-logs\") pod \"barbican-api-dd8b84bc8-9g27k\" (UID: \"ca62cb72-d397-4ee8-a7c4-15d831f87776\") " pod="openstack/barbican-api-dd8b84bc8-9g27k" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.753024 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca62cb72-d397-4ee8-a7c4-15d831f87776-config-data-custom\") pod \"barbican-api-dd8b84bc8-9g27k\" (UID: \"ca62cb72-d397-4ee8-a7c4-15d831f87776\") " pod="openstack/barbican-api-dd8b84bc8-9g27k" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.753942 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca62cb72-d397-4ee8-a7c4-15d831f87776-config-data\") pod \"barbican-api-dd8b84bc8-9g27k\" (UID: \"ca62cb72-d397-4ee8-a7c4-15d831f87776\") " pod="openstack/barbican-api-dd8b84bc8-9g27k" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.757434 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca62cb72-d397-4ee8-a7c4-15d831f87776-combined-ca-bundle\") pod \"barbican-api-dd8b84bc8-9g27k\" (UID: \"ca62cb72-d397-4ee8-a7c4-15d831f87776\") " pod="openstack/barbican-api-dd8b84bc8-9g27k" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.767521 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8ksw\" (UniqueName: \"kubernetes.io/projected/ca62cb72-d397-4ee8-a7c4-15d831f87776-kube-api-access-k8ksw\") pod \"barbican-api-dd8b84bc8-9g27k\" (UID: \"ca62cb72-d397-4ee8-a7c4-15d831f87776\") " pod="openstack/barbican-api-dd8b84bc8-9g27k" Nov 24 13:34:05 crc kubenswrapper[4790]: I1124 13:34:05.818823 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-dd8b84bc8-9g27k" Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.306961 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.338353 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8dvn5" Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.361576 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-58b5bcbcb-b6kbh"] Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.361627 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-845f9d57f-2s48z"] Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.366090 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-run-httpd\") pod \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\" (UID: \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\") " Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.366151 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-config-data\") pod \"3e5246df-ca01-43e6-bcd9-c8d9b6b78279\" (UID: \"3e5246df-ca01-43e6-bcd9-c8d9b6b78279\") " Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.366191 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w4sf\" (UniqueName: \"kubernetes.io/projected/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-kube-api-access-2w4sf\") pod \"3e5246df-ca01-43e6-bcd9-c8d9b6b78279\" (UID: \"3e5246df-ca01-43e6-bcd9-c8d9b6b78279\") " Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.366267 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-config-data\") pod \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\" (UID: \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\") " Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.366309 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-sg-core-conf-yaml\") pod \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\" (UID: \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\") " Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.366349 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-scripts\") pod \"3e5246df-ca01-43e6-bcd9-c8d9b6b78279\" (UID: \"3e5246df-ca01-43e6-bcd9-c8d9b6b78279\") " Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.366461 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-db-sync-config-data\") pod \"3e5246df-ca01-43e6-bcd9-c8d9b6b78279\" (UID: \"3e5246df-ca01-43e6-bcd9-c8d9b6b78279\") " Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.366493 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-combined-ca-bundle\") pod \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\" (UID: \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\") " Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.366518 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8m6zh\" (UniqueName: \"kubernetes.io/projected/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-kube-api-access-8m6zh\") pod \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\" (UID: \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\") " Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.366563 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-combined-ca-bundle\") pod \"3e5246df-ca01-43e6-bcd9-c8d9b6b78279\" (UID: \"3e5246df-ca01-43e6-bcd9-c8d9b6b78279\") " Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.366595 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-etc-machine-id\") pod \"3e5246df-ca01-43e6-bcd9-c8d9b6b78279\" (UID: \"3e5246df-ca01-43e6-bcd9-c8d9b6b78279\") " Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.366621 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-log-httpd\") pod \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\" (UID: \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\") " Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.366661 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-scripts\") pod \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\" (UID: \"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f\") " Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.368913 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4d574c9c-ecb5-4b8c-b0b3-a058c141c09f" (UID: "4d574c9c-ecb5-4b8c-b0b3-a058c141c09f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.371188 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "3e5246df-ca01-43e6-bcd9-c8d9b6b78279" (UID: "3e5246df-ca01-43e6-bcd9-c8d9b6b78279"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.371615 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4d574c9c-ecb5-4b8c-b0b3-a058c141c09f" (UID: "4d574c9c-ecb5-4b8c-b0b3-a058c141c09f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.374612 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-scripts" (OuterVolumeSpecName: "scripts") pod "3e5246df-ca01-43e6-bcd9-c8d9b6b78279" (UID: "3e5246df-ca01-43e6-bcd9-c8d9b6b78279"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.375940 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-scripts" (OuterVolumeSpecName: "scripts") pod "4d574c9c-ecb5-4b8c-b0b3-a058c141c09f" (UID: "4d574c9c-ecb5-4b8c-b0b3-a058c141c09f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.378234 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-kube-api-access-8m6zh" (OuterVolumeSpecName: "kube-api-access-8m6zh") pod "4d574c9c-ecb5-4b8c-b0b3-a058c141c09f" (UID: "4d574c9c-ecb5-4b8c-b0b3-a058c141c09f"). InnerVolumeSpecName "kube-api-access-8m6zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.378477 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-kube-api-access-2w4sf" (OuterVolumeSpecName: "kube-api-access-2w4sf") pod "3e5246df-ca01-43e6-bcd9-c8d9b6b78279" (UID: "3e5246df-ca01-43e6-bcd9-c8d9b6b78279"). InnerVolumeSpecName "kube-api-access-2w4sf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.384351 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "3e5246df-ca01-43e6-bcd9-c8d9b6b78279" (UID: "3e5246df-ca01-43e6-bcd9-c8d9b6b78279"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.424808 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3e5246df-ca01-43e6-bcd9-c8d9b6b78279" (UID: "3e5246df-ca01-43e6-bcd9-c8d9b6b78279"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.438559 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4d574c9c-ecb5-4b8c-b0b3-a058c141c09f" (UID: "4d574c9c-ecb5-4b8c-b0b3-a058c141c09f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.443586 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d574c9c-ecb5-4b8c-b0b3-a058c141c09f" (UID: "4d574c9c-ecb5-4b8c-b0b3-a058c141c09f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:06 crc kubenswrapper[4790]: W1124 13:34:06.450146 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca62cb72_d397_4ee8_a7c4_15d831f87776.slice/crio-8554416072f611c39dd5e98f8ba56cfcd1574c3ce771e9262cbf898851e86218 WatchSource:0}: Error finding container 8554416072f611c39dd5e98f8ba56cfcd1574c3ce771e9262cbf898851e86218: Status 404 returned error can't find the container with id 8554416072f611c39dd5e98f8ba56cfcd1574c3ce771e9262cbf898851e86218 Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.452947 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-dd8b84bc8-9g27k"] Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.455100 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-config-data" (OuterVolumeSpecName: "config-data") pod "3e5246df-ca01-43e6-bcd9-c8d9b6b78279" (UID: "3e5246df-ca01-43e6-bcd9-c8d9b6b78279"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.455402 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-config-data" (OuterVolumeSpecName: "config-data") pod "4d574c9c-ecb5-4b8c-b0b3-a058c141c09f" (UID: "4d574c9c-ecb5-4b8c-b0b3-a058c141c09f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:06 crc kubenswrapper[4790]: W1124 13:34:06.456762 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc96afe58_d2d2_4e9d_a1f7_67ecd78bf265.slice/crio-16addf5fde61c223e1beabcd551c64f32633b49ade27ffd3df96a8a0f315654e WatchSource:0}: Error finding container 16addf5fde61c223e1beabcd551c64f32633b49ade27ffd3df96a8a0f315654e: Status 404 returned error can't find the container with id 16addf5fde61c223e1beabcd551c64f32633b49ade27ffd3df96a8a0f315654e Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.462925 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-8hxrv"] Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.469459 4790 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.469488 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.469497 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8m6zh\" (UniqueName: \"kubernetes.io/projected/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-kube-api-access-8m6zh\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.469507 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.469516 4790 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.469525 4790 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.469534 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.469541 4790 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.469549 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.469557 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w4sf\" (UniqueName: \"kubernetes.io/projected/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-kube-api-access-2w4sf\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.469564 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.469572 4790 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.469580 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e5246df-ca01-43e6-bcd9-c8d9b6b78279-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.804714 4790 generic.go:334] "Generic (PLEG): container finished" podID="4d574c9c-ecb5-4b8c-b0b3-a058c141c09f" containerID="6a117358d00dc47eb45e9087790aa499d845a99631f8d556a818c3dbb36a7a89" exitCode=0 Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.804787 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f","Type":"ContainerDied","Data":"6a117358d00dc47eb45e9087790aa499d845a99631f8d556a818c3dbb36a7a89"} Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.804816 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d574c9c-ecb5-4b8c-b0b3-a058c141c09f","Type":"ContainerDied","Data":"35ba9ca9fd843161f03f5d502d310845016dd199433abf5abbe6cc64039bcf25"} Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.804834 4790 scope.go:117] "RemoveContainer" containerID="fdb5a55d080b1307b7fd39a3e10de2010b66b33a3318bca4aff4b9b6e6d39845" Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.804990 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.807416 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-845f9d57f-2s48z" event={"ID":"3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c","Type":"ContainerStarted","Data":"4484a8a800584754e0bc823ba32652af95c8926cd5358a8c3f54ea000f9ae805"} Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.811550 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8dvn5" event={"ID":"3e5246df-ca01-43e6-bcd9-c8d9b6b78279","Type":"ContainerDied","Data":"32a5e77128d6ef141be8ff5534229731c523fc643794e1407b8166941a35e49a"} Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.811591 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32a5e77128d6ef141be8ff5534229731c523fc643794e1407b8166941a35e49a" Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.811560 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8dvn5" Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.822425 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-dd8b84bc8-9g27k" event={"ID":"ca62cb72-d397-4ee8-a7c4-15d831f87776","Type":"ContainerStarted","Data":"1c160de17887f7283642017061e7f697b7b7ff536b2c1ca6125e5119a1a8b631"} Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.822479 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-dd8b84bc8-9g27k" event={"ID":"ca62cb72-d397-4ee8-a7c4-15d831f87776","Type":"ContainerStarted","Data":"8554416072f611c39dd5e98f8ba56cfcd1574c3ce771e9262cbf898851e86218"} Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.830946 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-8hxrv" event={"ID":"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265","Type":"ContainerStarted","Data":"9ac61b043b7a62382b037fd6716ed1875cc67c39b24ab18087b0d7f9ff34e2bf"} Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.831034 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-8hxrv" event={"ID":"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265","Type":"ContainerStarted","Data":"16addf5fde61c223e1beabcd551c64f32633b49ade27ffd3df96a8a0f315654e"} Nov 24 13:34:06 crc kubenswrapper[4790]: I1124 13:34:06.835064 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-58b5bcbcb-b6kbh" event={"ID":"a2ee8af7-26fb-4652-be37-594db62f1146","Type":"ContainerStarted","Data":"194a06eb24751737741f835e766a9edc998ff73eeed814aa6bd2b16d57534f4e"} Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:06.999460 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 13:34:07 crc kubenswrapper[4790]: E1124 13:34:07.000299 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d574c9c-ecb5-4b8c-b0b3-a058c141c09f" containerName="ceilometer-notification-agent" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.000316 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d574c9c-ecb5-4b8c-b0b3-a058c141c09f" containerName="ceilometer-notification-agent" Nov 24 13:34:07 crc kubenswrapper[4790]: E1124 13:34:07.000330 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d574c9c-ecb5-4b8c-b0b3-a058c141c09f" containerName="proxy-httpd" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.000338 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d574c9c-ecb5-4b8c-b0b3-a058c141c09f" containerName="proxy-httpd" Nov 24 13:34:07 crc kubenswrapper[4790]: E1124 13:34:07.000365 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e5246df-ca01-43e6-bcd9-c8d9b6b78279" containerName="cinder-db-sync" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.000374 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e5246df-ca01-43e6-bcd9-c8d9b6b78279" containerName="cinder-db-sync" Nov 24 13:34:07 crc kubenswrapper[4790]: E1124 13:34:07.000429 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d574c9c-ecb5-4b8c-b0b3-a058c141c09f" containerName="sg-core" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.000438 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d574c9c-ecb5-4b8c-b0b3-a058c141c09f" containerName="sg-core" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.000655 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d574c9c-ecb5-4b8c-b0b3-a058c141c09f" containerName="proxy-httpd" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.000682 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e5246df-ca01-43e6-bcd9-c8d9b6b78279" containerName="cinder-db-sync" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.000699 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d574c9c-ecb5-4b8c-b0b3-a058c141c09f" containerName="ceilometer-notification-agent" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.000719 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d574c9c-ecb5-4b8c-b0b3-a058c141c09f" containerName="sg-core" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.002187 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.012306 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.012466 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-bsm2t" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.012588 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.012671 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.018644 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.088338 4790 scope.go:117] "RemoveContainer" containerID="dc393809fb6808c6307e7e2fc40a54ca96fae2a33dd3a809b34d8dc7c75ce59b" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.088429 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55e60feb-8531-4c7b-9033-5d3c6a07258a-scripts\") pod \"cinder-scheduler-0\" (UID: \"55e60feb-8531-4c7b-9033-5d3c6a07258a\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.088479 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/55e60feb-8531-4c7b-9033-5d3c6a07258a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"55e60feb-8531-4c7b-9033-5d3c6a07258a\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.088536 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55e60feb-8531-4c7b-9033-5d3c6a07258a-config-data\") pod \"cinder-scheduler-0\" (UID: \"55e60feb-8531-4c7b-9033-5d3c6a07258a\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.088601 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55e60feb-8531-4c7b-9033-5d3c6a07258a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"55e60feb-8531-4c7b-9033-5d3c6a07258a\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.088655 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/55e60feb-8531-4c7b-9033-5d3c6a07258a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"55e60feb-8531-4c7b-9033-5d3c6a07258a\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.088771 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srth5\" (UniqueName: \"kubernetes.io/projected/55e60feb-8531-4c7b-9033-5d3c6a07258a-kube-api-access-srth5\") pod \"cinder-scheduler-0\" (UID: \"55e60feb-8531-4c7b-9033-5d3c6a07258a\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.132715 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-8hxrv"] Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.158937 4790 scope.go:117] "RemoveContainer" containerID="6a117358d00dc47eb45e9087790aa499d845a99631f8d556a818c3dbb36a7a89" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.190443 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55e60feb-8531-4c7b-9033-5d3c6a07258a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"55e60feb-8531-4c7b-9033-5d3c6a07258a\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.190835 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/55e60feb-8531-4c7b-9033-5d3c6a07258a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"55e60feb-8531-4c7b-9033-5d3c6a07258a\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.191005 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srth5\" (UniqueName: \"kubernetes.io/projected/55e60feb-8531-4c7b-9033-5d3c6a07258a-kube-api-access-srth5\") pod \"cinder-scheduler-0\" (UID: \"55e60feb-8531-4c7b-9033-5d3c6a07258a\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.191203 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55e60feb-8531-4c7b-9033-5d3c6a07258a-scripts\") pod \"cinder-scheduler-0\" (UID: \"55e60feb-8531-4c7b-9033-5d3c6a07258a\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.191374 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/55e60feb-8531-4c7b-9033-5d3c6a07258a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"55e60feb-8531-4c7b-9033-5d3c6a07258a\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.191459 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55e60feb-8531-4c7b-9033-5d3c6a07258a-config-data\") pod \"cinder-scheduler-0\" (UID: \"55e60feb-8531-4c7b-9033-5d3c6a07258a\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.193598 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/55e60feb-8531-4c7b-9033-5d3c6a07258a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"55e60feb-8531-4c7b-9033-5d3c6a07258a\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.198238 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-snppf"] Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.200067 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-snppf" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.205053 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/55e60feb-8531-4c7b-9033-5d3c6a07258a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"55e60feb-8531-4c7b-9033-5d3c6a07258a\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.215374 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55e60feb-8531-4c7b-9033-5d3c6a07258a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"55e60feb-8531-4c7b-9033-5d3c6a07258a\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.240588 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55e60feb-8531-4c7b-9033-5d3c6a07258a-scripts\") pod \"cinder-scheduler-0\" (UID: \"55e60feb-8531-4c7b-9033-5d3c6a07258a\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.240642 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55e60feb-8531-4c7b-9033-5d3c6a07258a-config-data\") pod \"cinder-scheduler-0\" (UID: \"55e60feb-8531-4c7b-9033-5d3c6a07258a\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.253673 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.270705 4790 scope.go:117] "RemoveContainer" containerID="fdb5a55d080b1307b7fd39a3e10de2010b66b33a3318bca4aff4b9b6e6d39845" Nov 24 13:34:07 crc kubenswrapper[4790]: E1124 13:34:07.272797 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdb5a55d080b1307b7fd39a3e10de2010b66b33a3318bca4aff4b9b6e6d39845\": container with ID starting with fdb5a55d080b1307b7fd39a3e10de2010b66b33a3318bca4aff4b9b6e6d39845 not found: ID does not exist" containerID="fdb5a55d080b1307b7fd39a3e10de2010b66b33a3318bca4aff4b9b6e6d39845" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.272839 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdb5a55d080b1307b7fd39a3e10de2010b66b33a3318bca4aff4b9b6e6d39845"} err="failed to get container status \"fdb5a55d080b1307b7fd39a3e10de2010b66b33a3318bca4aff4b9b6e6d39845\": rpc error: code = NotFound desc = could not find container \"fdb5a55d080b1307b7fd39a3e10de2010b66b33a3318bca4aff4b9b6e6d39845\": container with ID starting with fdb5a55d080b1307b7fd39a3e10de2010b66b33a3318bca4aff4b9b6e6d39845 not found: ID does not exist" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.272866 4790 scope.go:117] "RemoveContainer" containerID="dc393809fb6808c6307e7e2fc40a54ca96fae2a33dd3a809b34d8dc7c75ce59b" Nov 24 13:34:07 crc kubenswrapper[4790]: E1124 13:34:07.273100 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc393809fb6808c6307e7e2fc40a54ca96fae2a33dd3a809b34d8dc7c75ce59b\": container with ID starting with dc393809fb6808c6307e7e2fc40a54ca96fae2a33dd3a809b34d8dc7c75ce59b not found: ID does not exist" containerID="dc393809fb6808c6307e7e2fc40a54ca96fae2a33dd3a809b34d8dc7c75ce59b" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.273125 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc393809fb6808c6307e7e2fc40a54ca96fae2a33dd3a809b34d8dc7c75ce59b"} err="failed to get container status \"dc393809fb6808c6307e7e2fc40a54ca96fae2a33dd3a809b34d8dc7c75ce59b\": rpc error: code = NotFound desc = could not find container \"dc393809fb6808c6307e7e2fc40a54ca96fae2a33dd3a809b34d8dc7c75ce59b\": container with ID starting with dc393809fb6808c6307e7e2fc40a54ca96fae2a33dd3a809b34d8dc7c75ce59b not found: ID does not exist" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.273138 4790 scope.go:117] "RemoveContainer" containerID="6a117358d00dc47eb45e9087790aa499d845a99631f8d556a818c3dbb36a7a89" Nov 24 13:34:07 crc kubenswrapper[4790]: E1124 13:34:07.273304 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a117358d00dc47eb45e9087790aa499d845a99631f8d556a818c3dbb36a7a89\": container with ID starting with 6a117358d00dc47eb45e9087790aa499d845a99631f8d556a818c3dbb36a7a89 not found: ID does not exist" containerID="6a117358d00dc47eb45e9087790aa499d845a99631f8d556a818c3dbb36a7a89" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.273319 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a117358d00dc47eb45e9087790aa499d845a99631f8d556a818c3dbb36a7a89"} err="failed to get container status \"6a117358d00dc47eb45e9087790aa499d845a99631f8d556a818c3dbb36a7a89\": rpc error: code = NotFound desc = could not find container \"6a117358d00dc47eb45e9087790aa499d845a99631f8d556a818c3dbb36a7a89\": container with ID starting with 6a117358d00dc47eb45e9087790aa499d845a99631f8d556a818c3dbb36a7a89 not found: ID does not exist" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.278260 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srth5\" (UniqueName: \"kubernetes.io/projected/55e60feb-8531-4c7b-9033-5d3c6a07258a-kube-api-access-srth5\") pod \"cinder-scheduler-0\" (UID: \"55e60feb-8531-4c7b-9033-5d3c6a07258a\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.292935 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aabbe361-2ef8-444e-acf2-fe48abecf27b-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-snppf\" (UID: \"aabbe361-2ef8-444e-acf2-fe48abecf27b\") " pod="openstack/dnsmasq-dns-6578955fd5-snppf" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.293033 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aabbe361-2ef8-444e-acf2-fe48abecf27b-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-snppf\" (UID: \"aabbe361-2ef8-444e-acf2-fe48abecf27b\") " pod="openstack/dnsmasq-dns-6578955fd5-snppf" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.293061 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aabbe361-2ef8-444e-acf2-fe48abecf27b-config\") pod \"dnsmasq-dns-6578955fd5-snppf\" (UID: \"aabbe361-2ef8-444e-acf2-fe48abecf27b\") " pod="openstack/dnsmasq-dns-6578955fd5-snppf" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.293092 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-695lr\" (UniqueName: \"kubernetes.io/projected/aabbe361-2ef8-444e-acf2-fe48abecf27b-kube-api-access-695lr\") pod \"dnsmasq-dns-6578955fd5-snppf\" (UID: \"aabbe361-2ef8-444e-acf2-fe48abecf27b\") " pod="openstack/dnsmasq-dns-6578955fd5-snppf" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.293111 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aabbe361-2ef8-444e-acf2-fe48abecf27b-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-snppf\" (UID: \"aabbe361-2ef8-444e-acf2-fe48abecf27b\") " pod="openstack/dnsmasq-dns-6578955fd5-snppf" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.293134 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aabbe361-2ef8-444e-acf2-fe48abecf27b-dns-svc\") pod \"dnsmasq-dns-6578955fd5-snppf\" (UID: \"aabbe361-2ef8-444e-acf2-fe48abecf27b\") " pod="openstack/dnsmasq-dns-6578955fd5-snppf" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.342997 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.361589 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-snppf"] Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.378560 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.380929 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.383754 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.393738 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.393989 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.394148 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.395174 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8e48f476-da57-4a24-a043-babdc557f7e7-run-httpd\") pod \"ceilometer-0\" (UID: \"8e48f476-da57-4a24-a043-babdc557f7e7\") " pod="openstack/ceilometer-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.395216 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8e48f476-da57-4a24-a043-babdc557f7e7-log-httpd\") pod \"ceilometer-0\" (UID: \"8e48f476-da57-4a24-a043-babdc557f7e7\") " pod="openstack/ceilometer-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.395240 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-695lr\" (UniqueName: \"kubernetes.io/projected/aabbe361-2ef8-444e-acf2-fe48abecf27b-kube-api-access-695lr\") pod \"dnsmasq-dns-6578955fd5-snppf\" (UID: \"aabbe361-2ef8-444e-acf2-fe48abecf27b\") " pod="openstack/dnsmasq-dns-6578955fd5-snppf" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.395260 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8e48f476-da57-4a24-a043-babdc557f7e7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8e48f476-da57-4a24-a043-babdc557f7e7\") " pod="openstack/ceilometer-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.395280 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aabbe361-2ef8-444e-acf2-fe48abecf27b-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-snppf\" (UID: \"aabbe361-2ef8-444e-acf2-fe48abecf27b\") " pod="openstack/dnsmasq-dns-6578955fd5-snppf" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.395304 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8hcr\" (UniqueName: \"kubernetes.io/projected/8e48f476-da57-4a24-a043-babdc557f7e7-kube-api-access-z8hcr\") pod \"ceilometer-0\" (UID: \"8e48f476-da57-4a24-a043-babdc557f7e7\") " pod="openstack/ceilometer-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.395323 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aabbe361-2ef8-444e-acf2-fe48abecf27b-dns-svc\") pod \"dnsmasq-dns-6578955fd5-snppf\" (UID: \"aabbe361-2ef8-444e-acf2-fe48abecf27b\") " pod="openstack/dnsmasq-dns-6578955fd5-snppf" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.395359 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e48f476-da57-4a24-a043-babdc557f7e7-scripts\") pod \"ceilometer-0\" (UID: \"8e48f476-da57-4a24-a043-babdc557f7e7\") " pod="openstack/ceilometer-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.395394 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aabbe361-2ef8-444e-acf2-fe48abecf27b-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-snppf\" (UID: \"aabbe361-2ef8-444e-acf2-fe48abecf27b\") " pod="openstack/dnsmasq-dns-6578955fd5-snppf" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.395470 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aabbe361-2ef8-444e-acf2-fe48abecf27b-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-snppf\" (UID: \"aabbe361-2ef8-444e-acf2-fe48abecf27b\") " pod="openstack/dnsmasq-dns-6578955fd5-snppf" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.395490 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e48f476-da57-4a24-a043-babdc557f7e7-config-data\") pod \"ceilometer-0\" (UID: \"8e48f476-da57-4a24-a043-babdc557f7e7\") " pod="openstack/ceilometer-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.395505 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e48f476-da57-4a24-a043-babdc557f7e7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8e48f476-da57-4a24-a043-babdc557f7e7\") " pod="openstack/ceilometer-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.395519 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aabbe361-2ef8-444e-acf2-fe48abecf27b-config\") pod \"dnsmasq-dns-6578955fd5-snppf\" (UID: \"aabbe361-2ef8-444e-acf2-fe48abecf27b\") " pod="openstack/dnsmasq-dns-6578955fd5-snppf" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.396352 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aabbe361-2ef8-444e-acf2-fe48abecf27b-config\") pod \"dnsmasq-dns-6578955fd5-snppf\" (UID: \"aabbe361-2ef8-444e-acf2-fe48abecf27b\") " pod="openstack/dnsmasq-dns-6578955fd5-snppf" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.396817 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aabbe361-2ef8-444e-acf2-fe48abecf27b-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-snppf\" (UID: \"aabbe361-2ef8-444e-acf2-fe48abecf27b\") " pod="openstack/dnsmasq-dns-6578955fd5-snppf" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.396932 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aabbe361-2ef8-444e-acf2-fe48abecf27b-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-snppf\" (UID: \"aabbe361-2ef8-444e-acf2-fe48abecf27b\") " pod="openstack/dnsmasq-dns-6578955fd5-snppf" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.397564 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aabbe361-2ef8-444e-acf2-fe48abecf27b-dns-svc\") pod \"dnsmasq-dns-6578955fd5-snppf\" (UID: \"aabbe361-2ef8-444e-acf2-fe48abecf27b\") " pod="openstack/dnsmasq-dns-6578955fd5-snppf" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.409449 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aabbe361-2ef8-444e-acf2-fe48abecf27b-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-snppf\" (UID: \"aabbe361-2ef8-444e-acf2-fe48abecf27b\") " pod="openstack/dnsmasq-dns-6578955fd5-snppf" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.539307 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.548132 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.550131 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.551580 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.568875 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8e48f476-da57-4a24-a043-babdc557f7e7-run-httpd\") pod \"ceilometer-0\" (UID: \"8e48f476-da57-4a24-a043-babdc557f7e7\") " pod="openstack/ceilometer-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.569210 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8e48f476-da57-4a24-a043-babdc557f7e7-log-httpd\") pod \"ceilometer-0\" (UID: \"8e48f476-da57-4a24-a043-babdc557f7e7\") " pod="openstack/ceilometer-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.569322 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8e48f476-da57-4a24-a043-babdc557f7e7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8e48f476-da57-4a24-a043-babdc557f7e7\") " pod="openstack/ceilometer-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.569464 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8hcr\" (UniqueName: \"kubernetes.io/projected/8e48f476-da57-4a24-a043-babdc557f7e7-kube-api-access-z8hcr\") pod \"ceilometer-0\" (UID: \"8e48f476-da57-4a24-a043-babdc557f7e7\") " pod="openstack/ceilometer-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.569621 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e48f476-da57-4a24-a043-babdc557f7e7-scripts\") pod \"ceilometer-0\" (UID: \"8e48f476-da57-4a24-a043-babdc557f7e7\") " pod="openstack/ceilometer-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.569943 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e48f476-da57-4a24-a043-babdc557f7e7-config-data\") pod \"ceilometer-0\" (UID: \"8e48f476-da57-4a24-a043-babdc557f7e7\") " pod="openstack/ceilometer-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.570025 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e48f476-da57-4a24-a043-babdc557f7e7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8e48f476-da57-4a24-a043-babdc557f7e7\") " pod="openstack/ceilometer-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.582063 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8e48f476-da57-4a24-a043-babdc557f7e7-run-httpd\") pod \"ceilometer-0\" (UID: \"8e48f476-da57-4a24-a043-babdc557f7e7\") " pod="openstack/ceilometer-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.582371 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8e48f476-da57-4a24-a043-babdc557f7e7-log-httpd\") pod \"ceilometer-0\" (UID: \"8e48f476-da57-4a24-a043-babdc557f7e7\") " pod="openstack/ceilometer-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.588855 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e48f476-da57-4a24-a043-babdc557f7e7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8e48f476-da57-4a24-a043-babdc557f7e7\") " pod="openstack/ceilometer-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.600845 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8e48f476-da57-4a24-a043-babdc557f7e7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8e48f476-da57-4a24-a043-babdc557f7e7\") " pod="openstack/ceilometer-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.614833 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e48f476-da57-4a24-a043-babdc557f7e7-config-data\") pod \"ceilometer-0\" (UID: \"8e48f476-da57-4a24-a043-babdc557f7e7\") " pod="openstack/ceilometer-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.615933 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8hcr\" (UniqueName: \"kubernetes.io/projected/8e48f476-da57-4a24-a043-babdc557f7e7-kube-api-access-z8hcr\") pod \"ceilometer-0\" (UID: \"8e48f476-da57-4a24-a043-babdc557f7e7\") " pod="openstack/ceilometer-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.631095 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-695lr\" (UniqueName: \"kubernetes.io/projected/aabbe361-2ef8-444e-acf2-fe48abecf27b-kube-api-access-695lr\") pod \"dnsmasq-dns-6578955fd5-snppf\" (UID: \"aabbe361-2ef8-444e-acf2-fe48abecf27b\") " pod="openstack/dnsmasq-dns-6578955fd5-snppf" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.633656 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e48f476-da57-4a24-a043-babdc557f7e7-scripts\") pod \"ceilometer-0\" (UID: \"8e48f476-da57-4a24-a043-babdc557f7e7\") " pod="openstack/ceilometer-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.711356 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.787907 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3a796b0-1b3a-44e5-a300-d095012cc571-scripts\") pod \"cinder-api-0\" (UID: \"f3a796b0-1b3a-44e5-a300-d095012cc571\") " pod="openstack/cinder-api-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.787978 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3a796b0-1b3a-44e5-a300-d095012cc571-logs\") pod \"cinder-api-0\" (UID: \"f3a796b0-1b3a-44e5-a300-d095012cc571\") " pod="openstack/cinder-api-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.788011 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3a796b0-1b3a-44e5-a300-d095012cc571-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f3a796b0-1b3a-44e5-a300-d095012cc571\") " pod="openstack/cinder-api-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.788096 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f3a796b0-1b3a-44e5-a300-d095012cc571-config-data-custom\") pod \"cinder-api-0\" (UID: \"f3a796b0-1b3a-44e5-a300-d095012cc571\") " pod="openstack/cinder-api-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.788152 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f3a796b0-1b3a-44e5-a300-d095012cc571-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f3a796b0-1b3a-44e5-a300-d095012cc571\") " pod="openstack/cinder-api-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.788178 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gd7gq\" (UniqueName: \"kubernetes.io/projected/f3a796b0-1b3a-44e5-a300-d095012cc571-kube-api-access-gd7gq\") pod \"cinder-api-0\" (UID: \"f3a796b0-1b3a-44e5-a300-d095012cc571\") " pod="openstack/cinder-api-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.788202 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3a796b0-1b3a-44e5-a300-d095012cc571-config-data\") pod \"cinder-api-0\" (UID: \"f3a796b0-1b3a-44e5-a300-d095012cc571\") " pod="openstack/cinder-api-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.860362 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-dd8b84bc8-9g27k" event={"ID":"ca62cb72-d397-4ee8-a7c4-15d831f87776","Type":"ContainerStarted","Data":"d938a457c8fe17cde00b625b8f7c1a79193d9b6ab2bbe35882d849ca8039facf"} Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.860612 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-dd8b84bc8-9g27k" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.869938 4790 generic.go:334] "Generic (PLEG): container finished" podID="c96afe58-d2d2-4e9d-a1f7-67ecd78bf265" containerID="9ac61b043b7a62382b037fd6716ed1875cc67c39b24ab18087b0d7f9ff34e2bf" exitCode=0 Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.870083 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-848cf88cfc-8hxrv" podUID="c96afe58-d2d2-4e9d-a1f7-67ecd78bf265" containerName="dnsmasq-dns" containerID="cri-o://66fc4011cc52270b543feacd280ee055e4991d415cd3b36c27711b1e1cb82862" gracePeriod=10 Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.870408 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-8hxrv" event={"ID":"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265","Type":"ContainerDied","Data":"9ac61b043b7a62382b037fd6716ed1875cc67c39b24ab18087b0d7f9ff34e2bf"} Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.870445 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-848cf88cfc-8hxrv" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.870456 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-8hxrv" event={"ID":"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265","Type":"ContainerStarted","Data":"66fc4011cc52270b543feacd280ee055e4991d415cd3b36c27711b1e1cb82862"} Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.874409 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-snppf" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.888157 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-dd8b84bc8-9g27k" podStartSLOduration=2.888141456 podStartE2EDuration="2.888141456s" podCreationTimestamp="2025-11-24 13:34:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:34:07.878266365 +0000 UTC m=+1296.258160027" watchObservedRunningTime="2025-11-24 13:34:07.888141456 +0000 UTC m=+1296.268035118" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.900458 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f3a796b0-1b3a-44e5-a300-d095012cc571-config-data-custom\") pod \"cinder-api-0\" (UID: \"f3a796b0-1b3a-44e5-a300-d095012cc571\") " pod="openstack/cinder-api-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.900621 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f3a796b0-1b3a-44e5-a300-d095012cc571-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f3a796b0-1b3a-44e5-a300-d095012cc571\") " pod="openstack/cinder-api-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.900647 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gd7gq\" (UniqueName: \"kubernetes.io/projected/f3a796b0-1b3a-44e5-a300-d095012cc571-kube-api-access-gd7gq\") pod \"cinder-api-0\" (UID: \"f3a796b0-1b3a-44e5-a300-d095012cc571\") " pod="openstack/cinder-api-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.900676 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3a796b0-1b3a-44e5-a300-d095012cc571-config-data\") pod \"cinder-api-0\" (UID: \"f3a796b0-1b3a-44e5-a300-d095012cc571\") " pod="openstack/cinder-api-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.900793 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3a796b0-1b3a-44e5-a300-d095012cc571-scripts\") pod \"cinder-api-0\" (UID: \"f3a796b0-1b3a-44e5-a300-d095012cc571\") " pod="openstack/cinder-api-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.900857 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3a796b0-1b3a-44e5-a300-d095012cc571-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f3a796b0-1b3a-44e5-a300-d095012cc571\") " pod="openstack/cinder-api-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.900875 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3a796b0-1b3a-44e5-a300-d095012cc571-logs\") pod \"cinder-api-0\" (UID: \"f3a796b0-1b3a-44e5-a300-d095012cc571\") " pod="openstack/cinder-api-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.903443 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f3a796b0-1b3a-44e5-a300-d095012cc571-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f3a796b0-1b3a-44e5-a300-d095012cc571\") " pod="openstack/cinder-api-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.904542 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3a796b0-1b3a-44e5-a300-d095012cc571-logs\") pod \"cinder-api-0\" (UID: \"f3a796b0-1b3a-44e5-a300-d095012cc571\") " pod="openstack/cinder-api-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.911602 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3a796b0-1b3a-44e5-a300-d095012cc571-scripts\") pod \"cinder-api-0\" (UID: \"f3a796b0-1b3a-44e5-a300-d095012cc571\") " pod="openstack/cinder-api-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.911779 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f3a796b0-1b3a-44e5-a300-d095012cc571-config-data-custom\") pod \"cinder-api-0\" (UID: \"f3a796b0-1b3a-44e5-a300-d095012cc571\") " pod="openstack/cinder-api-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.915398 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3a796b0-1b3a-44e5-a300-d095012cc571-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f3a796b0-1b3a-44e5-a300-d095012cc571\") " pod="openstack/cinder-api-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.920346 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3a796b0-1b3a-44e5-a300-d095012cc571-config-data\") pod \"cinder-api-0\" (UID: \"f3a796b0-1b3a-44e5-a300-d095012cc571\") " pod="openstack/cinder-api-0" Nov 24 13:34:07 crc kubenswrapper[4790]: I1124 13:34:07.929418 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gd7gq\" (UniqueName: \"kubernetes.io/projected/f3a796b0-1b3a-44e5-a300-d095012cc571-kube-api-access-gd7gq\") pod \"cinder-api-0\" (UID: \"f3a796b0-1b3a-44e5-a300-d095012cc571\") " pod="openstack/cinder-api-0" Nov 24 13:34:08 crc kubenswrapper[4790]: I1124 13:34:08.028257 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 13:34:08 crc kubenswrapper[4790]: I1124 13:34:08.067361 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-848cf88cfc-8hxrv" podStartSLOduration=3.067338953 podStartE2EDuration="3.067338953s" podCreationTimestamp="2025-11-24 13:34:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:34:07.903785391 +0000 UTC m=+1296.283679053" watchObservedRunningTime="2025-11-24 13:34:08.067338953 +0000 UTC m=+1296.447232615" Nov 24 13:34:08 crc kubenswrapper[4790]: I1124 13:34:08.085762 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 13:34:08 crc kubenswrapper[4790]: I1124 13:34:08.135405 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:34:08 crc kubenswrapper[4790]: W1124 13:34:08.160712 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e48f476_da57_4a24_a043_babdc557f7e7.slice/crio-6e3d7ebb68ed6a22674d83e672a6047909c076e8a6b433bf10292081cfce42c0 WatchSource:0}: Error finding container 6e3d7ebb68ed6a22674d83e672a6047909c076e8a6b433bf10292081cfce42c0: Status 404 returned error can't find the container with id 6e3d7ebb68ed6a22674d83e672a6047909c076e8a6b433bf10292081cfce42c0 Nov 24 13:34:08 crc kubenswrapper[4790]: I1124 13:34:08.334305 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d574c9c-ecb5-4b8c-b0b3-a058c141c09f" path="/var/lib/kubelet/pods/4d574c9c-ecb5-4b8c-b0b3-a058c141c09f/volumes" Nov 24 13:34:08 crc kubenswrapper[4790]: I1124 13:34:08.432865 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-snppf"] Nov 24 13:34:08 crc kubenswrapper[4790]: I1124 13:34:08.564579 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 13:34:08 crc kubenswrapper[4790]: I1124 13:34:08.900574 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"55e60feb-8531-4c7b-9033-5d3c6a07258a","Type":"ContainerStarted","Data":"6b859b296a7f4f3fb03c2300829268a2d421476851806dc7575d838952eef3c1"} Nov 24 13:34:08 crc kubenswrapper[4790]: I1124 13:34:08.904919 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8e48f476-da57-4a24-a043-babdc557f7e7","Type":"ContainerStarted","Data":"6e3d7ebb68ed6a22674d83e672a6047909c076e8a6b433bf10292081cfce42c0"} Nov 24 13:34:08 crc kubenswrapper[4790]: I1124 13:34:08.906622 4790 generic.go:334] "Generic (PLEG): container finished" podID="c96afe58-d2d2-4e9d-a1f7-67ecd78bf265" containerID="66fc4011cc52270b543feacd280ee055e4991d415cd3b36c27711b1e1cb82862" exitCode=0 Nov 24 13:34:08 crc kubenswrapper[4790]: I1124 13:34:08.907580 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-8hxrv" event={"ID":"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265","Type":"ContainerDied","Data":"66fc4011cc52270b543feacd280ee055e4991d415cd3b36c27711b1e1cb82862"} Nov 24 13:34:08 crc kubenswrapper[4790]: I1124 13:34:08.907621 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-dd8b84bc8-9g27k" Nov 24 13:34:09 crc kubenswrapper[4790]: W1124 13:34:09.044324 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaabbe361_2ef8_444e_acf2_fe48abecf27b.slice/crio-6450b3b94bcfe1c5cb8992047032bb4da107a91cd1b01885dd410144995f21c8 WatchSource:0}: Error finding container 6450b3b94bcfe1c5cb8992047032bb4da107a91cd1b01885dd410144995f21c8: Status 404 returned error can't find the container with id 6450b3b94bcfe1c5cb8992047032bb4da107a91cd1b01885dd410144995f21c8 Nov 24 13:34:09 crc kubenswrapper[4790]: W1124 13:34:09.049706 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf3a796b0_1b3a_44e5_a300_d095012cc571.slice/crio-b1518b1522849af68833e3b723ba32be587f9bad00875343734c015ac3564cac WatchSource:0}: Error finding container b1518b1522849af68833e3b723ba32be587f9bad00875343734c015ac3564cac: Status 404 returned error can't find the container with id b1518b1522849af68833e3b723ba32be587f9bad00875343734c015ac3564cac Nov 24 13:34:09 crc kubenswrapper[4790]: I1124 13:34:09.185494 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-8hxrv" Nov 24 13:34:09 crc kubenswrapper[4790]: I1124 13:34:09.225811 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfn8l\" (UniqueName: \"kubernetes.io/projected/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-kube-api-access-tfn8l\") pod \"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265\" (UID: \"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265\") " Nov 24 13:34:09 crc kubenswrapper[4790]: I1124 13:34:09.226006 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-dns-svc\") pod \"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265\" (UID: \"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265\") " Nov 24 13:34:09 crc kubenswrapper[4790]: I1124 13:34:09.226128 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-config\") pod \"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265\" (UID: \"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265\") " Nov 24 13:34:09 crc kubenswrapper[4790]: I1124 13:34:09.226184 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-dns-swift-storage-0\") pod \"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265\" (UID: \"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265\") " Nov 24 13:34:09 crc kubenswrapper[4790]: I1124 13:34:09.226226 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-ovsdbserver-nb\") pod \"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265\" (UID: \"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265\") " Nov 24 13:34:09 crc kubenswrapper[4790]: I1124 13:34:09.226250 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-ovsdbserver-sb\") pod \"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265\" (UID: \"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265\") " Nov 24 13:34:09 crc kubenswrapper[4790]: I1124 13:34:09.239029 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-kube-api-access-tfn8l" (OuterVolumeSpecName: "kube-api-access-tfn8l") pod "c96afe58-d2d2-4e9d-a1f7-67ecd78bf265" (UID: "c96afe58-d2d2-4e9d-a1f7-67ecd78bf265"). InnerVolumeSpecName "kube-api-access-tfn8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:09 crc kubenswrapper[4790]: I1124 13:34:09.283691 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c96afe58-d2d2-4e9d-a1f7-67ecd78bf265" (UID: "c96afe58-d2d2-4e9d-a1f7-67ecd78bf265"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:09 crc kubenswrapper[4790]: I1124 13:34:09.286526 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c96afe58-d2d2-4e9d-a1f7-67ecd78bf265" (UID: "c96afe58-d2d2-4e9d-a1f7-67ecd78bf265"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:09 crc kubenswrapper[4790]: I1124 13:34:09.291729 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c96afe58-d2d2-4e9d-a1f7-67ecd78bf265" (UID: "c96afe58-d2d2-4e9d-a1f7-67ecd78bf265"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:09 crc kubenswrapper[4790]: I1124 13:34:09.293118 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-config" (OuterVolumeSpecName: "config") pod "c96afe58-d2d2-4e9d-a1f7-67ecd78bf265" (UID: "c96afe58-d2d2-4e9d-a1f7-67ecd78bf265"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:09 crc kubenswrapper[4790]: I1124 13:34:09.295900 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c96afe58-d2d2-4e9d-a1f7-67ecd78bf265" (UID: "c96afe58-d2d2-4e9d-a1f7-67ecd78bf265"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:09 crc kubenswrapper[4790]: I1124 13:34:09.327962 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:09 crc kubenswrapper[4790]: I1124 13:34:09.328001 4790 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:09 crc kubenswrapper[4790]: I1124 13:34:09.328017 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:09 crc kubenswrapper[4790]: I1124 13:34:09.328033 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:09 crc kubenswrapper[4790]: I1124 13:34:09.328045 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfn8l\" (UniqueName: \"kubernetes.io/projected/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-kube-api-access-tfn8l\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:09 crc kubenswrapper[4790]: I1124 13:34:09.328056 4790 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:09 crc kubenswrapper[4790]: I1124 13:34:09.927763 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-58b5bcbcb-b6kbh" event={"ID":"a2ee8af7-26fb-4652-be37-594db62f1146","Type":"ContainerStarted","Data":"2b38eafa7ed5a3895c1dd40a15a595d2c6fed37ac22fb3628de8e8a277d110ed"} Nov 24 13:34:09 crc kubenswrapper[4790]: I1124 13:34:09.932174 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8e48f476-da57-4a24-a043-babdc557f7e7","Type":"ContainerStarted","Data":"c8368f4f05ea5c50f6456f15ba36f75f7e4acde31bd265dde01d9744c3e3c454"} Nov 24 13:34:09 crc kubenswrapper[4790]: I1124 13:34:09.939920 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f3a796b0-1b3a-44e5-a300-d095012cc571","Type":"ContainerStarted","Data":"b1518b1522849af68833e3b723ba32be587f9bad00875343734c015ac3564cac"} Nov 24 13:34:09 crc kubenswrapper[4790]: I1124 13:34:09.944600 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-845f9d57f-2s48z" event={"ID":"3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c","Type":"ContainerStarted","Data":"9191be68775994f407e5cef4ebbca1e4aed9cfffa519ea7b28b2742579b483ad"} Nov 24 13:34:09 crc kubenswrapper[4790]: I1124 13:34:09.954016 4790 generic.go:334] "Generic (PLEG): container finished" podID="aabbe361-2ef8-444e-acf2-fe48abecf27b" containerID="2d446b0816edf4925a4d9a1f9987de79a211c49b58ca1683a329e8584d8616cb" exitCode=0 Nov 24 13:34:09 crc kubenswrapper[4790]: I1124 13:34:09.954115 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-snppf" event={"ID":"aabbe361-2ef8-444e-acf2-fe48abecf27b","Type":"ContainerDied","Data":"2d446b0816edf4925a4d9a1f9987de79a211c49b58ca1683a329e8584d8616cb"} Nov 24 13:34:09 crc kubenswrapper[4790]: I1124 13:34:09.954139 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-snppf" event={"ID":"aabbe361-2ef8-444e-acf2-fe48abecf27b","Type":"ContainerStarted","Data":"6450b3b94bcfe1c5cb8992047032bb4da107a91cd1b01885dd410144995f21c8"} Nov 24 13:34:09 crc kubenswrapper[4790]: I1124 13:34:09.959137 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-8hxrv" event={"ID":"c96afe58-d2d2-4e9d-a1f7-67ecd78bf265","Type":"ContainerDied","Data":"16addf5fde61c223e1beabcd551c64f32633b49ade27ffd3df96a8a0f315654e"} Nov 24 13:34:09 crc kubenswrapper[4790]: I1124 13:34:09.959193 4790 scope.go:117] "RemoveContainer" containerID="66fc4011cc52270b543feacd280ee055e4991d415cd3b36c27711b1e1cb82862" Nov 24 13:34:09 crc kubenswrapper[4790]: I1124 13:34:09.959818 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-8hxrv" Nov 24 13:34:10 crc kubenswrapper[4790]: I1124 13:34:10.010772 4790 scope.go:117] "RemoveContainer" containerID="9ac61b043b7a62382b037fd6716ed1875cc67c39b24ab18087b0d7f9ff34e2bf" Nov 24 13:34:10 crc kubenswrapper[4790]: I1124 13:34:10.018134 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-8hxrv"] Nov 24 13:34:10 crc kubenswrapper[4790]: I1124 13:34:10.027333 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-8hxrv"] Nov 24 13:34:10 crc kubenswrapper[4790]: I1124 13:34:10.350650 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c96afe58-d2d2-4e9d-a1f7-67ecd78bf265" path="/var/lib/kubelet/pods/c96afe58-d2d2-4e9d-a1f7-67ecd78bf265/volumes" Nov 24 13:34:10 crc kubenswrapper[4790]: I1124 13:34:10.592545 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 13:34:10 crc kubenswrapper[4790]: I1124 13:34:10.983011 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-58b5bcbcb-b6kbh" event={"ID":"a2ee8af7-26fb-4652-be37-594db62f1146","Type":"ContainerStarted","Data":"ef998419eb802553e1ad108205b17dff1eada33d519b61e4493aea07e0efefc0"} Nov 24 13:34:10 crc kubenswrapper[4790]: I1124 13:34:10.985279 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"55e60feb-8531-4c7b-9033-5d3c6a07258a","Type":"ContainerStarted","Data":"acbc1a5bb162c73d8e0386e58d22675b4882165da87252b0b4d8bb67dff52b0a"} Nov 24 13:34:10 crc kubenswrapper[4790]: I1124 13:34:10.992913 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8e48f476-da57-4a24-a043-babdc557f7e7","Type":"ContainerStarted","Data":"1d5fcf4525bd2dc5f89c29d6aa5e8981e4016fd7b3d6c292093e5f559c8905d7"} Nov 24 13:34:10 crc kubenswrapper[4790]: I1124 13:34:10.996676 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f3a796b0-1b3a-44e5-a300-d095012cc571","Type":"ContainerStarted","Data":"55c887f9a40216c24e10a416b3ebd75cb336ba29eebc061496145267c0e23866"} Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.001119 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-845f9d57f-2s48z" event={"ID":"3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c","Type":"ContainerStarted","Data":"32dece0aef9f5c5f700993f3c33e9dc9b09465378d0d7abd138bd90e7d5c6319"} Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.007828 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-snppf" event={"ID":"aabbe361-2ef8-444e-acf2-fe48abecf27b","Type":"ContainerStarted","Data":"6bee98f4451efc6d196589802f29d5f46905beb31c259a7f4c4ee20b26895450"} Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.008656 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6578955fd5-snppf" Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.013255 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-58b5bcbcb-b6kbh" podStartSLOduration=3.202191729 podStartE2EDuration="6.01323485s" podCreationTimestamp="2025-11-24 13:34:05 +0000 UTC" firstStartedPulling="2025-11-24 13:34:06.335619779 +0000 UTC m=+1294.715513441" lastFinishedPulling="2025-11-24 13:34:09.14666289 +0000 UTC m=+1297.526556562" observedRunningTime="2025-11-24 13:34:11.00341576 +0000 UTC m=+1299.383309442" watchObservedRunningTime="2025-11-24 13:34:11.01323485 +0000 UTC m=+1299.393128512" Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.030415 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-845f9d57f-2s48z" podStartSLOduration=3.20577693 podStartE2EDuration="6.030399608s" podCreationTimestamp="2025-11-24 13:34:05 +0000 UTC" firstStartedPulling="2025-11-24 13:34:06.333524369 +0000 UTC m=+1294.713418031" lastFinishedPulling="2025-11-24 13:34:09.158147047 +0000 UTC m=+1297.538040709" observedRunningTime="2025-11-24 13:34:11.028683639 +0000 UTC m=+1299.408577301" watchObservedRunningTime="2025-11-24 13:34:11.030399608 +0000 UTC m=+1299.410293270" Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.071808 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6578955fd5-snppf" podStartSLOduration=4.071790695 podStartE2EDuration="4.071790695s" podCreationTimestamp="2025-11-24 13:34:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:34:11.070984252 +0000 UTC m=+1299.450877914" watchObservedRunningTime="2025-11-24 13:34:11.071790695 +0000 UTC m=+1299.451684357" Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.727641 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7d69688466-r2bq9"] Nov 24 13:34:11 crc kubenswrapper[4790]: E1124 13:34:11.728373 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c96afe58-d2d2-4e9d-a1f7-67ecd78bf265" containerName="dnsmasq-dns" Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.728396 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="c96afe58-d2d2-4e9d-a1f7-67ecd78bf265" containerName="dnsmasq-dns" Nov 24 13:34:11 crc kubenswrapper[4790]: E1124 13:34:11.728428 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c96afe58-d2d2-4e9d-a1f7-67ecd78bf265" containerName="init" Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.728435 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="c96afe58-d2d2-4e9d-a1f7-67ecd78bf265" containerName="init" Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.728594 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="c96afe58-d2d2-4e9d-a1f7-67ecd78bf265" containerName="dnsmasq-dns" Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.729606 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7d69688466-r2bq9" Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.733385 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.743744 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.765953 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7d69688466-r2bq9"] Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.806828 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8311dfa1-6e87-493d-92e0-4eb69c792afe-config-data\") pod \"barbican-api-7d69688466-r2bq9\" (UID: \"8311dfa1-6e87-493d-92e0-4eb69c792afe\") " pod="openstack/barbican-api-7d69688466-r2bq9" Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.806896 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8311dfa1-6e87-493d-92e0-4eb69c792afe-logs\") pod \"barbican-api-7d69688466-r2bq9\" (UID: \"8311dfa1-6e87-493d-92e0-4eb69c792afe\") " pod="openstack/barbican-api-7d69688466-r2bq9" Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.806951 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8311dfa1-6e87-493d-92e0-4eb69c792afe-combined-ca-bundle\") pod \"barbican-api-7d69688466-r2bq9\" (UID: \"8311dfa1-6e87-493d-92e0-4eb69c792afe\") " pod="openstack/barbican-api-7d69688466-r2bq9" Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.807007 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8311dfa1-6e87-493d-92e0-4eb69c792afe-internal-tls-certs\") pod \"barbican-api-7d69688466-r2bq9\" (UID: \"8311dfa1-6e87-493d-92e0-4eb69c792afe\") " pod="openstack/barbican-api-7d69688466-r2bq9" Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.807031 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nr64t\" (UniqueName: \"kubernetes.io/projected/8311dfa1-6e87-493d-92e0-4eb69c792afe-kube-api-access-nr64t\") pod \"barbican-api-7d69688466-r2bq9\" (UID: \"8311dfa1-6e87-493d-92e0-4eb69c792afe\") " pod="openstack/barbican-api-7d69688466-r2bq9" Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.807078 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8311dfa1-6e87-493d-92e0-4eb69c792afe-config-data-custom\") pod \"barbican-api-7d69688466-r2bq9\" (UID: \"8311dfa1-6e87-493d-92e0-4eb69c792afe\") " pod="openstack/barbican-api-7d69688466-r2bq9" Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.807095 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8311dfa1-6e87-493d-92e0-4eb69c792afe-public-tls-certs\") pod \"barbican-api-7d69688466-r2bq9\" (UID: \"8311dfa1-6e87-493d-92e0-4eb69c792afe\") " pod="openstack/barbican-api-7d69688466-r2bq9" Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.908457 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8311dfa1-6e87-493d-92e0-4eb69c792afe-config-data-custom\") pod \"barbican-api-7d69688466-r2bq9\" (UID: \"8311dfa1-6e87-493d-92e0-4eb69c792afe\") " pod="openstack/barbican-api-7d69688466-r2bq9" Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.908503 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8311dfa1-6e87-493d-92e0-4eb69c792afe-public-tls-certs\") pod \"barbican-api-7d69688466-r2bq9\" (UID: \"8311dfa1-6e87-493d-92e0-4eb69c792afe\") " pod="openstack/barbican-api-7d69688466-r2bq9" Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.908572 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8311dfa1-6e87-493d-92e0-4eb69c792afe-config-data\") pod \"barbican-api-7d69688466-r2bq9\" (UID: \"8311dfa1-6e87-493d-92e0-4eb69c792afe\") " pod="openstack/barbican-api-7d69688466-r2bq9" Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.908615 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8311dfa1-6e87-493d-92e0-4eb69c792afe-logs\") pod \"barbican-api-7d69688466-r2bq9\" (UID: \"8311dfa1-6e87-493d-92e0-4eb69c792afe\") " pod="openstack/barbican-api-7d69688466-r2bq9" Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.908666 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8311dfa1-6e87-493d-92e0-4eb69c792afe-combined-ca-bundle\") pod \"barbican-api-7d69688466-r2bq9\" (UID: \"8311dfa1-6e87-493d-92e0-4eb69c792afe\") " pod="openstack/barbican-api-7d69688466-r2bq9" Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.908703 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8311dfa1-6e87-493d-92e0-4eb69c792afe-internal-tls-certs\") pod \"barbican-api-7d69688466-r2bq9\" (UID: \"8311dfa1-6e87-493d-92e0-4eb69c792afe\") " pod="openstack/barbican-api-7d69688466-r2bq9" Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.908730 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nr64t\" (UniqueName: \"kubernetes.io/projected/8311dfa1-6e87-493d-92e0-4eb69c792afe-kube-api-access-nr64t\") pod \"barbican-api-7d69688466-r2bq9\" (UID: \"8311dfa1-6e87-493d-92e0-4eb69c792afe\") " pod="openstack/barbican-api-7d69688466-r2bq9" Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.909454 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8311dfa1-6e87-493d-92e0-4eb69c792afe-logs\") pod \"barbican-api-7d69688466-r2bq9\" (UID: \"8311dfa1-6e87-493d-92e0-4eb69c792afe\") " pod="openstack/barbican-api-7d69688466-r2bq9" Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.913068 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8311dfa1-6e87-493d-92e0-4eb69c792afe-config-data-custom\") pod \"barbican-api-7d69688466-r2bq9\" (UID: \"8311dfa1-6e87-493d-92e0-4eb69c792afe\") " pod="openstack/barbican-api-7d69688466-r2bq9" Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.913584 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8311dfa1-6e87-493d-92e0-4eb69c792afe-public-tls-certs\") pod \"barbican-api-7d69688466-r2bq9\" (UID: \"8311dfa1-6e87-493d-92e0-4eb69c792afe\") " pod="openstack/barbican-api-7d69688466-r2bq9" Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.914798 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8311dfa1-6e87-493d-92e0-4eb69c792afe-internal-tls-certs\") pod \"barbican-api-7d69688466-r2bq9\" (UID: \"8311dfa1-6e87-493d-92e0-4eb69c792afe\") " pod="openstack/barbican-api-7d69688466-r2bq9" Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.916945 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8311dfa1-6e87-493d-92e0-4eb69c792afe-config-data\") pod \"barbican-api-7d69688466-r2bq9\" (UID: \"8311dfa1-6e87-493d-92e0-4eb69c792afe\") " pod="openstack/barbican-api-7d69688466-r2bq9" Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.926140 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8311dfa1-6e87-493d-92e0-4eb69c792afe-combined-ca-bundle\") pod \"barbican-api-7d69688466-r2bq9\" (UID: \"8311dfa1-6e87-493d-92e0-4eb69c792afe\") " pod="openstack/barbican-api-7d69688466-r2bq9" Nov 24 13:34:11 crc kubenswrapper[4790]: I1124 13:34:11.929673 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nr64t\" (UniqueName: \"kubernetes.io/projected/8311dfa1-6e87-493d-92e0-4eb69c792afe-kube-api-access-nr64t\") pod \"barbican-api-7d69688466-r2bq9\" (UID: \"8311dfa1-6e87-493d-92e0-4eb69c792afe\") " pod="openstack/barbican-api-7d69688466-r2bq9" Nov 24 13:34:12 crc kubenswrapper[4790]: I1124 13:34:12.018179 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"55e60feb-8531-4c7b-9033-5d3c6a07258a","Type":"ContainerStarted","Data":"31df2983c371d070c392694c8f305220de213a9f32dff5c05e09472f74b9a653"} Nov 24 13:34:12 crc kubenswrapper[4790]: I1124 13:34:12.022508 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f3a796b0-1b3a-44e5-a300-d095012cc571","Type":"ContainerStarted","Data":"b3c13279d4abc89b6ca528942a7b0cacfd3ae91cdf24bc9ecb3b5d28fa99b218"} Nov 24 13:34:12 crc kubenswrapper[4790]: I1124 13:34:12.023579 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="f3a796b0-1b3a-44e5-a300-d095012cc571" containerName="cinder-api" containerID="cri-o://b3c13279d4abc89b6ca528942a7b0cacfd3ae91cdf24bc9ecb3b5d28fa99b218" gracePeriod=30 Nov 24 13:34:12 crc kubenswrapper[4790]: I1124 13:34:12.023559 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="f3a796b0-1b3a-44e5-a300-d095012cc571" containerName="cinder-api-log" containerID="cri-o://55c887f9a40216c24e10a416b3ebd75cb336ba29eebc061496145267c0e23866" gracePeriod=30 Nov 24 13:34:12 crc kubenswrapper[4790]: I1124 13:34:12.039027 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.346152467 podStartE2EDuration="6.039005755s" podCreationTimestamp="2025-11-24 13:34:06 +0000 UTC" firstStartedPulling="2025-11-24 13:34:08.134964086 +0000 UTC m=+1296.514857748" lastFinishedPulling="2025-11-24 13:34:09.827817374 +0000 UTC m=+1298.207711036" observedRunningTime="2025-11-24 13:34:12.037725039 +0000 UTC m=+1300.417618701" watchObservedRunningTime="2025-11-24 13:34:12.039005755 +0000 UTC m=+1300.418899417" Nov 24 13:34:12 crc kubenswrapper[4790]: I1124 13:34:12.061325 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.061305419 podStartE2EDuration="5.061305419s" podCreationTimestamp="2025-11-24 13:34:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:34:12.054815835 +0000 UTC m=+1300.434709497" watchObservedRunningTime="2025-11-24 13:34:12.061305419 +0000 UTC m=+1300.441199081" Nov 24 13:34:12 crc kubenswrapper[4790]: I1124 13:34:12.064515 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7d69688466-r2bq9" Nov 24 13:34:12 crc kubenswrapper[4790]: I1124 13:34:12.394420 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 24 13:34:12 crc kubenswrapper[4790]: I1124 13:34:12.579082 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7d69688466-r2bq9"] Nov 24 13:34:12 crc kubenswrapper[4790]: I1124 13:34:12.670130 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 13:34:12 crc kubenswrapper[4790]: I1124 13:34:12.846718 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f3a796b0-1b3a-44e5-a300-d095012cc571-config-data-custom\") pod \"f3a796b0-1b3a-44e5-a300-d095012cc571\" (UID: \"f3a796b0-1b3a-44e5-a300-d095012cc571\") " Nov 24 13:34:12 crc kubenswrapper[4790]: I1124 13:34:12.846768 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3a796b0-1b3a-44e5-a300-d095012cc571-combined-ca-bundle\") pod \"f3a796b0-1b3a-44e5-a300-d095012cc571\" (UID: \"f3a796b0-1b3a-44e5-a300-d095012cc571\") " Nov 24 13:34:12 crc kubenswrapper[4790]: I1124 13:34:12.846880 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3a796b0-1b3a-44e5-a300-d095012cc571-logs\") pod \"f3a796b0-1b3a-44e5-a300-d095012cc571\" (UID: \"f3a796b0-1b3a-44e5-a300-d095012cc571\") " Nov 24 13:34:12 crc kubenswrapper[4790]: I1124 13:34:12.846945 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3a796b0-1b3a-44e5-a300-d095012cc571-config-data\") pod \"f3a796b0-1b3a-44e5-a300-d095012cc571\" (UID: \"f3a796b0-1b3a-44e5-a300-d095012cc571\") " Nov 24 13:34:12 crc kubenswrapper[4790]: I1124 13:34:12.846986 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gd7gq\" (UniqueName: \"kubernetes.io/projected/f3a796b0-1b3a-44e5-a300-d095012cc571-kube-api-access-gd7gq\") pod \"f3a796b0-1b3a-44e5-a300-d095012cc571\" (UID: \"f3a796b0-1b3a-44e5-a300-d095012cc571\") " Nov 24 13:34:12 crc kubenswrapper[4790]: I1124 13:34:12.847090 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f3a796b0-1b3a-44e5-a300-d095012cc571-etc-machine-id\") pod \"f3a796b0-1b3a-44e5-a300-d095012cc571\" (UID: \"f3a796b0-1b3a-44e5-a300-d095012cc571\") " Nov 24 13:34:12 crc kubenswrapper[4790]: I1124 13:34:12.847233 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3a796b0-1b3a-44e5-a300-d095012cc571-scripts\") pod \"f3a796b0-1b3a-44e5-a300-d095012cc571\" (UID: \"f3a796b0-1b3a-44e5-a300-d095012cc571\") " Nov 24 13:34:12 crc kubenswrapper[4790]: I1124 13:34:12.847251 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3a796b0-1b3a-44e5-a300-d095012cc571-logs" (OuterVolumeSpecName: "logs") pod "f3a796b0-1b3a-44e5-a300-d095012cc571" (UID: "f3a796b0-1b3a-44e5-a300-d095012cc571"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:34:12 crc kubenswrapper[4790]: I1124 13:34:12.847260 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f3a796b0-1b3a-44e5-a300-d095012cc571-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f3a796b0-1b3a-44e5-a300-d095012cc571" (UID: "f3a796b0-1b3a-44e5-a300-d095012cc571"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:34:12 crc kubenswrapper[4790]: I1124 13:34:12.847757 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3a796b0-1b3a-44e5-a300-d095012cc571-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:12 crc kubenswrapper[4790]: I1124 13:34:12.847783 4790 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f3a796b0-1b3a-44e5-a300-d095012cc571-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:12 crc kubenswrapper[4790]: I1124 13:34:12.850847 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3a796b0-1b3a-44e5-a300-d095012cc571-kube-api-access-gd7gq" (OuterVolumeSpecName: "kube-api-access-gd7gq") pod "f3a796b0-1b3a-44e5-a300-d095012cc571" (UID: "f3a796b0-1b3a-44e5-a300-d095012cc571"). InnerVolumeSpecName "kube-api-access-gd7gq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:12 crc kubenswrapper[4790]: I1124 13:34:12.851381 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3a796b0-1b3a-44e5-a300-d095012cc571-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f3a796b0-1b3a-44e5-a300-d095012cc571" (UID: "f3a796b0-1b3a-44e5-a300-d095012cc571"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:12 crc kubenswrapper[4790]: I1124 13:34:12.854458 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3a796b0-1b3a-44e5-a300-d095012cc571-scripts" (OuterVolumeSpecName: "scripts") pod "f3a796b0-1b3a-44e5-a300-d095012cc571" (UID: "f3a796b0-1b3a-44e5-a300-d095012cc571"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:12 crc kubenswrapper[4790]: I1124 13:34:12.902411 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3a796b0-1b3a-44e5-a300-d095012cc571-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f3a796b0-1b3a-44e5-a300-d095012cc571" (UID: "f3a796b0-1b3a-44e5-a300-d095012cc571"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:12 crc kubenswrapper[4790]: I1124 13:34:12.911596 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3a796b0-1b3a-44e5-a300-d095012cc571-config-data" (OuterVolumeSpecName: "config-data") pod "f3a796b0-1b3a-44e5-a300-d095012cc571" (UID: "f3a796b0-1b3a-44e5-a300-d095012cc571"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:12 crc kubenswrapper[4790]: I1124 13:34:12.949152 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3a796b0-1b3a-44e5-a300-d095012cc571-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:12 crc kubenswrapper[4790]: I1124 13:34:12.949186 4790 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f3a796b0-1b3a-44e5-a300-d095012cc571-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:12 crc kubenswrapper[4790]: I1124 13:34:12.949199 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3a796b0-1b3a-44e5-a300-d095012cc571-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:12 crc kubenswrapper[4790]: I1124 13:34:12.949209 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3a796b0-1b3a-44e5-a300-d095012cc571-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:12 crc kubenswrapper[4790]: I1124 13:34:12.949218 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gd7gq\" (UniqueName: \"kubernetes.io/projected/f3a796b0-1b3a-44e5-a300-d095012cc571-kube-api-access-gd7gq\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.046965 4790 generic.go:334] "Generic (PLEG): container finished" podID="f3a796b0-1b3a-44e5-a300-d095012cc571" containerID="b3c13279d4abc89b6ca528942a7b0cacfd3ae91cdf24bc9ecb3b5d28fa99b218" exitCode=0 Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.047003 4790 generic.go:334] "Generic (PLEG): container finished" podID="f3a796b0-1b3a-44e5-a300-d095012cc571" containerID="55c887f9a40216c24e10a416b3ebd75cb336ba29eebc061496145267c0e23866" exitCode=143 Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.047050 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f3a796b0-1b3a-44e5-a300-d095012cc571","Type":"ContainerDied","Data":"b3c13279d4abc89b6ca528942a7b0cacfd3ae91cdf24bc9ecb3b5d28fa99b218"} Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.047082 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f3a796b0-1b3a-44e5-a300-d095012cc571","Type":"ContainerDied","Data":"55c887f9a40216c24e10a416b3ebd75cb336ba29eebc061496145267c0e23866"} Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.047096 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f3a796b0-1b3a-44e5-a300-d095012cc571","Type":"ContainerDied","Data":"b1518b1522849af68833e3b723ba32be587f9bad00875343734c015ac3564cac"} Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.047114 4790 scope.go:117] "RemoveContainer" containerID="b3c13279d4abc89b6ca528942a7b0cacfd3ae91cdf24bc9ecb3b5d28fa99b218" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.047256 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.052914 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7d69688466-r2bq9" event={"ID":"8311dfa1-6e87-493d-92e0-4eb69c792afe","Type":"ContainerStarted","Data":"5a96182e01bd62e46435ce7bc8d7492b1b623ec41db4e9fa84ec0792133db77d"} Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.053243 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7d69688466-r2bq9" event={"ID":"8311dfa1-6e87-493d-92e0-4eb69c792afe","Type":"ContainerStarted","Data":"fd82fc411383d3237786b63929394f537a76680df3850407c10f53055832007b"} Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.058052 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8e48f476-da57-4a24-a043-babdc557f7e7","Type":"ContainerStarted","Data":"f1b98a50d873e693fb3407704285e40ec164a5b3e980a457fef5e24e36ab451f"} Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.080767 4790 scope.go:117] "RemoveContainer" containerID="55c887f9a40216c24e10a416b3ebd75cb336ba29eebc061496145267c0e23866" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.098386 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.118040 4790 scope.go:117] "RemoveContainer" containerID="b3c13279d4abc89b6ca528942a7b0cacfd3ae91cdf24bc9ecb3b5d28fa99b218" Nov 24 13:34:13 crc kubenswrapper[4790]: E1124 13:34:13.124551 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3c13279d4abc89b6ca528942a7b0cacfd3ae91cdf24bc9ecb3b5d28fa99b218\": container with ID starting with b3c13279d4abc89b6ca528942a7b0cacfd3ae91cdf24bc9ecb3b5d28fa99b218 not found: ID does not exist" containerID="b3c13279d4abc89b6ca528942a7b0cacfd3ae91cdf24bc9ecb3b5d28fa99b218" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.124603 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3c13279d4abc89b6ca528942a7b0cacfd3ae91cdf24bc9ecb3b5d28fa99b218"} err="failed to get container status \"b3c13279d4abc89b6ca528942a7b0cacfd3ae91cdf24bc9ecb3b5d28fa99b218\": rpc error: code = NotFound desc = could not find container \"b3c13279d4abc89b6ca528942a7b0cacfd3ae91cdf24bc9ecb3b5d28fa99b218\": container with ID starting with b3c13279d4abc89b6ca528942a7b0cacfd3ae91cdf24bc9ecb3b5d28fa99b218 not found: ID does not exist" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.124629 4790 scope.go:117] "RemoveContainer" containerID="55c887f9a40216c24e10a416b3ebd75cb336ba29eebc061496145267c0e23866" Nov 24 13:34:13 crc kubenswrapper[4790]: E1124 13:34:13.125067 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55c887f9a40216c24e10a416b3ebd75cb336ba29eebc061496145267c0e23866\": container with ID starting with 55c887f9a40216c24e10a416b3ebd75cb336ba29eebc061496145267c0e23866 not found: ID does not exist" containerID="55c887f9a40216c24e10a416b3ebd75cb336ba29eebc061496145267c0e23866" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.125094 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55c887f9a40216c24e10a416b3ebd75cb336ba29eebc061496145267c0e23866"} err="failed to get container status \"55c887f9a40216c24e10a416b3ebd75cb336ba29eebc061496145267c0e23866\": rpc error: code = NotFound desc = could not find container \"55c887f9a40216c24e10a416b3ebd75cb336ba29eebc061496145267c0e23866\": container with ID starting with 55c887f9a40216c24e10a416b3ebd75cb336ba29eebc061496145267c0e23866 not found: ID does not exist" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.125106 4790 scope.go:117] "RemoveContainer" containerID="b3c13279d4abc89b6ca528942a7b0cacfd3ae91cdf24bc9ecb3b5d28fa99b218" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.125314 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3c13279d4abc89b6ca528942a7b0cacfd3ae91cdf24bc9ecb3b5d28fa99b218"} err="failed to get container status \"b3c13279d4abc89b6ca528942a7b0cacfd3ae91cdf24bc9ecb3b5d28fa99b218\": rpc error: code = NotFound desc = could not find container \"b3c13279d4abc89b6ca528942a7b0cacfd3ae91cdf24bc9ecb3b5d28fa99b218\": container with ID starting with b3c13279d4abc89b6ca528942a7b0cacfd3ae91cdf24bc9ecb3b5d28fa99b218 not found: ID does not exist" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.125328 4790 scope.go:117] "RemoveContainer" containerID="55c887f9a40216c24e10a416b3ebd75cb336ba29eebc061496145267c0e23866" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.125919 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55c887f9a40216c24e10a416b3ebd75cb336ba29eebc061496145267c0e23866"} err="failed to get container status \"55c887f9a40216c24e10a416b3ebd75cb336ba29eebc061496145267c0e23866\": rpc error: code = NotFound desc = could not find container \"55c887f9a40216c24e10a416b3ebd75cb336ba29eebc061496145267c0e23866\": container with ID starting with 55c887f9a40216c24e10a416b3ebd75cb336ba29eebc061496145267c0e23866 not found: ID does not exist" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.125954 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.136749 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 24 13:34:13 crc kubenswrapper[4790]: E1124 13:34:13.137248 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3a796b0-1b3a-44e5-a300-d095012cc571" containerName="cinder-api-log" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.137270 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3a796b0-1b3a-44e5-a300-d095012cc571" containerName="cinder-api-log" Nov 24 13:34:13 crc kubenswrapper[4790]: E1124 13:34:13.137306 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3a796b0-1b3a-44e5-a300-d095012cc571" containerName="cinder-api" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.137313 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3a796b0-1b3a-44e5-a300-d095012cc571" containerName="cinder-api" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.137515 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3a796b0-1b3a-44e5-a300-d095012cc571" containerName="cinder-api" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.137546 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3a796b0-1b3a-44e5-a300-d095012cc571" containerName="cinder-api-log" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.138498 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.141984 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.142196 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.142314 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.162176 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.253499 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e6d696a7-d618-4416-b499-aae08088b079-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " pod="openstack/cinder-api-0" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.253614 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-scripts\") pod \"cinder-api-0\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " pod="openstack/cinder-api-0" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.253641 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6d696a7-d618-4416-b499-aae08088b079-logs\") pod \"cinder-api-0\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " pod="openstack/cinder-api-0" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.253690 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdxq7\" (UniqueName: \"kubernetes.io/projected/e6d696a7-d618-4416-b499-aae08088b079-kube-api-access-mdxq7\") pod \"cinder-api-0\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " pod="openstack/cinder-api-0" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.253734 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " pod="openstack/cinder-api-0" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.253789 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-config-data\") pod \"cinder-api-0\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " pod="openstack/cinder-api-0" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.253909 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " pod="openstack/cinder-api-0" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.254004 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " pod="openstack/cinder-api-0" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.255979 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-config-data-custom\") pod \"cinder-api-0\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " pod="openstack/cinder-api-0" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.358682 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-config-data-custom\") pod \"cinder-api-0\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " pod="openstack/cinder-api-0" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.358733 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e6d696a7-d618-4416-b499-aae08088b079-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " pod="openstack/cinder-api-0" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.358772 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-scripts\") pod \"cinder-api-0\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " pod="openstack/cinder-api-0" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.358788 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6d696a7-d618-4416-b499-aae08088b079-logs\") pod \"cinder-api-0\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " pod="openstack/cinder-api-0" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.358814 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdxq7\" (UniqueName: \"kubernetes.io/projected/e6d696a7-d618-4416-b499-aae08088b079-kube-api-access-mdxq7\") pod \"cinder-api-0\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " pod="openstack/cinder-api-0" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.358848 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " pod="openstack/cinder-api-0" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.358874 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-config-data\") pod \"cinder-api-0\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " pod="openstack/cinder-api-0" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.358922 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " pod="openstack/cinder-api-0" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.358957 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " pod="openstack/cinder-api-0" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.370673 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " pod="openstack/cinder-api-0" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.371187 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e6d696a7-d618-4416-b499-aae08088b079-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " pod="openstack/cinder-api-0" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.371662 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-config-data\") pod \"cinder-api-0\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " pod="openstack/cinder-api-0" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.376056 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6d696a7-d618-4416-b499-aae08088b079-logs\") pod \"cinder-api-0\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " pod="openstack/cinder-api-0" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.380363 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " pod="openstack/cinder-api-0" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.389404 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-scripts\") pod \"cinder-api-0\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " pod="openstack/cinder-api-0" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.390761 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-config-data-custom\") pod \"cinder-api-0\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " pod="openstack/cinder-api-0" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.391228 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " pod="openstack/cinder-api-0" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.391234 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdxq7\" (UniqueName: \"kubernetes.io/projected/e6d696a7-d618-4416-b499-aae08088b079-kube-api-access-mdxq7\") pod \"cinder-api-0\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " pod="openstack/cinder-api-0" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.471315 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 13:34:13 crc kubenswrapper[4790]: I1124 13:34:13.938353 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 13:34:14 crc kubenswrapper[4790]: I1124 13:34:14.069565 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e6d696a7-d618-4416-b499-aae08088b079","Type":"ContainerStarted","Data":"2ce5dd9abe30c2b33028a60ed6174141ddfe8cc25750d7d125c466efb0a51375"} Nov 24 13:34:14 crc kubenswrapper[4790]: I1124 13:34:14.072563 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7d69688466-r2bq9" event={"ID":"8311dfa1-6e87-493d-92e0-4eb69c792afe","Type":"ContainerStarted","Data":"d07411834060975e29533a55785dd1993dfe65d5e3ce92d8eb62b08cba9dc0d1"} Nov 24 13:34:14 crc kubenswrapper[4790]: I1124 13:34:14.072696 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7d69688466-r2bq9" Nov 24 13:34:14 crc kubenswrapper[4790]: I1124 13:34:14.075778 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8e48f476-da57-4a24-a043-babdc557f7e7","Type":"ContainerStarted","Data":"e658fe14b9d40b94b379fe01c57f1fecccce64f3cf8d0ce939df9c88cb6a031e"} Nov 24 13:34:14 crc kubenswrapper[4790]: I1124 13:34:14.076355 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 13:34:14 crc kubenswrapper[4790]: I1124 13:34:14.103004 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7d69688466-r2bq9" podStartSLOduration=3.102975518 podStartE2EDuration="3.102975518s" podCreationTimestamp="2025-11-24 13:34:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:34:14.096304818 +0000 UTC m=+1302.476198500" watchObservedRunningTime="2025-11-24 13:34:14.102975518 +0000 UTC m=+1302.482869200" Nov 24 13:34:14 crc kubenswrapper[4790]: I1124 13:34:14.126109 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.022124849 podStartE2EDuration="7.126088145s" podCreationTimestamp="2025-11-24 13:34:07 +0000 UTC" firstStartedPulling="2025-11-24 13:34:08.17340819 +0000 UTC m=+1296.553301852" lastFinishedPulling="2025-11-24 13:34:13.277371486 +0000 UTC m=+1301.657265148" observedRunningTime="2025-11-24 13:34:14.117717997 +0000 UTC m=+1302.497611659" watchObservedRunningTime="2025-11-24 13:34:14.126088145 +0000 UTC m=+1302.505981797" Nov 24 13:34:14 crc kubenswrapper[4790]: I1124 13:34:14.340605 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3a796b0-1b3a-44e5-a300-d095012cc571" path="/var/lib/kubelet/pods/f3a796b0-1b3a-44e5-a300-d095012cc571/volumes" Nov 24 13:34:14 crc kubenswrapper[4790]: I1124 13:34:14.444267 4790 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","podbe3f173b-f907-4f35-bb20-3afe1e874583"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort podbe3f173b-f907-4f35-bb20-3afe1e874583] : Timed out while waiting for systemd to remove kubepods-besteffort-podbe3f173b_f907_4f35_bb20_3afe1e874583.slice" Nov 24 13:34:14 crc kubenswrapper[4790]: E1124 13:34:14.444436 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort podbe3f173b-f907-4f35-bb20-3afe1e874583] : unable to destroy cgroup paths for cgroup [kubepods besteffort podbe3f173b-f907-4f35-bb20-3afe1e874583] : Timed out while waiting for systemd to remove kubepods-besteffort-podbe3f173b_f907_4f35_bb20_3afe1e874583.slice" pod="openstack/glance-default-internal-api-0" podUID="be3f173b-f907-4f35-bb20-3afe1e874583" Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.091645 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.091681 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e6d696a7-d618-4416-b499-aae08088b079","Type":"ContainerStarted","Data":"7d8bdf173f30451f767a3b85fb31741d1552f46cecc70197217878b14326622c"} Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.092706 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7d69688466-r2bq9" Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.115963 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.125292 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.138346 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.140127 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.142924 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.143019 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.146917 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.293983 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.294062 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-logs\") pod \"glance-default-internal-api-0\" (UID: \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.294105 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.294166 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.294195 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.294260 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.294296 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fht4d\" (UniqueName: \"kubernetes.io/projected/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-kube-api-access-fht4d\") pod \"glance-default-internal-api-0\" (UID: \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.294330 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.395817 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.395909 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.395969 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-logs\") pod \"glance-default-internal-api-0\" (UID: \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.396014 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.396082 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.396122 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.396184 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.396233 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fht4d\" (UniqueName: \"kubernetes.io/projected/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-kube-api-access-fht4d\") pod \"glance-default-internal-api-0\" (UID: \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.396363 4790 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.396675 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.396719 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-logs\") pod \"glance-default-internal-api-0\" (UID: \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.403629 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.404588 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.405622 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.408122 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.423402 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fht4d\" (UniqueName: \"kubernetes.io/projected/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-kube-api-access-fht4d\") pod \"glance-default-internal-api-0\" (UID: \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.465243 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:15 crc kubenswrapper[4790]: I1124 13:34:15.767197 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:16 crc kubenswrapper[4790]: I1124 13:34:16.109089 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e6d696a7-d618-4416-b499-aae08088b079","Type":"ContainerStarted","Data":"002e858954a60800481c9819989042b1ad82aae0f1762f13550da43a5f262b74"} Nov 24 13:34:16 crc kubenswrapper[4790]: I1124 13:34:16.109931 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 24 13:34:16 crc kubenswrapper[4790]: I1124 13:34:16.134915 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.134879619 podStartE2EDuration="3.134879619s" podCreationTimestamp="2025-11-24 13:34:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:34:16.128435016 +0000 UTC m=+1304.508328708" watchObservedRunningTime="2025-11-24 13:34:16.134879619 +0000 UTC m=+1304.514773281" Nov 24 13:34:16 crc kubenswrapper[4790]: I1124 13:34:16.324282 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be3f173b-f907-4f35-bb20-3afe1e874583" path="/var/lib/kubelet/pods/be3f173b-f907-4f35-bb20-3afe1e874583/volumes" Nov 24 13:34:16 crc kubenswrapper[4790]: I1124 13:34:16.346018 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:34:16 crc kubenswrapper[4790]: W1124 13:34:16.355317 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7fe37e44_47d4_4a0e_9cc1_4920db57f7d5.slice/crio-a1ba81ff268aaa8250fac541a25b59fd744152f7a345e29acd289db285ddd7d6 WatchSource:0}: Error finding container a1ba81ff268aaa8250fac541a25b59fd744152f7a345e29acd289db285ddd7d6: Status 404 returned error can't find the container with id a1ba81ff268aaa8250fac541a25b59fd744152f7a345e29acd289db285ddd7d6 Nov 24 13:34:17 crc kubenswrapper[4790]: I1124 13:34:17.118051 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5","Type":"ContainerStarted","Data":"eecb4a557c0f4d6e2a5c95fa519ba7f6cd2b3df6ed7cb80d7301eb06b36f512d"} Nov 24 13:34:17 crc kubenswrapper[4790]: I1124 13:34:17.118350 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5","Type":"ContainerStarted","Data":"a1ba81ff268aaa8250fac541a25b59fd744152f7a345e29acd289db285ddd7d6"} Nov 24 13:34:17 crc kubenswrapper[4790]: I1124 13:34:17.437325 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-dd8b84bc8-9g27k" Nov 24 13:34:17 crc kubenswrapper[4790]: I1124 13:34:17.492611 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-dd8b84bc8-9g27k" Nov 24 13:34:17 crc kubenswrapper[4790]: I1124 13:34:17.744205 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 24 13:34:17 crc kubenswrapper[4790]: I1124 13:34:17.802998 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 13:34:17 crc kubenswrapper[4790]: I1124 13:34:17.876011 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6578955fd5-snppf" Nov 24 13:34:17 crc kubenswrapper[4790]: I1124 13:34:17.944208 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-z76jp"] Nov 24 13:34:17 crc kubenswrapper[4790]: I1124 13:34:17.944480 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7b667979-z76jp" podUID="7294e855-6ce3-4807-9871-2fd2efbcfd0f" containerName="dnsmasq-dns" containerID="cri-o://a66746afa43133bf3b2cc2c02574b9bde384aeefd099c3f8982e8a6cf882206d" gracePeriod=10 Nov 24 13:34:18 crc kubenswrapper[4790]: I1124 13:34:18.156185 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5","Type":"ContainerStarted","Data":"ad8dcf46bbcdce25e8690aef88e66ffa33d828aa18a8a5458f8950ce7d7f7349"} Nov 24 13:34:18 crc kubenswrapper[4790]: I1124 13:34:18.177152 4790 generic.go:334] "Generic (PLEG): container finished" podID="7294e855-6ce3-4807-9871-2fd2efbcfd0f" containerID="a66746afa43133bf3b2cc2c02574b9bde384aeefd099c3f8982e8a6cf882206d" exitCode=0 Nov 24 13:34:18 crc kubenswrapper[4790]: I1124 13:34:18.177361 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="55e60feb-8531-4c7b-9033-5d3c6a07258a" containerName="cinder-scheduler" containerID="cri-o://acbc1a5bb162c73d8e0386e58d22675b4882165da87252b0b4d8bb67dff52b0a" gracePeriod=30 Nov 24 13:34:18 crc kubenswrapper[4790]: I1124 13:34:18.177663 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-z76jp" event={"ID":"7294e855-6ce3-4807-9871-2fd2efbcfd0f","Type":"ContainerDied","Data":"a66746afa43133bf3b2cc2c02574b9bde384aeefd099c3f8982e8a6cf882206d"} Nov 24 13:34:18 crc kubenswrapper[4790]: I1124 13:34:18.177990 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="55e60feb-8531-4c7b-9033-5d3c6a07258a" containerName="probe" containerID="cri-o://31df2983c371d070c392694c8f305220de213a9f32dff5c05e09472f74b9a653" gracePeriod=30 Nov 24 13:34:18 crc kubenswrapper[4790]: I1124 13:34:18.196900 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.196863366 podStartE2EDuration="3.196863366s" podCreationTimestamp="2025-11-24 13:34:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:34:18.181206251 +0000 UTC m=+1306.561099923" watchObservedRunningTime="2025-11-24 13:34:18.196863366 +0000 UTC m=+1306.576757028" Nov 24 13:34:18 crc kubenswrapper[4790]: I1124 13:34:18.442425 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-z76jp" Nov 24 13:34:18 crc kubenswrapper[4790]: I1124 13:34:18.566006 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7294e855-6ce3-4807-9871-2fd2efbcfd0f-ovsdbserver-sb\") pod \"7294e855-6ce3-4807-9871-2fd2efbcfd0f\" (UID: \"7294e855-6ce3-4807-9871-2fd2efbcfd0f\") " Nov 24 13:34:18 crc kubenswrapper[4790]: I1124 13:34:18.566059 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7294e855-6ce3-4807-9871-2fd2efbcfd0f-ovsdbserver-nb\") pod \"7294e855-6ce3-4807-9871-2fd2efbcfd0f\" (UID: \"7294e855-6ce3-4807-9871-2fd2efbcfd0f\") " Nov 24 13:34:18 crc kubenswrapper[4790]: I1124 13:34:18.566105 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7294e855-6ce3-4807-9871-2fd2efbcfd0f-dns-swift-storage-0\") pod \"7294e855-6ce3-4807-9871-2fd2efbcfd0f\" (UID: \"7294e855-6ce3-4807-9871-2fd2efbcfd0f\") " Nov 24 13:34:18 crc kubenswrapper[4790]: I1124 13:34:18.566187 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z58nr\" (UniqueName: \"kubernetes.io/projected/7294e855-6ce3-4807-9871-2fd2efbcfd0f-kube-api-access-z58nr\") pod \"7294e855-6ce3-4807-9871-2fd2efbcfd0f\" (UID: \"7294e855-6ce3-4807-9871-2fd2efbcfd0f\") " Nov 24 13:34:18 crc kubenswrapper[4790]: I1124 13:34:18.566232 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7294e855-6ce3-4807-9871-2fd2efbcfd0f-config\") pod \"7294e855-6ce3-4807-9871-2fd2efbcfd0f\" (UID: \"7294e855-6ce3-4807-9871-2fd2efbcfd0f\") " Nov 24 13:34:18 crc kubenswrapper[4790]: I1124 13:34:18.566326 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7294e855-6ce3-4807-9871-2fd2efbcfd0f-dns-svc\") pod \"7294e855-6ce3-4807-9871-2fd2efbcfd0f\" (UID: \"7294e855-6ce3-4807-9871-2fd2efbcfd0f\") " Nov 24 13:34:18 crc kubenswrapper[4790]: I1124 13:34:18.582898 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7294e855-6ce3-4807-9871-2fd2efbcfd0f-kube-api-access-z58nr" (OuterVolumeSpecName: "kube-api-access-z58nr") pod "7294e855-6ce3-4807-9871-2fd2efbcfd0f" (UID: "7294e855-6ce3-4807-9871-2fd2efbcfd0f"). InnerVolumeSpecName "kube-api-access-z58nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:18 crc kubenswrapper[4790]: I1124 13:34:18.633188 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7294e855-6ce3-4807-9871-2fd2efbcfd0f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7294e855-6ce3-4807-9871-2fd2efbcfd0f" (UID: "7294e855-6ce3-4807-9871-2fd2efbcfd0f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:18 crc kubenswrapper[4790]: I1124 13:34:18.651001 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7294e855-6ce3-4807-9871-2fd2efbcfd0f-config" (OuterVolumeSpecName: "config") pod "7294e855-6ce3-4807-9871-2fd2efbcfd0f" (UID: "7294e855-6ce3-4807-9871-2fd2efbcfd0f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:18 crc kubenswrapper[4790]: I1124 13:34:18.655334 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7294e855-6ce3-4807-9871-2fd2efbcfd0f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7294e855-6ce3-4807-9871-2fd2efbcfd0f" (UID: "7294e855-6ce3-4807-9871-2fd2efbcfd0f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:18 crc kubenswrapper[4790]: I1124 13:34:18.662161 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7294e855-6ce3-4807-9871-2fd2efbcfd0f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7294e855-6ce3-4807-9871-2fd2efbcfd0f" (UID: "7294e855-6ce3-4807-9871-2fd2efbcfd0f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:18 crc kubenswrapper[4790]: I1124 13:34:18.668921 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z58nr\" (UniqueName: \"kubernetes.io/projected/7294e855-6ce3-4807-9871-2fd2efbcfd0f-kube-api-access-z58nr\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:18 crc kubenswrapper[4790]: I1124 13:34:18.668969 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7294e855-6ce3-4807-9871-2fd2efbcfd0f-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:18 crc kubenswrapper[4790]: I1124 13:34:18.668981 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7294e855-6ce3-4807-9871-2fd2efbcfd0f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:18 crc kubenswrapper[4790]: I1124 13:34:18.668989 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7294e855-6ce3-4807-9871-2fd2efbcfd0f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:18 crc kubenswrapper[4790]: I1124 13:34:18.668999 4790 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7294e855-6ce3-4807-9871-2fd2efbcfd0f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:18 crc kubenswrapper[4790]: I1124 13:34:18.670723 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7294e855-6ce3-4807-9871-2fd2efbcfd0f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7294e855-6ce3-4807-9871-2fd2efbcfd0f" (UID: "7294e855-6ce3-4807-9871-2fd2efbcfd0f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:18 crc kubenswrapper[4790]: I1124 13:34:18.771396 4790 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7294e855-6ce3-4807-9871-2fd2efbcfd0f-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:19 crc kubenswrapper[4790]: I1124 13:34:19.109162 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7d69688466-r2bq9" Nov 24 13:34:19 crc kubenswrapper[4790]: I1124 13:34:19.189376 4790 generic.go:334] "Generic (PLEG): container finished" podID="55e60feb-8531-4c7b-9033-5d3c6a07258a" containerID="31df2983c371d070c392694c8f305220de213a9f32dff5c05e09472f74b9a653" exitCode=0 Nov 24 13:34:19 crc kubenswrapper[4790]: I1124 13:34:19.189445 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"55e60feb-8531-4c7b-9033-5d3c6a07258a","Type":"ContainerDied","Data":"31df2983c371d070c392694c8f305220de213a9f32dff5c05e09472f74b9a653"} Nov 24 13:34:19 crc kubenswrapper[4790]: I1124 13:34:19.192256 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-z76jp" event={"ID":"7294e855-6ce3-4807-9871-2fd2efbcfd0f","Type":"ContainerDied","Data":"382b2a2e0f19b90b7f32fc8330f9439f63f759dc23f5e33b3f177f5df34060bc"} Nov 24 13:34:19 crc kubenswrapper[4790]: I1124 13:34:19.192309 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-z76jp" Nov 24 13:34:19 crc kubenswrapper[4790]: I1124 13:34:19.192320 4790 scope.go:117] "RemoveContainer" containerID="a66746afa43133bf3b2cc2c02574b9bde384aeefd099c3f8982e8a6cf882206d" Nov 24 13:34:19 crc kubenswrapper[4790]: I1124 13:34:19.215519 4790 scope.go:117] "RemoveContainer" containerID="70fd1853a5a4f323a22d1dbf5921aaad40f430d3febd19818b17b8ed831a66a4" Nov 24 13:34:19 crc kubenswrapper[4790]: I1124 13:34:19.253943 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-z76jp"] Nov 24 13:34:19 crc kubenswrapper[4790]: I1124 13:34:19.263122 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-z76jp"] Nov 24 13:34:20 crc kubenswrapper[4790]: I1124 13:34:20.328601 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7294e855-6ce3-4807-9871-2fd2efbcfd0f" path="/var/lib/kubelet/pods/7294e855-6ce3-4807-9871-2fd2efbcfd0f/volumes" Nov 24 13:34:20 crc kubenswrapper[4790]: I1124 13:34:20.740401 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7d69688466-r2bq9" Nov 24 13:34:20 crc kubenswrapper[4790]: I1124 13:34:20.857362 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-dd8b84bc8-9g27k"] Nov 24 13:34:20 crc kubenswrapper[4790]: I1124 13:34:20.857579 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-dd8b84bc8-9g27k" podUID="ca62cb72-d397-4ee8-a7c4-15d831f87776" containerName="barbican-api-log" containerID="cri-o://1c160de17887f7283642017061e7f697b7b7ff536b2c1ca6125e5119a1a8b631" gracePeriod=30 Nov 24 13:34:20 crc kubenswrapper[4790]: I1124 13:34:20.858023 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-dd8b84bc8-9g27k" podUID="ca62cb72-d397-4ee8-a7c4-15d831f87776" containerName="barbican-api" containerID="cri-o://d938a457c8fe17cde00b625b8f7c1a79193d9b6ab2bbe35882d849ca8039facf" gracePeriod=30 Nov 24 13:34:20 crc kubenswrapper[4790]: I1124 13:34:20.869272 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-dd8b84bc8-9g27k" podUID="ca62cb72-d397-4ee8-a7c4-15d831f87776" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.155:9311/healthcheck\": EOF" Nov 24 13:34:21 crc kubenswrapper[4790]: I1124 13:34:21.221873 4790 generic.go:334] "Generic (PLEG): container finished" podID="ca62cb72-d397-4ee8-a7c4-15d831f87776" containerID="1c160de17887f7283642017061e7f697b7b7ff536b2c1ca6125e5119a1a8b631" exitCode=143 Nov 24 13:34:21 crc kubenswrapper[4790]: I1124 13:34:21.221921 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-dd8b84bc8-9g27k" event={"ID":"ca62cb72-d397-4ee8-a7c4-15d831f87776","Type":"ContainerDied","Data":"1c160de17887f7283642017061e7f697b7b7ff536b2c1ca6125e5119a1a8b631"} Nov 24 13:34:21 crc kubenswrapper[4790]: I1124 13:34:21.480031 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-7ddd868cf8-lc6qt" Nov 24 13:34:22 crc kubenswrapper[4790]: I1124 13:34:22.125551 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7957b6846-f8n7f" Nov 24 13:34:22 crc kubenswrapper[4790]: I1124 13:34:22.144179 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-69f849cd44-8v55p" Nov 24 13:34:22 crc kubenswrapper[4790]: I1124 13:34:22.223709 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7957b6846-f8n7f" Nov 24 13:34:22 crc kubenswrapper[4790]: I1124 13:34:22.259925 4790 generic.go:334] "Generic (PLEG): container finished" podID="55e60feb-8531-4c7b-9033-5d3c6a07258a" containerID="acbc1a5bb162c73d8e0386e58d22675b4882165da87252b0b4d8bb67dff52b0a" exitCode=0 Nov 24 13:34:22 crc kubenswrapper[4790]: I1124 13:34:22.286496 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"55e60feb-8531-4c7b-9033-5d3c6a07258a","Type":"ContainerDied","Data":"acbc1a5bb162c73d8e0386e58d22675b4882165da87252b0b4d8bb67dff52b0a"} Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.085732 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.173580 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55e60feb-8531-4c7b-9033-5d3c6a07258a-scripts\") pod \"55e60feb-8531-4c7b-9033-5d3c6a07258a\" (UID: \"55e60feb-8531-4c7b-9033-5d3c6a07258a\") " Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.173640 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55e60feb-8531-4c7b-9033-5d3c6a07258a-config-data\") pod \"55e60feb-8531-4c7b-9033-5d3c6a07258a\" (UID: \"55e60feb-8531-4c7b-9033-5d3c6a07258a\") " Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.173771 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/55e60feb-8531-4c7b-9033-5d3c6a07258a-etc-machine-id\") pod \"55e60feb-8531-4c7b-9033-5d3c6a07258a\" (UID: \"55e60feb-8531-4c7b-9033-5d3c6a07258a\") " Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.173809 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55e60feb-8531-4c7b-9033-5d3c6a07258a-combined-ca-bundle\") pod \"55e60feb-8531-4c7b-9033-5d3c6a07258a\" (UID: \"55e60feb-8531-4c7b-9033-5d3c6a07258a\") " Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.173913 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/55e60feb-8531-4c7b-9033-5d3c6a07258a-config-data-custom\") pod \"55e60feb-8531-4c7b-9033-5d3c6a07258a\" (UID: \"55e60feb-8531-4c7b-9033-5d3c6a07258a\") " Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.173972 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srth5\" (UniqueName: \"kubernetes.io/projected/55e60feb-8531-4c7b-9033-5d3c6a07258a-kube-api-access-srth5\") pod \"55e60feb-8531-4c7b-9033-5d3c6a07258a\" (UID: \"55e60feb-8531-4c7b-9033-5d3c6a07258a\") " Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.174978 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/55e60feb-8531-4c7b-9033-5d3c6a07258a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "55e60feb-8531-4c7b-9033-5d3c6a07258a" (UID: "55e60feb-8531-4c7b-9033-5d3c6a07258a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.189753 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55e60feb-8531-4c7b-9033-5d3c6a07258a-kube-api-access-srth5" (OuterVolumeSpecName: "kube-api-access-srth5") pod "55e60feb-8531-4c7b-9033-5d3c6a07258a" (UID: "55e60feb-8531-4c7b-9033-5d3c6a07258a"). InnerVolumeSpecName "kube-api-access-srth5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.206016 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55e60feb-8531-4c7b-9033-5d3c6a07258a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "55e60feb-8531-4c7b-9033-5d3c6a07258a" (UID: "55e60feb-8531-4c7b-9033-5d3c6a07258a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.206101 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55e60feb-8531-4c7b-9033-5d3c6a07258a-scripts" (OuterVolumeSpecName: "scripts") pod "55e60feb-8531-4c7b-9033-5d3c6a07258a" (UID: "55e60feb-8531-4c7b-9033-5d3c6a07258a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.275733 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srth5\" (UniqueName: \"kubernetes.io/projected/55e60feb-8531-4c7b-9033-5d3c6a07258a-kube-api-access-srth5\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.275761 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55e60feb-8531-4c7b-9033-5d3c6a07258a-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.275774 4790 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/55e60feb-8531-4c7b-9033-5d3c6a07258a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.275783 4790 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/55e60feb-8531-4c7b-9033-5d3c6a07258a-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.281389 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55e60feb-8531-4c7b-9033-5d3c6a07258a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "55e60feb-8531-4c7b-9033-5d3c6a07258a" (UID: "55e60feb-8531-4c7b-9033-5d3c6a07258a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.303114 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"55e60feb-8531-4c7b-9033-5d3c6a07258a","Type":"ContainerDied","Data":"6b859b296a7f4f3fb03c2300829268a2d421476851806dc7575d838952eef3c1"} Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.303160 4790 scope.go:117] "RemoveContainer" containerID="31df2983c371d070c392694c8f305220de213a9f32dff5c05e09472f74b9a653" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.303276 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.324387 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55e60feb-8531-4c7b-9033-5d3c6a07258a-config-data" (OuterVolumeSpecName: "config-data") pod "55e60feb-8531-4c7b-9033-5d3c6a07258a" (UID: "55e60feb-8531-4c7b-9033-5d3c6a07258a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.377381 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55e60feb-8531-4c7b-9033-5d3c6a07258a-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.377426 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55e60feb-8531-4c7b-9033-5d3c6a07258a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.385402 4790 scope.go:117] "RemoveContainer" containerID="acbc1a5bb162c73d8e0386e58d22675b4882165da87252b0b4d8bb67dff52b0a" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.642533 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.662896 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.667046 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 13:34:23 crc kubenswrapper[4790]: E1124 13:34:23.667508 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7294e855-6ce3-4807-9871-2fd2efbcfd0f" containerName="dnsmasq-dns" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.667525 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="7294e855-6ce3-4807-9871-2fd2efbcfd0f" containerName="dnsmasq-dns" Nov 24 13:34:23 crc kubenswrapper[4790]: E1124 13:34:23.667537 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55e60feb-8531-4c7b-9033-5d3c6a07258a" containerName="cinder-scheduler" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.667543 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="55e60feb-8531-4c7b-9033-5d3c6a07258a" containerName="cinder-scheduler" Nov 24 13:34:23 crc kubenswrapper[4790]: E1124 13:34:23.667568 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7294e855-6ce3-4807-9871-2fd2efbcfd0f" containerName="init" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.667573 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="7294e855-6ce3-4807-9871-2fd2efbcfd0f" containerName="init" Nov 24 13:34:23 crc kubenswrapper[4790]: E1124 13:34:23.667587 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55e60feb-8531-4c7b-9033-5d3c6a07258a" containerName="probe" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.667593 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="55e60feb-8531-4c7b-9033-5d3c6a07258a" containerName="probe" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.667739 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="55e60feb-8531-4c7b-9033-5d3c6a07258a" containerName="probe" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.667763 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="7294e855-6ce3-4807-9871-2fd2efbcfd0f" containerName="dnsmasq-dns" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.667779 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="55e60feb-8531-4c7b-9033-5d3c6a07258a" containerName="cinder-scheduler" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.668659 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.681668 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.695772 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.786763 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d0fa96c6-35fb-4e66-a7dc-31f6fa589132\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.786847 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d0fa96c6-35fb-4e66-a7dc-31f6fa589132\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.786919 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-scripts\") pod \"cinder-scheduler-0\" (UID: \"d0fa96c6-35fb-4e66-a7dc-31f6fa589132\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.786945 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdzgl\" (UniqueName: \"kubernetes.io/projected/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-kube-api-access-mdzgl\") pod \"cinder-scheduler-0\" (UID: \"d0fa96c6-35fb-4e66-a7dc-31f6fa589132\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.786991 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d0fa96c6-35fb-4e66-a7dc-31f6fa589132\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.787031 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-config-data\") pod \"cinder-scheduler-0\" (UID: \"d0fa96c6-35fb-4e66-a7dc-31f6fa589132\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.889465 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d0fa96c6-35fb-4e66-a7dc-31f6fa589132\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.889789 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-config-data\") pod \"cinder-scheduler-0\" (UID: \"d0fa96c6-35fb-4e66-a7dc-31f6fa589132\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.890133 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d0fa96c6-35fb-4e66-a7dc-31f6fa589132\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.890648 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d0fa96c6-35fb-4e66-a7dc-31f6fa589132\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.890737 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-scripts\") pod \"cinder-scheduler-0\" (UID: \"d0fa96c6-35fb-4e66-a7dc-31f6fa589132\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.890791 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdzgl\" (UniqueName: \"kubernetes.io/projected/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-kube-api-access-mdzgl\") pod \"cinder-scheduler-0\" (UID: \"d0fa96c6-35fb-4e66-a7dc-31f6fa589132\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.890218 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d0fa96c6-35fb-4e66-a7dc-31f6fa589132\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.893096 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d0fa96c6-35fb-4e66-a7dc-31f6fa589132\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.896813 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-scripts\") pod \"cinder-scheduler-0\" (UID: \"d0fa96c6-35fb-4e66-a7dc-31f6fa589132\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.897732 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d0fa96c6-35fb-4e66-a7dc-31f6fa589132\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.898673 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-config-data\") pod \"cinder-scheduler-0\" (UID: \"d0fa96c6-35fb-4e66-a7dc-31f6fa589132\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:23 crc kubenswrapper[4790]: I1124 13:34:23.917500 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdzgl\" (UniqueName: \"kubernetes.io/projected/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-kube-api-access-mdzgl\") pod \"cinder-scheduler-0\" (UID: \"d0fa96c6-35fb-4e66-a7dc-31f6fa589132\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:24 crc kubenswrapper[4790]: I1124 13:34:24.022734 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 13:34:24 crc kubenswrapper[4790]: I1124 13:34:24.259436 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-78cd478f4f-bqm94" Nov 24 13:34:24 crc kubenswrapper[4790]: I1124 13:34:24.329117 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55e60feb-8531-4c7b-9033-5d3c6a07258a" path="/var/lib/kubelet/pods/55e60feb-8531-4c7b-9033-5d3c6a07258a/volumes" Nov 24 13:34:24 crc kubenswrapper[4790]: I1124 13:34:24.334039 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-69f849cd44-8v55p"] Nov 24 13:34:24 crc kubenswrapper[4790]: I1124 13:34:24.334303 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-69f849cd44-8v55p" podUID="001f4ed0-f6e6-440e-a878-2bca6982ea47" containerName="neutron-api" containerID="cri-o://db4596a34402f208f174677329719e056fd084851d4e5a37a88604237aa9cfd2" gracePeriod=30 Nov 24 13:34:24 crc kubenswrapper[4790]: I1124 13:34:24.334613 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-69f849cd44-8v55p" podUID="001f4ed0-f6e6-440e-a878-2bca6982ea47" containerName="neutron-httpd" containerID="cri-o://6537ec4e93e7d7e332d59fb429a20d95718baf70f342e2c89ccf3f495ddb3c80" gracePeriod=30 Nov 24 13:34:24 crc kubenswrapper[4790]: I1124 13:34:24.341324 4790 generic.go:334] "Generic (PLEG): container finished" podID="ca62cb72-d397-4ee8-a7c4-15d831f87776" containerID="d938a457c8fe17cde00b625b8f7c1a79193d9b6ab2bbe35882d849ca8039facf" exitCode=0 Nov 24 13:34:24 crc kubenswrapper[4790]: I1124 13:34:24.341370 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-dd8b84bc8-9g27k" event={"ID":"ca62cb72-d397-4ee8-a7c4-15d831f87776","Type":"ContainerDied","Data":"d938a457c8fe17cde00b625b8f7c1a79193d9b6ab2bbe35882d849ca8039facf"} Nov 24 13:34:24 crc kubenswrapper[4790]: I1124 13:34:24.509808 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 13:34:24 crc kubenswrapper[4790]: I1124 13:34:24.636698 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-dd8b84bc8-9g27k" Nov 24 13:34:24 crc kubenswrapper[4790]: I1124 13:34:24.715169 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca62cb72-d397-4ee8-a7c4-15d831f87776-config-data-custom\") pod \"ca62cb72-d397-4ee8-a7c4-15d831f87776\" (UID: \"ca62cb72-d397-4ee8-a7c4-15d831f87776\") " Nov 24 13:34:24 crc kubenswrapper[4790]: I1124 13:34:24.715331 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8ksw\" (UniqueName: \"kubernetes.io/projected/ca62cb72-d397-4ee8-a7c4-15d831f87776-kube-api-access-k8ksw\") pod \"ca62cb72-d397-4ee8-a7c4-15d831f87776\" (UID: \"ca62cb72-d397-4ee8-a7c4-15d831f87776\") " Nov 24 13:34:24 crc kubenswrapper[4790]: I1124 13:34:24.715417 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca62cb72-d397-4ee8-a7c4-15d831f87776-config-data\") pod \"ca62cb72-d397-4ee8-a7c4-15d831f87776\" (UID: \"ca62cb72-d397-4ee8-a7c4-15d831f87776\") " Nov 24 13:34:24 crc kubenswrapper[4790]: I1124 13:34:24.715525 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca62cb72-d397-4ee8-a7c4-15d831f87776-combined-ca-bundle\") pod \"ca62cb72-d397-4ee8-a7c4-15d831f87776\" (UID: \"ca62cb72-d397-4ee8-a7c4-15d831f87776\") " Nov 24 13:34:24 crc kubenswrapper[4790]: I1124 13:34:24.715577 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca62cb72-d397-4ee8-a7c4-15d831f87776-logs\") pod \"ca62cb72-d397-4ee8-a7c4-15d831f87776\" (UID: \"ca62cb72-d397-4ee8-a7c4-15d831f87776\") " Nov 24 13:34:24 crc kubenswrapper[4790]: I1124 13:34:24.716691 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca62cb72-d397-4ee8-a7c4-15d831f87776-logs" (OuterVolumeSpecName: "logs") pod "ca62cb72-d397-4ee8-a7c4-15d831f87776" (UID: "ca62cb72-d397-4ee8-a7c4-15d831f87776"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:34:24 crc kubenswrapper[4790]: I1124 13:34:24.718302 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca62cb72-d397-4ee8-a7c4-15d831f87776-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:24 crc kubenswrapper[4790]: I1124 13:34:24.731044 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca62cb72-d397-4ee8-a7c4-15d831f87776-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ca62cb72-d397-4ee8-a7c4-15d831f87776" (UID: "ca62cb72-d397-4ee8-a7c4-15d831f87776"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:24 crc kubenswrapper[4790]: I1124 13:34:24.731170 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca62cb72-d397-4ee8-a7c4-15d831f87776-kube-api-access-k8ksw" (OuterVolumeSpecName: "kube-api-access-k8ksw") pod "ca62cb72-d397-4ee8-a7c4-15d831f87776" (UID: "ca62cb72-d397-4ee8-a7c4-15d831f87776"). InnerVolumeSpecName "kube-api-access-k8ksw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:24 crc kubenswrapper[4790]: I1124 13:34:24.739122 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca62cb72-d397-4ee8-a7c4-15d831f87776-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ca62cb72-d397-4ee8-a7c4-15d831f87776" (UID: "ca62cb72-d397-4ee8-a7c4-15d831f87776"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:24 crc kubenswrapper[4790]: I1124 13:34:24.797057 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca62cb72-d397-4ee8-a7c4-15d831f87776-config-data" (OuterVolumeSpecName: "config-data") pod "ca62cb72-d397-4ee8-a7c4-15d831f87776" (UID: "ca62cb72-d397-4ee8-a7c4-15d831f87776"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:24 crc kubenswrapper[4790]: I1124 13:34:24.820384 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca62cb72-d397-4ee8-a7c4-15d831f87776-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:24 crc kubenswrapper[4790]: I1124 13:34:24.820418 4790 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca62cb72-d397-4ee8-a7c4-15d831f87776-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:24 crc kubenswrapper[4790]: I1124 13:34:24.820431 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8ksw\" (UniqueName: \"kubernetes.io/projected/ca62cb72-d397-4ee8-a7c4-15d831f87776-kube-api-access-k8ksw\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:24 crc kubenswrapper[4790]: I1124 13:34:24.820442 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca62cb72-d397-4ee8-a7c4-15d831f87776-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.379142 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-dd8b84bc8-9g27k" event={"ID":"ca62cb72-d397-4ee8-a7c4-15d831f87776","Type":"ContainerDied","Data":"8554416072f611c39dd5e98f8ba56cfcd1574c3ce771e9262cbf898851e86218"} Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.379200 4790 scope.go:117] "RemoveContainer" containerID="d938a457c8fe17cde00b625b8f7c1a79193d9b6ab2bbe35882d849ca8039facf" Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.379342 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-dd8b84bc8-9g27k" Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.391219 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d0fa96c6-35fb-4e66-a7dc-31f6fa589132","Type":"ContainerStarted","Data":"92ffb21f402a267f3a3f49d256f427879130f4670ece784d82a260628c06ed0d"} Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.391278 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d0fa96c6-35fb-4e66-a7dc-31f6fa589132","Type":"ContainerStarted","Data":"5519e85bf1eaf040d0b597f6c16478311b609bfa922465b207a9b70aea556814"} Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.396602 4790 generic.go:334] "Generic (PLEG): container finished" podID="001f4ed0-f6e6-440e-a878-2bca6982ea47" containerID="6537ec4e93e7d7e332d59fb429a20d95718baf70f342e2c89ccf3f495ddb3c80" exitCode=0 Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.396643 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69f849cd44-8v55p" event={"ID":"001f4ed0-f6e6-440e-a878-2bca6982ea47","Type":"ContainerDied","Data":"6537ec4e93e7d7e332d59fb429a20d95718baf70f342e2c89ccf3f495ddb3c80"} Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.417456 4790 scope.go:117] "RemoveContainer" containerID="1c160de17887f7283642017061e7f697b7b7ff536b2c1ca6125e5119a1a8b631" Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.454570 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-dd8b84bc8-9g27k"] Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.469630 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-dd8b84bc8-9g27k"] Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.612145 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 24 13:34:25 crc kubenswrapper[4790]: E1124 13:34:25.612587 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca62cb72-d397-4ee8-a7c4-15d831f87776" containerName="barbican-api" Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.612609 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca62cb72-d397-4ee8-a7c4-15d831f87776" containerName="barbican-api" Nov 24 13:34:25 crc kubenswrapper[4790]: E1124 13:34:25.612647 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca62cb72-d397-4ee8-a7c4-15d831f87776" containerName="barbican-api-log" Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.612655 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca62cb72-d397-4ee8-a7c4-15d831f87776" containerName="barbican-api-log" Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.612866 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca62cb72-d397-4ee8-a7c4-15d831f87776" containerName="barbican-api-log" Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.612918 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca62cb72-d397-4ee8-a7c4-15d831f87776" containerName="barbican-api" Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.613678 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.616525 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-qvbg5" Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.616776 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.617187 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.633634 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1cbce39d-eeed-48b4-bea3-64d23c6ffe4f-openstack-config-secret\") pod \"openstackclient\" (UID: \"1cbce39d-eeed-48b4-bea3-64d23c6ffe4f\") " pod="openstack/openstackclient" Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.633682 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1cbce39d-eeed-48b4-bea3-64d23c6ffe4f-openstack-config\") pod \"openstackclient\" (UID: \"1cbce39d-eeed-48b4-bea3-64d23c6ffe4f\") " pod="openstack/openstackclient" Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.633753 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cbce39d-eeed-48b4-bea3-64d23c6ffe4f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1cbce39d-eeed-48b4-bea3-64d23c6ffe4f\") " pod="openstack/openstackclient" Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.633843 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-md64h\" (UniqueName: \"kubernetes.io/projected/1cbce39d-eeed-48b4-bea3-64d23c6ffe4f-kube-api-access-md64h\") pod \"openstackclient\" (UID: \"1cbce39d-eeed-48b4-bea3-64d23c6ffe4f\") " pod="openstack/openstackclient" Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.665282 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.735271 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1cbce39d-eeed-48b4-bea3-64d23c6ffe4f-openstack-config-secret\") pod \"openstackclient\" (UID: \"1cbce39d-eeed-48b4-bea3-64d23c6ffe4f\") " pod="openstack/openstackclient" Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.735319 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1cbce39d-eeed-48b4-bea3-64d23c6ffe4f-openstack-config\") pod \"openstackclient\" (UID: \"1cbce39d-eeed-48b4-bea3-64d23c6ffe4f\") " pod="openstack/openstackclient" Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.735366 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cbce39d-eeed-48b4-bea3-64d23c6ffe4f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1cbce39d-eeed-48b4-bea3-64d23c6ffe4f\") " pod="openstack/openstackclient" Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.735390 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-md64h\" (UniqueName: \"kubernetes.io/projected/1cbce39d-eeed-48b4-bea3-64d23c6ffe4f-kube-api-access-md64h\") pod \"openstackclient\" (UID: \"1cbce39d-eeed-48b4-bea3-64d23c6ffe4f\") " pod="openstack/openstackclient" Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.737039 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1cbce39d-eeed-48b4-bea3-64d23c6ffe4f-openstack-config\") pod \"openstackclient\" (UID: \"1cbce39d-eeed-48b4-bea3-64d23c6ffe4f\") " pod="openstack/openstackclient" Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.741138 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cbce39d-eeed-48b4-bea3-64d23c6ffe4f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1cbce39d-eeed-48b4-bea3-64d23c6ffe4f\") " pod="openstack/openstackclient" Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.747403 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1cbce39d-eeed-48b4-bea3-64d23c6ffe4f-openstack-config-secret\") pod \"openstackclient\" (UID: \"1cbce39d-eeed-48b4-bea3-64d23c6ffe4f\") " pod="openstack/openstackclient" Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.756190 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-md64h\" (UniqueName: \"kubernetes.io/projected/1cbce39d-eeed-48b4-bea3-64d23c6ffe4f-kube-api-access-md64h\") pod \"openstackclient\" (UID: \"1cbce39d-eeed-48b4-bea3-64d23c6ffe4f\") " pod="openstack/openstackclient" Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.768493 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.768545 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.814570 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.825950 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.826627 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.837665 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.908716 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.910102 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.921288 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 13:34:25 crc kubenswrapper[4790]: I1124 13:34:25.922058 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:26 crc kubenswrapper[4790]: E1124 13:34:26.043546 4790 log.go:32] "RunPodSandbox from runtime service failed" err=< Nov 24 13:34:26 crc kubenswrapper[4790]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_1cbce39d-eeed-48b4-bea3-64d23c6ffe4f_0(5bf84cf567d6d41c19e953ccf3ebcfa724fc83d166c9117714fe97221f2b8e6e): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"5bf84cf567d6d41c19e953ccf3ebcfa724fc83d166c9117714fe97221f2b8e6e" Netns:"/var/run/netns/9d3ef40f-1fe0-4b27-a5cd-2940fc4a4835" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=5bf84cf567d6d41c19e953ccf3ebcfa724fc83d166c9117714fe97221f2b8e6e;K8S_POD_UID=1cbce39d-eeed-48b4-bea3-64d23c6ffe4f" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/1cbce39d-eeed-48b4-bea3-64d23c6ffe4f]: expected pod UID "1cbce39d-eeed-48b4-bea3-64d23c6ffe4f" but got "f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c" from Kube API Nov 24 13:34:26 crc kubenswrapper[4790]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Nov 24 13:34:26 crc kubenswrapper[4790]: > Nov 24 13:34:26 crc kubenswrapper[4790]: E1124 13:34:26.043614 4790 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Nov 24 13:34:26 crc kubenswrapper[4790]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_1cbce39d-eeed-48b4-bea3-64d23c6ffe4f_0(5bf84cf567d6d41c19e953ccf3ebcfa724fc83d166c9117714fe97221f2b8e6e): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"5bf84cf567d6d41c19e953ccf3ebcfa724fc83d166c9117714fe97221f2b8e6e" Netns:"/var/run/netns/9d3ef40f-1fe0-4b27-a5cd-2940fc4a4835" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=5bf84cf567d6d41c19e953ccf3ebcfa724fc83d166c9117714fe97221f2b8e6e;K8S_POD_UID=1cbce39d-eeed-48b4-bea3-64d23c6ffe4f" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/1cbce39d-eeed-48b4-bea3-64d23c6ffe4f]: expected pod UID "1cbce39d-eeed-48b4-bea3-64d23c6ffe4f" but got "f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c" from Kube API Nov 24 13:34:26 crc kubenswrapper[4790]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Nov 24 13:34:26 crc kubenswrapper[4790]: > pod="openstack/openstackclient" Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.048865 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5t4x\" (UniqueName: \"kubernetes.io/projected/f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c-kube-api-access-q5t4x\") pod \"openstackclient\" (UID: \"f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c\") " pod="openstack/openstackclient" Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.048990 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c\") " pod="openstack/openstackclient" Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.049127 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c-openstack-config\") pod \"openstackclient\" (UID: \"f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c\") " pod="openstack/openstackclient" Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.049236 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c-openstack-config-secret\") pod \"openstackclient\" (UID: \"f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c\") " pod="openstack/openstackclient" Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.138413 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.151162 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5t4x\" (UniqueName: \"kubernetes.io/projected/f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c-kube-api-access-q5t4x\") pod \"openstackclient\" (UID: \"f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c\") " pod="openstack/openstackclient" Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.151318 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c\") " pod="openstack/openstackclient" Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.151425 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c-openstack-config\") pod \"openstackclient\" (UID: \"f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c\") " pod="openstack/openstackclient" Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.151495 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c-openstack-config-secret\") pod \"openstackclient\" (UID: \"f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c\") " pod="openstack/openstackclient" Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.152341 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c-openstack-config\") pod \"openstackclient\" (UID: \"f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c\") " pod="openstack/openstackclient" Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.156082 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c-openstack-config-secret\") pod \"openstackclient\" (UID: \"f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c\") " pod="openstack/openstackclient" Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.156953 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c\") " pod="openstack/openstackclient" Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.184441 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5t4x\" (UniqueName: \"kubernetes.io/projected/f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c-kube-api-access-q5t4x\") pod \"openstackclient\" (UID: \"f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c\") " pod="openstack/openstackclient" Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.243568 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.337780 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca62cb72-d397-4ee8-a7c4-15d831f87776" path="/var/lib/kubelet/pods/ca62cb72-d397-4ee8-a7c4-15d831f87776/volumes" Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.422220 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d0fa96c6-35fb-4e66-a7dc-31f6fa589132","Type":"ContainerStarted","Data":"fb2d4795a70584cebf1c4bedf0675e9e61a8fae6696ec9e5217fb4161288c248"} Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.423370 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.423394 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.423424 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.438335 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.449272 4790 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="1cbce39d-eeed-48b4-bea3-64d23c6ffe4f" podUID="f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c" Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.451293 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.451271738 podStartE2EDuration="3.451271738s" podCreationTimestamp="2025-11-24 13:34:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:34:26.442689644 +0000 UTC m=+1314.822583336" watchObservedRunningTime="2025-11-24 13:34:26.451271738 +0000 UTC m=+1314.831165410" Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.558460 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-md64h\" (UniqueName: \"kubernetes.io/projected/1cbce39d-eeed-48b4-bea3-64d23c6ffe4f-kube-api-access-md64h\") pod \"1cbce39d-eeed-48b4-bea3-64d23c6ffe4f\" (UID: \"1cbce39d-eeed-48b4-bea3-64d23c6ffe4f\") " Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.558517 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1cbce39d-eeed-48b4-bea3-64d23c6ffe4f-openstack-config\") pod \"1cbce39d-eeed-48b4-bea3-64d23c6ffe4f\" (UID: \"1cbce39d-eeed-48b4-bea3-64d23c6ffe4f\") " Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.558689 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1cbce39d-eeed-48b4-bea3-64d23c6ffe4f-openstack-config-secret\") pod \"1cbce39d-eeed-48b4-bea3-64d23c6ffe4f\" (UID: \"1cbce39d-eeed-48b4-bea3-64d23c6ffe4f\") " Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.558747 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cbce39d-eeed-48b4-bea3-64d23c6ffe4f-combined-ca-bundle\") pod \"1cbce39d-eeed-48b4-bea3-64d23c6ffe4f\" (UID: \"1cbce39d-eeed-48b4-bea3-64d23c6ffe4f\") " Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.559205 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1cbce39d-eeed-48b4-bea3-64d23c6ffe4f-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "1cbce39d-eeed-48b4-bea3-64d23c6ffe4f" (UID: "1cbce39d-eeed-48b4-bea3-64d23c6ffe4f"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.563824 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cbce39d-eeed-48b4-bea3-64d23c6ffe4f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1cbce39d-eeed-48b4-bea3-64d23c6ffe4f" (UID: "1cbce39d-eeed-48b4-bea3-64d23c6ffe4f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.578349 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1cbce39d-eeed-48b4-bea3-64d23c6ffe4f-kube-api-access-md64h" (OuterVolumeSpecName: "kube-api-access-md64h") pod "1cbce39d-eeed-48b4-bea3-64d23c6ffe4f" (UID: "1cbce39d-eeed-48b4-bea3-64d23c6ffe4f"). InnerVolumeSpecName "kube-api-access-md64h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.579095 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cbce39d-eeed-48b4-bea3-64d23c6ffe4f-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "1cbce39d-eeed-48b4-bea3-64d23c6ffe4f" (UID: "1cbce39d-eeed-48b4-bea3-64d23c6ffe4f"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.660611 4790 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1cbce39d-eeed-48b4-bea3-64d23c6ffe4f-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.660952 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cbce39d-eeed-48b4-bea3-64d23c6ffe4f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.660964 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-md64h\" (UniqueName: \"kubernetes.io/projected/1cbce39d-eeed-48b4-bea3-64d23c6ffe4f-kube-api-access-md64h\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.660976 4790 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1cbce39d-eeed-48b4-bea3-64d23c6ffe4f-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:26 crc kubenswrapper[4790]: I1124 13:34:26.736742 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 13:34:26 crc kubenswrapper[4790]: W1124 13:34:26.738283 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf8be67b1_1f0a_4ddb_82f1_906fbbf6ac7c.slice/crio-e092b99a7e8455dce908d6c9ab098352d707824f634865301badf37c9c1105d5 WatchSource:0}: Error finding container e092b99a7e8455dce908d6c9ab098352d707824f634865301badf37c9c1105d5: Status 404 returned error can't find the container with id e092b99a7e8455dce908d6c9ab098352d707824f634865301badf37c9c1105d5 Nov 24 13:34:27 crc kubenswrapper[4790]: I1124 13:34:27.440021 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 13:34:27 crc kubenswrapper[4790]: I1124 13:34:27.443997 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c","Type":"ContainerStarted","Data":"e092b99a7e8455dce908d6c9ab098352d707824f634865301badf37c9c1105d5"} Nov 24 13:34:27 crc kubenswrapper[4790]: I1124 13:34:27.456185 4790 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="1cbce39d-eeed-48b4-bea3-64d23c6ffe4f" podUID="f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c" Nov 24 13:34:28 crc kubenswrapper[4790]: I1124 13:34:28.331309 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1cbce39d-eeed-48b4-bea3-64d23c6ffe4f" path="/var/lib/kubelet/pods/1cbce39d-eeed-48b4-bea3-64d23c6ffe4f/volumes" Nov 24 13:34:28 crc kubenswrapper[4790]: I1124 13:34:28.451898 4790 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 13:34:28 crc kubenswrapper[4790]: I1124 13:34:28.451929 4790 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 13:34:28 crc kubenswrapper[4790]: I1124 13:34:28.559051 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:28 crc kubenswrapper[4790]: I1124 13:34:28.573782 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:29 crc kubenswrapper[4790]: I1124 13:34:29.023614 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 24 13:34:29 crc kubenswrapper[4790]: I1124 13:34:29.991605 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-bf58ddff5-swc8b"] Nov 24 13:34:29 crc kubenswrapper[4790]: I1124 13:34:29.997248 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-bf58ddff5-swc8b" Nov 24 13:34:29 crc kubenswrapper[4790]: I1124 13:34:29.999941 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Nov 24 13:34:30 crc kubenswrapper[4790]: I1124 13:34:30.000173 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 24 13:34:30 crc kubenswrapper[4790]: I1124 13:34:30.000327 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Nov 24 13:34:30 crc kubenswrapper[4790]: I1124 13:34:30.034584 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-bf58ddff5-swc8b"] Nov 24 13:34:30 crc kubenswrapper[4790]: I1124 13:34:30.143853 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9362fee7-a172-4948-8721-b4c83bf04a30-config-data\") pod \"swift-proxy-bf58ddff5-swc8b\" (UID: \"9362fee7-a172-4948-8721-b4c83bf04a30\") " pod="openstack/swift-proxy-bf58ddff5-swc8b" Nov 24 13:34:30 crc kubenswrapper[4790]: I1124 13:34:30.143920 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9362fee7-a172-4948-8721-b4c83bf04a30-internal-tls-certs\") pod \"swift-proxy-bf58ddff5-swc8b\" (UID: \"9362fee7-a172-4948-8721-b4c83bf04a30\") " pod="openstack/swift-proxy-bf58ddff5-swc8b" Nov 24 13:34:30 crc kubenswrapper[4790]: I1124 13:34:30.144064 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9362fee7-a172-4948-8721-b4c83bf04a30-public-tls-certs\") pod \"swift-proxy-bf58ddff5-swc8b\" (UID: \"9362fee7-a172-4948-8721-b4c83bf04a30\") " pod="openstack/swift-proxy-bf58ddff5-swc8b" Nov 24 13:34:30 crc kubenswrapper[4790]: I1124 13:34:30.144146 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9362fee7-a172-4948-8721-b4c83bf04a30-combined-ca-bundle\") pod \"swift-proxy-bf58ddff5-swc8b\" (UID: \"9362fee7-a172-4948-8721-b4c83bf04a30\") " pod="openstack/swift-proxy-bf58ddff5-swc8b" Nov 24 13:34:30 crc kubenswrapper[4790]: I1124 13:34:30.144291 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9362fee7-a172-4948-8721-b4c83bf04a30-run-httpd\") pod \"swift-proxy-bf58ddff5-swc8b\" (UID: \"9362fee7-a172-4948-8721-b4c83bf04a30\") " pod="openstack/swift-proxy-bf58ddff5-swc8b" Nov 24 13:34:30 crc kubenswrapper[4790]: I1124 13:34:30.144329 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9362fee7-a172-4948-8721-b4c83bf04a30-etc-swift\") pod \"swift-proxy-bf58ddff5-swc8b\" (UID: \"9362fee7-a172-4948-8721-b4c83bf04a30\") " pod="openstack/swift-proxy-bf58ddff5-swc8b" Nov 24 13:34:30 crc kubenswrapper[4790]: I1124 13:34:30.144461 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9362fee7-a172-4948-8721-b4c83bf04a30-log-httpd\") pod \"swift-proxy-bf58ddff5-swc8b\" (UID: \"9362fee7-a172-4948-8721-b4c83bf04a30\") " pod="openstack/swift-proxy-bf58ddff5-swc8b" Nov 24 13:34:30 crc kubenswrapper[4790]: I1124 13:34:30.144555 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rqhq\" (UniqueName: \"kubernetes.io/projected/9362fee7-a172-4948-8721-b4c83bf04a30-kube-api-access-7rqhq\") pod \"swift-proxy-bf58ddff5-swc8b\" (UID: \"9362fee7-a172-4948-8721-b4c83bf04a30\") " pod="openstack/swift-proxy-bf58ddff5-swc8b" Nov 24 13:34:30 crc kubenswrapper[4790]: I1124 13:34:30.246303 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9362fee7-a172-4948-8721-b4c83bf04a30-run-httpd\") pod \"swift-proxy-bf58ddff5-swc8b\" (UID: \"9362fee7-a172-4948-8721-b4c83bf04a30\") " pod="openstack/swift-proxy-bf58ddff5-swc8b" Nov 24 13:34:30 crc kubenswrapper[4790]: I1124 13:34:30.246350 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9362fee7-a172-4948-8721-b4c83bf04a30-etc-swift\") pod \"swift-proxy-bf58ddff5-swc8b\" (UID: \"9362fee7-a172-4948-8721-b4c83bf04a30\") " pod="openstack/swift-proxy-bf58ddff5-swc8b" Nov 24 13:34:30 crc kubenswrapper[4790]: I1124 13:34:30.246402 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9362fee7-a172-4948-8721-b4c83bf04a30-log-httpd\") pod \"swift-proxy-bf58ddff5-swc8b\" (UID: \"9362fee7-a172-4948-8721-b4c83bf04a30\") " pod="openstack/swift-proxy-bf58ddff5-swc8b" Nov 24 13:34:30 crc kubenswrapper[4790]: I1124 13:34:30.246437 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rqhq\" (UniqueName: \"kubernetes.io/projected/9362fee7-a172-4948-8721-b4c83bf04a30-kube-api-access-7rqhq\") pod \"swift-proxy-bf58ddff5-swc8b\" (UID: \"9362fee7-a172-4948-8721-b4c83bf04a30\") " pod="openstack/swift-proxy-bf58ddff5-swc8b" Nov 24 13:34:30 crc kubenswrapper[4790]: I1124 13:34:30.246472 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9362fee7-a172-4948-8721-b4c83bf04a30-config-data\") pod \"swift-proxy-bf58ddff5-swc8b\" (UID: \"9362fee7-a172-4948-8721-b4c83bf04a30\") " pod="openstack/swift-proxy-bf58ddff5-swc8b" Nov 24 13:34:30 crc kubenswrapper[4790]: I1124 13:34:30.246492 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9362fee7-a172-4948-8721-b4c83bf04a30-internal-tls-certs\") pod \"swift-proxy-bf58ddff5-swc8b\" (UID: \"9362fee7-a172-4948-8721-b4c83bf04a30\") " pod="openstack/swift-proxy-bf58ddff5-swc8b" Nov 24 13:34:30 crc kubenswrapper[4790]: I1124 13:34:30.246534 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9362fee7-a172-4948-8721-b4c83bf04a30-public-tls-certs\") pod \"swift-proxy-bf58ddff5-swc8b\" (UID: \"9362fee7-a172-4948-8721-b4c83bf04a30\") " pod="openstack/swift-proxy-bf58ddff5-swc8b" Nov 24 13:34:30 crc kubenswrapper[4790]: I1124 13:34:30.246553 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9362fee7-a172-4948-8721-b4c83bf04a30-combined-ca-bundle\") pod \"swift-proxy-bf58ddff5-swc8b\" (UID: \"9362fee7-a172-4948-8721-b4c83bf04a30\") " pod="openstack/swift-proxy-bf58ddff5-swc8b" Nov 24 13:34:30 crc kubenswrapper[4790]: I1124 13:34:30.248077 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9362fee7-a172-4948-8721-b4c83bf04a30-run-httpd\") pod \"swift-proxy-bf58ddff5-swc8b\" (UID: \"9362fee7-a172-4948-8721-b4c83bf04a30\") " pod="openstack/swift-proxy-bf58ddff5-swc8b" Nov 24 13:34:30 crc kubenswrapper[4790]: I1124 13:34:30.249331 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9362fee7-a172-4948-8721-b4c83bf04a30-log-httpd\") pod \"swift-proxy-bf58ddff5-swc8b\" (UID: \"9362fee7-a172-4948-8721-b4c83bf04a30\") " pod="openstack/swift-proxy-bf58ddff5-swc8b" Nov 24 13:34:30 crc kubenswrapper[4790]: I1124 13:34:30.253803 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9362fee7-a172-4948-8721-b4c83bf04a30-combined-ca-bundle\") pod \"swift-proxy-bf58ddff5-swc8b\" (UID: \"9362fee7-a172-4948-8721-b4c83bf04a30\") " pod="openstack/swift-proxy-bf58ddff5-swc8b" Nov 24 13:34:30 crc kubenswrapper[4790]: I1124 13:34:30.254751 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9362fee7-a172-4948-8721-b4c83bf04a30-etc-swift\") pod \"swift-proxy-bf58ddff5-swc8b\" (UID: \"9362fee7-a172-4948-8721-b4c83bf04a30\") " pod="openstack/swift-proxy-bf58ddff5-swc8b" Nov 24 13:34:30 crc kubenswrapper[4790]: I1124 13:34:30.273580 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9362fee7-a172-4948-8721-b4c83bf04a30-internal-tls-certs\") pod \"swift-proxy-bf58ddff5-swc8b\" (UID: \"9362fee7-a172-4948-8721-b4c83bf04a30\") " pod="openstack/swift-proxy-bf58ddff5-swc8b" Nov 24 13:34:30 crc kubenswrapper[4790]: I1124 13:34:30.277626 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9362fee7-a172-4948-8721-b4c83bf04a30-public-tls-certs\") pod \"swift-proxy-bf58ddff5-swc8b\" (UID: \"9362fee7-a172-4948-8721-b4c83bf04a30\") " pod="openstack/swift-proxy-bf58ddff5-swc8b" Nov 24 13:34:30 crc kubenswrapper[4790]: I1124 13:34:30.278421 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9362fee7-a172-4948-8721-b4c83bf04a30-config-data\") pod \"swift-proxy-bf58ddff5-swc8b\" (UID: \"9362fee7-a172-4948-8721-b4c83bf04a30\") " pod="openstack/swift-proxy-bf58ddff5-swc8b" Nov 24 13:34:30 crc kubenswrapper[4790]: I1124 13:34:30.281584 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rqhq\" (UniqueName: \"kubernetes.io/projected/9362fee7-a172-4948-8721-b4c83bf04a30-kube-api-access-7rqhq\") pod \"swift-proxy-bf58ddff5-swc8b\" (UID: \"9362fee7-a172-4948-8721-b4c83bf04a30\") " pod="openstack/swift-proxy-bf58ddff5-swc8b" Nov 24 13:34:30 crc kubenswrapper[4790]: I1124 13:34:30.317919 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-bf58ddff5-swc8b" Nov 24 13:34:30 crc kubenswrapper[4790]: I1124 13:34:30.906829 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-bf58ddff5-swc8b"] Nov 24 13:34:31 crc kubenswrapper[4790]: I1124 13:34:31.493443 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-bf58ddff5-swc8b" event={"ID":"9362fee7-a172-4948-8721-b4c83bf04a30","Type":"ContainerStarted","Data":"2893ed9292031995f05a0b830dc78e77b0e8abbf5b78c8c24b7cdaa2e9f045de"} Nov 24 13:34:31 crc kubenswrapper[4790]: I1124 13:34:31.493736 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-bf58ddff5-swc8b" event={"ID":"9362fee7-a172-4948-8721-b4c83bf04a30","Type":"ContainerStarted","Data":"3898a01261d357b913fd416fce34429e84d9cd9629907d6bc5716e0919fddbbd"} Nov 24 13:34:31 crc kubenswrapper[4790]: I1124 13:34:31.493750 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-bf58ddff5-swc8b" event={"ID":"9362fee7-a172-4948-8721-b4c83bf04a30","Type":"ContainerStarted","Data":"a19ffc076b57ad785b185695066fd43a17b7546b1762de4a4f70d8894cb8ee26"} Nov 24 13:34:31 crc kubenswrapper[4790]: I1124 13:34:31.494027 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-bf58ddff5-swc8b" Nov 24 13:34:31 crc kubenswrapper[4790]: I1124 13:34:31.494362 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-bf58ddff5-swc8b" Nov 24 13:34:32 crc kubenswrapper[4790]: I1124 13:34:32.342135 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-bf58ddff5-swc8b" podStartSLOduration=3.342119105 podStartE2EDuration="3.342119105s" podCreationTimestamp="2025-11-24 13:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:34:31.520189807 +0000 UTC m=+1319.900083469" watchObservedRunningTime="2025-11-24 13:34:32.342119105 +0000 UTC m=+1320.722012767" Nov 24 13:34:34 crc kubenswrapper[4790]: I1124 13:34:34.298966 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 24 13:34:34 crc kubenswrapper[4790]: I1124 13:34:34.638919 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:34:34 crc kubenswrapper[4790]: I1124 13:34:34.639472 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8e48f476-da57-4a24-a043-babdc557f7e7" containerName="ceilometer-central-agent" containerID="cri-o://c8368f4f05ea5c50f6456f15ba36f75f7e4acde31bd265dde01d9744c3e3c454" gracePeriod=30 Nov 24 13:34:34 crc kubenswrapper[4790]: I1124 13:34:34.639628 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8e48f476-da57-4a24-a043-babdc557f7e7" containerName="ceilometer-notification-agent" containerID="cri-o://1d5fcf4525bd2dc5f89c29d6aa5e8981e4016fd7b3d6c292093e5f559c8905d7" gracePeriod=30 Nov 24 13:34:34 crc kubenswrapper[4790]: I1124 13:34:34.639638 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8e48f476-da57-4a24-a043-babdc557f7e7" containerName="proxy-httpd" containerID="cri-o://e658fe14b9d40b94b379fe01c57f1fecccce64f3cf8d0ce939df9c88cb6a031e" gracePeriod=30 Nov 24 13:34:34 crc kubenswrapper[4790]: I1124 13:34:34.639659 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8e48f476-da57-4a24-a043-babdc557f7e7" containerName="sg-core" containerID="cri-o://f1b98a50d873e693fb3407704285e40ec164a5b3e980a457fef5e24e36ab451f" gracePeriod=30 Nov 24 13:34:34 crc kubenswrapper[4790]: I1124 13:34:34.646285 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 13:34:35 crc kubenswrapper[4790]: I1124 13:34:35.541331 4790 generic.go:334] "Generic (PLEG): container finished" podID="8e48f476-da57-4a24-a043-babdc557f7e7" containerID="e658fe14b9d40b94b379fe01c57f1fecccce64f3cf8d0ce939df9c88cb6a031e" exitCode=0 Nov 24 13:34:35 crc kubenswrapper[4790]: I1124 13:34:35.541361 4790 generic.go:334] "Generic (PLEG): container finished" podID="8e48f476-da57-4a24-a043-babdc557f7e7" containerID="f1b98a50d873e693fb3407704285e40ec164a5b3e980a457fef5e24e36ab451f" exitCode=2 Nov 24 13:34:35 crc kubenswrapper[4790]: I1124 13:34:35.541371 4790 generic.go:334] "Generic (PLEG): container finished" podID="8e48f476-da57-4a24-a043-babdc557f7e7" containerID="c8368f4f05ea5c50f6456f15ba36f75f7e4acde31bd265dde01d9744c3e3c454" exitCode=0 Nov 24 13:34:35 crc kubenswrapper[4790]: I1124 13:34:35.541390 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8e48f476-da57-4a24-a043-babdc557f7e7","Type":"ContainerDied","Data":"e658fe14b9d40b94b379fe01c57f1fecccce64f3cf8d0ce939df9c88cb6a031e"} Nov 24 13:34:35 crc kubenswrapper[4790]: I1124 13:34:35.541413 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8e48f476-da57-4a24-a043-babdc557f7e7","Type":"ContainerDied","Data":"f1b98a50d873e693fb3407704285e40ec164a5b3e980a457fef5e24e36ab451f"} Nov 24 13:34:35 crc kubenswrapper[4790]: I1124 13:34:35.541424 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8e48f476-da57-4a24-a043-babdc557f7e7","Type":"ContainerDied","Data":"c8368f4f05ea5c50f6456f15ba36f75f7e4acde31bd265dde01d9744c3e3c454"} Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.412456 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69f849cd44-8v55p" Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.528685 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/001f4ed0-f6e6-440e-a878-2bca6982ea47-ovndb-tls-certs\") pod \"001f4ed0-f6e6-440e-a878-2bca6982ea47\" (UID: \"001f4ed0-f6e6-440e-a878-2bca6982ea47\") " Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.528871 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/001f4ed0-f6e6-440e-a878-2bca6982ea47-httpd-config\") pod \"001f4ed0-f6e6-440e-a878-2bca6982ea47\" (UID: \"001f4ed0-f6e6-440e-a878-2bca6982ea47\") " Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.528917 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/001f4ed0-f6e6-440e-a878-2bca6982ea47-combined-ca-bundle\") pod \"001f4ed0-f6e6-440e-a878-2bca6982ea47\" (UID: \"001f4ed0-f6e6-440e-a878-2bca6982ea47\") " Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.529044 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjlcb\" (UniqueName: \"kubernetes.io/projected/001f4ed0-f6e6-440e-a878-2bca6982ea47-kube-api-access-kjlcb\") pod \"001f4ed0-f6e6-440e-a878-2bca6982ea47\" (UID: \"001f4ed0-f6e6-440e-a878-2bca6982ea47\") " Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.529083 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/001f4ed0-f6e6-440e-a878-2bca6982ea47-config\") pod \"001f4ed0-f6e6-440e-a878-2bca6982ea47\" (UID: \"001f4ed0-f6e6-440e-a878-2bca6982ea47\") " Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.552163 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/001f4ed0-f6e6-440e-a878-2bca6982ea47-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "001f4ed0-f6e6-440e-a878-2bca6982ea47" (UID: "001f4ed0-f6e6-440e-a878-2bca6982ea47"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.574078 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/001f4ed0-f6e6-440e-a878-2bca6982ea47-kube-api-access-kjlcb" (OuterVolumeSpecName: "kube-api-access-kjlcb") pod "001f4ed0-f6e6-440e-a878-2bca6982ea47" (UID: "001f4ed0-f6e6-440e-a878-2bca6982ea47"). InnerVolumeSpecName "kube-api-access-kjlcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.615546 4790 generic.go:334] "Generic (PLEG): container finished" podID="001f4ed0-f6e6-440e-a878-2bca6982ea47" containerID="db4596a34402f208f174677329719e056fd084851d4e5a37a88604237aa9cfd2" exitCode=0 Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.615627 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69f849cd44-8v55p" event={"ID":"001f4ed0-f6e6-440e-a878-2bca6982ea47","Type":"ContainerDied","Data":"db4596a34402f208f174677329719e056fd084851d4e5a37a88604237aa9cfd2"} Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.615657 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69f849cd44-8v55p" event={"ID":"001f4ed0-f6e6-440e-a878-2bca6982ea47","Type":"ContainerDied","Data":"f09ab5eeebda745a1632f857ef38ee2cf3134e76bbb87bc46078ee01b4771c6f"} Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.615674 4790 scope.go:117] "RemoveContainer" containerID="6537ec4e93e7d7e332d59fb429a20d95718baf70f342e2c89ccf3f495ddb3c80" Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.615796 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69f849cd44-8v55p" Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.621629 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c","Type":"ContainerStarted","Data":"0981452a5c31b7ac7be6aca3f35f652875423629b90d147c20e6f8b123ad006d"} Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.630928 4790 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/001f4ed0-f6e6-440e-a878-2bca6982ea47-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.630957 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjlcb\" (UniqueName: \"kubernetes.io/projected/001f4ed0-f6e6-440e-a878-2bca6982ea47-kube-api-access-kjlcb\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.641484 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/001f4ed0-f6e6-440e-a878-2bca6982ea47-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "001f4ed0-f6e6-440e-a878-2bca6982ea47" (UID: "001f4ed0-f6e6-440e-a878-2bca6982ea47"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.642528 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.297205446 podStartE2EDuration="12.642507578s" podCreationTimestamp="2025-11-24 13:34:25 +0000 UTC" firstStartedPulling="2025-11-24 13:34:26.738750693 +0000 UTC m=+1315.118644355" lastFinishedPulling="2025-11-24 13:34:37.084052825 +0000 UTC m=+1325.463946487" observedRunningTime="2025-11-24 13:34:37.640733378 +0000 UTC m=+1326.020627030" watchObservedRunningTime="2025-11-24 13:34:37.642507578 +0000 UTC m=+1326.022401240" Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.645545 4790 scope.go:117] "RemoveContainer" containerID="db4596a34402f208f174677329719e056fd084851d4e5a37a88604237aa9cfd2" Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.653391 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/001f4ed0-f6e6-440e-a878-2bca6982ea47-config" (OuterVolumeSpecName: "config") pod "001f4ed0-f6e6-440e-a878-2bca6982ea47" (UID: "001f4ed0-f6e6-440e-a878-2bca6982ea47"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.668064 4790 scope.go:117] "RemoveContainer" containerID="6537ec4e93e7d7e332d59fb429a20d95718baf70f342e2c89ccf3f495ddb3c80" Nov 24 13:34:37 crc kubenswrapper[4790]: E1124 13:34:37.668559 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6537ec4e93e7d7e332d59fb429a20d95718baf70f342e2c89ccf3f495ddb3c80\": container with ID starting with 6537ec4e93e7d7e332d59fb429a20d95718baf70f342e2c89ccf3f495ddb3c80 not found: ID does not exist" containerID="6537ec4e93e7d7e332d59fb429a20d95718baf70f342e2c89ccf3f495ddb3c80" Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.668594 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6537ec4e93e7d7e332d59fb429a20d95718baf70f342e2c89ccf3f495ddb3c80"} err="failed to get container status \"6537ec4e93e7d7e332d59fb429a20d95718baf70f342e2c89ccf3f495ddb3c80\": rpc error: code = NotFound desc = could not find container \"6537ec4e93e7d7e332d59fb429a20d95718baf70f342e2c89ccf3f495ddb3c80\": container with ID starting with 6537ec4e93e7d7e332d59fb429a20d95718baf70f342e2c89ccf3f495ddb3c80 not found: ID does not exist" Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.668646 4790 scope.go:117] "RemoveContainer" containerID="db4596a34402f208f174677329719e056fd084851d4e5a37a88604237aa9cfd2" Nov 24 13:34:37 crc kubenswrapper[4790]: E1124 13:34:37.669113 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db4596a34402f208f174677329719e056fd084851d4e5a37a88604237aa9cfd2\": container with ID starting with db4596a34402f208f174677329719e056fd084851d4e5a37a88604237aa9cfd2 not found: ID does not exist" containerID="db4596a34402f208f174677329719e056fd084851d4e5a37a88604237aa9cfd2" Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.669162 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db4596a34402f208f174677329719e056fd084851d4e5a37a88604237aa9cfd2"} err="failed to get container status \"db4596a34402f208f174677329719e056fd084851d4e5a37a88604237aa9cfd2\": rpc error: code = NotFound desc = could not find container \"db4596a34402f208f174677329719e056fd084851d4e5a37a88604237aa9cfd2\": container with ID starting with db4596a34402f208f174677329719e056fd084851d4e5a37a88604237aa9cfd2 not found: ID does not exist" Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.671018 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/001f4ed0-f6e6-440e-a878-2bca6982ea47-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "001f4ed0-f6e6-440e-a878-2bca6982ea47" (UID: "001f4ed0-f6e6-440e-a878-2bca6982ea47"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.713145 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="8e48f476-da57-4a24-a043-babdc557f7e7" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.158:3000/\": dial tcp 10.217.0.158:3000: connect: connection refused" Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.733235 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/001f4ed0-f6e6-440e-a878-2bca6982ea47-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.733263 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/001f4ed0-f6e6-440e-a878-2bca6982ea47-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.733272 4790 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/001f4ed0-f6e6-440e-a878-2bca6982ea47-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.845052 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.845926 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="69617248-5d4b-4027-a2a9-e972278e9d5e" containerName="glance-log" containerID="cri-o://590c7623f08a06bb7521b7040e8300b1caee6cdbd6b60c1d3823cf11d4cb0fbb" gracePeriod=30 Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.846209 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="69617248-5d4b-4027-a2a9-e972278e9d5e" containerName="glance-httpd" containerID="cri-o://1783160f79b92abc9fcc7d3d325d99cf4fd66d34a919d93503b4b9eb6f006447" gracePeriod=30 Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.945833 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-69f849cd44-8v55p"] Nov 24 13:34:37 crc kubenswrapper[4790]: I1124 13:34:37.962591 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-69f849cd44-8v55p"] Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.303307 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.323916 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="001f4ed0-f6e6-440e-a878-2bca6982ea47" path="/var/lib/kubelet/pods/001f4ed0-f6e6-440e-a878-2bca6982ea47/volumes" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.450920 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e48f476-da57-4a24-a043-babdc557f7e7-combined-ca-bundle\") pod \"8e48f476-da57-4a24-a043-babdc557f7e7\" (UID: \"8e48f476-da57-4a24-a043-babdc557f7e7\") " Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.451016 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8e48f476-da57-4a24-a043-babdc557f7e7-log-httpd\") pod \"8e48f476-da57-4a24-a043-babdc557f7e7\" (UID: \"8e48f476-da57-4a24-a043-babdc557f7e7\") " Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.451083 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8e48f476-da57-4a24-a043-babdc557f7e7-sg-core-conf-yaml\") pod \"8e48f476-da57-4a24-a043-babdc557f7e7\" (UID: \"8e48f476-da57-4a24-a043-babdc557f7e7\") " Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.451100 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e48f476-da57-4a24-a043-babdc557f7e7-config-data\") pod \"8e48f476-da57-4a24-a043-babdc557f7e7\" (UID: \"8e48f476-da57-4a24-a043-babdc557f7e7\") " Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.451141 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8e48f476-da57-4a24-a043-babdc557f7e7-run-httpd\") pod \"8e48f476-da57-4a24-a043-babdc557f7e7\" (UID: \"8e48f476-da57-4a24-a043-babdc557f7e7\") " Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.451184 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8hcr\" (UniqueName: \"kubernetes.io/projected/8e48f476-da57-4a24-a043-babdc557f7e7-kube-api-access-z8hcr\") pod \"8e48f476-da57-4a24-a043-babdc557f7e7\" (UID: \"8e48f476-da57-4a24-a043-babdc557f7e7\") " Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.451211 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e48f476-da57-4a24-a043-babdc557f7e7-scripts\") pod \"8e48f476-da57-4a24-a043-babdc557f7e7\" (UID: \"8e48f476-da57-4a24-a043-babdc557f7e7\") " Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.452715 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e48f476-da57-4a24-a043-babdc557f7e7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8e48f476-da57-4a24-a043-babdc557f7e7" (UID: "8e48f476-da57-4a24-a043-babdc557f7e7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.452997 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e48f476-da57-4a24-a043-babdc557f7e7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8e48f476-da57-4a24-a043-babdc557f7e7" (UID: "8e48f476-da57-4a24-a043-babdc557f7e7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.456118 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e48f476-da57-4a24-a043-babdc557f7e7-scripts" (OuterVolumeSpecName: "scripts") pod "8e48f476-da57-4a24-a043-babdc557f7e7" (UID: "8e48f476-da57-4a24-a043-babdc557f7e7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.473267 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e48f476-da57-4a24-a043-babdc557f7e7-kube-api-access-z8hcr" (OuterVolumeSpecName: "kube-api-access-z8hcr") pod "8e48f476-da57-4a24-a043-babdc557f7e7" (UID: "8e48f476-da57-4a24-a043-babdc557f7e7"). InnerVolumeSpecName "kube-api-access-z8hcr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.484029 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e48f476-da57-4a24-a043-babdc557f7e7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8e48f476-da57-4a24-a043-babdc557f7e7" (UID: "8e48f476-da57-4a24-a043-babdc557f7e7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.534407 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e48f476-da57-4a24-a043-babdc557f7e7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8e48f476-da57-4a24-a043-babdc557f7e7" (UID: "8e48f476-da57-4a24-a043-babdc557f7e7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.553199 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e48f476-da57-4a24-a043-babdc557f7e7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.553227 4790 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8e48f476-da57-4a24-a043-babdc557f7e7-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.553236 4790 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8e48f476-da57-4a24-a043-babdc557f7e7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.553244 4790 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8e48f476-da57-4a24-a043-babdc557f7e7-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.553252 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8hcr\" (UniqueName: \"kubernetes.io/projected/8e48f476-da57-4a24-a043-babdc557f7e7-kube-api-access-z8hcr\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.553262 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e48f476-da57-4a24-a043-babdc557f7e7-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.553418 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e48f476-da57-4a24-a043-babdc557f7e7-config-data" (OuterVolumeSpecName: "config-data") pod "8e48f476-da57-4a24-a043-babdc557f7e7" (UID: "8e48f476-da57-4a24-a043-babdc557f7e7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.634281 4790 generic.go:334] "Generic (PLEG): container finished" podID="69617248-5d4b-4027-a2a9-e972278e9d5e" containerID="590c7623f08a06bb7521b7040e8300b1caee6cdbd6b60c1d3823cf11d4cb0fbb" exitCode=143 Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.634368 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"69617248-5d4b-4027-a2a9-e972278e9d5e","Type":"ContainerDied","Data":"590c7623f08a06bb7521b7040e8300b1caee6cdbd6b60c1d3823cf11d4cb0fbb"} Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.636728 4790 generic.go:334] "Generic (PLEG): container finished" podID="8e48f476-da57-4a24-a043-babdc557f7e7" containerID="1d5fcf4525bd2dc5f89c29d6aa5e8981e4016fd7b3d6c292093e5f559c8905d7" exitCode=0 Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.636770 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8e48f476-da57-4a24-a043-babdc557f7e7","Type":"ContainerDied","Data":"1d5fcf4525bd2dc5f89c29d6aa5e8981e4016fd7b3d6c292093e5f559c8905d7"} Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.636790 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8e48f476-da57-4a24-a043-babdc557f7e7","Type":"ContainerDied","Data":"6e3d7ebb68ed6a22674d83e672a6047909c076e8a6b433bf10292081cfce42c0"} Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.636807 4790 scope.go:117] "RemoveContainer" containerID="e658fe14b9d40b94b379fe01c57f1fecccce64f3cf8d0ce939df9c88cb6a031e" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.636953 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.654451 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e48f476-da57-4a24-a043-babdc557f7e7-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.671114 4790 scope.go:117] "RemoveContainer" containerID="f1b98a50d873e693fb3407704285e40ec164a5b3e980a457fef5e24e36ab451f" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.676060 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.684690 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.714208 4790 scope.go:117] "RemoveContainer" containerID="1d5fcf4525bd2dc5f89c29d6aa5e8981e4016fd7b3d6c292093e5f559c8905d7" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.718118 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:34:38 crc kubenswrapper[4790]: E1124 13:34:38.718456 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e48f476-da57-4a24-a043-babdc557f7e7" containerName="ceilometer-notification-agent" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.718471 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e48f476-da57-4a24-a043-babdc557f7e7" containerName="ceilometer-notification-agent" Nov 24 13:34:38 crc kubenswrapper[4790]: E1124 13:34:38.718486 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e48f476-da57-4a24-a043-babdc557f7e7" containerName="ceilometer-central-agent" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.718492 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e48f476-da57-4a24-a043-babdc557f7e7" containerName="ceilometer-central-agent" Nov 24 13:34:38 crc kubenswrapper[4790]: E1124 13:34:38.718502 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="001f4ed0-f6e6-440e-a878-2bca6982ea47" containerName="neutron-api" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.718508 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="001f4ed0-f6e6-440e-a878-2bca6982ea47" containerName="neutron-api" Nov 24 13:34:38 crc kubenswrapper[4790]: E1124 13:34:38.718520 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e48f476-da57-4a24-a043-babdc557f7e7" containerName="sg-core" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.718526 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e48f476-da57-4a24-a043-babdc557f7e7" containerName="sg-core" Nov 24 13:34:38 crc kubenswrapper[4790]: E1124 13:34:38.718543 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e48f476-da57-4a24-a043-babdc557f7e7" containerName="proxy-httpd" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.718549 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e48f476-da57-4a24-a043-babdc557f7e7" containerName="proxy-httpd" Nov 24 13:34:38 crc kubenswrapper[4790]: E1124 13:34:38.718572 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="001f4ed0-f6e6-440e-a878-2bca6982ea47" containerName="neutron-httpd" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.718577 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="001f4ed0-f6e6-440e-a878-2bca6982ea47" containerName="neutron-httpd" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.718771 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="001f4ed0-f6e6-440e-a878-2bca6982ea47" containerName="neutron-api" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.718785 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e48f476-da57-4a24-a043-babdc557f7e7" containerName="ceilometer-notification-agent" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.718797 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="001f4ed0-f6e6-440e-a878-2bca6982ea47" containerName="neutron-httpd" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.718807 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e48f476-da57-4a24-a043-babdc557f7e7" containerName="sg-core" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.718818 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e48f476-da57-4a24-a043-babdc557f7e7" containerName="ceilometer-central-agent" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.718829 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e48f476-da57-4a24-a043-babdc557f7e7" containerName="proxy-httpd" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.722065 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.726136 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.726372 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.732055 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.757495 4790 scope.go:117] "RemoveContainer" containerID="c8368f4f05ea5c50f6456f15ba36f75f7e4acde31bd265dde01d9744c3e3c454" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.760205 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/63ac46eb-bf90-4a9f-84ec-35f5b8079766-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\") " pod="openstack/ceilometer-0" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.760337 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63ac46eb-bf90-4a9f-84ec-35f5b8079766-run-httpd\") pod \"ceilometer-0\" (UID: \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\") " pod="openstack/ceilometer-0" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.760373 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jkv6\" (UniqueName: \"kubernetes.io/projected/63ac46eb-bf90-4a9f-84ec-35f5b8079766-kube-api-access-7jkv6\") pod \"ceilometer-0\" (UID: \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\") " pod="openstack/ceilometer-0" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.760388 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63ac46eb-bf90-4a9f-84ec-35f5b8079766-config-data\") pod \"ceilometer-0\" (UID: \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\") " pod="openstack/ceilometer-0" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.760423 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63ac46eb-bf90-4a9f-84ec-35f5b8079766-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\") " pod="openstack/ceilometer-0" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.760447 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63ac46eb-bf90-4a9f-84ec-35f5b8079766-scripts\") pod \"ceilometer-0\" (UID: \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\") " pod="openstack/ceilometer-0" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.760477 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63ac46eb-bf90-4a9f-84ec-35f5b8079766-log-httpd\") pod \"ceilometer-0\" (UID: \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\") " pod="openstack/ceilometer-0" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.776752 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.777382 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7fe37e44-47d4-4a0e-9cc1-4920db57f7d5" containerName="glance-log" containerID="cri-o://eecb4a557c0f4d6e2a5c95fa519ba7f6cd2b3df6ed7cb80d7301eb06b36f512d" gracePeriod=30 Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.777832 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7fe37e44-47d4-4a0e-9cc1-4920db57f7d5" containerName="glance-httpd" containerID="cri-o://ad8dcf46bbcdce25e8690aef88e66ffa33d828aa18a8a5458f8950ce7d7f7349" gracePeriod=30 Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.799120 4790 scope.go:117] "RemoveContainer" containerID="e658fe14b9d40b94b379fe01c57f1fecccce64f3cf8d0ce939df9c88cb6a031e" Nov 24 13:34:38 crc kubenswrapper[4790]: E1124 13:34:38.799680 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e658fe14b9d40b94b379fe01c57f1fecccce64f3cf8d0ce939df9c88cb6a031e\": container with ID starting with e658fe14b9d40b94b379fe01c57f1fecccce64f3cf8d0ce939df9c88cb6a031e not found: ID does not exist" containerID="e658fe14b9d40b94b379fe01c57f1fecccce64f3cf8d0ce939df9c88cb6a031e" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.799723 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e658fe14b9d40b94b379fe01c57f1fecccce64f3cf8d0ce939df9c88cb6a031e"} err="failed to get container status \"e658fe14b9d40b94b379fe01c57f1fecccce64f3cf8d0ce939df9c88cb6a031e\": rpc error: code = NotFound desc = could not find container \"e658fe14b9d40b94b379fe01c57f1fecccce64f3cf8d0ce939df9c88cb6a031e\": container with ID starting with e658fe14b9d40b94b379fe01c57f1fecccce64f3cf8d0ce939df9c88cb6a031e not found: ID does not exist" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.799749 4790 scope.go:117] "RemoveContainer" containerID="f1b98a50d873e693fb3407704285e40ec164a5b3e980a457fef5e24e36ab451f" Nov 24 13:34:38 crc kubenswrapper[4790]: E1124 13:34:38.800069 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1b98a50d873e693fb3407704285e40ec164a5b3e980a457fef5e24e36ab451f\": container with ID starting with f1b98a50d873e693fb3407704285e40ec164a5b3e980a457fef5e24e36ab451f not found: ID does not exist" containerID="f1b98a50d873e693fb3407704285e40ec164a5b3e980a457fef5e24e36ab451f" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.800114 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1b98a50d873e693fb3407704285e40ec164a5b3e980a457fef5e24e36ab451f"} err="failed to get container status \"f1b98a50d873e693fb3407704285e40ec164a5b3e980a457fef5e24e36ab451f\": rpc error: code = NotFound desc = could not find container \"f1b98a50d873e693fb3407704285e40ec164a5b3e980a457fef5e24e36ab451f\": container with ID starting with f1b98a50d873e693fb3407704285e40ec164a5b3e980a457fef5e24e36ab451f not found: ID does not exist" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.800127 4790 scope.go:117] "RemoveContainer" containerID="1d5fcf4525bd2dc5f89c29d6aa5e8981e4016fd7b3d6c292093e5f559c8905d7" Nov 24 13:34:38 crc kubenswrapper[4790]: E1124 13:34:38.800361 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d5fcf4525bd2dc5f89c29d6aa5e8981e4016fd7b3d6c292093e5f559c8905d7\": container with ID starting with 1d5fcf4525bd2dc5f89c29d6aa5e8981e4016fd7b3d6c292093e5f559c8905d7 not found: ID does not exist" containerID="1d5fcf4525bd2dc5f89c29d6aa5e8981e4016fd7b3d6c292093e5f559c8905d7" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.800379 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d5fcf4525bd2dc5f89c29d6aa5e8981e4016fd7b3d6c292093e5f559c8905d7"} err="failed to get container status \"1d5fcf4525bd2dc5f89c29d6aa5e8981e4016fd7b3d6c292093e5f559c8905d7\": rpc error: code = NotFound desc = could not find container \"1d5fcf4525bd2dc5f89c29d6aa5e8981e4016fd7b3d6c292093e5f559c8905d7\": container with ID starting with 1d5fcf4525bd2dc5f89c29d6aa5e8981e4016fd7b3d6c292093e5f559c8905d7 not found: ID does not exist" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.800390 4790 scope.go:117] "RemoveContainer" containerID="c8368f4f05ea5c50f6456f15ba36f75f7e4acde31bd265dde01d9744c3e3c454" Nov 24 13:34:38 crc kubenswrapper[4790]: E1124 13:34:38.800741 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8368f4f05ea5c50f6456f15ba36f75f7e4acde31bd265dde01d9744c3e3c454\": container with ID starting with c8368f4f05ea5c50f6456f15ba36f75f7e4acde31bd265dde01d9744c3e3c454 not found: ID does not exist" containerID="c8368f4f05ea5c50f6456f15ba36f75f7e4acde31bd265dde01d9744c3e3c454" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.800763 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8368f4f05ea5c50f6456f15ba36f75f7e4acde31bd265dde01d9744c3e3c454"} err="failed to get container status \"c8368f4f05ea5c50f6456f15ba36f75f7e4acde31bd265dde01d9744c3e3c454\": rpc error: code = NotFound desc = could not find container \"c8368f4f05ea5c50f6456f15ba36f75f7e4acde31bd265dde01d9744c3e3c454\": container with ID starting with c8368f4f05ea5c50f6456f15ba36f75f7e4acde31bd265dde01d9744c3e3c454 not found: ID does not exist" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.861823 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/63ac46eb-bf90-4a9f-84ec-35f5b8079766-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\") " pod="openstack/ceilometer-0" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.861920 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63ac46eb-bf90-4a9f-84ec-35f5b8079766-run-httpd\") pod \"ceilometer-0\" (UID: \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\") " pod="openstack/ceilometer-0" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.861955 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jkv6\" (UniqueName: \"kubernetes.io/projected/63ac46eb-bf90-4a9f-84ec-35f5b8079766-kube-api-access-7jkv6\") pod \"ceilometer-0\" (UID: \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\") " pod="openstack/ceilometer-0" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.861972 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63ac46eb-bf90-4a9f-84ec-35f5b8079766-config-data\") pod \"ceilometer-0\" (UID: \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\") " pod="openstack/ceilometer-0" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.861997 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63ac46eb-bf90-4a9f-84ec-35f5b8079766-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\") " pod="openstack/ceilometer-0" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.862017 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63ac46eb-bf90-4a9f-84ec-35f5b8079766-scripts\") pod \"ceilometer-0\" (UID: \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\") " pod="openstack/ceilometer-0" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.862035 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63ac46eb-bf90-4a9f-84ec-35f5b8079766-log-httpd\") pod \"ceilometer-0\" (UID: \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\") " pod="openstack/ceilometer-0" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.862422 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63ac46eb-bf90-4a9f-84ec-35f5b8079766-log-httpd\") pod \"ceilometer-0\" (UID: \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\") " pod="openstack/ceilometer-0" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.862858 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63ac46eb-bf90-4a9f-84ec-35f5b8079766-run-httpd\") pod \"ceilometer-0\" (UID: \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\") " pod="openstack/ceilometer-0" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.867592 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/63ac46eb-bf90-4a9f-84ec-35f5b8079766-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\") " pod="openstack/ceilometer-0" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.867704 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63ac46eb-bf90-4a9f-84ec-35f5b8079766-config-data\") pod \"ceilometer-0\" (UID: \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\") " pod="openstack/ceilometer-0" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.868173 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63ac46eb-bf90-4a9f-84ec-35f5b8079766-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\") " pod="openstack/ceilometer-0" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.872650 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63ac46eb-bf90-4a9f-84ec-35f5b8079766-scripts\") pod \"ceilometer-0\" (UID: \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\") " pod="openstack/ceilometer-0" Nov 24 13:34:38 crc kubenswrapper[4790]: I1124 13:34:38.881034 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jkv6\" (UniqueName: \"kubernetes.io/projected/63ac46eb-bf90-4a9f-84ec-35f5b8079766-kube-api-access-7jkv6\") pod \"ceilometer-0\" (UID: \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\") " pod="openstack/ceilometer-0" Nov 24 13:34:39 crc kubenswrapper[4790]: I1124 13:34:39.060792 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:34:39 crc kubenswrapper[4790]: I1124 13:34:39.386202 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:34:39 crc kubenswrapper[4790]: I1124 13:34:39.474855 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:34:39 crc kubenswrapper[4790]: I1124 13:34:39.652784 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"63ac46eb-bf90-4a9f-84ec-35f5b8079766","Type":"ContainerStarted","Data":"e7aedbd2d9c4f164cd5f9c561b07d58f822634c8f673d5db2747ba10a75d637d"} Nov 24 13:34:39 crc kubenswrapper[4790]: I1124 13:34:39.658402 4790 generic.go:334] "Generic (PLEG): container finished" podID="7fe37e44-47d4-4a0e-9cc1-4920db57f7d5" containerID="eecb4a557c0f4d6e2a5c95fa519ba7f6cd2b3df6ed7cb80d7301eb06b36f512d" exitCode=143 Nov 24 13:34:39 crc kubenswrapper[4790]: I1124 13:34:39.658447 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5","Type":"ContainerDied","Data":"eecb4a557c0f4d6e2a5c95fa519ba7f6cd2b3df6ed7cb80d7301eb06b36f512d"} Nov 24 13:34:40 crc kubenswrapper[4790]: I1124 13:34:40.329517 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e48f476-da57-4a24-a043-babdc557f7e7" path="/var/lib/kubelet/pods/8e48f476-da57-4a24-a043-babdc557f7e7/volumes" Nov 24 13:34:40 crc kubenswrapper[4790]: I1124 13:34:40.330816 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-bf58ddff5-swc8b" Nov 24 13:34:40 crc kubenswrapper[4790]: I1124 13:34:40.330848 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-bf58ddff5-swc8b" Nov 24 13:34:40 crc kubenswrapper[4790]: I1124 13:34:40.670907 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"63ac46eb-bf90-4a9f-84ec-35f5b8079766","Type":"ContainerStarted","Data":"fd2b559f111df55674a77621a3c07fbb31acd9726dd77eb7839df232e0182d8b"} Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.491214 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.609924 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/69617248-5d4b-4027-a2a9-e972278e9d5e-httpd-run\") pod \"69617248-5d4b-4027-a2a9-e972278e9d5e\" (UID: \"69617248-5d4b-4027-a2a9-e972278e9d5e\") " Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.610298 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69617248-5d4b-4027-a2a9-e972278e9d5e-combined-ca-bundle\") pod \"69617248-5d4b-4027-a2a9-e972278e9d5e\" (UID: \"69617248-5d4b-4027-a2a9-e972278e9d5e\") " Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.610323 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69617248-5d4b-4027-a2a9-e972278e9d5e-scripts\") pod \"69617248-5d4b-4027-a2a9-e972278e9d5e\" (UID: \"69617248-5d4b-4027-a2a9-e972278e9d5e\") " Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.610369 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/69617248-5d4b-4027-a2a9-e972278e9d5e-public-tls-certs\") pod \"69617248-5d4b-4027-a2a9-e972278e9d5e\" (UID: \"69617248-5d4b-4027-a2a9-e972278e9d5e\") " Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.610395 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69617248-5d4b-4027-a2a9-e972278e9d5e-config-data\") pod \"69617248-5d4b-4027-a2a9-e972278e9d5e\" (UID: \"69617248-5d4b-4027-a2a9-e972278e9d5e\") " Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.610446 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thlrp\" (UniqueName: \"kubernetes.io/projected/69617248-5d4b-4027-a2a9-e972278e9d5e-kube-api-access-thlrp\") pod \"69617248-5d4b-4027-a2a9-e972278e9d5e\" (UID: \"69617248-5d4b-4027-a2a9-e972278e9d5e\") " Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.610471 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"69617248-5d4b-4027-a2a9-e972278e9d5e\" (UID: \"69617248-5d4b-4027-a2a9-e972278e9d5e\") " Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.610502 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69617248-5d4b-4027-a2a9-e972278e9d5e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "69617248-5d4b-4027-a2a9-e972278e9d5e" (UID: "69617248-5d4b-4027-a2a9-e972278e9d5e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.610534 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69617248-5d4b-4027-a2a9-e972278e9d5e-logs\") pod \"69617248-5d4b-4027-a2a9-e972278e9d5e\" (UID: \"69617248-5d4b-4027-a2a9-e972278e9d5e\") " Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.610832 4790 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/69617248-5d4b-4027-a2a9-e972278e9d5e-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.611147 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69617248-5d4b-4027-a2a9-e972278e9d5e-logs" (OuterVolumeSpecName: "logs") pod "69617248-5d4b-4027-a2a9-e972278e9d5e" (UID: "69617248-5d4b-4027-a2a9-e972278e9d5e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.615548 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69617248-5d4b-4027-a2a9-e972278e9d5e-scripts" (OuterVolumeSpecName: "scripts") pod "69617248-5d4b-4027-a2a9-e972278e9d5e" (UID: "69617248-5d4b-4027-a2a9-e972278e9d5e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.615613 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69617248-5d4b-4027-a2a9-e972278e9d5e-kube-api-access-thlrp" (OuterVolumeSpecName: "kube-api-access-thlrp") pod "69617248-5d4b-4027-a2a9-e972278e9d5e" (UID: "69617248-5d4b-4027-a2a9-e972278e9d5e"). InnerVolumeSpecName "kube-api-access-thlrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.618067 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "69617248-5d4b-4027-a2a9-e972278e9d5e" (UID: "69617248-5d4b-4027-a2a9-e972278e9d5e"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.638264 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69617248-5d4b-4027-a2a9-e972278e9d5e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "69617248-5d4b-4027-a2a9-e972278e9d5e" (UID: "69617248-5d4b-4027-a2a9-e972278e9d5e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.664994 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69617248-5d4b-4027-a2a9-e972278e9d5e-config-data" (OuterVolumeSpecName: "config-data") pod "69617248-5d4b-4027-a2a9-e972278e9d5e" (UID: "69617248-5d4b-4027-a2a9-e972278e9d5e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.667650 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69617248-5d4b-4027-a2a9-e972278e9d5e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "69617248-5d4b-4027-a2a9-e972278e9d5e" (UID: "69617248-5d4b-4027-a2a9-e972278e9d5e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.680116 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"63ac46eb-bf90-4a9f-84ec-35f5b8079766","Type":"ContainerStarted","Data":"382f16954cb5e0b4469a639f71a43a8f6acd9877668741eb6de89a93416768d1"} Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.680169 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"63ac46eb-bf90-4a9f-84ec-35f5b8079766","Type":"ContainerStarted","Data":"01cbbfdf1ad3f830c64fd15d6ba213032f264a89ad6cbfc71d2a6609c93e194f"} Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.682253 4790 generic.go:334] "Generic (PLEG): container finished" podID="69617248-5d4b-4027-a2a9-e972278e9d5e" containerID="1783160f79b92abc9fcc7d3d325d99cf4fd66d34a919d93503b4b9eb6f006447" exitCode=0 Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.682286 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"69617248-5d4b-4027-a2a9-e972278e9d5e","Type":"ContainerDied","Data":"1783160f79b92abc9fcc7d3d325d99cf4fd66d34a919d93503b4b9eb6f006447"} Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.682307 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"69617248-5d4b-4027-a2a9-e972278e9d5e","Type":"ContainerDied","Data":"16744171db16176fa3a2049b1e44d1307eb50cc23f3c2b08a0fd5cff09d24569"} Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.682308 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.682322 4790 scope.go:117] "RemoveContainer" containerID="1783160f79b92abc9fcc7d3d325d99cf4fd66d34a919d93503b4b9eb6f006447" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.712391 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69617248-5d4b-4027-a2a9-e972278e9d5e-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.712421 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69617248-5d4b-4027-a2a9-e972278e9d5e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.712433 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69617248-5d4b-4027-a2a9-e972278e9d5e-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.712445 4790 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/69617248-5d4b-4027-a2a9-e972278e9d5e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.712563 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69617248-5d4b-4027-a2a9-e972278e9d5e-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.712577 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thlrp\" (UniqueName: \"kubernetes.io/projected/69617248-5d4b-4027-a2a9-e972278e9d5e-kube-api-access-thlrp\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.712632 4790 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.718057 4790 scope.go:117] "RemoveContainer" containerID="590c7623f08a06bb7521b7040e8300b1caee6cdbd6b60c1d3823cf11d4cb0fbb" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.736957 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.738066 4790 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.747754 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.752393 4790 scope.go:117] "RemoveContainer" containerID="1783160f79b92abc9fcc7d3d325d99cf4fd66d34a919d93503b4b9eb6f006447" Nov 24 13:34:41 crc kubenswrapper[4790]: E1124 13:34:41.753653 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1783160f79b92abc9fcc7d3d325d99cf4fd66d34a919d93503b4b9eb6f006447\": container with ID starting with 1783160f79b92abc9fcc7d3d325d99cf4fd66d34a919d93503b4b9eb6f006447 not found: ID does not exist" containerID="1783160f79b92abc9fcc7d3d325d99cf4fd66d34a919d93503b4b9eb6f006447" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.753699 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1783160f79b92abc9fcc7d3d325d99cf4fd66d34a919d93503b4b9eb6f006447"} err="failed to get container status \"1783160f79b92abc9fcc7d3d325d99cf4fd66d34a919d93503b4b9eb6f006447\": rpc error: code = NotFound desc = could not find container \"1783160f79b92abc9fcc7d3d325d99cf4fd66d34a919d93503b4b9eb6f006447\": container with ID starting with 1783160f79b92abc9fcc7d3d325d99cf4fd66d34a919d93503b4b9eb6f006447 not found: ID does not exist" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.753729 4790 scope.go:117] "RemoveContainer" containerID="590c7623f08a06bb7521b7040e8300b1caee6cdbd6b60c1d3823cf11d4cb0fbb" Nov 24 13:34:41 crc kubenswrapper[4790]: E1124 13:34:41.754179 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"590c7623f08a06bb7521b7040e8300b1caee6cdbd6b60c1d3823cf11d4cb0fbb\": container with ID starting with 590c7623f08a06bb7521b7040e8300b1caee6cdbd6b60c1d3823cf11d4cb0fbb not found: ID does not exist" containerID="590c7623f08a06bb7521b7040e8300b1caee6cdbd6b60c1d3823cf11d4cb0fbb" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.754209 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"590c7623f08a06bb7521b7040e8300b1caee6cdbd6b60c1d3823cf11d4cb0fbb"} err="failed to get container status \"590c7623f08a06bb7521b7040e8300b1caee6cdbd6b60c1d3823cf11d4cb0fbb\": rpc error: code = NotFound desc = could not find container \"590c7623f08a06bb7521b7040e8300b1caee6cdbd6b60c1d3823cf11d4cb0fbb\": container with ID starting with 590c7623f08a06bb7521b7040e8300b1caee6cdbd6b60c1d3823cf11d4cb0fbb not found: ID does not exist" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.756433 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 13:34:41 crc kubenswrapper[4790]: E1124 13:34:41.757009 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69617248-5d4b-4027-a2a9-e972278e9d5e" containerName="glance-httpd" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.757030 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="69617248-5d4b-4027-a2a9-e972278e9d5e" containerName="glance-httpd" Nov 24 13:34:41 crc kubenswrapper[4790]: E1124 13:34:41.757054 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69617248-5d4b-4027-a2a9-e972278e9d5e" containerName="glance-log" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.757062 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="69617248-5d4b-4027-a2a9-e972278e9d5e" containerName="glance-log" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.757259 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="69617248-5d4b-4027-a2a9-e972278e9d5e" containerName="glance-httpd" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.757281 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="69617248-5d4b-4027-a2a9-e972278e9d5e" containerName="glance-log" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.758434 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.760818 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.762436 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.768374 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.813879 4790 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.915856 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cbd6e74-a758-4bbf-81eb-acffdfa56955-config-data\") pod \"glance-default-external-api-0\" (UID: \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.915919 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8cbd6e74-a758-4bbf-81eb-acffdfa56955-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.915948 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cbd6e74-a758-4bbf-81eb-acffdfa56955-scripts\") pod \"glance-default-external-api-0\" (UID: \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.915978 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cbd6e74-a758-4bbf-81eb-acffdfa56955-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.916006 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cbd6e74-a758-4bbf-81eb-acffdfa56955-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.916268 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cbd6e74-a758-4bbf-81eb-acffdfa56955-logs\") pod \"glance-default-external-api-0\" (UID: \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.916340 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:41 crc kubenswrapper[4790]: I1124 13:34:41.916372 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsvp6\" (UniqueName: \"kubernetes.io/projected/8cbd6e74-a758-4bbf-81eb-acffdfa56955-kube-api-access-xsvp6\") pod \"glance-default-external-api-0\" (UID: \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:42 crc kubenswrapper[4790]: I1124 13:34:42.018463 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cbd6e74-a758-4bbf-81eb-acffdfa56955-logs\") pod \"glance-default-external-api-0\" (UID: \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:42 crc kubenswrapper[4790]: I1124 13:34:42.018516 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:42 crc kubenswrapper[4790]: I1124 13:34:42.018545 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsvp6\" (UniqueName: \"kubernetes.io/projected/8cbd6e74-a758-4bbf-81eb-acffdfa56955-kube-api-access-xsvp6\") pod \"glance-default-external-api-0\" (UID: \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:42 crc kubenswrapper[4790]: I1124 13:34:42.018666 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cbd6e74-a758-4bbf-81eb-acffdfa56955-config-data\") pod \"glance-default-external-api-0\" (UID: \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:42 crc kubenswrapper[4790]: I1124 13:34:42.018697 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8cbd6e74-a758-4bbf-81eb-acffdfa56955-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:42 crc kubenswrapper[4790]: I1124 13:34:42.018722 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cbd6e74-a758-4bbf-81eb-acffdfa56955-scripts\") pod \"glance-default-external-api-0\" (UID: \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:42 crc kubenswrapper[4790]: I1124 13:34:42.018752 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cbd6e74-a758-4bbf-81eb-acffdfa56955-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:42 crc kubenswrapper[4790]: I1124 13:34:42.018776 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cbd6e74-a758-4bbf-81eb-acffdfa56955-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:42 crc kubenswrapper[4790]: I1124 13:34:42.019734 4790 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Nov 24 13:34:42 crc kubenswrapper[4790]: I1124 13:34:42.020189 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cbd6e74-a758-4bbf-81eb-acffdfa56955-logs\") pod \"glance-default-external-api-0\" (UID: \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:42 crc kubenswrapper[4790]: I1124 13:34:42.020254 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8cbd6e74-a758-4bbf-81eb-acffdfa56955-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:42 crc kubenswrapper[4790]: I1124 13:34:42.023488 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cbd6e74-a758-4bbf-81eb-acffdfa56955-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:42 crc kubenswrapper[4790]: I1124 13:34:42.024722 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cbd6e74-a758-4bbf-81eb-acffdfa56955-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:42 crc kubenswrapper[4790]: I1124 13:34:42.027948 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cbd6e74-a758-4bbf-81eb-acffdfa56955-config-data\") pod \"glance-default-external-api-0\" (UID: \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:42 crc kubenswrapper[4790]: I1124 13:34:42.033710 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cbd6e74-a758-4bbf-81eb-acffdfa56955-scripts\") pod \"glance-default-external-api-0\" (UID: \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:42 crc kubenswrapper[4790]: I1124 13:34:42.046563 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsvp6\" (UniqueName: \"kubernetes.io/projected/8cbd6e74-a758-4bbf-81eb-acffdfa56955-kube-api-access-xsvp6\") pod \"glance-default-external-api-0\" (UID: \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:42 crc kubenswrapper[4790]: I1124 13:34:42.048084 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:42 crc kubenswrapper[4790]: I1124 13:34:42.089146 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 13:34:42 crc kubenswrapper[4790]: I1124 13:34:42.346797 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69617248-5d4b-4027-a2a9-e972278e9d5e" path="/var/lib/kubelet/pods/69617248-5d4b-4027-a2a9-e972278e9d5e/volumes" Nov 24 13:34:42 crc kubenswrapper[4790]: I1124 13:34:42.599677 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 13:34:42 crc kubenswrapper[4790]: W1124 13:34:42.612463 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cbd6e74_a758_4bbf_81eb_acffdfa56955.slice/crio-9fff344ac0471c407ef0a075373fb989e75368750bedada8926fbfb9ee7adfe0 WatchSource:0}: Error finding container 9fff344ac0471c407ef0a075373fb989e75368750bedada8926fbfb9ee7adfe0: Status 404 returned error can't find the container with id 9fff344ac0471c407ef0a075373fb989e75368750bedada8926fbfb9ee7adfe0 Nov 24 13:34:42 crc kubenswrapper[4790]: I1124 13:34:42.703199 4790 generic.go:334] "Generic (PLEG): container finished" podID="7fe37e44-47d4-4a0e-9cc1-4920db57f7d5" containerID="ad8dcf46bbcdce25e8690aef88e66ffa33d828aa18a8a5458f8950ce7d7f7349" exitCode=0 Nov 24 13:34:42 crc kubenswrapper[4790]: I1124 13:34:42.703309 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5","Type":"ContainerDied","Data":"ad8dcf46bbcdce25e8690aef88e66ffa33d828aa18a8a5458f8950ce7d7f7349"} Nov 24 13:34:42 crc kubenswrapper[4790]: I1124 13:34:42.709734 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8cbd6e74-a758-4bbf-81eb-acffdfa56955","Type":"ContainerStarted","Data":"9fff344ac0471c407ef0a075373fb989e75368750bedada8926fbfb9ee7adfe0"} Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.039214 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.143103 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-config-data\") pod \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\" (UID: \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\") " Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.143197 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-httpd-run\") pod \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\" (UID: \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\") " Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.143236 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\" (UID: \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\") " Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.143257 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-scripts\") pod \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\" (UID: \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\") " Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.143284 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-internal-tls-certs\") pod \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\" (UID: \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\") " Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.143311 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fht4d\" (UniqueName: \"kubernetes.io/projected/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-kube-api-access-fht4d\") pod \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\" (UID: \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\") " Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.143335 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-logs\") pod \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\" (UID: \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\") " Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.143361 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-combined-ca-bundle\") pod \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\" (UID: \"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5\") " Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.143643 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7fe37e44-47d4-4a0e-9cc1-4920db57f7d5" (UID: "7fe37e44-47d4-4a0e-9cc1-4920db57f7d5"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.148056 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-logs" (OuterVolumeSpecName: "logs") pod "7fe37e44-47d4-4a0e-9cc1-4920db57f7d5" (UID: "7fe37e44-47d4-4a0e-9cc1-4920db57f7d5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.152096 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-kube-api-access-fht4d" (OuterVolumeSpecName: "kube-api-access-fht4d") pod "7fe37e44-47d4-4a0e-9cc1-4920db57f7d5" (UID: "7fe37e44-47d4-4a0e-9cc1-4920db57f7d5"). InnerVolumeSpecName "kube-api-access-fht4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.163062 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-scripts" (OuterVolumeSpecName: "scripts") pod "7fe37e44-47d4-4a0e-9cc1-4920db57f7d5" (UID: "7fe37e44-47d4-4a0e-9cc1-4920db57f7d5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.180786 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "7fe37e44-47d4-4a0e-9cc1-4920db57f7d5" (UID: "7fe37e44-47d4-4a0e-9cc1-4920db57f7d5"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.193596 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7fe37e44-47d4-4a0e-9cc1-4920db57f7d5" (UID: "7fe37e44-47d4-4a0e-9cc1-4920db57f7d5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.245340 4790 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.245397 4790 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.245410 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.245420 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fht4d\" (UniqueName: \"kubernetes.io/projected/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-kube-api-access-fht4d\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.245433 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.245444 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.264197 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "7fe37e44-47d4-4a0e-9cc1-4920db57f7d5" (UID: "7fe37e44-47d4-4a0e-9cc1-4920db57f7d5"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.271670 4790 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.290631 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-config-data" (OuterVolumeSpecName: "config-data") pod "7fe37e44-47d4-4a0e-9cc1-4920db57f7d5" (UID: "7fe37e44-47d4-4a0e-9cc1-4920db57f7d5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.347189 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.347219 4790 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.347230 4790 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.734567 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="63ac46eb-bf90-4a9f-84ec-35f5b8079766" containerName="ceilometer-central-agent" containerID="cri-o://fd2b559f111df55674a77621a3c07fbb31acd9726dd77eb7839df232e0182d8b" gracePeriod=30 Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.734986 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"63ac46eb-bf90-4a9f-84ec-35f5b8079766","Type":"ContainerStarted","Data":"1c6a3e2ef7bb1b7101738cbe22d63a5a3a00eaef3483d8ddf6735fc036fda3eb"} Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.735038 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.735174 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="63ac46eb-bf90-4a9f-84ec-35f5b8079766" containerName="proxy-httpd" containerID="cri-o://1c6a3e2ef7bb1b7101738cbe22d63a5a3a00eaef3483d8ddf6735fc036fda3eb" gracePeriod=30 Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.735312 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="63ac46eb-bf90-4a9f-84ec-35f5b8079766" containerName="ceilometer-notification-agent" containerID="cri-o://01cbbfdf1ad3f830c64fd15d6ba213032f264a89ad6cbfc71d2a6609c93e194f" gracePeriod=30 Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.735335 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="63ac46eb-bf90-4a9f-84ec-35f5b8079766" containerName="sg-core" containerID="cri-o://382f16954cb5e0b4469a639f71a43a8f6acd9877668741eb6de89a93416768d1" gracePeriod=30 Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.744013 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8cbd6e74-a758-4bbf-81eb-acffdfa56955","Type":"ContainerStarted","Data":"1d279a19d65b0962ffcf048273f881a8a609e8c6c381d7ee412f985feb7393fb"} Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.758206 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7fe37e44-47d4-4a0e-9cc1-4920db57f7d5","Type":"ContainerDied","Data":"a1ba81ff268aaa8250fac541a25b59fd744152f7a345e29acd289db285ddd7d6"} Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.758260 4790 scope.go:117] "RemoveContainer" containerID="ad8dcf46bbcdce25e8690aef88e66ffa33d828aa18a8a5458f8950ce7d7f7349" Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.758451 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.783908 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.077525625 podStartE2EDuration="5.783868021s" podCreationTimestamp="2025-11-24 13:34:38 +0000 UTC" firstStartedPulling="2025-11-24 13:34:39.481259736 +0000 UTC m=+1327.861153398" lastFinishedPulling="2025-11-24 13:34:43.187602132 +0000 UTC m=+1331.567495794" observedRunningTime="2025-11-24 13:34:43.783723567 +0000 UTC m=+1332.163617229" watchObservedRunningTime="2025-11-24 13:34:43.783868021 +0000 UTC m=+1332.163761683" Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.831270 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.838156 4790 scope.go:117] "RemoveContainer" containerID="eecb4a557c0f4d6e2a5c95fa519ba7f6cd2b3df6ed7cb80d7301eb06b36f512d" Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.847751 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.888738 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:34:43 crc kubenswrapper[4790]: E1124 13:34:43.889200 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fe37e44-47d4-4a0e-9cc1-4920db57f7d5" containerName="glance-httpd" Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.889216 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fe37e44-47d4-4a0e-9cc1-4920db57f7d5" containerName="glance-httpd" Nov 24 13:34:43 crc kubenswrapper[4790]: E1124 13:34:43.889230 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fe37e44-47d4-4a0e-9cc1-4920db57f7d5" containerName="glance-log" Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.889236 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fe37e44-47d4-4a0e-9cc1-4920db57f7d5" containerName="glance-log" Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.889402 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fe37e44-47d4-4a0e-9cc1-4920db57f7d5" containerName="glance-httpd" Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.889427 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fe37e44-47d4-4a0e-9cc1-4920db57f7d5" containerName="glance-log" Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.890380 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.900121 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.900652 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 13:34:43 crc kubenswrapper[4790]: I1124 13:34:43.903786 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.069847 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.070225 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.070261 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h829l\" (UniqueName: \"kubernetes.io/projected/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-kube-api-access-h829l\") pod \"glance-default-internal-api-0\" (UID: \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.070292 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.070336 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-logs\") pod \"glance-default-internal-api-0\" (UID: \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.070473 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.070612 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.070705 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.172762 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.172816 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.172850 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h829l\" (UniqueName: \"kubernetes.io/projected/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-kube-api-access-h829l\") pod \"glance-default-internal-api-0\" (UID: \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.172883 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.172952 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-logs\") pod \"glance-default-internal-api-0\" (UID: \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.172991 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.173053 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.173139 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.173218 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.173746 4790 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.173906 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-logs\") pod \"glance-default-internal-api-0\" (UID: \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.179601 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.179601 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.179990 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.181572 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.209378 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h829l\" (UniqueName: \"kubernetes.io/projected/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-kube-api-access-h829l\") pod \"glance-default-internal-api-0\" (UID: \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.213453 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.245232 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.327218 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fe37e44-47d4-4a0e-9cc1-4920db57f7d5" path="/var/lib/kubelet/pods/7fe37e44-47d4-4a0e-9cc1-4920db57f7d5/volumes" Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.770272 4790 generic.go:334] "Generic (PLEG): container finished" podID="63ac46eb-bf90-4a9f-84ec-35f5b8079766" containerID="1c6a3e2ef7bb1b7101738cbe22d63a5a3a00eaef3483d8ddf6735fc036fda3eb" exitCode=0 Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.770534 4790 generic.go:334] "Generic (PLEG): container finished" podID="63ac46eb-bf90-4a9f-84ec-35f5b8079766" containerID="382f16954cb5e0b4469a639f71a43a8f6acd9877668741eb6de89a93416768d1" exitCode=2 Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.770547 4790 generic.go:334] "Generic (PLEG): container finished" podID="63ac46eb-bf90-4a9f-84ec-35f5b8079766" containerID="01cbbfdf1ad3f830c64fd15d6ba213032f264a89ad6cbfc71d2a6609c93e194f" exitCode=0 Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.770348 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"63ac46eb-bf90-4a9f-84ec-35f5b8079766","Type":"ContainerDied","Data":"1c6a3e2ef7bb1b7101738cbe22d63a5a3a00eaef3483d8ddf6735fc036fda3eb"} Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.770612 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"63ac46eb-bf90-4a9f-84ec-35f5b8079766","Type":"ContainerDied","Data":"382f16954cb5e0b4469a639f71a43a8f6acd9877668741eb6de89a93416768d1"} Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.770631 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"63ac46eb-bf90-4a9f-84ec-35f5b8079766","Type":"ContainerDied","Data":"01cbbfdf1ad3f830c64fd15d6ba213032f264a89ad6cbfc71d2a6609c93e194f"} Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.772560 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8cbd6e74-a758-4bbf-81eb-acffdfa56955","Type":"ContainerStarted","Data":"8b87dbc9b7ac42a8486a73feacf6d51b270b4dd6af8a81539a4ea922eb8a0e1b"} Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.808324 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.8083005070000002 podStartE2EDuration="3.808300507s" podCreationTimestamp="2025-11-24 13:34:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:34:44.807539666 +0000 UTC m=+1333.187433318" watchObservedRunningTime="2025-11-24 13:34:44.808300507 +0000 UTC m=+1333.188194169" Nov 24 13:34:44 crc kubenswrapper[4790]: I1124 13:34:44.909937 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:34:44 crc kubenswrapper[4790]: W1124 13:34:44.915386 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f8fe1f4_7b63_455c_97d1_d3c78826c3c4.slice/crio-189abe70c405aaf3f7415caeafe81469f1527b66c9fcf4bdd2d54cc3e040a222 WatchSource:0}: Error finding container 189abe70c405aaf3f7415caeafe81469f1527b66c9fcf4bdd2d54cc3e040a222: Status 404 returned error can't find the container with id 189abe70c405aaf3f7415caeafe81469f1527b66c9fcf4bdd2d54cc3e040a222 Nov 24 13:34:45 crc kubenswrapper[4790]: I1124 13:34:45.803325 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4","Type":"ContainerStarted","Data":"3e9b86f86ff0501669479b2ac2723096fea52e20d1dedb6c13ee82635ab6f1aa"} Nov 24 13:34:45 crc kubenswrapper[4790]: I1124 13:34:45.803655 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4","Type":"ContainerStarted","Data":"189abe70c405aaf3f7415caeafe81469f1527b66c9fcf4bdd2d54cc3e040a222"} Nov 24 13:34:46 crc kubenswrapper[4790]: I1124 13:34:46.812179 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4","Type":"ContainerStarted","Data":"77d68f10a4550d5451c7aeeccb710c9bb4cd97f4057890444850ad09cff25f6b"} Nov 24 13:34:46 crc kubenswrapper[4790]: I1124 13:34:46.833276 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.833257472 podStartE2EDuration="3.833257472s" podCreationTimestamp="2025-11-24 13:34:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:34:46.829044402 +0000 UTC m=+1335.208938074" watchObservedRunningTime="2025-11-24 13:34:46.833257472 +0000 UTC m=+1335.213151124" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.414029 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-4qj2r"] Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.415743 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4qj2r" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.452736 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-4qj2r"] Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.517787 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-94scw"] Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.525278 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-94scw" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.541377 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9cef4bc7-1efa-4f25-a73e-9cff140e0e09-operator-scripts\") pod \"nova-api-db-create-4qj2r\" (UID: \"9cef4bc7-1efa-4f25-a73e-9cff140e0e09\") " pod="openstack/nova-api-db-create-4qj2r" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.541932 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgw2x\" (UniqueName: \"kubernetes.io/projected/9cef4bc7-1efa-4f25-a73e-9cff140e0e09-kube-api-access-dgw2x\") pod \"nova-api-db-create-4qj2r\" (UID: \"9cef4bc7-1efa-4f25-a73e-9cff140e0e09\") " pod="openstack/nova-api-db-create-4qj2r" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.543031 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-94scw"] Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.623719 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-d63e-account-create-vhfd4"] Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.625037 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-d63e-account-create-vhfd4" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.627970 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.630352 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-rncp4"] Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.631551 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-rncp4" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.646654 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/36466fc4-4db9-473c-9372-dce30dbac728-operator-scripts\") pod \"nova-cell0-db-create-94scw\" (UID: \"36466fc4-4db9-473c-9372-dce30dbac728\") " pod="openstack/nova-cell0-db-create-94scw" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.646714 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2xjp\" (UniqueName: \"kubernetes.io/projected/8451cd35-fbf0-429e-a323-c6534b64e2d6-kube-api-access-l2xjp\") pod \"nova-cell1-db-create-rncp4\" (UID: \"8451cd35-fbf0-429e-a323-c6534b64e2d6\") " pod="openstack/nova-cell1-db-create-rncp4" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.646783 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgw2x\" (UniqueName: \"kubernetes.io/projected/9cef4bc7-1efa-4f25-a73e-9cff140e0e09-kube-api-access-dgw2x\") pod \"nova-api-db-create-4qj2r\" (UID: \"9cef4bc7-1efa-4f25-a73e-9cff140e0e09\") " pod="openstack/nova-api-db-create-4qj2r" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.646864 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cq672\" (UniqueName: \"kubernetes.io/projected/db12b3e5-072f-4b17-8ffe-f2861edc47e8-kube-api-access-cq672\") pod \"nova-api-d63e-account-create-vhfd4\" (UID: \"db12b3e5-072f-4b17-8ffe-f2861edc47e8\") " pod="openstack/nova-api-d63e-account-create-vhfd4" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.647034 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db12b3e5-072f-4b17-8ffe-f2861edc47e8-operator-scripts\") pod \"nova-api-d63e-account-create-vhfd4\" (UID: \"db12b3e5-072f-4b17-8ffe-f2861edc47e8\") " pod="openstack/nova-api-d63e-account-create-vhfd4" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.647201 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9fcm\" (UniqueName: \"kubernetes.io/projected/36466fc4-4db9-473c-9372-dce30dbac728-kube-api-access-j9fcm\") pod \"nova-cell0-db-create-94scw\" (UID: \"36466fc4-4db9-473c-9372-dce30dbac728\") " pod="openstack/nova-cell0-db-create-94scw" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.647353 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9cef4bc7-1efa-4f25-a73e-9cff140e0e09-operator-scripts\") pod \"nova-api-db-create-4qj2r\" (UID: \"9cef4bc7-1efa-4f25-a73e-9cff140e0e09\") " pod="openstack/nova-api-db-create-4qj2r" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.647475 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8451cd35-fbf0-429e-a323-c6534b64e2d6-operator-scripts\") pod \"nova-cell1-db-create-rncp4\" (UID: \"8451cd35-fbf0-429e-a323-c6534b64e2d6\") " pod="openstack/nova-cell1-db-create-rncp4" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.650791 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9cef4bc7-1efa-4f25-a73e-9cff140e0e09-operator-scripts\") pod \"nova-api-db-create-4qj2r\" (UID: \"9cef4bc7-1efa-4f25-a73e-9cff140e0e09\") " pod="openstack/nova-api-db-create-4qj2r" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.671792 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgw2x\" (UniqueName: \"kubernetes.io/projected/9cef4bc7-1efa-4f25-a73e-9cff140e0e09-kube-api-access-dgw2x\") pod \"nova-api-db-create-4qj2r\" (UID: \"9cef4bc7-1efa-4f25-a73e-9cff140e0e09\") " pod="openstack/nova-api-db-create-4qj2r" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.681736 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-d63e-account-create-vhfd4"] Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.700904 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-rncp4"] Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.744859 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4qj2r" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.748511 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9fcm\" (UniqueName: \"kubernetes.io/projected/36466fc4-4db9-473c-9372-dce30dbac728-kube-api-access-j9fcm\") pod \"nova-cell0-db-create-94scw\" (UID: \"36466fc4-4db9-473c-9372-dce30dbac728\") " pod="openstack/nova-cell0-db-create-94scw" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.748599 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8451cd35-fbf0-429e-a323-c6534b64e2d6-operator-scripts\") pod \"nova-cell1-db-create-rncp4\" (UID: \"8451cd35-fbf0-429e-a323-c6534b64e2d6\") " pod="openstack/nova-cell1-db-create-rncp4" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.748639 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/36466fc4-4db9-473c-9372-dce30dbac728-operator-scripts\") pod \"nova-cell0-db-create-94scw\" (UID: \"36466fc4-4db9-473c-9372-dce30dbac728\") " pod="openstack/nova-cell0-db-create-94scw" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.748664 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2xjp\" (UniqueName: \"kubernetes.io/projected/8451cd35-fbf0-429e-a323-c6534b64e2d6-kube-api-access-l2xjp\") pod \"nova-cell1-db-create-rncp4\" (UID: \"8451cd35-fbf0-429e-a323-c6534b64e2d6\") " pod="openstack/nova-cell1-db-create-rncp4" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.748930 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cq672\" (UniqueName: \"kubernetes.io/projected/db12b3e5-072f-4b17-8ffe-f2861edc47e8-kube-api-access-cq672\") pod \"nova-api-d63e-account-create-vhfd4\" (UID: \"db12b3e5-072f-4b17-8ffe-f2861edc47e8\") " pod="openstack/nova-api-d63e-account-create-vhfd4" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.748964 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db12b3e5-072f-4b17-8ffe-f2861edc47e8-operator-scripts\") pod \"nova-api-d63e-account-create-vhfd4\" (UID: \"db12b3e5-072f-4b17-8ffe-f2861edc47e8\") " pod="openstack/nova-api-d63e-account-create-vhfd4" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.750182 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db12b3e5-072f-4b17-8ffe-f2861edc47e8-operator-scripts\") pod \"nova-api-d63e-account-create-vhfd4\" (UID: \"db12b3e5-072f-4b17-8ffe-f2861edc47e8\") " pod="openstack/nova-api-d63e-account-create-vhfd4" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.750362 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8451cd35-fbf0-429e-a323-c6534b64e2d6-operator-scripts\") pod \"nova-cell1-db-create-rncp4\" (UID: \"8451cd35-fbf0-429e-a323-c6534b64e2d6\") " pod="openstack/nova-cell1-db-create-rncp4" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.753658 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/36466fc4-4db9-473c-9372-dce30dbac728-operator-scripts\") pod \"nova-cell0-db-create-94scw\" (UID: \"36466fc4-4db9-473c-9372-dce30dbac728\") " pod="openstack/nova-cell0-db-create-94scw" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.770301 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2xjp\" (UniqueName: \"kubernetes.io/projected/8451cd35-fbf0-429e-a323-c6534b64e2d6-kube-api-access-l2xjp\") pod \"nova-cell1-db-create-rncp4\" (UID: \"8451cd35-fbf0-429e-a323-c6534b64e2d6\") " pod="openstack/nova-cell1-db-create-rncp4" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.771730 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9fcm\" (UniqueName: \"kubernetes.io/projected/36466fc4-4db9-473c-9372-dce30dbac728-kube-api-access-j9fcm\") pod \"nova-cell0-db-create-94scw\" (UID: \"36466fc4-4db9-473c-9372-dce30dbac728\") " pod="openstack/nova-cell0-db-create-94scw" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.774662 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cq672\" (UniqueName: \"kubernetes.io/projected/db12b3e5-072f-4b17-8ffe-f2861edc47e8-kube-api-access-cq672\") pod \"nova-api-d63e-account-create-vhfd4\" (UID: \"db12b3e5-072f-4b17-8ffe-f2861edc47e8\") " pod="openstack/nova-api-d63e-account-create-vhfd4" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.848403 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-94scw" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.849061 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-0c43-account-create-l9mrr"] Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.850368 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0c43-account-create-l9mrr" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.852843 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.870777 4790 generic.go:334] "Generic (PLEG): container finished" podID="63ac46eb-bf90-4a9f-84ec-35f5b8079766" containerID="fd2b559f111df55674a77621a3c07fbb31acd9726dd77eb7839df232e0182d8b" exitCode=0 Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.871057 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"63ac46eb-bf90-4a9f-84ec-35f5b8079766","Type":"ContainerDied","Data":"fd2b559f111df55674a77621a3c07fbb31acd9726dd77eb7839df232e0182d8b"} Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.886473 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-0c43-account-create-l9mrr"] Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.946249 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.951395 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97cce111-7b68-4ff8-bafe-e76b7f0da75a-operator-scripts\") pod \"nova-cell0-0c43-account-create-l9mrr\" (UID: \"97cce111-7b68-4ff8-bafe-e76b7f0da75a\") " pod="openstack/nova-cell0-0c43-account-create-l9mrr" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.951445 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wt7dd\" (UniqueName: \"kubernetes.io/projected/97cce111-7b68-4ff8-bafe-e76b7f0da75a-kube-api-access-wt7dd\") pod \"nova-cell0-0c43-account-create-l9mrr\" (UID: \"97cce111-7b68-4ff8-bafe-e76b7f0da75a\") " pod="openstack/nova-cell0-0c43-account-create-l9mrr" Nov 24 13:34:50 crc kubenswrapper[4790]: I1124 13:34:50.974651 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-d63e-account-create-vhfd4" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.034016 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-964b-account-create-f25wh"] Nov 24 13:34:51 crc kubenswrapper[4790]: E1124 13:34:51.034451 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63ac46eb-bf90-4a9f-84ec-35f5b8079766" containerName="ceilometer-central-agent" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.034466 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="63ac46eb-bf90-4a9f-84ec-35f5b8079766" containerName="ceilometer-central-agent" Nov 24 13:34:51 crc kubenswrapper[4790]: E1124 13:34:51.034477 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63ac46eb-bf90-4a9f-84ec-35f5b8079766" containerName="proxy-httpd" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.034484 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="63ac46eb-bf90-4a9f-84ec-35f5b8079766" containerName="proxy-httpd" Nov 24 13:34:51 crc kubenswrapper[4790]: E1124 13:34:51.034515 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63ac46eb-bf90-4a9f-84ec-35f5b8079766" containerName="sg-core" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.034524 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="63ac46eb-bf90-4a9f-84ec-35f5b8079766" containerName="sg-core" Nov 24 13:34:51 crc kubenswrapper[4790]: E1124 13:34:51.034547 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63ac46eb-bf90-4a9f-84ec-35f5b8079766" containerName="ceilometer-notification-agent" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.034554 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="63ac46eb-bf90-4a9f-84ec-35f5b8079766" containerName="ceilometer-notification-agent" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.035930 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="63ac46eb-bf90-4a9f-84ec-35f5b8079766" containerName="ceilometer-central-agent" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.035961 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="63ac46eb-bf90-4a9f-84ec-35f5b8079766" containerName="sg-core" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.035974 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="63ac46eb-bf90-4a9f-84ec-35f5b8079766" containerName="proxy-httpd" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.036004 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="63ac46eb-bf90-4a9f-84ec-35f5b8079766" containerName="ceilometer-notification-agent" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.037219 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-964b-account-create-f25wh" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.040208 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-rncp4" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.041112 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.047264 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-964b-account-create-f25wh"] Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.052500 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63ac46eb-bf90-4a9f-84ec-35f5b8079766-config-data\") pod \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\" (UID: \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\") " Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.052583 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/63ac46eb-bf90-4a9f-84ec-35f5b8079766-sg-core-conf-yaml\") pod \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\" (UID: \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\") " Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.052648 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jkv6\" (UniqueName: \"kubernetes.io/projected/63ac46eb-bf90-4a9f-84ec-35f5b8079766-kube-api-access-7jkv6\") pod \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\" (UID: \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\") " Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.052716 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63ac46eb-bf90-4a9f-84ec-35f5b8079766-combined-ca-bundle\") pod \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\" (UID: \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\") " Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.052766 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63ac46eb-bf90-4a9f-84ec-35f5b8079766-scripts\") pod \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\" (UID: \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\") " Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.052832 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63ac46eb-bf90-4a9f-84ec-35f5b8079766-run-httpd\") pod \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\" (UID: \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\") " Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.053095 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63ac46eb-bf90-4a9f-84ec-35f5b8079766-log-httpd\") pod \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\" (UID: \"63ac46eb-bf90-4a9f-84ec-35f5b8079766\") " Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.053401 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97cce111-7b68-4ff8-bafe-e76b7f0da75a-operator-scripts\") pod \"nova-cell0-0c43-account-create-l9mrr\" (UID: \"97cce111-7b68-4ff8-bafe-e76b7f0da75a\") " pod="openstack/nova-cell0-0c43-account-create-l9mrr" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.053447 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wt7dd\" (UniqueName: \"kubernetes.io/projected/97cce111-7b68-4ff8-bafe-e76b7f0da75a-kube-api-access-wt7dd\") pod \"nova-cell0-0c43-account-create-l9mrr\" (UID: \"97cce111-7b68-4ff8-bafe-e76b7f0da75a\") " pod="openstack/nova-cell0-0c43-account-create-l9mrr" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.054188 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63ac46eb-bf90-4a9f-84ec-35f5b8079766-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "63ac46eb-bf90-4a9f-84ec-35f5b8079766" (UID: "63ac46eb-bf90-4a9f-84ec-35f5b8079766"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.054511 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63ac46eb-bf90-4a9f-84ec-35f5b8079766-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "63ac46eb-bf90-4a9f-84ec-35f5b8079766" (UID: "63ac46eb-bf90-4a9f-84ec-35f5b8079766"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.055114 4790 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63ac46eb-bf90-4a9f-84ec-35f5b8079766-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.055955 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97cce111-7b68-4ff8-bafe-e76b7f0da75a-operator-scripts\") pod \"nova-cell0-0c43-account-create-l9mrr\" (UID: \"97cce111-7b68-4ff8-bafe-e76b7f0da75a\") " pod="openstack/nova-cell0-0c43-account-create-l9mrr" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.066167 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63ac46eb-bf90-4a9f-84ec-35f5b8079766-kube-api-access-7jkv6" (OuterVolumeSpecName: "kube-api-access-7jkv6") pod "63ac46eb-bf90-4a9f-84ec-35f5b8079766" (UID: "63ac46eb-bf90-4a9f-84ec-35f5b8079766"). InnerVolumeSpecName "kube-api-access-7jkv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.066640 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63ac46eb-bf90-4a9f-84ec-35f5b8079766-scripts" (OuterVolumeSpecName: "scripts") pod "63ac46eb-bf90-4a9f-84ec-35f5b8079766" (UID: "63ac46eb-bf90-4a9f-84ec-35f5b8079766"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.073726 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wt7dd\" (UniqueName: \"kubernetes.io/projected/97cce111-7b68-4ff8-bafe-e76b7f0da75a-kube-api-access-wt7dd\") pod \"nova-cell0-0c43-account-create-l9mrr\" (UID: \"97cce111-7b68-4ff8-bafe-e76b7f0da75a\") " pod="openstack/nova-cell0-0c43-account-create-l9mrr" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.098058 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63ac46eb-bf90-4a9f-84ec-35f5b8079766-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "63ac46eb-bf90-4a9f-84ec-35f5b8079766" (UID: "63ac46eb-bf90-4a9f-84ec-35f5b8079766"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.157032 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l74v4\" (UniqueName: \"kubernetes.io/projected/5deb4907-763e-4f05-8901-4b577c8802cf-kube-api-access-l74v4\") pod \"nova-cell1-964b-account-create-f25wh\" (UID: \"5deb4907-763e-4f05-8901-4b577c8802cf\") " pod="openstack/nova-cell1-964b-account-create-f25wh" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.157451 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5deb4907-763e-4f05-8901-4b577c8802cf-operator-scripts\") pod \"nova-cell1-964b-account-create-f25wh\" (UID: \"5deb4907-763e-4f05-8901-4b577c8802cf\") " pod="openstack/nova-cell1-964b-account-create-f25wh" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.157586 4790 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63ac46eb-bf90-4a9f-84ec-35f5b8079766-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.157677 4790 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/63ac46eb-bf90-4a9f-84ec-35f5b8079766-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.157713 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jkv6\" (UniqueName: \"kubernetes.io/projected/63ac46eb-bf90-4a9f-84ec-35f5b8079766-kube-api-access-7jkv6\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.157727 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63ac46eb-bf90-4a9f-84ec-35f5b8079766-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.172865 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63ac46eb-bf90-4a9f-84ec-35f5b8079766-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "63ac46eb-bf90-4a9f-84ec-35f5b8079766" (UID: "63ac46eb-bf90-4a9f-84ec-35f5b8079766"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.186593 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63ac46eb-bf90-4a9f-84ec-35f5b8079766-config-data" (OuterVolumeSpecName: "config-data") pod "63ac46eb-bf90-4a9f-84ec-35f5b8079766" (UID: "63ac46eb-bf90-4a9f-84ec-35f5b8079766"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.213428 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0c43-account-create-l9mrr" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.259443 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l74v4\" (UniqueName: \"kubernetes.io/projected/5deb4907-763e-4f05-8901-4b577c8802cf-kube-api-access-l74v4\") pod \"nova-cell1-964b-account-create-f25wh\" (UID: \"5deb4907-763e-4f05-8901-4b577c8802cf\") " pod="openstack/nova-cell1-964b-account-create-f25wh" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.259570 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5deb4907-763e-4f05-8901-4b577c8802cf-operator-scripts\") pod \"nova-cell1-964b-account-create-f25wh\" (UID: \"5deb4907-763e-4f05-8901-4b577c8802cf\") " pod="openstack/nova-cell1-964b-account-create-f25wh" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.259641 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63ac46eb-bf90-4a9f-84ec-35f5b8079766-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.259656 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63ac46eb-bf90-4a9f-84ec-35f5b8079766-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.260862 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5deb4907-763e-4f05-8901-4b577c8802cf-operator-scripts\") pod \"nova-cell1-964b-account-create-f25wh\" (UID: \"5deb4907-763e-4f05-8901-4b577c8802cf\") " pod="openstack/nova-cell1-964b-account-create-f25wh" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.283492 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l74v4\" (UniqueName: \"kubernetes.io/projected/5deb4907-763e-4f05-8901-4b577c8802cf-kube-api-access-l74v4\") pod \"nova-cell1-964b-account-create-f25wh\" (UID: \"5deb4907-763e-4f05-8901-4b577c8802cf\") " pod="openstack/nova-cell1-964b-account-create-f25wh" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.321978 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-4qj2r"] Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.356135 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-964b-account-create-f25wh" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.469212 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-94scw"] Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.537259 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-d63e-account-create-vhfd4"] Nov 24 13:34:51 crc kubenswrapper[4790]: W1124 13:34:51.542144 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb12b3e5_072f_4b17_8ffe_f2861edc47e8.slice/crio-2e3c975edb2d81886f009e1fcdd73477065e00cf37589d7750e9257b7166964b WatchSource:0}: Error finding container 2e3c975edb2d81886f009e1fcdd73477065e00cf37589d7750e9257b7166964b: Status 404 returned error can't find the container with id 2e3c975edb2d81886f009e1fcdd73477065e00cf37589d7750e9257b7166964b Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.645705 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-rncp4"] Nov 24 13:34:51 crc kubenswrapper[4790]: W1124 13:34:51.655366 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8451cd35_fbf0_429e_a323_c6534b64e2d6.slice/crio-903f5ef9c8e9ac143f0d92d925b27271d05a88e37be1d9e5be9bc6eabe6f5bb2 WatchSource:0}: Error finding container 903f5ef9c8e9ac143f0d92d925b27271d05a88e37be1d9e5be9bc6eabe6f5bb2: Status 404 returned error can't find the container with id 903f5ef9c8e9ac143f0d92d925b27271d05a88e37be1d9e5be9bc6eabe6f5bb2 Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.738312 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-0c43-account-create-l9mrr"] Nov 24 13:34:51 crc kubenswrapper[4790]: W1124 13:34:51.754213 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97cce111_7b68_4ff8_bafe_e76b7f0da75a.slice/crio-db893833069094a6ca75271d82d64395a0d70ea2e0d647c91bbc5c61feb7cd10 WatchSource:0}: Error finding container db893833069094a6ca75271d82d64395a0d70ea2e0d647c91bbc5c61feb7cd10: Status 404 returned error can't find the container with id db893833069094a6ca75271d82d64395a0d70ea2e0d647c91bbc5c61feb7cd10 Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.881313 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-0c43-account-create-l9mrr" event={"ID":"97cce111-7b68-4ff8-bafe-e76b7f0da75a","Type":"ContainerStarted","Data":"db893833069094a6ca75271d82d64395a0d70ea2e0d647c91bbc5c61feb7cd10"} Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.882427 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-d63e-account-create-vhfd4" event={"ID":"db12b3e5-072f-4b17-8ffe-f2861edc47e8","Type":"ContainerStarted","Data":"2e3c975edb2d81886f009e1fcdd73477065e00cf37589d7750e9257b7166964b"} Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.884100 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-94scw" event={"ID":"36466fc4-4db9-473c-9372-dce30dbac728","Type":"ContainerStarted","Data":"4b2f07e17afec3de426c05a71735b69dd3969df664dcf3337683a38ebb86ccde"} Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.886098 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4qj2r" event={"ID":"9cef4bc7-1efa-4f25-a73e-9cff140e0e09","Type":"ContainerStarted","Data":"559b39f1cb8cb24bfe5aee6eebf0bd4b5f7ea54f034e3fab965ea795a0c3e81d"} Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.886128 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4qj2r" event={"ID":"9cef4bc7-1efa-4f25-a73e-9cff140e0e09","Type":"ContainerStarted","Data":"6532a1875e01f28c4d11a739bc7d36e083da8f6608b939461a21a996fdd3be43"} Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.887793 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-rncp4" event={"ID":"8451cd35-fbf0-429e-a323-c6534b64e2d6","Type":"ContainerStarted","Data":"903f5ef9c8e9ac143f0d92d925b27271d05a88e37be1d9e5be9bc6eabe6f5bb2"} Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.890991 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"63ac46eb-bf90-4a9f-84ec-35f5b8079766","Type":"ContainerDied","Data":"e7aedbd2d9c4f164cd5f9c561b07d58f822634c8f673d5db2747ba10a75d637d"} Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.891033 4790 scope.go:117] "RemoveContainer" containerID="1c6a3e2ef7bb1b7101738cbe22d63a5a3a00eaef3483d8ddf6735fc036fda3eb" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.891150 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.908674 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-4qj2r" podStartSLOduration=1.908653726 podStartE2EDuration="1.908653726s" podCreationTimestamp="2025-11-24 13:34:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:34:51.901329827 +0000 UTC m=+1340.281223489" watchObservedRunningTime="2025-11-24 13:34:51.908653726 +0000 UTC m=+1340.288547388" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.927068 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-964b-account-create-f25wh"] Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.938500 4790 scope.go:117] "RemoveContainer" containerID="382f16954cb5e0b4469a639f71a43a8f6acd9877668741eb6de89a93416768d1" Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.946096 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:34:51 crc kubenswrapper[4790]: W1124 13:34:51.954517 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5deb4907_763e_4f05_8901_4b577c8802cf.slice/crio-7688cdaba28e6c3b0aca784edb62a8f2d807a6b74f7c96720aa1f78c6142b303 WatchSource:0}: Error finding container 7688cdaba28e6c3b0aca784edb62a8f2d807a6b74f7c96720aa1f78c6142b303: Status 404 returned error can't find the container with id 7688cdaba28e6c3b0aca784edb62a8f2d807a6b74f7c96720aa1f78c6142b303 Nov 24 13:34:51 crc kubenswrapper[4790]: I1124 13:34:51.956508 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.018250 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.026276 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.029237 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.051623 4790 scope.go:117] "RemoveContainer" containerID="01cbbfdf1ad3f830c64fd15d6ba213032f264a89ad6cbfc71d2a6609c93e194f" Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.051794 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.095311 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-scripts\") pod \"ceilometer-0\" (UID: \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\") " pod="openstack/ceilometer-0" Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.095439 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\") " pod="openstack/ceilometer-0" Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.095494 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwljz\" (UniqueName: \"kubernetes.io/projected/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-kube-api-access-gwljz\") pod \"ceilometer-0\" (UID: \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\") " pod="openstack/ceilometer-0" Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.095612 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-log-httpd\") pod \"ceilometer-0\" (UID: \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\") " pod="openstack/ceilometer-0" Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.095659 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-config-data\") pod \"ceilometer-0\" (UID: \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\") " pod="openstack/ceilometer-0" Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.095736 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-run-httpd\") pod \"ceilometer-0\" (UID: \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\") " pod="openstack/ceilometer-0" Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.095785 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\") " pod="openstack/ceilometer-0" Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.096115 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.096796 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.102131 4790 scope.go:117] "RemoveContainer" containerID="fd2b559f111df55674a77621a3c07fbb31acd9726dd77eb7839df232e0182d8b" Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.114951 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.144454 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.149151 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.199088 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-log-httpd\") pod \"ceilometer-0\" (UID: \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\") " pod="openstack/ceilometer-0" Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.199162 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-config-data\") pod \"ceilometer-0\" (UID: \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\") " pod="openstack/ceilometer-0" Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.199214 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-run-httpd\") pod \"ceilometer-0\" (UID: \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\") " pod="openstack/ceilometer-0" Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.199286 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\") " pod="openstack/ceilometer-0" Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.199494 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-scripts\") pod \"ceilometer-0\" (UID: \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\") " pod="openstack/ceilometer-0" Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.199539 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\") " pod="openstack/ceilometer-0" Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.199577 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwljz\" (UniqueName: \"kubernetes.io/projected/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-kube-api-access-gwljz\") pod \"ceilometer-0\" (UID: \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\") " pod="openstack/ceilometer-0" Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.199805 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-log-httpd\") pod \"ceilometer-0\" (UID: \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\") " pod="openstack/ceilometer-0" Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.202485 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-run-httpd\") pod \"ceilometer-0\" (UID: \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\") " pod="openstack/ceilometer-0" Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.208575 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-scripts\") pod \"ceilometer-0\" (UID: \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\") " pod="openstack/ceilometer-0" Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.211161 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\") " pod="openstack/ceilometer-0" Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.213580 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-config-data\") pod \"ceilometer-0\" (UID: \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\") " pod="openstack/ceilometer-0" Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.217325 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\") " pod="openstack/ceilometer-0" Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.220048 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwljz\" (UniqueName: \"kubernetes.io/projected/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-kube-api-access-gwljz\") pod \"ceilometer-0\" (UID: \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\") " pod="openstack/ceilometer-0" Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.331180 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63ac46eb-bf90-4a9f-84ec-35f5b8079766" path="/var/lib/kubelet/pods/63ac46eb-bf90-4a9f-84ec-35f5b8079766/volumes" Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.438329 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.826744 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.912723 4790 generic.go:334] "Generic (PLEG): container finished" podID="97cce111-7b68-4ff8-bafe-e76b7f0da75a" containerID="1ce47c3d1e1960a319da58bba01d387b5ca2f97570e10c6078a4e5e9f4e33040" exitCode=0 Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.913186 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-0c43-account-create-l9mrr" event={"ID":"97cce111-7b68-4ff8-bafe-e76b7f0da75a","Type":"ContainerDied","Data":"1ce47c3d1e1960a319da58bba01d387b5ca2f97570e10c6078a4e5e9f4e33040"} Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.923337 4790 generic.go:334] "Generic (PLEG): container finished" podID="db12b3e5-072f-4b17-8ffe-f2861edc47e8" containerID="f9e4f39302ec3fef535c4d941a4deb560f0999311d673edd1adcbdd80b3f09fa" exitCode=0 Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.923405 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-d63e-account-create-vhfd4" event={"ID":"db12b3e5-072f-4b17-8ffe-f2861edc47e8","Type":"ContainerDied","Data":"f9e4f39302ec3fef535c4d941a4deb560f0999311d673edd1adcbdd80b3f09fa"} Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.926718 4790 generic.go:334] "Generic (PLEG): container finished" podID="36466fc4-4db9-473c-9372-dce30dbac728" containerID="4c149f4ac8815779e6d7cb3c9896f38e1386454d59e01a601935ffd9f4da470b" exitCode=0 Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.926767 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-94scw" event={"ID":"36466fc4-4db9-473c-9372-dce30dbac728","Type":"ContainerDied","Data":"4c149f4ac8815779e6d7cb3c9896f38e1386454d59e01a601935ffd9f4da470b"} Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.928574 4790 generic.go:334] "Generic (PLEG): container finished" podID="9cef4bc7-1efa-4f25-a73e-9cff140e0e09" containerID="559b39f1cb8cb24bfe5aee6eebf0bd4b5f7ea54f034e3fab965ea795a0c3e81d" exitCode=0 Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.928611 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4qj2r" event={"ID":"9cef4bc7-1efa-4f25-a73e-9cff140e0e09","Type":"ContainerDied","Data":"559b39f1cb8cb24bfe5aee6eebf0bd4b5f7ea54f034e3fab965ea795a0c3e81d"} Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.933085 4790 generic.go:334] "Generic (PLEG): container finished" podID="8451cd35-fbf0-429e-a323-c6534b64e2d6" containerID="9a7848c9a709ae9ddd3f799bb95363637b474062de0cad438d73ae38726dea36" exitCode=0 Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.933155 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-rncp4" event={"ID":"8451cd35-fbf0-429e-a323-c6534b64e2d6","Type":"ContainerDied","Data":"9a7848c9a709ae9ddd3f799bb95363637b474062de0cad438d73ae38726dea36"} Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.938384 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.939537 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-964b-account-create-f25wh" event={"ID":"5deb4907-763e-4f05-8901-4b577c8802cf","Type":"ContainerStarted","Data":"99ebe4f359d29dd9dadccf947e629f98fdb917d1ee6164cc3c18632420cc9128"} Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.939563 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-964b-account-create-f25wh" event={"ID":"5deb4907-763e-4f05-8901-4b577c8802cf","Type":"ContainerStarted","Data":"7688cdaba28e6c3b0aca784edb62a8f2d807a6b74f7c96720aa1f78c6142b303"} Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.940423 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 13:34:52 crc kubenswrapper[4790]: I1124 13:34:52.940446 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 13:34:53 crc kubenswrapper[4790]: I1124 13:34:53.034928 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-964b-account-create-f25wh" podStartSLOduration=2.034909469 podStartE2EDuration="2.034909469s" podCreationTimestamp="2025-11-24 13:34:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:34:53.011185504 +0000 UTC m=+1341.391079166" watchObservedRunningTime="2025-11-24 13:34:53.034909469 +0000 UTC m=+1341.414803141" Nov 24 13:34:53 crc kubenswrapper[4790]: I1124 13:34:53.952622 4790 generic.go:334] "Generic (PLEG): container finished" podID="5deb4907-763e-4f05-8901-4b577c8802cf" containerID="99ebe4f359d29dd9dadccf947e629f98fdb917d1ee6164cc3c18632420cc9128" exitCode=0 Nov 24 13:34:53 crc kubenswrapper[4790]: I1124 13:34:53.952832 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-964b-account-create-f25wh" event={"ID":"5deb4907-763e-4f05-8901-4b577c8802cf","Type":"ContainerDied","Data":"99ebe4f359d29dd9dadccf947e629f98fdb917d1ee6164cc3c18632420cc9128"} Nov 24 13:34:53 crc kubenswrapper[4790]: I1124 13:34:53.956607 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681","Type":"ContainerStarted","Data":"8d8eb81bfcbfb49c45181dbfad85f0fcca486f70ef864c14f5cb8598f920cd81"} Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.246243 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.246513 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.296215 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.306654 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.391480 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-d63e-account-create-vhfd4" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.446636 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db12b3e5-072f-4b17-8ffe-f2861edc47e8-operator-scripts\") pod \"db12b3e5-072f-4b17-8ffe-f2861edc47e8\" (UID: \"db12b3e5-072f-4b17-8ffe-f2861edc47e8\") " Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.446806 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cq672\" (UniqueName: \"kubernetes.io/projected/db12b3e5-072f-4b17-8ffe-f2861edc47e8-kube-api-access-cq672\") pod \"db12b3e5-072f-4b17-8ffe-f2861edc47e8\" (UID: \"db12b3e5-072f-4b17-8ffe-f2861edc47e8\") " Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.447423 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db12b3e5-072f-4b17-8ffe-f2861edc47e8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "db12b3e5-072f-4b17-8ffe-f2861edc47e8" (UID: "db12b3e5-072f-4b17-8ffe-f2861edc47e8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.456717 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db12b3e5-072f-4b17-8ffe-f2861edc47e8-kube-api-access-cq672" (OuterVolumeSpecName: "kube-api-access-cq672") pod "db12b3e5-072f-4b17-8ffe-f2861edc47e8" (UID: "db12b3e5-072f-4b17-8ffe-f2861edc47e8"). InnerVolumeSpecName "kube-api-access-cq672". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.471167 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0c43-account-create-l9mrr" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.489083 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4qj2r" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.509164 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-94scw" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.511437 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-rncp4" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.548449 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wt7dd\" (UniqueName: \"kubernetes.io/projected/97cce111-7b68-4ff8-bafe-e76b7f0da75a-kube-api-access-wt7dd\") pod \"97cce111-7b68-4ff8-bafe-e76b7f0da75a\" (UID: \"97cce111-7b68-4ff8-bafe-e76b7f0da75a\") " Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.548830 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgw2x\" (UniqueName: \"kubernetes.io/projected/9cef4bc7-1efa-4f25-a73e-9cff140e0e09-kube-api-access-dgw2x\") pod \"9cef4bc7-1efa-4f25-a73e-9cff140e0e09\" (UID: \"9cef4bc7-1efa-4f25-a73e-9cff140e0e09\") " Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.548985 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97cce111-7b68-4ff8-bafe-e76b7f0da75a-operator-scripts\") pod \"97cce111-7b68-4ff8-bafe-e76b7f0da75a\" (UID: \"97cce111-7b68-4ff8-bafe-e76b7f0da75a\") " Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.549670 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97cce111-7b68-4ff8-bafe-e76b7f0da75a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "97cce111-7b68-4ff8-bafe-e76b7f0da75a" (UID: "97cce111-7b68-4ff8-bafe-e76b7f0da75a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.549976 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9cef4bc7-1efa-4f25-a73e-9cff140e0e09-operator-scripts\") pod \"9cef4bc7-1efa-4f25-a73e-9cff140e0e09\" (UID: \"9cef4bc7-1efa-4f25-a73e-9cff140e0e09\") " Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.550530 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cq672\" (UniqueName: \"kubernetes.io/projected/db12b3e5-072f-4b17-8ffe-f2861edc47e8-kube-api-access-cq672\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.550549 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97cce111-7b68-4ff8-bafe-e76b7f0da75a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.550557 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db12b3e5-072f-4b17-8ffe-f2861edc47e8-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.550914 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9cef4bc7-1efa-4f25-a73e-9cff140e0e09-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9cef4bc7-1efa-4f25-a73e-9cff140e0e09" (UID: "9cef4bc7-1efa-4f25-a73e-9cff140e0e09"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.552610 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97cce111-7b68-4ff8-bafe-e76b7f0da75a-kube-api-access-wt7dd" (OuterVolumeSpecName: "kube-api-access-wt7dd") pod "97cce111-7b68-4ff8-bafe-e76b7f0da75a" (UID: "97cce111-7b68-4ff8-bafe-e76b7f0da75a"). InnerVolumeSpecName "kube-api-access-wt7dd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.556815 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cef4bc7-1efa-4f25-a73e-9cff140e0e09-kube-api-access-dgw2x" (OuterVolumeSpecName: "kube-api-access-dgw2x") pod "9cef4bc7-1efa-4f25-a73e-9cff140e0e09" (UID: "9cef4bc7-1efa-4f25-a73e-9cff140e0e09"). InnerVolumeSpecName "kube-api-access-dgw2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.652049 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/36466fc4-4db9-473c-9372-dce30dbac728-operator-scripts\") pod \"36466fc4-4db9-473c-9372-dce30dbac728\" (UID: \"36466fc4-4db9-473c-9372-dce30dbac728\") " Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.652105 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8451cd35-fbf0-429e-a323-c6534b64e2d6-operator-scripts\") pod \"8451cd35-fbf0-429e-a323-c6534b64e2d6\" (UID: \"8451cd35-fbf0-429e-a323-c6534b64e2d6\") " Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.652239 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2xjp\" (UniqueName: \"kubernetes.io/projected/8451cd35-fbf0-429e-a323-c6534b64e2d6-kube-api-access-l2xjp\") pod \"8451cd35-fbf0-429e-a323-c6534b64e2d6\" (UID: \"8451cd35-fbf0-429e-a323-c6534b64e2d6\") " Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.652423 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9fcm\" (UniqueName: \"kubernetes.io/projected/36466fc4-4db9-473c-9372-dce30dbac728-kube-api-access-j9fcm\") pod \"36466fc4-4db9-473c-9372-dce30dbac728\" (UID: \"36466fc4-4db9-473c-9372-dce30dbac728\") " Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.652807 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36466fc4-4db9-473c-9372-dce30dbac728-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "36466fc4-4db9-473c-9372-dce30dbac728" (UID: "36466fc4-4db9-473c-9372-dce30dbac728"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.653071 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8451cd35-fbf0-429e-a323-c6534b64e2d6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8451cd35-fbf0-429e-a323-c6534b64e2d6" (UID: "8451cd35-fbf0-429e-a323-c6534b64e2d6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.653382 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgw2x\" (UniqueName: \"kubernetes.io/projected/9cef4bc7-1efa-4f25-a73e-9cff140e0e09-kube-api-access-dgw2x\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.653406 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9cef4bc7-1efa-4f25-a73e-9cff140e0e09-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.653419 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wt7dd\" (UniqueName: \"kubernetes.io/projected/97cce111-7b68-4ff8-bafe-e76b7f0da75a-kube-api-access-wt7dd\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.653432 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/36466fc4-4db9-473c-9372-dce30dbac728-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.653443 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8451cd35-fbf0-429e-a323-c6534b64e2d6-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.656133 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36466fc4-4db9-473c-9372-dce30dbac728-kube-api-access-j9fcm" (OuterVolumeSpecName: "kube-api-access-j9fcm") pod "36466fc4-4db9-473c-9372-dce30dbac728" (UID: "36466fc4-4db9-473c-9372-dce30dbac728"). InnerVolumeSpecName "kube-api-access-j9fcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.656932 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8451cd35-fbf0-429e-a323-c6534b64e2d6-kube-api-access-l2xjp" (OuterVolumeSpecName: "kube-api-access-l2xjp") pod "8451cd35-fbf0-429e-a323-c6534b64e2d6" (UID: "8451cd35-fbf0-429e-a323-c6534b64e2d6"). InnerVolumeSpecName "kube-api-access-l2xjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.755448 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9fcm\" (UniqueName: \"kubernetes.io/projected/36466fc4-4db9-473c-9372-dce30dbac728-kube-api-access-j9fcm\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.755785 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2xjp\" (UniqueName: \"kubernetes.io/projected/8451cd35-fbf0-429e-a323-c6534b64e2d6-kube-api-access-l2xjp\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.964053 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681","Type":"ContainerStarted","Data":"745f3d1295cb0862befcfc61cf938b4759974af5280b36246b2dd03515b5389e"} Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.964094 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681","Type":"ContainerStarted","Data":"88d86e7122598d0c47ead3ff237534c57ec8c976adc9689dcaf211afaab01df3"} Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.965213 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-0c43-account-create-l9mrr" event={"ID":"97cce111-7b68-4ff8-bafe-e76b7f0da75a","Type":"ContainerDied","Data":"db893833069094a6ca75271d82d64395a0d70ea2e0d647c91bbc5c61feb7cd10"} Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.965248 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db893833069094a6ca75271d82d64395a0d70ea2e0d647c91bbc5c61feb7cd10" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.965299 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0c43-account-create-l9mrr" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.972814 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-d63e-account-create-vhfd4" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.972775 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-d63e-account-create-vhfd4" event={"ID":"db12b3e5-072f-4b17-8ffe-f2861edc47e8","Type":"ContainerDied","Data":"2e3c975edb2d81886f009e1fcdd73477065e00cf37589d7750e9257b7166964b"} Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.973003 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e3c975edb2d81886f009e1fcdd73477065e00cf37589d7750e9257b7166964b" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.977423 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-94scw" event={"ID":"36466fc4-4db9-473c-9372-dce30dbac728","Type":"ContainerDied","Data":"4b2f07e17afec3de426c05a71735b69dd3969df664dcf3337683a38ebb86ccde"} Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.977474 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b2f07e17afec3de426c05a71735b69dd3969df664dcf3337683a38ebb86ccde" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.977440 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-94scw" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.982895 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4qj2r" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.982860 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4qj2r" event={"ID":"9cef4bc7-1efa-4f25-a73e-9cff140e0e09","Type":"ContainerDied","Data":"6532a1875e01f28c4d11a739bc7d36e083da8f6608b939461a21a996fdd3be43"} Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.982988 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6532a1875e01f28c4d11a739bc7d36e083da8f6608b939461a21a996fdd3be43" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.987723 4790 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.987745 4790 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.987974 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-rncp4" event={"ID":"8451cd35-fbf0-429e-a323-c6534b64e2d6","Type":"ContainerDied","Data":"903f5ef9c8e9ac143f0d92d925b27271d05a88e37be1d9e5be9bc6eabe6f5bb2"} Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.988014 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="903f5ef9c8e9ac143f0d92d925b27271d05a88e37be1d9e5be9bc6eabe6f5bb2" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.988087 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-rncp4" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.989532 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:54 crc kubenswrapper[4790]: I1124 13:34:54.989667 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:55 crc kubenswrapper[4790]: I1124 13:34:55.050974 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 13:34:55 crc kubenswrapper[4790]: I1124 13:34:55.072505 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 13:34:55 crc kubenswrapper[4790]: I1124 13:34:55.421972 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-964b-account-create-f25wh" Nov 24 13:34:55 crc kubenswrapper[4790]: I1124 13:34:55.497116 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l74v4\" (UniqueName: \"kubernetes.io/projected/5deb4907-763e-4f05-8901-4b577c8802cf-kube-api-access-l74v4\") pod \"5deb4907-763e-4f05-8901-4b577c8802cf\" (UID: \"5deb4907-763e-4f05-8901-4b577c8802cf\") " Nov 24 13:34:55 crc kubenswrapper[4790]: I1124 13:34:55.497349 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5deb4907-763e-4f05-8901-4b577c8802cf-operator-scripts\") pod \"5deb4907-763e-4f05-8901-4b577c8802cf\" (UID: \"5deb4907-763e-4f05-8901-4b577c8802cf\") " Nov 24 13:34:55 crc kubenswrapper[4790]: I1124 13:34:55.498469 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5deb4907-763e-4f05-8901-4b577c8802cf-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5deb4907-763e-4f05-8901-4b577c8802cf" (UID: "5deb4907-763e-4f05-8901-4b577c8802cf"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:55 crc kubenswrapper[4790]: I1124 13:34:55.533805 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5deb4907-763e-4f05-8901-4b577c8802cf-kube-api-access-l74v4" (OuterVolumeSpecName: "kube-api-access-l74v4") pod "5deb4907-763e-4f05-8901-4b577c8802cf" (UID: "5deb4907-763e-4f05-8901-4b577c8802cf"). InnerVolumeSpecName "kube-api-access-l74v4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:55 crc kubenswrapper[4790]: I1124 13:34:55.599253 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l74v4\" (UniqueName: \"kubernetes.io/projected/5deb4907-763e-4f05-8901-4b577c8802cf-kube-api-access-l74v4\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:55 crc kubenswrapper[4790]: I1124 13:34:55.599291 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5deb4907-763e-4f05-8901-4b577c8802cf-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.000147 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681","Type":"ContainerStarted","Data":"9689802d8b7838f31aa896e0bd57aa2180198393ea4527e352c87bab4e8d865a"} Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.004027 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-964b-account-create-f25wh" Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.009492 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-964b-account-create-f25wh" event={"ID":"5deb4907-763e-4f05-8901-4b577c8802cf","Type":"ContainerDied","Data":"7688cdaba28e6c3b0aca784edb62a8f2d807a6b74f7c96720aa1f78c6142b303"} Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.009532 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7688cdaba28e6c3b0aca784edb62a8f2d807a6b74f7c96720aa1f78c6142b303" Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.211529 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-s2mrt"] Nov 24 13:34:56 crc kubenswrapper[4790]: E1124 13:34:56.211970 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5deb4907-763e-4f05-8901-4b577c8802cf" containerName="mariadb-account-create" Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.211992 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="5deb4907-763e-4f05-8901-4b577c8802cf" containerName="mariadb-account-create" Nov 24 13:34:56 crc kubenswrapper[4790]: E1124 13:34:56.212011 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8451cd35-fbf0-429e-a323-c6534b64e2d6" containerName="mariadb-database-create" Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.212019 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="8451cd35-fbf0-429e-a323-c6534b64e2d6" containerName="mariadb-database-create" Nov 24 13:34:56 crc kubenswrapper[4790]: E1124 13:34:56.212030 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db12b3e5-072f-4b17-8ffe-f2861edc47e8" containerName="mariadb-account-create" Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.212037 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="db12b3e5-072f-4b17-8ffe-f2861edc47e8" containerName="mariadb-account-create" Nov 24 13:34:56 crc kubenswrapper[4790]: E1124 13:34:56.212048 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97cce111-7b68-4ff8-bafe-e76b7f0da75a" containerName="mariadb-account-create" Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.212057 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="97cce111-7b68-4ff8-bafe-e76b7f0da75a" containerName="mariadb-account-create" Nov 24 13:34:56 crc kubenswrapper[4790]: E1124 13:34:56.212082 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cef4bc7-1efa-4f25-a73e-9cff140e0e09" containerName="mariadb-database-create" Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.212088 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cef4bc7-1efa-4f25-a73e-9cff140e0e09" containerName="mariadb-database-create" Nov 24 13:34:56 crc kubenswrapper[4790]: E1124 13:34:56.212104 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36466fc4-4db9-473c-9372-dce30dbac728" containerName="mariadb-database-create" Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.212113 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="36466fc4-4db9-473c-9372-dce30dbac728" containerName="mariadb-database-create" Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.212334 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="8451cd35-fbf0-429e-a323-c6534b64e2d6" containerName="mariadb-database-create" Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.212350 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cef4bc7-1efa-4f25-a73e-9cff140e0e09" containerName="mariadb-database-create" Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.212365 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="36466fc4-4db9-473c-9372-dce30dbac728" containerName="mariadb-database-create" Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.212375 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="97cce111-7b68-4ff8-bafe-e76b7f0da75a" containerName="mariadb-account-create" Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.212418 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="5deb4907-763e-4f05-8901-4b577c8802cf" containerName="mariadb-account-create" Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.212430 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="db12b3e5-072f-4b17-8ffe-f2861edc47e8" containerName="mariadb-account-create" Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.213147 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-s2mrt" Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.226003 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-294qh" Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.226087 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.226265 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.323903 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-s2mrt"] Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.324026 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnkw7\" (UniqueName: \"kubernetes.io/projected/76801ca0-a42b-4a25-84b8-43a689e929ff-kube-api-access-jnkw7\") pod \"nova-cell0-conductor-db-sync-s2mrt\" (UID: \"76801ca0-a42b-4a25-84b8-43a689e929ff\") " pod="openstack/nova-cell0-conductor-db-sync-s2mrt" Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.324324 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76801ca0-a42b-4a25-84b8-43a689e929ff-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-s2mrt\" (UID: \"76801ca0-a42b-4a25-84b8-43a689e929ff\") " pod="openstack/nova-cell0-conductor-db-sync-s2mrt" Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.324357 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76801ca0-a42b-4a25-84b8-43a689e929ff-config-data\") pod \"nova-cell0-conductor-db-sync-s2mrt\" (UID: \"76801ca0-a42b-4a25-84b8-43a689e929ff\") " pod="openstack/nova-cell0-conductor-db-sync-s2mrt" Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.324489 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76801ca0-a42b-4a25-84b8-43a689e929ff-scripts\") pod \"nova-cell0-conductor-db-sync-s2mrt\" (UID: \"76801ca0-a42b-4a25-84b8-43a689e929ff\") " pod="openstack/nova-cell0-conductor-db-sync-s2mrt" Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.425812 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76801ca0-a42b-4a25-84b8-43a689e929ff-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-s2mrt\" (UID: \"76801ca0-a42b-4a25-84b8-43a689e929ff\") " pod="openstack/nova-cell0-conductor-db-sync-s2mrt" Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.425860 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76801ca0-a42b-4a25-84b8-43a689e929ff-config-data\") pod \"nova-cell0-conductor-db-sync-s2mrt\" (UID: \"76801ca0-a42b-4a25-84b8-43a689e929ff\") " pod="openstack/nova-cell0-conductor-db-sync-s2mrt" Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.426009 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76801ca0-a42b-4a25-84b8-43a689e929ff-scripts\") pod \"nova-cell0-conductor-db-sync-s2mrt\" (UID: \"76801ca0-a42b-4a25-84b8-43a689e929ff\") " pod="openstack/nova-cell0-conductor-db-sync-s2mrt" Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.426070 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnkw7\" (UniqueName: \"kubernetes.io/projected/76801ca0-a42b-4a25-84b8-43a689e929ff-kube-api-access-jnkw7\") pod \"nova-cell0-conductor-db-sync-s2mrt\" (UID: \"76801ca0-a42b-4a25-84b8-43a689e929ff\") " pod="openstack/nova-cell0-conductor-db-sync-s2mrt" Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.432859 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76801ca0-a42b-4a25-84b8-43a689e929ff-config-data\") pod \"nova-cell0-conductor-db-sync-s2mrt\" (UID: \"76801ca0-a42b-4a25-84b8-43a689e929ff\") " pod="openstack/nova-cell0-conductor-db-sync-s2mrt" Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.436558 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76801ca0-a42b-4a25-84b8-43a689e929ff-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-s2mrt\" (UID: \"76801ca0-a42b-4a25-84b8-43a689e929ff\") " pod="openstack/nova-cell0-conductor-db-sync-s2mrt" Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.460530 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnkw7\" (UniqueName: \"kubernetes.io/projected/76801ca0-a42b-4a25-84b8-43a689e929ff-kube-api-access-jnkw7\") pod \"nova-cell0-conductor-db-sync-s2mrt\" (UID: \"76801ca0-a42b-4a25-84b8-43a689e929ff\") " pod="openstack/nova-cell0-conductor-db-sync-s2mrt" Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.463353 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76801ca0-a42b-4a25-84b8-43a689e929ff-scripts\") pod \"nova-cell0-conductor-db-sync-s2mrt\" (UID: \"76801ca0-a42b-4a25-84b8-43a689e929ff\") " pod="openstack/nova-cell0-conductor-db-sync-s2mrt" Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.527667 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-s2mrt" Nov 24 13:34:56 crc kubenswrapper[4790]: W1124 13:34:56.978464 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76801ca0_a42b_4a25_84b8_43a689e929ff.slice/crio-bc2bffe906e252d6e0b71a7e74bfa4a43861e50d36cc805c33e3e339261eb2fe WatchSource:0}: Error finding container bc2bffe906e252d6e0b71a7e74bfa4a43861e50d36cc805c33e3e339261eb2fe: Status 404 returned error can't find the container with id bc2bffe906e252d6e0b71a7e74bfa4a43861e50d36cc805c33e3e339261eb2fe Nov 24 13:34:56 crc kubenswrapper[4790]: I1124 13:34:56.989729 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-s2mrt"] Nov 24 13:34:57 crc kubenswrapper[4790]: I1124 13:34:57.012550 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-s2mrt" event={"ID":"76801ca0-a42b-4a25-84b8-43a689e929ff","Type":"ContainerStarted","Data":"bc2bffe906e252d6e0b71a7e74bfa4a43861e50d36cc805c33e3e339261eb2fe"} Nov 24 13:34:57 crc kubenswrapper[4790]: I1124 13:34:57.262247 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:57 crc kubenswrapper[4790]: I1124 13:34:57.262370 4790 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 13:34:57 crc kubenswrapper[4790]: I1124 13:34:57.323105 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 13:35:05 crc kubenswrapper[4790]: I1124 13:35:05.103943 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681","Type":"ContainerStarted","Data":"74fa6831af4f662dc1d64fe0d79dfdaa11bcaf6438540da07535916e9c3be8b5"} Nov 24 13:35:05 crc kubenswrapper[4790]: I1124 13:35:05.104302 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7b3e66b8-f5c2-4b15-b5f7-52182f2c6681" containerName="ceilometer-central-agent" containerID="cri-o://88d86e7122598d0c47ead3ff237534c57ec8c976adc9689dcaf211afaab01df3" gracePeriod=30 Nov 24 13:35:05 crc kubenswrapper[4790]: I1124 13:35:05.104355 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7b3e66b8-f5c2-4b15-b5f7-52182f2c6681" containerName="proxy-httpd" containerID="cri-o://74fa6831af4f662dc1d64fe0d79dfdaa11bcaf6438540da07535916e9c3be8b5" gracePeriod=30 Nov 24 13:35:05 crc kubenswrapper[4790]: I1124 13:35:05.104397 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7b3e66b8-f5c2-4b15-b5f7-52182f2c6681" containerName="ceilometer-notification-agent" containerID="cri-o://745f3d1295cb0862befcfc61cf938b4759974af5280b36246b2dd03515b5389e" gracePeriod=30 Nov 24 13:35:05 crc kubenswrapper[4790]: I1124 13:35:05.104411 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7b3e66b8-f5c2-4b15-b5f7-52182f2c6681" containerName="sg-core" containerID="cri-o://9689802d8b7838f31aa896e0bd57aa2180198393ea4527e352c87bab4e8d865a" gracePeriod=30 Nov 24 13:35:05 crc kubenswrapper[4790]: I1124 13:35:05.104599 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 13:35:05 crc kubenswrapper[4790]: I1124 13:35:05.129142 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.100364483 podStartE2EDuration="14.129122252s" podCreationTimestamp="2025-11-24 13:34:51 +0000 UTC" firstStartedPulling="2025-11-24 13:34:52.955878841 +0000 UTC m=+1341.335772503" lastFinishedPulling="2025-11-24 13:35:03.98463661 +0000 UTC m=+1352.364530272" observedRunningTime="2025-11-24 13:35:05.121777453 +0000 UTC m=+1353.501671115" watchObservedRunningTime="2025-11-24 13:35:05.129122252 +0000 UTC m=+1353.509015914" Nov 24 13:35:06 crc kubenswrapper[4790]: I1124 13:35:06.114291 4790 generic.go:334] "Generic (PLEG): container finished" podID="7b3e66b8-f5c2-4b15-b5f7-52182f2c6681" containerID="74fa6831af4f662dc1d64fe0d79dfdaa11bcaf6438540da07535916e9c3be8b5" exitCode=0 Nov 24 13:35:06 crc kubenswrapper[4790]: I1124 13:35:06.114567 4790 generic.go:334] "Generic (PLEG): container finished" podID="7b3e66b8-f5c2-4b15-b5f7-52182f2c6681" containerID="9689802d8b7838f31aa896e0bd57aa2180198393ea4527e352c87bab4e8d865a" exitCode=2 Nov 24 13:35:06 crc kubenswrapper[4790]: I1124 13:35:06.114332 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681","Type":"ContainerDied","Data":"74fa6831af4f662dc1d64fe0d79dfdaa11bcaf6438540da07535916e9c3be8b5"} Nov 24 13:35:06 crc kubenswrapper[4790]: I1124 13:35:06.114633 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681","Type":"ContainerDied","Data":"9689802d8b7838f31aa896e0bd57aa2180198393ea4527e352c87bab4e8d865a"} Nov 24 13:35:07 crc kubenswrapper[4790]: I1124 13:35:07.130847 4790 generic.go:334] "Generic (PLEG): container finished" podID="7b3e66b8-f5c2-4b15-b5f7-52182f2c6681" containerID="745f3d1295cb0862befcfc61cf938b4759974af5280b36246b2dd03515b5389e" exitCode=0 Nov 24 13:35:07 crc kubenswrapper[4790]: I1124 13:35:07.130959 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681","Type":"ContainerDied","Data":"745f3d1295cb0862befcfc61cf938b4759974af5280b36246b2dd03515b5389e"} Nov 24 13:35:08 crc kubenswrapper[4790]: I1124 13:35:08.146450 4790 generic.go:334] "Generic (PLEG): container finished" podID="7b3e66b8-f5c2-4b15-b5f7-52182f2c6681" containerID="88d86e7122598d0c47ead3ff237534c57ec8c976adc9689dcaf211afaab01df3" exitCode=0 Nov 24 13:35:08 crc kubenswrapper[4790]: I1124 13:35:08.146543 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681","Type":"ContainerDied","Data":"88d86e7122598d0c47ead3ff237534c57ec8c976adc9689dcaf211afaab01df3"} Nov 24 13:35:11 crc kubenswrapper[4790]: I1124 13:35:11.436642 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:35:11 crc kubenswrapper[4790]: I1124 13:35:11.598561 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-config-data\") pod \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\" (UID: \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\") " Nov 24 13:35:11 crc kubenswrapper[4790]: I1124 13:35:11.598662 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-log-httpd\") pod \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\" (UID: \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\") " Nov 24 13:35:11 crc kubenswrapper[4790]: I1124 13:35:11.598776 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-scripts\") pod \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\" (UID: \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\") " Nov 24 13:35:11 crc kubenswrapper[4790]: I1124 13:35:11.598796 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-run-httpd\") pod \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\" (UID: \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\") " Nov 24 13:35:11 crc kubenswrapper[4790]: I1124 13:35:11.598853 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-sg-core-conf-yaml\") pod \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\" (UID: \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\") " Nov 24 13:35:11 crc kubenswrapper[4790]: I1124 13:35:11.598959 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-combined-ca-bundle\") pod \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\" (UID: \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\") " Nov 24 13:35:11 crc kubenswrapper[4790]: I1124 13:35:11.598997 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwljz\" (UniqueName: \"kubernetes.io/projected/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-kube-api-access-gwljz\") pod \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\" (UID: \"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681\") " Nov 24 13:35:11 crc kubenswrapper[4790]: I1124 13:35:11.599699 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7b3e66b8-f5c2-4b15-b5f7-52182f2c6681" (UID: "7b3e66b8-f5c2-4b15-b5f7-52182f2c6681"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:35:11 crc kubenswrapper[4790]: I1124 13:35:11.599722 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7b3e66b8-f5c2-4b15-b5f7-52182f2c6681" (UID: "7b3e66b8-f5c2-4b15-b5f7-52182f2c6681"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:35:11 crc kubenswrapper[4790]: I1124 13:35:11.606111 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-kube-api-access-gwljz" (OuterVolumeSpecName: "kube-api-access-gwljz") pod "7b3e66b8-f5c2-4b15-b5f7-52182f2c6681" (UID: "7b3e66b8-f5c2-4b15-b5f7-52182f2c6681"). InnerVolumeSpecName "kube-api-access-gwljz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:35:11 crc kubenswrapper[4790]: I1124 13:35:11.607054 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-scripts" (OuterVolumeSpecName: "scripts") pod "7b3e66b8-f5c2-4b15-b5f7-52182f2c6681" (UID: "7b3e66b8-f5c2-4b15-b5f7-52182f2c6681"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:11 crc kubenswrapper[4790]: I1124 13:35:11.633892 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7b3e66b8-f5c2-4b15-b5f7-52182f2c6681" (UID: "7b3e66b8-f5c2-4b15-b5f7-52182f2c6681"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:11 crc kubenswrapper[4790]: I1124 13:35:11.665452 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7b3e66b8-f5c2-4b15-b5f7-52182f2c6681" (UID: "7b3e66b8-f5c2-4b15-b5f7-52182f2c6681"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:11 crc kubenswrapper[4790]: I1124 13:35:11.701305 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:11 crc kubenswrapper[4790]: I1124 13:35:11.701343 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwljz\" (UniqueName: \"kubernetes.io/projected/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-kube-api-access-gwljz\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:11 crc kubenswrapper[4790]: I1124 13:35:11.701358 4790 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:11 crc kubenswrapper[4790]: I1124 13:35:11.701369 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:11 crc kubenswrapper[4790]: I1124 13:35:11.701380 4790 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:11 crc kubenswrapper[4790]: I1124 13:35:11.701390 4790 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:11 crc kubenswrapper[4790]: I1124 13:35:11.704041 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-config-data" (OuterVolumeSpecName: "config-data") pod "7b3e66b8-f5c2-4b15-b5f7-52182f2c6681" (UID: "7b3e66b8-f5c2-4b15-b5f7-52182f2c6681"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:11 crc kubenswrapper[4790]: I1124 13:35:11.803538 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.182567 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-s2mrt" event={"ID":"76801ca0-a42b-4a25-84b8-43a689e929ff","Type":"ContainerStarted","Data":"82a2835a200a5782ef25e715c28b47b86306977b719a0b3e260fe10df86d647c"} Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.185363 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7b3e66b8-f5c2-4b15-b5f7-52182f2c6681","Type":"ContainerDied","Data":"8d8eb81bfcbfb49c45181dbfad85f0fcca486f70ef864c14f5cb8598f920cd81"} Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.185436 4790 scope.go:117] "RemoveContainer" containerID="74fa6831af4f662dc1d64fe0d79dfdaa11bcaf6438540da07535916e9c3be8b5" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.185561 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.217370 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-s2mrt" podStartSLOduration=2.027822025 podStartE2EDuration="16.217298893s" podCreationTimestamp="2025-11-24 13:34:56 +0000 UTC" firstStartedPulling="2025-11-24 13:34:56.980103597 +0000 UTC m=+1345.359997249" lastFinishedPulling="2025-11-24 13:35:11.169580455 +0000 UTC m=+1359.549474117" observedRunningTime="2025-11-24 13:35:12.211830336 +0000 UTC m=+1360.591723998" watchObservedRunningTime="2025-11-24 13:35:12.217298893 +0000 UTC m=+1360.597192555" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.232115 4790 scope.go:117] "RemoveContainer" containerID="9689802d8b7838f31aa896e0bd57aa2180198393ea4527e352c87bab4e8d865a" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.241740 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.250393 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.257788 4790 scope.go:117] "RemoveContainer" containerID="745f3d1295cb0862befcfc61cf938b4759974af5280b36246b2dd03515b5389e" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.268903 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:35:12 crc kubenswrapper[4790]: E1124 13:35:12.269382 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b3e66b8-f5c2-4b15-b5f7-52182f2c6681" containerName="ceilometer-central-agent" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.269409 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b3e66b8-f5c2-4b15-b5f7-52182f2c6681" containerName="ceilometer-central-agent" Nov 24 13:35:12 crc kubenswrapper[4790]: E1124 13:35:12.269442 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b3e66b8-f5c2-4b15-b5f7-52182f2c6681" containerName="sg-core" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.269451 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b3e66b8-f5c2-4b15-b5f7-52182f2c6681" containerName="sg-core" Nov 24 13:35:12 crc kubenswrapper[4790]: E1124 13:35:12.269465 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b3e66b8-f5c2-4b15-b5f7-52182f2c6681" containerName="ceilometer-notification-agent" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.269474 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b3e66b8-f5c2-4b15-b5f7-52182f2c6681" containerName="ceilometer-notification-agent" Nov 24 13:35:12 crc kubenswrapper[4790]: E1124 13:35:12.269502 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b3e66b8-f5c2-4b15-b5f7-52182f2c6681" containerName="proxy-httpd" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.269510 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b3e66b8-f5c2-4b15-b5f7-52182f2c6681" containerName="proxy-httpd" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.269712 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b3e66b8-f5c2-4b15-b5f7-52182f2c6681" containerName="sg-core" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.269736 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b3e66b8-f5c2-4b15-b5f7-52182f2c6681" containerName="proxy-httpd" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.269747 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b3e66b8-f5c2-4b15-b5f7-52182f2c6681" containerName="ceilometer-notification-agent" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.269760 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b3e66b8-f5c2-4b15-b5f7-52182f2c6681" containerName="ceilometer-central-agent" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.271957 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.276430 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.276745 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.285461 4790 scope.go:117] "RemoveContainer" containerID="88d86e7122598d0c47ead3ff237534c57ec8c976adc9689dcaf211afaab01df3" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.288800 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.339852 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b3e66b8-f5c2-4b15-b5f7-52182f2c6681" path="/var/lib/kubelet/pods/7b3e66b8-f5c2-4b15-b5f7-52182f2c6681/volumes" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.420750 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/237e7166-6c17-46d8-a18b-2af671d75e28-log-httpd\") pod \"ceilometer-0\" (UID: \"237e7166-6c17-46d8-a18b-2af671d75e28\") " pod="openstack/ceilometer-0" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.420896 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/237e7166-6c17-46d8-a18b-2af671d75e28-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"237e7166-6c17-46d8-a18b-2af671d75e28\") " pod="openstack/ceilometer-0" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.420959 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/237e7166-6c17-46d8-a18b-2af671d75e28-run-httpd\") pod \"ceilometer-0\" (UID: \"237e7166-6c17-46d8-a18b-2af671d75e28\") " pod="openstack/ceilometer-0" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.421016 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pz2k4\" (UniqueName: \"kubernetes.io/projected/237e7166-6c17-46d8-a18b-2af671d75e28-kube-api-access-pz2k4\") pod \"ceilometer-0\" (UID: \"237e7166-6c17-46d8-a18b-2af671d75e28\") " pod="openstack/ceilometer-0" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.421048 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/237e7166-6c17-46d8-a18b-2af671d75e28-config-data\") pod \"ceilometer-0\" (UID: \"237e7166-6c17-46d8-a18b-2af671d75e28\") " pod="openstack/ceilometer-0" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.421065 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/237e7166-6c17-46d8-a18b-2af671d75e28-scripts\") pod \"ceilometer-0\" (UID: \"237e7166-6c17-46d8-a18b-2af671d75e28\") " pod="openstack/ceilometer-0" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.421103 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/237e7166-6c17-46d8-a18b-2af671d75e28-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"237e7166-6c17-46d8-a18b-2af671d75e28\") " pod="openstack/ceilometer-0" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.522742 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/237e7166-6c17-46d8-a18b-2af671d75e28-run-httpd\") pod \"ceilometer-0\" (UID: \"237e7166-6c17-46d8-a18b-2af671d75e28\") " pod="openstack/ceilometer-0" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.522821 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pz2k4\" (UniqueName: \"kubernetes.io/projected/237e7166-6c17-46d8-a18b-2af671d75e28-kube-api-access-pz2k4\") pod \"ceilometer-0\" (UID: \"237e7166-6c17-46d8-a18b-2af671d75e28\") " pod="openstack/ceilometer-0" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.522857 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/237e7166-6c17-46d8-a18b-2af671d75e28-config-data\") pod \"ceilometer-0\" (UID: \"237e7166-6c17-46d8-a18b-2af671d75e28\") " pod="openstack/ceilometer-0" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.522887 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/237e7166-6c17-46d8-a18b-2af671d75e28-scripts\") pod \"ceilometer-0\" (UID: \"237e7166-6c17-46d8-a18b-2af671d75e28\") " pod="openstack/ceilometer-0" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.522925 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/237e7166-6c17-46d8-a18b-2af671d75e28-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"237e7166-6c17-46d8-a18b-2af671d75e28\") " pod="openstack/ceilometer-0" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.522953 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/237e7166-6c17-46d8-a18b-2af671d75e28-log-httpd\") pod \"ceilometer-0\" (UID: \"237e7166-6c17-46d8-a18b-2af671d75e28\") " pod="openstack/ceilometer-0" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.522993 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/237e7166-6c17-46d8-a18b-2af671d75e28-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"237e7166-6c17-46d8-a18b-2af671d75e28\") " pod="openstack/ceilometer-0" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.523294 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/237e7166-6c17-46d8-a18b-2af671d75e28-run-httpd\") pod \"ceilometer-0\" (UID: \"237e7166-6c17-46d8-a18b-2af671d75e28\") " pod="openstack/ceilometer-0" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.524094 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/237e7166-6c17-46d8-a18b-2af671d75e28-log-httpd\") pod \"ceilometer-0\" (UID: \"237e7166-6c17-46d8-a18b-2af671d75e28\") " pod="openstack/ceilometer-0" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.527119 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/237e7166-6c17-46d8-a18b-2af671d75e28-scripts\") pod \"ceilometer-0\" (UID: \"237e7166-6c17-46d8-a18b-2af671d75e28\") " pod="openstack/ceilometer-0" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.527181 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/237e7166-6c17-46d8-a18b-2af671d75e28-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"237e7166-6c17-46d8-a18b-2af671d75e28\") " pod="openstack/ceilometer-0" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.527907 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/237e7166-6c17-46d8-a18b-2af671d75e28-config-data\") pod \"ceilometer-0\" (UID: \"237e7166-6c17-46d8-a18b-2af671d75e28\") " pod="openstack/ceilometer-0" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.531202 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/237e7166-6c17-46d8-a18b-2af671d75e28-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"237e7166-6c17-46d8-a18b-2af671d75e28\") " pod="openstack/ceilometer-0" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.538715 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pz2k4\" (UniqueName: \"kubernetes.io/projected/237e7166-6c17-46d8-a18b-2af671d75e28-kube-api-access-pz2k4\") pod \"ceilometer-0\" (UID: \"237e7166-6c17-46d8-a18b-2af671d75e28\") " pod="openstack/ceilometer-0" Nov 24 13:35:12 crc kubenswrapper[4790]: I1124 13:35:12.645901 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:35:13 crc kubenswrapper[4790]: I1124 13:35:13.069613 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:35:13 crc kubenswrapper[4790]: W1124 13:35:13.070687 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod237e7166_6c17_46d8_a18b_2af671d75e28.slice/crio-4def04435b8fa9bcf11de020fdd00930a2050f00a84542747dad5156b9bea006 WatchSource:0}: Error finding container 4def04435b8fa9bcf11de020fdd00930a2050f00a84542747dad5156b9bea006: Status 404 returned error can't find the container with id 4def04435b8fa9bcf11de020fdd00930a2050f00a84542747dad5156b9bea006 Nov 24 13:35:13 crc kubenswrapper[4790]: I1124 13:35:13.196101 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"237e7166-6c17-46d8-a18b-2af671d75e28","Type":"ContainerStarted","Data":"4def04435b8fa9bcf11de020fdd00930a2050f00a84542747dad5156b9bea006"} Nov 24 13:35:14 crc kubenswrapper[4790]: I1124 13:35:14.213042 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"237e7166-6c17-46d8-a18b-2af671d75e28","Type":"ContainerStarted","Data":"6b70cf6fea320d4bb5acf67752d705bc8f4e8b0eea3d825bd9b466ef671d7cee"} Nov 24 13:35:15 crc kubenswrapper[4790]: I1124 13:35:15.224679 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"237e7166-6c17-46d8-a18b-2af671d75e28","Type":"ContainerStarted","Data":"04275059fed1ba9eabf4afa369dbc5147006448caf80244a63fecf325ccc57e7"} Nov 24 13:35:15 crc kubenswrapper[4790]: I1124 13:35:15.224971 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"237e7166-6c17-46d8-a18b-2af671d75e28","Type":"ContainerStarted","Data":"da4959e122210172c5153b9713ea054958931f764b05982716558bf2d9239f5f"} Nov 24 13:35:17 crc kubenswrapper[4790]: I1124 13:35:17.242461 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"237e7166-6c17-46d8-a18b-2af671d75e28","Type":"ContainerStarted","Data":"243b98c6ff9675ec9019b09fe6a4ee32b15708475572e6be2dd633baf9614988"} Nov 24 13:35:17 crc kubenswrapper[4790]: I1124 13:35:17.243068 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 13:35:17 crc kubenswrapper[4790]: I1124 13:35:17.270202 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.24293686 podStartE2EDuration="5.2701853s" podCreationTimestamp="2025-11-24 13:35:12 +0000 UTC" firstStartedPulling="2025-11-24 13:35:13.073790375 +0000 UTC m=+1361.453684037" lastFinishedPulling="2025-11-24 13:35:16.101038815 +0000 UTC m=+1364.480932477" observedRunningTime="2025-11-24 13:35:17.260398151 +0000 UTC m=+1365.640291833" watchObservedRunningTime="2025-11-24 13:35:17.2701853 +0000 UTC m=+1365.650078962" Nov 24 13:35:21 crc kubenswrapper[4790]: I1124 13:35:21.284089 4790 generic.go:334] "Generic (PLEG): container finished" podID="76801ca0-a42b-4a25-84b8-43a689e929ff" containerID="82a2835a200a5782ef25e715c28b47b86306977b719a0b3e260fe10df86d647c" exitCode=0 Nov 24 13:35:21 crc kubenswrapper[4790]: I1124 13:35:21.284655 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-s2mrt" event={"ID":"76801ca0-a42b-4a25-84b8-43a689e929ff","Type":"ContainerDied","Data":"82a2835a200a5782ef25e715c28b47b86306977b719a0b3e260fe10df86d647c"} Nov 24 13:35:22 crc kubenswrapper[4790]: I1124 13:35:22.641753 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-s2mrt" Nov 24 13:35:22 crc kubenswrapper[4790]: I1124 13:35:22.727547 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76801ca0-a42b-4a25-84b8-43a689e929ff-config-data\") pod \"76801ca0-a42b-4a25-84b8-43a689e929ff\" (UID: \"76801ca0-a42b-4a25-84b8-43a689e929ff\") " Nov 24 13:35:22 crc kubenswrapper[4790]: I1124 13:35:22.727623 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76801ca0-a42b-4a25-84b8-43a689e929ff-scripts\") pod \"76801ca0-a42b-4a25-84b8-43a689e929ff\" (UID: \"76801ca0-a42b-4a25-84b8-43a689e929ff\") " Nov 24 13:35:22 crc kubenswrapper[4790]: I1124 13:35:22.727669 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnkw7\" (UniqueName: \"kubernetes.io/projected/76801ca0-a42b-4a25-84b8-43a689e929ff-kube-api-access-jnkw7\") pod \"76801ca0-a42b-4a25-84b8-43a689e929ff\" (UID: \"76801ca0-a42b-4a25-84b8-43a689e929ff\") " Nov 24 13:35:22 crc kubenswrapper[4790]: I1124 13:35:22.727711 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76801ca0-a42b-4a25-84b8-43a689e929ff-combined-ca-bundle\") pod \"76801ca0-a42b-4a25-84b8-43a689e929ff\" (UID: \"76801ca0-a42b-4a25-84b8-43a689e929ff\") " Nov 24 13:35:22 crc kubenswrapper[4790]: I1124 13:35:22.734036 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76801ca0-a42b-4a25-84b8-43a689e929ff-scripts" (OuterVolumeSpecName: "scripts") pod "76801ca0-a42b-4a25-84b8-43a689e929ff" (UID: "76801ca0-a42b-4a25-84b8-43a689e929ff"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:22 crc kubenswrapper[4790]: I1124 13:35:22.734496 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76801ca0-a42b-4a25-84b8-43a689e929ff-kube-api-access-jnkw7" (OuterVolumeSpecName: "kube-api-access-jnkw7") pod "76801ca0-a42b-4a25-84b8-43a689e929ff" (UID: "76801ca0-a42b-4a25-84b8-43a689e929ff"). InnerVolumeSpecName "kube-api-access-jnkw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:35:22 crc kubenswrapper[4790]: I1124 13:35:22.756194 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76801ca0-a42b-4a25-84b8-43a689e929ff-config-data" (OuterVolumeSpecName: "config-data") pod "76801ca0-a42b-4a25-84b8-43a689e929ff" (UID: "76801ca0-a42b-4a25-84b8-43a689e929ff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:22 crc kubenswrapper[4790]: I1124 13:35:22.759495 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76801ca0-a42b-4a25-84b8-43a689e929ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "76801ca0-a42b-4a25-84b8-43a689e929ff" (UID: "76801ca0-a42b-4a25-84b8-43a689e929ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:22 crc kubenswrapper[4790]: I1124 13:35:22.830347 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnkw7\" (UniqueName: \"kubernetes.io/projected/76801ca0-a42b-4a25-84b8-43a689e929ff-kube-api-access-jnkw7\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:22 crc kubenswrapper[4790]: I1124 13:35:22.830399 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76801ca0-a42b-4a25-84b8-43a689e929ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:22 crc kubenswrapper[4790]: I1124 13:35:22.830408 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76801ca0-a42b-4a25-84b8-43a689e929ff-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:22 crc kubenswrapper[4790]: I1124 13:35:22.830418 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76801ca0-a42b-4a25-84b8-43a689e929ff-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:23 crc kubenswrapper[4790]: I1124 13:35:23.307399 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-s2mrt" event={"ID":"76801ca0-a42b-4a25-84b8-43a689e929ff","Type":"ContainerDied","Data":"bc2bffe906e252d6e0b71a7e74bfa4a43861e50d36cc805c33e3e339261eb2fe"} Nov 24 13:35:23 crc kubenswrapper[4790]: I1124 13:35:23.307432 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc2bffe906e252d6e0b71a7e74bfa4a43861e50d36cc805c33e3e339261eb2fe" Nov 24 13:35:23 crc kubenswrapper[4790]: I1124 13:35:23.307438 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-s2mrt" Nov 24 13:35:23 crc kubenswrapper[4790]: I1124 13:35:23.392503 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 13:35:23 crc kubenswrapper[4790]: E1124 13:35:23.392938 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76801ca0-a42b-4a25-84b8-43a689e929ff" containerName="nova-cell0-conductor-db-sync" Nov 24 13:35:23 crc kubenswrapper[4790]: I1124 13:35:23.393198 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="76801ca0-a42b-4a25-84b8-43a689e929ff" containerName="nova-cell0-conductor-db-sync" Nov 24 13:35:23 crc kubenswrapper[4790]: I1124 13:35:23.393388 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="76801ca0-a42b-4a25-84b8-43a689e929ff" containerName="nova-cell0-conductor-db-sync" Nov 24 13:35:23 crc kubenswrapper[4790]: I1124 13:35:23.394014 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 13:35:23 crc kubenswrapper[4790]: I1124 13:35:23.396490 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 13:35:23 crc kubenswrapper[4790]: I1124 13:35:23.396714 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-294qh" Nov 24 13:35:23 crc kubenswrapper[4790]: I1124 13:35:23.403715 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 13:35:23 crc kubenswrapper[4790]: I1124 13:35:23.540161 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fdd5969-f22b-41e6-a0d9-f3eeb16689fd-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"7fdd5969-f22b-41e6-a0d9-f3eeb16689fd\") " pod="openstack/nova-cell0-conductor-0" Nov 24 13:35:23 crc kubenswrapper[4790]: I1124 13:35:23.540511 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fdd5969-f22b-41e6-a0d9-f3eeb16689fd-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"7fdd5969-f22b-41e6-a0d9-f3eeb16689fd\") " pod="openstack/nova-cell0-conductor-0" Nov 24 13:35:23 crc kubenswrapper[4790]: I1124 13:35:23.540616 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-852vf\" (UniqueName: \"kubernetes.io/projected/7fdd5969-f22b-41e6-a0d9-f3eeb16689fd-kube-api-access-852vf\") pod \"nova-cell0-conductor-0\" (UID: \"7fdd5969-f22b-41e6-a0d9-f3eeb16689fd\") " pod="openstack/nova-cell0-conductor-0" Nov 24 13:35:23 crc kubenswrapper[4790]: I1124 13:35:23.642245 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fdd5969-f22b-41e6-a0d9-f3eeb16689fd-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"7fdd5969-f22b-41e6-a0d9-f3eeb16689fd\") " pod="openstack/nova-cell0-conductor-0" Nov 24 13:35:23 crc kubenswrapper[4790]: I1124 13:35:23.642355 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fdd5969-f22b-41e6-a0d9-f3eeb16689fd-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"7fdd5969-f22b-41e6-a0d9-f3eeb16689fd\") " pod="openstack/nova-cell0-conductor-0" Nov 24 13:35:23 crc kubenswrapper[4790]: I1124 13:35:23.642401 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-852vf\" (UniqueName: \"kubernetes.io/projected/7fdd5969-f22b-41e6-a0d9-f3eeb16689fd-kube-api-access-852vf\") pod \"nova-cell0-conductor-0\" (UID: \"7fdd5969-f22b-41e6-a0d9-f3eeb16689fd\") " pod="openstack/nova-cell0-conductor-0" Nov 24 13:35:23 crc kubenswrapper[4790]: I1124 13:35:23.649616 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fdd5969-f22b-41e6-a0d9-f3eeb16689fd-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"7fdd5969-f22b-41e6-a0d9-f3eeb16689fd\") " pod="openstack/nova-cell0-conductor-0" Nov 24 13:35:23 crc kubenswrapper[4790]: I1124 13:35:23.652314 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fdd5969-f22b-41e6-a0d9-f3eeb16689fd-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"7fdd5969-f22b-41e6-a0d9-f3eeb16689fd\") " pod="openstack/nova-cell0-conductor-0" Nov 24 13:35:23 crc kubenswrapper[4790]: I1124 13:35:23.658140 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-852vf\" (UniqueName: \"kubernetes.io/projected/7fdd5969-f22b-41e6-a0d9-f3eeb16689fd-kube-api-access-852vf\") pod \"nova-cell0-conductor-0\" (UID: \"7fdd5969-f22b-41e6-a0d9-f3eeb16689fd\") " pod="openstack/nova-cell0-conductor-0" Nov 24 13:35:23 crc kubenswrapper[4790]: I1124 13:35:23.709852 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 13:35:24 crc kubenswrapper[4790]: I1124 13:35:24.145603 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 13:35:24 crc kubenswrapper[4790]: I1124 13:35:24.326664 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"7fdd5969-f22b-41e6-a0d9-f3eeb16689fd","Type":"ContainerStarted","Data":"e05aa9b0eb724d4a31ae22ff76a39eeca75170714d76458a140c2bb64d160a41"} Nov 24 13:35:25 crc kubenswrapper[4790]: I1124 13:35:25.333810 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"7fdd5969-f22b-41e6-a0d9-f3eeb16689fd","Type":"ContainerStarted","Data":"d8eab96ea490fbfa774abdd6a272fffab4b7cacb733249bf3f282fa0dfcf8521"} Nov 24 13:35:25 crc kubenswrapper[4790]: I1124 13:35:25.336055 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 24 13:35:33 crc kubenswrapper[4790]: I1124 13:35:33.735098 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 24 13:35:33 crc kubenswrapper[4790]: I1124 13:35:33.770372 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=10.770348892 podStartE2EDuration="10.770348892s" podCreationTimestamp="2025-11-24 13:35:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:35:25.356806782 +0000 UTC m=+1373.736700444" watchObservedRunningTime="2025-11-24 13:35:33.770348892 +0000 UTC m=+1382.150242554" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.252707 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-gjs8m"] Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.254261 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-gjs8m" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.257106 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.264543 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-gjs8m"] Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.278553 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.342400 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkj8h\" (UniqueName: \"kubernetes.io/projected/fddd7928-5ec3-4605-94fb-a663c7f410e3-kube-api-access-xkj8h\") pod \"nova-cell0-cell-mapping-gjs8m\" (UID: \"fddd7928-5ec3-4605-94fb-a663c7f410e3\") " pod="openstack/nova-cell0-cell-mapping-gjs8m" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.342809 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fddd7928-5ec3-4605-94fb-a663c7f410e3-config-data\") pod \"nova-cell0-cell-mapping-gjs8m\" (UID: \"fddd7928-5ec3-4605-94fb-a663c7f410e3\") " pod="openstack/nova-cell0-cell-mapping-gjs8m" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.343057 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fddd7928-5ec3-4605-94fb-a663c7f410e3-scripts\") pod \"nova-cell0-cell-mapping-gjs8m\" (UID: \"fddd7928-5ec3-4605-94fb-a663c7f410e3\") " pod="openstack/nova-cell0-cell-mapping-gjs8m" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.343164 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fddd7928-5ec3-4605-94fb-a663c7f410e3-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-gjs8m\" (UID: \"fddd7928-5ec3-4605-94fb-a663c7f410e3\") " pod="openstack/nova-cell0-cell-mapping-gjs8m" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.445053 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkj8h\" (UniqueName: \"kubernetes.io/projected/fddd7928-5ec3-4605-94fb-a663c7f410e3-kube-api-access-xkj8h\") pod \"nova-cell0-cell-mapping-gjs8m\" (UID: \"fddd7928-5ec3-4605-94fb-a663c7f410e3\") " pod="openstack/nova-cell0-cell-mapping-gjs8m" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.445119 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fddd7928-5ec3-4605-94fb-a663c7f410e3-config-data\") pod \"nova-cell0-cell-mapping-gjs8m\" (UID: \"fddd7928-5ec3-4605-94fb-a663c7f410e3\") " pod="openstack/nova-cell0-cell-mapping-gjs8m" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.445183 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fddd7928-5ec3-4605-94fb-a663c7f410e3-scripts\") pod \"nova-cell0-cell-mapping-gjs8m\" (UID: \"fddd7928-5ec3-4605-94fb-a663c7f410e3\") " pod="openstack/nova-cell0-cell-mapping-gjs8m" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.445229 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fddd7928-5ec3-4605-94fb-a663c7f410e3-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-gjs8m\" (UID: \"fddd7928-5ec3-4605-94fb-a663c7f410e3\") " pod="openstack/nova-cell0-cell-mapping-gjs8m" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.455697 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fddd7928-5ec3-4605-94fb-a663c7f410e3-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-gjs8m\" (UID: \"fddd7928-5ec3-4605-94fb-a663c7f410e3\") " pod="openstack/nova-cell0-cell-mapping-gjs8m" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.462511 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fddd7928-5ec3-4605-94fb-a663c7f410e3-config-data\") pod \"nova-cell0-cell-mapping-gjs8m\" (UID: \"fddd7928-5ec3-4605-94fb-a663c7f410e3\") " pod="openstack/nova-cell0-cell-mapping-gjs8m" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.468509 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fddd7928-5ec3-4605-94fb-a663c7f410e3-scripts\") pod \"nova-cell0-cell-mapping-gjs8m\" (UID: \"fddd7928-5ec3-4605-94fb-a663c7f410e3\") " pod="openstack/nova-cell0-cell-mapping-gjs8m" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.538374 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.542576 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkj8h\" (UniqueName: \"kubernetes.io/projected/fddd7928-5ec3-4605-94fb-a663c7f410e3-kube-api-access-xkj8h\") pod \"nova-cell0-cell-mapping-gjs8m\" (UID: \"fddd7928-5ec3-4605-94fb-a663c7f410e3\") " pod="openstack/nova-cell0-cell-mapping-gjs8m" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.549061 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.570275 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.579770 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.630222 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-gjs8m" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.651217 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38092ab0-6779-4823-b8f0-7955f7e436ee-logs\") pod \"nova-metadata-0\" (UID: \"38092ab0-6779-4823-b8f0-7955f7e436ee\") " pod="openstack/nova-metadata-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.651260 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcwk5\" (UniqueName: \"kubernetes.io/projected/38092ab0-6779-4823-b8f0-7955f7e436ee-kube-api-access-gcwk5\") pod \"nova-metadata-0\" (UID: \"38092ab0-6779-4823-b8f0-7955f7e436ee\") " pod="openstack/nova-metadata-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.651298 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38092ab0-6779-4823-b8f0-7955f7e436ee-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"38092ab0-6779-4823-b8f0-7955f7e436ee\") " pod="openstack/nova-metadata-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.651367 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38092ab0-6779-4823-b8f0-7955f7e436ee-config-data\") pod \"nova-metadata-0\" (UID: \"38092ab0-6779-4823-b8f0-7955f7e436ee\") " pod="openstack/nova-metadata-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.671966 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.673070 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.680217 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.715954 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.742398 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.744099 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.758277 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.758340 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.758364 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38092ab0-6779-4823-b8f0-7955f7e436ee-config-data\") pod \"nova-metadata-0\" (UID: \"38092ab0-6779-4823-b8f0-7955f7e436ee\") " pod="openstack/nova-metadata-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.758435 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4krqq\" (UniqueName: \"kubernetes.io/projected/5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0-kube-api-access-4krqq\") pod \"nova-cell1-novncproxy-0\" (UID: \"5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.758468 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38092ab0-6779-4823-b8f0-7955f7e436ee-logs\") pod \"nova-metadata-0\" (UID: \"38092ab0-6779-4823-b8f0-7955f7e436ee\") " pod="openstack/nova-metadata-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.758496 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcwk5\" (UniqueName: \"kubernetes.io/projected/38092ab0-6779-4823-b8f0-7955f7e436ee-kube-api-access-gcwk5\") pod \"nova-metadata-0\" (UID: \"38092ab0-6779-4823-b8f0-7955f7e436ee\") " pod="openstack/nova-metadata-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.758529 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38092ab0-6779-4823-b8f0-7955f7e436ee-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"38092ab0-6779-4823-b8f0-7955f7e436ee\") " pod="openstack/nova-metadata-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.761612 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38092ab0-6779-4823-b8f0-7955f7e436ee-logs\") pod \"nova-metadata-0\" (UID: \"38092ab0-6779-4823-b8f0-7955f7e436ee\") " pod="openstack/nova-metadata-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.762886 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.769361 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38092ab0-6779-4823-b8f0-7955f7e436ee-config-data\") pod \"nova-metadata-0\" (UID: \"38092ab0-6779-4823-b8f0-7955f7e436ee\") " pod="openstack/nova-metadata-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.769789 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38092ab0-6779-4823-b8f0-7955f7e436ee-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"38092ab0-6779-4823-b8f0-7955f7e436ee\") " pod="openstack/nova-metadata-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.791096 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.813992 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.815783 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.824745 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.825195 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcwk5\" (UniqueName: \"kubernetes.io/projected/38092ab0-6779-4823-b8f0-7955f7e436ee-kube-api-access-gcwk5\") pod \"nova-metadata-0\" (UID: \"38092ab0-6779-4823-b8f0-7955f7e436ee\") " pod="openstack/nova-metadata-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.842010 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-skvpz"] Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.843856 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-skvpz" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.861494 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd9e4394-01fa-485a-9e85-05b2178445e8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fd9e4394-01fa-485a-9e85-05b2178445e8\") " pod="openstack/nova-api-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.861540 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd9e4394-01fa-485a-9e85-05b2178445e8-config-data\") pod \"nova-api-0\" (UID: \"fd9e4394-01fa-485a-9e85-05b2178445e8\") " pod="openstack/nova-api-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.861596 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-skvpz\" (UID: \"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412\") " pod="openstack/dnsmasq-dns-bccf8f775-skvpz" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.861633 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.861655 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e988296-b952-43b2-bad8-39c2916bcecd-config-data\") pod \"nova-scheduler-0\" (UID: \"7e988296-b952-43b2-bad8-39c2916bcecd\") " pod="openstack/nova-scheduler-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.861711 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.861765 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e988296-b952-43b2-bad8-39c2916bcecd-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7e988296-b952-43b2-bad8-39c2916bcecd\") " pod="openstack/nova-scheduler-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.861818 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd9e4394-01fa-485a-9e85-05b2178445e8-logs\") pod \"nova-api-0\" (UID: \"fd9e4394-01fa-485a-9e85-05b2178445e8\") " pod="openstack/nova-api-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.861848 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-skvpz\" (UID: \"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412\") " pod="openstack/dnsmasq-dns-bccf8f775-skvpz" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.861918 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hw7cx\" (UniqueName: \"kubernetes.io/projected/fd9e4394-01fa-485a-9e85-05b2178445e8-kube-api-access-hw7cx\") pod \"nova-api-0\" (UID: \"fd9e4394-01fa-485a-9e85-05b2178445e8\") " pod="openstack/nova-api-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.861967 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4krqq\" (UniqueName: \"kubernetes.io/projected/5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0-kube-api-access-4krqq\") pod \"nova-cell1-novncproxy-0\" (UID: \"5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.861994 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-dns-svc\") pod \"dnsmasq-dns-bccf8f775-skvpz\" (UID: \"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412\") " pod="openstack/dnsmasq-dns-bccf8f775-skvpz" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.862022 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-config\") pod \"dnsmasq-dns-bccf8f775-skvpz\" (UID: \"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412\") " pod="openstack/dnsmasq-dns-bccf8f775-skvpz" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.862098 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-skvpz\" (UID: \"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412\") " pod="openstack/dnsmasq-dns-bccf8f775-skvpz" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.862123 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzk5l\" (UniqueName: \"kubernetes.io/projected/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-kube-api-access-gzk5l\") pod \"dnsmasq-dns-bccf8f775-skvpz\" (UID: \"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412\") " pod="openstack/dnsmasq-dns-bccf8f775-skvpz" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.862156 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhk95\" (UniqueName: \"kubernetes.io/projected/7e988296-b952-43b2-bad8-39c2916bcecd-kube-api-access-jhk95\") pod \"nova-scheduler-0\" (UID: \"7e988296-b952-43b2-bad8-39c2916bcecd\") " pod="openstack/nova-scheduler-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.864523 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.880419 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.882979 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.907319 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-skvpz"] Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.912494 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4krqq\" (UniqueName: \"kubernetes.io/projected/5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0-kube-api-access-4krqq\") pod \"nova-cell1-novncproxy-0\" (UID: \"5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.975599 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-skvpz\" (UID: \"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412\") " pod="openstack/dnsmasq-dns-bccf8f775-skvpz" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.975657 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e988296-b952-43b2-bad8-39c2916bcecd-config-data\") pod \"nova-scheduler-0\" (UID: \"7e988296-b952-43b2-bad8-39c2916bcecd\") " pod="openstack/nova-scheduler-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.975731 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e988296-b952-43b2-bad8-39c2916bcecd-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7e988296-b952-43b2-bad8-39c2916bcecd\") " pod="openstack/nova-scheduler-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.975782 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd9e4394-01fa-485a-9e85-05b2178445e8-logs\") pod \"nova-api-0\" (UID: \"fd9e4394-01fa-485a-9e85-05b2178445e8\") " pod="openstack/nova-api-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.975806 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-skvpz\" (UID: \"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412\") " pod="openstack/dnsmasq-dns-bccf8f775-skvpz" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.975849 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hw7cx\" (UniqueName: \"kubernetes.io/projected/fd9e4394-01fa-485a-9e85-05b2178445e8-kube-api-access-hw7cx\") pod \"nova-api-0\" (UID: \"fd9e4394-01fa-485a-9e85-05b2178445e8\") " pod="openstack/nova-api-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.975898 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-dns-svc\") pod \"dnsmasq-dns-bccf8f775-skvpz\" (UID: \"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412\") " pod="openstack/dnsmasq-dns-bccf8f775-skvpz" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.975926 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-config\") pod \"dnsmasq-dns-bccf8f775-skvpz\" (UID: \"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412\") " pod="openstack/dnsmasq-dns-bccf8f775-skvpz" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.975976 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-skvpz\" (UID: \"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412\") " pod="openstack/dnsmasq-dns-bccf8f775-skvpz" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.975997 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzk5l\" (UniqueName: \"kubernetes.io/projected/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-kube-api-access-gzk5l\") pod \"dnsmasq-dns-bccf8f775-skvpz\" (UID: \"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412\") " pod="openstack/dnsmasq-dns-bccf8f775-skvpz" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.976027 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhk95\" (UniqueName: \"kubernetes.io/projected/7e988296-b952-43b2-bad8-39c2916bcecd-kube-api-access-jhk95\") pod \"nova-scheduler-0\" (UID: \"7e988296-b952-43b2-bad8-39c2916bcecd\") " pod="openstack/nova-scheduler-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.976080 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd9e4394-01fa-485a-9e85-05b2178445e8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fd9e4394-01fa-485a-9e85-05b2178445e8\") " pod="openstack/nova-api-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.976102 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd9e4394-01fa-485a-9e85-05b2178445e8-config-data\") pod \"nova-api-0\" (UID: \"fd9e4394-01fa-485a-9e85-05b2178445e8\") " pod="openstack/nova-api-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.977240 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd9e4394-01fa-485a-9e85-05b2178445e8-logs\") pod \"nova-api-0\" (UID: \"fd9e4394-01fa-485a-9e85-05b2178445e8\") " pod="openstack/nova-api-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.977324 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-skvpz\" (UID: \"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412\") " pod="openstack/dnsmasq-dns-bccf8f775-skvpz" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.978842 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-skvpz\" (UID: \"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412\") " pod="openstack/dnsmasq-dns-bccf8f775-skvpz" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.980593 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-config\") pod \"dnsmasq-dns-bccf8f775-skvpz\" (UID: \"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412\") " pod="openstack/dnsmasq-dns-bccf8f775-skvpz" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.981778 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-dns-svc\") pod \"dnsmasq-dns-bccf8f775-skvpz\" (UID: \"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412\") " pod="openstack/dnsmasq-dns-bccf8f775-skvpz" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.980639 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-skvpz\" (UID: \"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412\") " pod="openstack/dnsmasq-dns-bccf8f775-skvpz" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.986001 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd9e4394-01fa-485a-9e85-05b2178445e8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fd9e4394-01fa-485a-9e85-05b2178445e8\") " pod="openstack/nova-api-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.986353 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e988296-b952-43b2-bad8-39c2916bcecd-config-data\") pod \"nova-scheduler-0\" (UID: \"7e988296-b952-43b2-bad8-39c2916bcecd\") " pod="openstack/nova-scheduler-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.986463 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e988296-b952-43b2-bad8-39c2916bcecd-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7e988296-b952-43b2-bad8-39c2916bcecd\") " pod="openstack/nova-scheduler-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.995667 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hw7cx\" (UniqueName: \"kubernetes.io/projected/fd9e4394-01fa-485a-9e85-05b2178445e8-kube-api-access-hw7cx\") pod \"nova-api-0\" (UID: \"fd9e4394-01fa-485a-9e85-05b2178445e8\") " pod="openstack/nova-api-0" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.997330 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzk5l\" (UniqueName: \"kubernetes.io/projected/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-kube-api-access-gzk5l\") pod \"dnsmasq-dns-bccf8f775-skvpz\" (UID: \"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412\") " pod="openstack/dnsmasq-dns-bccf8f775-skvpz" Nov 24 13:35:34 crc kubenswrapper[4790]: I1124 13:35:34.999438 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhk95\" (UniqueName: \"kubernetes.io/projected/7e988296-b952-43b2-bad8-39c2916bcecd-kube-api-access-jhk95\") pod \"nova-scheduler-0\" (UID: \"7e988296-b952-43b2-bad8-39c2916bcecd\") " pod="openstack/nova-scheduler-0" Nov 24 13:35:35 crc kubenswrapper[4790]: I1124 13:35:35.003632 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd9e4394-01fa-485a-9e85-05b2178445e8-config-data\") pod \"nova-api-0\" (UID: \"fd9e4394-01fa-485a-9e85-05b2178445e8\") " pod="openstack/nova-api-0" Nov 24 13:35:35 crc kubenswrapper[4790]: I1124 13:35:35.069234 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 13:35:35 crc kubenswrapper[4790]: I1124 13:35:35.107079 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:35:35 crc kubenswrapper[4790]: I1124 13:35:35.179363 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 13:35:35 crc kubenswrapper[4790]: I1124 13:35:35.210257 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 13:35:35 crc kubenswrapper[4790]: I1124 13:35:35.227394 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-skvpz" Nov 24 13:35:35 crc kubenswrapper[4790]: I1124 13:35:35.336804 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-gjs8m"] Nov 24 13:35:35 crc kubenswrapper[4790]: I1124 13:35:35.446283 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-gjs8m" event={"ID":"fddd7928-5ec3-4605-94fb-a663c7f410e3","Type":"ContainerStarted","Data":"af844a721b94b278c61acc496d4f4abd074b70acf255cb8a3ea9be743d48c49f"} Nov 24 13:35:35 crc kubenswrapper[4790]: I1124 13:35:35.457220 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-f6lwg"] Nov 24 13:35:35 crc kubenswrapper[4790]: I1124 13:35:35.458767 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-f6lwg" Nov 24 13:35:35 crc kubenswrapper[4790]: I1124 13:35:35.463215 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 24 13:35:35 crc kubenswrapper[4790]: I1124 13:35:35.463456 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 13:35:35 crc kubenswrapper[4790]: I1124 13:35:35.467446 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-f6lwg"] Nov 24 13:35:35 crc kubenswrapper[4790]: I1124 13:35:35.496186 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkglg\" (UniqueName: \"kubernetes.io/projected/6803c7f5-eace-4b73-8ed7-39609fc9334a-kube-api-access-pkglg\") pod \"nova-cell1-conductor-db-sync-f6lwg\" (UID: \"6803c7f5-eace-4b73-8ed7-39609fc9334a\") " pod="openstack/nova-cell1-conductor-db-sync-f6lwg" Nov 24 13:35:35 crc kubenswrapper[4790]: I1124 13:35:35.496228 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6803c7f5-eace-4b73-8ed7-39609fc9334a-config-data\") pod \"nova-cell1-conductor-db-sync-f6lwg\" (UID: \"6803c7f5-eace-4b73-8ed7-39609fc9334a\") " pod="openstack/nova-cell1-conductor-db-sync-f6lwg" Nov 24 13:35:35 crc kubenswrapper[4790]: I1124 13:35:35.496276 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6803c7f5-eace-4b73-8ed7-39609fc9334a-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-f6lwg\" (UID: \"6803c7f5-eace-4b73-8ed7-39609fc9334a\") " pod="openstack/nova-cell1-conductor-db-sync-f6lwg" Nov 24 13:35:35 crc kubenswrapper[4790]: I1124 13:35:35.496369 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6803c7f5-eace-4b73-8ed7-39609fc9334a-scripts\") pod \"nova-cell1-conductor-db-sync-f6lwg\" (UID: \"6803c7f5-eace-4b73-8ed7-39609fc9334a\") " pod="openstack/nova-cell1-conductor-db-sync-f6lwg" Nov 24 13:35:35 crc kubenswrapper[4790]: I1124 13:35:35.594111 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:35:35 crc kubenswrapper[4790]: I1124 13:35:35.598311 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6803c7f5-eace-4b73-8ed7-39609fc9334a-scripts\") pod \"nova-cell1-conductor-db-sync-f6lwg\" (UID: \"6803c7f5-eace-4b73-8ed7-39609fc9334a\") " pod="openstack/nova-cell1-conductor-db-sync-f6lwg" Nov 24 13:35:35 crc kubenswrapper[4790]: I1124 13:35:35.598379 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkglg\" (UniqueName: \"kubernetes.io/projected/6803c7f5-eace-4b73-8ed7-39609fc9334a-kube-api-access-pkglg\") pod \"nova-cell1-conductor-db-sync-f6lwg\" (UID: \"6803c7f5-eace-4b73-8ed7-39609fc9334a\") " pod="openstack/nova-cell1-conductor-db-sync-f6lwg" Nov 24 13:35:35 crc kubenswrapper[4790]: I1124 13:35:35.600563 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6803c7f5-eace-4b73-8ed7-39609fc9334a-config-data\") pod \"nova-cell1-conductor-db-sync-f6lwg\" (UID: \"6803c7f5-eace-4b73-8ed7-39609fc9334a\") " pod="openstack/nova-cell1-conductor-db-sync-f6lwg" Nov 24 13:35:35 crc kubenswrapper[4790]: I1124 13:35:35.600620 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6803c7f5-eace-4b73-8ed7-39609fc9334a-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-f6lwg\" (UID: \"6803c7f5-eace-4b73-8ed7-39609fc9334a\") " pod="openstack/nova-cell1-conductor-db-sync-f6lwg" Nov 24 13:35:35 crc kubenswrapper[4790]: I1124 13:35:35.608812 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6803c7f5-eace-4b73-8ed7-39609fc9334a-scripts\") pod \"nova-cell1-conductor-db-sync-f6lwg\" (UID: \"6803c7f5-eace-4b73-8ed7-39609fc9334a\") " pod="openstack/nova-cell1-conductor-db-sync-f6lwg" Nov 24 13:35:35 crc kubenswrapper[4790]: I1124 13:35:35.609232 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6803c7f5-eace-4b73-8ed7-39609fc9334a-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-f6lwg\" (UID: \"6803c7f5-eace-4b73-8ed7-39609fc9334a\") " pod="openstack/nova-cell1-conductor-db-sync-f6lwg" Nov 24 13:35:35 crc kubenswrapper[4790]: I1124 13:35:35.614963 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6803c7f5-eace-4b73-8ed7-39609fc9334a-config-data\") pod \"nova-cell1-conductor-db-sync-f6lwg\" (UID: \"6803c7f5-eace-4b73-8ed7-39609fc9334a\") " pod="openstack/nova-cell1-conductor-db-sync-f6lwg" Nov 24 13:35:35 crc kubenswrapper[4790]: I1124 13:35:35.619778 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkglg\" (UniqueName: \"kubernetes.io/projected/6803c7f5-eace-4b73-8ed7-39609fc9334a-kube-api-access-pkglg\") pod \"nova-cell1-conductor-db-sync-f6lwg\" (UID: \"6803c7f5-eace-4b73-8ed7-39609fc9334a\") " pod="openstack/nova-cell1-conductor-db-sync-f6lwg" Nov 24 13:35:35 crc kubenswrapper[4790]: I1124 13:35:35.675479 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 13:35:35 crc kubenswrapper[4790]: I1124 13:35:35.768774 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:35:35 crc kubenswrapper[4790]: W1124 13:35:35.777331 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd9e4394_01fa_485a_9e85_05b2178445e8.slice/crio-881171b3621bf54b14036d2b0347475c6c3360777aa49b39971c14ff9178cbb0 WatchSource:0}: Error finding container 881171b3621bf54b14036d2b0347475c6c3360777aa49b39971c14ff9178cbb0: Status 404 returned error can't find the container with id 881171b3621bf54b14036d2b0347475c6c3360777aa49b39971c14ff9178cbb0 Nov 24 13:35:35 crc kubenswrapper[4790]: I1124 13:35:35.782974 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 13:35:35 crc kubenswrapper[4790]: I1124 13:35:35.875743 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-f6lwg" Nov 24 13:35:35 crc kubenswrapper[4790]: I1124 13:35:35.980127 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-skvpz"] Nov 24 13:35:35 crc kubenswrapper[4790]: W1124 13:35:35.993072 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b922e9a_ffc7_43ed_ad5c_d5d7f2ac3412.slice/crio-0d68d75d9bcf475988f9548ceb92947f84704648eeb7fb199bb0afbcefa283a3 WatchSource:0}: Error finding container 0d68d75d9bcf475988f9548ceb92947f84704648eeb7fb199bb0afbcefa283a3: Status 404 returned error can't find the container with id 0d68d75d9bcf475988f9548ceb92947f84704648eeb7fb199bb0afbcefa283a3 Nov 24 13:35:36 crc kubenswrapper[4790]: I1124 13:35:36.343691 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-f6lwg"] Nov 24 13:35:36 crc kubenswrapper[4790]: W1124 13:35:36.350830 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6803c7f5_eace_4b73_8ed7_39609fc9334a.slice/crio-a35414d917640009d30748ea0c5f39c81f4319468ae27a6fd2dab106f7c071b1 WatchSource:0}: Error finding container a35414d917640009d30748ea0c5f39c81f4319468ae27a6fd2dab106f7c071b1: Status 404 returned error can't find the container with id a35414d917640009d30748ea0c5f39c81f4319468ae27a6fd2dab106f7c071b1 Nov 24 13:35:36 crc kubenswrapper[4790]: I1124 13:35:36.527643 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0","Type":"ContainerStarted","Data":"dab33147fa8de28ae5aa9c9af8db4f22d54ae18131601532ffcfb4240b37cbca"} Nov 24 13:35:36 crc kubenswrapper[4790]: I1124 13:35:36.529804 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-gjs8m" event={"ID":"fddd7928-5ec3-4605-94fb-a663c7f410e3","Type":"ContainerStarted","Data":"73738b88616d24d9d38a7dde3fbe8536abf120812c9d699ed3f1dae25824b738"} Nov 24 13:35:36 crc kubenswrapper[4790]: I1124 13:35:36.530664 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-f6lwg" event={"ID":"6803c7f5-eace-4b73-8ed7-39609fc9334a","Type":"ContainerStarted","Data":"a35414d917640009d30748ea0c5f39c81f4319468ae27a6fd2dab106f7c071b1"} Nov 24 13:35:36 crc kubenswrapper[4790]: I1124 13:35:36.532149 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fd9e4394-01fa-485a-9e85-05b2178445e8","Type":"ContainerStarted","Data":"881171b3621bf54b14036d2b0347475c6c3360777aa49b39971c14ff9178cbb0"} Nov 24 13:35:36 crc kubenswrapper[4790]: I1124 13:35:36.535234 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7e988296-b952-43b2-bad8-39c2916bcecd","Type":"ContainerStarted","Data":"f22bd6efe39521b9e1992013ea3771042237c9a20af4ad476a5e69d3c0a5f2e2"} Nov 24 13:35:36 crc kubenswrapper[4790]: I1124 13:35:36.536962 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"38092ab0-6779-4823-b8f0-7955f7e436ee","Type":"ContainerStarted","Data":"b58d2d3fc049387b87e266c5ded97ab9a7fa13307b0bcd9211d0b896111dce3d"} Nov 24 13:35:36 crc kubenswrapper[4790]: I1124 13:35:36.540757 4790 generic.go:334] "Generic (PLEG): container finished" podID="6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412" containerID="1239012cbfa2d3cd8a5ff484e5663a42f10626d77202e5759db80d62f1dc224f" exitCode=0 Nov 24 13:35:36 crc kubenswrapper[4790]: I1124 13:35:36.540811 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-skvpz" event={"ID":"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412","Type":"ContainerDied","Data":"1239012cbfa2d3cd8a5ff484e5663a42f10626d77202e5759db80d62f1dc224f"} Nov 24 13:35:36 crc kubenswrapper[4790]: I1124 13:35:36.540836 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-skvpz" event={"ID":"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412","Type":"ContainerStarted","Data":"0d68d75d9bcf475988f9548ceb92947f84704648eeb7fb199bb0afbcefa283a3"} Nov 24 13:35:36 crc kubenswrapper[4790]: I1124 13:35:36.551271 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-gjs8m" podStartSLOduration=2.551255451 podStartE2EDuration="2.551255451s" podCreationTimestamp="2025-11-24 13:35:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:35:36.54456561 +0000 UTC m=+1384.924459292" watchObservedRunningTime="2025-11-24 13:35:36.551255451 +0000 UTC m=+1384.931149113" Nov 24 13:35:37 crc kubenswrapper[4790]: I1124 13:35:37.561494 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-f6lwg" event={"ID":"6803c7f5-eace-4b73-8ed7-39609fc9334a","Type":"ContainerStarted","Data":"13ff2a6bf7913ecdacaf058d5f469f70a2b6c6e055b55607d3a365e30aaf58d9"} Nov 24 13:35:37 crc kubenswrapper[4790]: I1124 13:35:37.567221 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-skvpz" event={"ID":"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412","Type":"ContainerStarted","Data":"56bf93c28a64fc47955b7be4f3cb20d93bfc8f764601c7c707b0215042bde17d"} Nov 24 13:35:37 crc kubenswrapper[4790]: I1124 13:35:37.579761 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-f6lwg" podStartSLOduration=2.579744459 podStartE2EDuration="2.579744459s" podCreationTimestamp="2025-11-24 13:35:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:35:37.57732253 +0000 UTC m=+1385.957216192" watchObservedRunningTime="2025-11-24 13:35:37.579744459 +0000 UTC m=+1385.959638121" Nov 24 13:35:37 crc kubenswrapper[4790]: I1124 13:35:37.599698 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bccf8f775-skvpz" podStartSLOduration=3.599678699 podStartE2EDuration="3.599678699s" podCreationTimestamp="2025-11-24 13:35:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:35:37.595655254 +0000 UTC m=+1385.975548926" watchObservedRunningTime="2025-11-24 13:35:37.599678699 +0000 UTC m=+1385.979572351" Nov 24 13:35:38 crc kubenswrapper[4790]: I1124 13:35:38.576856 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bccf8f775-skvpz" Nov 24 13:35:38 crc kubenswrapper[4790]: I1124 13:35:38.673420 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 13:35:38 crc kubenswrapper[4790]: I1124 13:35:38.684833 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:35:39 crc kubenswrapper[4790]: I1124 13:35:39.589609 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fd9e4394-01fa-485a-9e85-05b2178445e8","Type":"ContainerStarted","Data":"a0ecd98ed58aec4504de27538119d64f6110d3754aa6003370bad164ec160264"} Nov 24 13:35:39 crc kubenswrapper[4790]: I1124 13:35:39.589971 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fd9e4394-01fa-485a-9e85-05b2178445e8","Type":"ContainerStarted","Data":"ac41cac3e8aa088d4a24a51d2af71fb2a17b3b8ba1cdb4e868128f884132244e"} Nov 24 13:35:39 crc kubenswrapper[4790]: I1124 13:35:39.592417 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7e988296-b952-43b2-bad8-39c2916bcecd","Type":"ContainerStarted","Data":"33636b9e3d7522fa5b65628a38a9b55c3049d7876c4ab7273d5e0270bede393e"} Nov 24 13:35:39 crc kubenswrapper[4790]: I1124 13:35:39.595125 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"38092ab0-6779-4823-b8f0-7955f7e436ee","Type":"ContainerStarted","Data":"481b593fdcac4a22b729f6fff7271f2fcb1469af401697bae50cee0a2c51287e"} Nov 24 13:35:39 crc kubenswrapper[4790]: I1124 13:35:39.595155 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"38092ab0-6779-4823-b8f0-7955f7e436ee","Type":"ContainerStarted","Data":"d1461458ef51b1e1cb71bb529472da13f5e7c8f13add50fdf893bebde5f346d1"} Nov 24 13:35:39 crc kubenswrapper[4790]: I1124 13:35:39.595186 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="38092ab0-6779-4823-b8f0-7955f7e436ee" containerName="nova-metadata-log" containerID="cri-o://d1461458ef51b1e1cb71bb529472da13f5e7c8f13add50fdf893bebde5f346d1" gracePeriod=30 Nov 24 13:35:39 crc kubenswrapper[4790]: I1124 13:35:39.595213 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="38092ab0-6779-4823-b8f0-7955f7e436ee" containerName="nova-metadata-metadata" containerID="cri-o://481b593fdcac4a22b729f6fff7271f2fcb1469af401697bae50cee0a2c51287e" gracePeriod=30 Nov 24 13:35:39 crc kubenswrapper[4790]: I1124 13:35:39.597562 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0","Type":"ContainerStarted","Data":"aba5dac078aa2c293348629db3cab9f01c0d490999bb82d7e064e7ea72ffe77c"} Nov 24 13:35:39 crc kubenswrapper[4790]: I1124 13:35:39.597765 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://aba5dac078aa2c293348629db3cab9f01c0d490999bb82d7e064e7ea72ffe77c" gracePeriod=30 Nov 24 13:35:39 crc kubenswrapper[4790]: I1124 13:35:39.619229 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.090455008 podStartE2EDuration="5.619208883s" podCreationTimestamp="2025-11-24 13:35:34 +0000 UTC" firstStartedPulling="2025-11-24 13:35:35.778993605 +0000 UTC m=+1384.158887267" lastFinishedPulling="2025-11-24 13:35:38.30774748 +0000 UTC m=+1386.687641142" observedRunningTime="2025-11-24 13:35:39.607390165 +0000 UTC m=+1387.987283827" watchObservedRunningTime="2025-11-24 13:35:39.619208883 +0000 UTC m=+1387.999102545" Nov 24 13:35:39 crc kubenswrapper[4790]: I1124 13:35:39.629325 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.9198354220000002 podStartE2EDuration="5.629301491s" podCreationTimestamp="2025-11-24 13:35:34 +0000 UTC" firstStartedPulling="2025-11-24 13:35:35.596597143 +0000 UTC m=+1383.976490805" lastFinishedPulling="2025-11-24 13:35:38.306063212 +0000 UTC m=+1386.685956874" observedRunningTime="2025-11-24 13:35:39.625584965 +0000 UTC m=+1388.005478637" watchObservedRunningTime="2025-11-24 13:35:39.629301491 +0000 UTC m=+1388.009195153" Nov 24 13:35:39 crc kubenswrapper[4790]: I1124 13:35:39.646258 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.134795994 podStartE2EDuration="5.646238375s" podCreationTimestamp="2025-11-24 13:35:34 +0000 UTC" firstStartedPulling="2025-11-24 13:35:35.795741303 +0000 UTC m=+1384.175634965" lastFinishedPulling="2025-11-24 13:35:38.307183684 +0000 UTC m=+1386.687077346" observedRunningTime="2025-11-24 13:35:39.641066348 +0000 UTC m=+1388.020960000" watchObservedRunningTime="2025-11-24 13:35:39.646238375 +0000 UTC m=+1388.026132037" Nov 24 13:35:39 crc kubenswrapper[4790]: I1124 13:35:39.663597 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.032709418 podStartE2EDuration="5.663581161s" podCreationTimestamp="2025-11-24 13:35:34 +0000 UTC" firstStartedPulling="2025-11-24 13:35:35.68048742 +0000 UTC m=+1384.060381092" lastFinishedPulling="2025-11-24 13:35:38.311359173 +0000 UTC m=+1386.691252835" observedRunningTime="2025-11-24 13:35:39.659550866 +0000 UTC m=+1388.039444538" watchObservedRunningTime="2025-11-24 13:35:39.663581161 +0000 UTC m=+1388.043474823" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.070634 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.071086 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.107269 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.162472 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.189119 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcwk5\" (UniqueName: \"kubernetes.io/projected/38092ab0-6779-4823-b8f0-7955f7e436ee-kube-api-access-gcwk5\") pod \"38092ab0-6779-4823-b8f0-7955f7e436ee\" (UID: \"38092ab0-6779-4823-b8f0-7955f7e436ee\") " Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.189217 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38092ab0-6779-4823-b8f0-7955f7e436ee-logs\") pod \"38092ab0-6779-4823-b8f0-7955f7e436ee\" (UID: \"38092ab0-6779-4823-b8f0-7955f7e436ee\") " Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.189343 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38092ab0-6779-4823-b8f0-7955f7e436ee-config-data\") pod \"38092ab0-6779-4823-b8f0-7955f7e436ee\" (UID: \"38092ab0-6779-4823-b8f0-7955f7e436ee\") " Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.189474 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38092ab0-6779-4823-b8f0-7955f7e436ee-combined-ca-bundle\") pod \"38092ab0-6779-4823-b8f0-7955f7e436ee\" (UID: \"38092ab0-6779-4823-b8f0-7955f7e436ee\") " Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.189806 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38092ab0-6779-4823-b8f0-7955f7e436ee-logs" (OuterVolumeSpecName: "logs") pod "38092ab0-6779-4823-b8f0-7955f7e436ee" (UID: "38092ab0-6779-4823-b8f0-7955f7e436ee"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.190744 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38092ab0-6779-4823-b8f0-7955f7e436ee-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.197347 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38092ab0-6779-4823-b8f0-7955f7e436ee-kube-api-access-gcwk5" (OuterVolumeSpecName: "kube-api-access-gcwk5") pod "38092ab0-6779-4823-b8f0-7955f7e436ee" (UID: "38092ab0-6779-4823-b8f0-7955f7e436ee"). InnerVolumeSpecName "kube-api-access-gcwk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.211339 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.225336 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38092ab0-6779-4823-b8f0-7955f7e436ee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "38092ab0-6779-4823-b8f0-7955f7e436ee" (UID: "38092ab0-6779-4823-b8f0-7955f7e436ee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.225919 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38092ab0-6779-4823-b8f0-7955f7e436ee-config-data" (OuterVolumeSpecName: "config-data") pod "38092ab0-6779-4823-b8f0-7955f7e436ee" (UID: "38092ab0-6779-4823-b8f0-7955f7e436ee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.292753 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38092ab0-6779-4823-b8f0-7955f7e436ee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.292794 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcwk5\" (UniqueName: \"kubernetes.io/projected/38092ab0-6779-4823-b8f0-7955f7e436ee-kube-api-access-gcwk5\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.292810 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38092ab0-6779-4823-b8f0-7955f7e436ee-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.608623 4790 generic.go:334] "Generic (PLEG): container finished" podID="38092ab0-6779-4823-b8f0-7955f7e436ee" containerID="481b593fdcac4a22b729f6fff7271f2fcb1469af401697bae50cee0a2c51287e" exitCode=0 Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.608659 4790 generic.go:334] "Generic (PLEG): container finished" podID="38092ab0-6779-4823-b8f0-7955f7e436ee" containerID="d1461458ef51b1e1cb71bb529472da13f5e7c8f13add50fdf893bebde5f346d1" exitCode=143 Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.608673 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"38092ab0-6779-4823-b8f0-7955f7e436ee","Type":"ContainerDied","Data":"481b593fdcac4a22b729f6fff7271f2fcb1469af401697bae50cee0a2c51287e"} Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.608715 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.608746 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"38092ab0-6779-4823-b8f0-7955f7e436ee","Type":"ContainerDied","Data":"d1461458ef51b1e1cb71bb529472da13f5e7c8f13add50fdf893bebde5f346d1"} Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.608759 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"38092ab0-6779-4823-b8f0-7955f7e436ee","Type":"ContainerDied","Data":"b58d2d3fc049387b87e266c5ded97ab9a7fa13307b0bcd9211d0b896111dce3d"} Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.608785 4790 scope.go:117] "RemoveContainer" containerID="481b593fdcac4a22b729f6fff7271f2fcb1469af401697bae50cee0a2c51287e" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.631277 4790 scope.go:117] "RemoveContainer" containerID="d1461458ef51b1e1cb71bb529472da13f5e7c8f13add50fdf893bebde5f346d1" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.647701 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.662090 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.665225 4790 scope.go:117] "RemoveContainer" containerID="481b593fdcac4a22b729f6fff7271f2fcb1469af401697bae50cee0a2c51287e" Nov 24 13:35:40 crc kubenswrapper[4790]: E1124 13:35:40.665630 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"481b593fdcac4a22b729f6fff7271f2fcb1469af401697bae50cee0a2c51287e\": container with ID starting with 481b593fdcac4a22b729f6fff7271f2fcb1469af401697bae50cee0a2c51287e not found: ID does not exist" containerID="481b593fdcac4a22b729f6fff7271f2fcb1469af401697bae50cee0a2c51287e" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.665668 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"481b593fdcac4a22b729f6fff7271f2fcb1469af401697bae50cee0a2c51287e"} err="failed to get container status \"481b593fdcac4a22b729f6fff7271f2fcb1469af401697bae50cee0a2c51287e\": rpc error: code = NotFound desc = could not find container \"481b593fdcac4a22b729f6fff7271f2fcb1469af401697bae50cee0a2c51287e\": container with ID starting with 481b593fdcac4a22b729f6fff7271f2fcb1469af401697bae50cee0a2c51287e not found: ID does not exist" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.665696 4790 scope.go:117] "RemoveContainer" containerID="d1461458ef51b1e1cb71bb529472da13f5e7c8f13add50fdf893bebde5f346d1" Nov 24 13:35:40 crc kubenswrapper[4790]: E1124 13:35:40.666464 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1461458ef51b1e1cb71bb529472da13f5e7c8f13add50fdf893bebde5f346d1\": container with ID starting with d1461458ef51b1e1cb71bb529472da13f5e7c8f13add50fdf893bebde5f346d1 not found: ID does not exist" containerID="d1461458ef51b1e1cb71bb529472da13f5e7c8f13add50fdf893bebde5f346d1" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.666500 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1461458ef51b1e1cb71bb529472da13f5e7c8f13add50fdf893bebde5f346d1"} err="failed to get container status \"d1461458ef51b1e1cb71bb529472da13f5e7c8f13add50fdf893bebde5f346d1\": rpc error: code = NotFound desc = could not find container \"d1461458ef51b1e1cb71bb529472da13f5e7c8f13add50fdf893bebde5f346d1\": container with ID starting with d1461458ef51b1e1cb71bb529472da13f5e7c8f13add50fdf893bebde5f346d1 not found: ID does not exist" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.666522 4790 scope.go:117] "RemoveContainer" containerID="481b593fdcac4a22b729f6fff7271f2fcb1469af401697bae50cee0a2c51287e" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.667371 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"481b593fdcac4a22b729f6fff7271f2fcb1469af401697bae50cee0a2c51287e"} err="failed to get container status \"481b593fdcac4a22b729f6fff7271f2fcb1469af401697bae50cee0a2c51287e\": rpc error: code = NotFound desc = could not find container \"481b593fdcac4a22b729f6fff7271f2fcb1469af401697bae50cee0a2c51287e\": container with ID starting with 481b593fdcac4a22b729f6fff7271f2fcb1469af401697bae50cee0a2c51287e not found: ID does not exist" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.667399 4790 scope.go:117] "RemoveContainer" containerID="d1461458ef51b1e1cb71bb529472da13f5e7c8f13add50fdf893bebde5f346d1" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.668415 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1461458ef51b1e1cb71bb529472da13f5e7c8f13add50fdf893bebde5f346d1"} err="failed to get container status \"d1461458ef51b1e1cb71bb529472da13f5e7c8f13add50fdf893bebde5f346d1\": rpc error: code = NotFound desc = could not find container \"d1461458ef51b1e1cb71bb529472da13f5e7c8f13add50fdf893bebde5f346d1\": container with ID starting with d1461458ef51b1e1cb71bb529472da13f5e7c8f13add50fdf893bebde5f346d1 not found: ID does not exist" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.676370 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:35:40 crc kubenswrapper[4790]: E1124 13:35:40.676737 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38092ab0-6779-4823-b8f0-7955f7e436ee" containerName="nova-metadata-metadata" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.676749 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="38092ab0-6779-4823-b8f0-7955f7e436ee" containerName="nova-metadata-metadata" Nov 24 13:35:40 crc kubenswrapper[4790]: E1124 13:35:40.676772 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38092ab0-6779-4823-b8f0-7955f7e436ee" containerName="nova-metadata-log" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.676781 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="38092ab0-6779-4823-b8f0-7955f7e436ee" containerName="nova-metadata-log" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.676989 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="38092ab0-6779-4823-b8f0-7955f7e436ee" containerName="nova-metadata-metadata" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.677017 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="38092ab0-6779-4823-b8f0-7955f7e436ee" containerName="nova-metadata-log" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.677995 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.692421 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.692512 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.692647 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.804173 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tngwk\" (UniqueName: \"kubernetes.io/projected/adc9b1fa-6906-429f-8238-32798c6f563a-kube-api-access-tngwk\") pod \"nova-metadata-0\" (UID: \"adc9b1fa-6906-429f-8238-32798c6f563a\") " pod="openstack/nova-metadata-0" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.804345 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/adc9b1fa-6906-429f-8238-32798c6f563a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"adc9b1fa-6906-429f-8238-32798c6f563a\") " pod="openstack/nova-metadata-0" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.804471 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adc9b1fa-6906-429f-8238-32798c6f563a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"adc9b1fa-6906-429f-8238-32798c6f563a\") " pod="openstack/nova-metadata-0" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.804522 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adc9b1fa-6906-429f-8238-32798c6f563a-logs\") pod \"nova-metadata-0\" (UID: \"adc9b1fa-6906-429f-8238-32798c6f563a\") " pod="openstack/nova-metadata-0" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.804590 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adc9b1fa-6906-429f-8238-32798c6f563a-config-data\") pod \"nova-metadata-0\" (UID: \"adc9b1fa-6906-429f-8238-32798c6f563a\") " pod="openstack/nova-metadata-0" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.906536 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adc9b1fa-6906-429f-8238-32798c6f563a-logs\") pod \"nova-metadata-0\" (UID: \"adc9b1fa-6906-429f-8238-32798c6f563a\") " pod="openstack/nova-metadata-0" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.906612 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adc9b1fa-6906-429f-8238-32798c6f563a-config-data\") pod \"nova-metadata-0\" (UID: \"adc9b1fa-6906-429f-8238-32798c6f563a\") " pod="openstack/nova-metadata-0" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.906735 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tngwk\" (UniqueName: \"kubernetes.io/projected/adc9b1fa-6906-429f-8238-32798c6f563a-kube-api-access-tngwk\") pod \"nova-metadata-0\" (UID: \"adc9b1fa-6906-429f-8238-32798c6f563a\") " pod="openstack/nova-metadata-0" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.906807 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/adc9b1fa-6906-429f-8238-32798c6f563a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"adc9b1fa-6906-429f-8238-32798c6f563a\") " pod="openstack/nova-metadata-0" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.906917 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adc9b1fa-6906-429f-8238-32798c6f563a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"adc9b1fa-6906-429f-8238-32798c6f563a\") " pod="openstack/nova-metadata-0" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.907170 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adc9b1fa-6906-429f-8238-32798c6f563a-logs\") pod \"nova-metadata-0\" (UID: \"adc9b1fa-6906-429f-8238-32798c6f563a\") " pod="openstack/nova-metadata-0" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.910996 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adc9b1fa-6906-429f-8238-32798c6f563a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"adc9b1fa-6906-429f-8238-32798c6f563a\") " pod="openstack/nova-metadata-0" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.918439 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/adc9b1fa-6906-429f-8238-32798c6f563a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"adc9b1fa-6906-429f-8238-32798c6f563a\") " pod="openstack/nova-metadata-0" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.922214 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adc9b1fa-6906-429f-8238-32798c6f563a-config-data\") pod \"nova-metadata-0\" (UID: \"adc9b1fa-6906-429f-8238-32798c6f563a\") " pod="openstack/nova-metadata-0" Nov 24 13:35:40 crc kubenswrapper[4790]: I1124 13:35:40.926569 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tngwk\" (UniqueName: \"kubernetes.io/projected/adc9b1fa-6906-429f-8238-32798c6f563a-kube-api-access-tngwk\") pod \"nova-metadata-0\" (UID: \"adc9b1fa-6906-429f-8238-32798c6f563a\") " pod="openstack/nova-metadata-0" Nov 24 13:35:41 crc kubenswrapper[4790]: I1124 13:35:41.011843 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 13:35:41 crc kubenswrapper[4790]: W1124 13:35:41.490230 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podadc9b1fa_6906_429f_8238_32798c6f563a.slice/crio-8f950e5cbd0e7faca115a9a2f61ec2a33ed9d1e104a01467e44b6523a7c4368a WatchSource:0}: Error finding container 8f950e5cbd0e7faca115a9a2f61ec2a33ed9d1e104a01467e44b6523a7c4368a: Status 404 returned error can't find the container with id 8f950e5cbd0e7faca115a9a2f61ec2a33ed9d1e104a01467e44b6523a7c4368a Nov 24 13:35:41 crc kubenswrapper[4790]: I1124 13:35:41.505485 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:35:41 crc kubenswrapper[4790]: I1124 13:35:41.617968 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"adc9b1fa-6906-429f-8238-32798c6f563a","Type":"ContainerStarted","Data":"8f950e5cbd0e7faca115a9a2f61ec2a33ed9d1e104a01467e44b6523a7c4368a"} Nov 24 13:35:42 crc kubenswrapper[4790]: I1124 13:35:42.336116 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38092ab0-6779-4823-b8f0-7955f7e436ee" path="/var/lib/kubelet/pods/38092ab0-6779-4823-b8f0-7955f7e436ee/volumes" Nov 24 13:35:42 crc kubenswrapper[4790]: I1124 13:35:42.630926 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"adc9b1fa-6906-429f-8238-32798c6f563a","Type":"ContainerStarted","Data":"247fde78a48a936a512702ebf27b9f4fbd05111a078a61ed16aa6dfb332af6ce"} Nov 24 13:35:42 crc kubenswrapper[4790]: I1124 13:35:42.630978 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"adc9b1fa-6906-429f-8238-32798c6f563a","Type":"ContainerStarted","Data":"841f6a28de463438bab63ae94767350083795ba72c0122aa68ad0bbb8d5fea31"} Nov 24 13:35:42 crc kubenswrapper[4790]: I1124 13:35:42.658038 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 13:35:42 crc kubenswrapper[4790]: I1124 13:35:42.664841 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.664818036 podStartE2EDuration="2.664818036s" podCreationTimestamp="2025-11-24 13:35:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:35:42.655772928 +0000 UTC m=+1391.035666610" watchObservedRunningTime="2025-11-24 13:35:42.664818036 +0000 UTC m=+1391.044711718" Nov 24 13:35:43 crc kubenswrapper[4790]: I1124 13:35:43.640195 4790 generic.go:334] "Generic (PLEG): container finished" podID="fddd7928-5ec3-4605-94fb-a663c7f410e3" containerID="73738b88616d24d9d38a7dde3fbe8536abf120812c9d699ed3f1dae25824b738" exitCode=0 Nov 24 13:35:43 crc kubenswrapper[4790]: I1124 13:35:43.640285 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-gjs8m" event={"ID":"fddd7928-5ec3-4605-94fb-a663c7f410e3","Type":"ContainerDied","Data":"73738b88616d24d9d38a7dde3fbe8536abf120812c9d699ed3f1dae25824b738"} Nov 24 13:35:44 crc kubenswrapper[4790]: I1124 13:35:44.649866 4790 generic.go:334] "Generic (PLEG): container finished" podID="6803c7f5-eace-4b73-8ed7-39609fc9334a" containerID="13ff2a6bf7913ecdacaf058d5f469f70a2b6c6e055b55607d3a365e30aaf58d9" exitCode=0 Nov 24 13:35:44 crc kubenswrapper[4790]: I1124 13:35:44.649912 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-f6lwg" event={"ID":"6803c7f5-eace-4b73-8ed7-39609fc9334a","Type":"ContainerDied","Data":"13ff2a6bf7913ecdacaf058d5f469f70a2b6c6e055b55607d3a365e30aaf58d9"} Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.065974 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-gjs8m" Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.181762 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.181831 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.189770 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkj8h\" (UniqueName: \"kubernetes.io/projected/fddd7928-5ec3-4605-94fb-a663c7f410e3-kube-api-access-xkj8h\") pod \"fddd7928-5ec3-4605-94fb-a663c7f410e3\" (UID: \"fddd7928-5ec3-4605-94fb-a663c7f410e3\") " Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.189817 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fddd7928-5ec3-4605-94fb-a663c7f410e3-combined-ca-bundle\") pod \"fddd7928-5ec3-4605-94fb-a663c7f410e3\" (UID: \"fddd7928-5ec3-4605-94fb-a663c7f410e3\") " Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.190003 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fddd7928-5ec3-4605-94fb-a663c7f410e3-scripts\") pod \"fddd7928-5ec3-4605-94fb-a663c7f410e3\" (UID: \"fddd7928-5ec3-4605-94fb-a663c7f410e3\") " Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.190078 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fddd7928-5ec3-4605-94fb-a663c7f410e3-config-data\") pod \"fddd7928-5ec3-4605-94fb-a663c7f410e3\" (UID: \"fddd7928-5ec3-4605-94fb-a663c7f410e3\") " Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.198015 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fddd7928-5ec3-4605-94fb-a663c7f410e3-scripts" (OuterVolumeSpecName: "scripts") pod "fddd7928-5ec3-4605-94fb-a663c7f410e3" (UID: "fddd7928-5ec3-4605-94fb-a663c7f410e3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.211838 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.215150 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fddd7928-5ec3-4605-94fb-a663c7f410e3-kube-api-access-xkj8h" (OuterVolumeSpecName: "kube-api-access-xkj8h") pod "fddd7928-5ec3-4605-94fb-a663c7f410e3" (UID: "fddd7928-5ec3-4605-94fb-a663c7f410e3"). InnerVolumeSpecName "kube-api-access-xkj8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.225760 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fddd7928-5ec3-4605-94fb-a663c7f410e3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fddd7928-5ec3-4605-94fb-a663c7f410e3" (UID: "fddd7928-5ec3-4605-94fb-a663c7f410e3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.229989 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fddd7928-5ec3-4605-94fb-a663c7f410e3-config-data" (OuterVolumeSpecName: "config-data") pod "fddd7928-5ec3-4605-94fb-a663c7f410e3" (UID: "fddd7928-5ec3-4605-94fb-a663c7f410e3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.230061 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bccf8f775-skvpz" Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.239696 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.303877 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkj8h\" (UniqueName: \"kubernetes.io/projected/fddd7928-5ec3-4605-94fb-a663c7f410e3-kube-api-access-xkj8h\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.303921 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fddd7928-5ec3-4605-94fb-a663c7f410e3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.303930 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fddd7928-5ec3-4605-94fb-a663c7f410e3-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.303939 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fddd7928-5ec3-4605-94fb-a663c7f410e3-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.306055 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-snppf"] Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.306338 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6578955fd5-snppf" podUID="aabbe361-2ef8-444e-acf2-fe48abecf27b" containerName="dnsmasq-dns" containerID="cri-o://6bee98f4451efc6d196589802f29d5f46905beb31c259a7f4c4ee20b26895450" gracePeriod=10 Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.664596 4790 generic.go:334] "Generic (PLEG): container finished" podID="aabbe361-2ef8-444e-acf2-fe48abecf27b" containerID="6bee98f4451efc6d196589802f29d5f46905beb31c259a7f4c4ee20b26895450" exitCode=0 Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.664949 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-snppf" event={"ID":"aabbe361-2ef8-444e-acf2-fe48abecf27b","Type":"ContainerDied","Data":"6bee98f4451efc6d196589802f29d5f46905beb31c259a7f4c4ee20b26895450"} Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.667161 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-gjs8m" Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.667199 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-gjs8m" event={"ID":"fddd7928-5ec3-4605-94fb-a663c7f410e3","Type":"ContainerDied","Data":"af844a721b94b278c61acc496d4f4abd074b70acf255cb8a3ea9be743d48c49f"} Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.667220 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af844a721b94b278c61acc496d4f4abd074b70acf255cb8a3ea9be743d48c49f" Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.717497 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.757023 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-snppf" Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.817496 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aabbe361-2ef8-444e-acf2-fe48abecf27b-dns-swift-storage-0\") pod \"aabbe361-2ef8-444e-acf2-fe48abecf27b\" (UID: \"aabbe361-2ef8-444e-acf2-fe48abecf27b\") " Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.817569 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-695lr\" (UniqueName: \"kubernetes.io/projected/aabbe361-2ef8-444e-acf2-fe48abecf27b-kube-api-access-695lr\") pod \"aabbe361-2ef8-444e-acf2-fe48abecf27b\" (UID: \"aabbe361-2ef8-444e-acf2-fe48abecf27b\") " Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.817598 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aabbe361-2ef8-444e-acf2-fe48abecf27b-dns-svc\") pod \"aabbe361-2ef8-444e-acf2-fe48abecf27b\" (UID: \"aabbe361-2ef8-444e-acf2-fe48abecf27b\") " Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.817622 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aabbe361-2ef8-444e-acf2-fe48abecf27b-ovsdbserver-nb\") pod \"aabbe361-2ef8-444e-acf2-fe48abecf27b\" (UID: \"aabbe361-2ef8-444e-acf2-fe48abecf27b\") " Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.817708 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aabbe361-2ef8-444e-acf2-fe48abecf27b-ovsdbserver-sb\") pod \"aabbe361-2ef8-444e-acf2-fe48abecf27b\" (UID: \"aabbe361-2ef8-444e-acf2-fe48abecf27b\") " Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.817807 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aabbe361-2ef8-444e-acf2-fe48abecf27b-config\") pod \"aabbe361-2ef8-444e-acf2-fe48abecf27b\" (UID: \"aabbe361-2ef8-444e-acf2-fe48abecf27b\") " Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.822486 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aabbe361-2ef8-444e-acf2-fe48abecf27b-kube-api-access-695lr" (OuterVolumeSpecName: "kube-api-access-695lr") pod "aabbe361-2ef8-444e-acf2-fe48abecf27b" (UID: "aabbe361-2ef8-444e-acf2-fe48abecf27b"). InnerVolumeSpecName "kube-api-access-695lr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.877403 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aabbe361-2ef8-444e-acf2-fe48abecf27b-config" (OuterVolumeSpecName: "config") pod "aabbe361-2ef8-444e-acf2-fe48abecf27b" (UID: "aabbe361-2ef8-444e-acf2-fe48abecf27b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.908276 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aabbe361-2ef8-444e-acf2-fe48abecf27b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "aabbe361-2ef8-444e-acf2-fe48abecf27b" (UID: "aabbe361-2ef8-444e-acf2-fe48abecf27b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.920774 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-695lr\" (UniqueName: \"kubernetes.io/projected/aabbe361-2ef8-444e-acf2-fe48abecf27b-kube-api-access-695lr\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.920807 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aabbe361-2ef8-444e-acf2-fe48abecf27b-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.920823 4790 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aabbe361-2ef8-444e-acf2-fe48abecf27b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.925529 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aabbe361-2ef8-444e-acf2-fe48abecf27b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "aabbe361-2ef8-444e-acf2-fe48abecf27b" (UID: "aabbe361-2ef8-444e-acf2-fe48abecf27b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.970666 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aabbe361-2ef8-444e-acf2-fe48abecf27b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "aabbe361-2ef8-444e-acf2-fe48abecf27b" (UID: "aabbe361-2ef8-444e-acf2-fe48abecf27b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.971320 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aabbe361-2ef8-444e-acf2-fe48abecf27b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "aabbe361-2ef8-444e-acf2-fe48abecf27b" (UID: "aabbe361-2ef8-444e-acf2-fe48abecf27b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.978388 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.978600 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="fd9e4394-01fa-485a-9e85-05b2178445e8" containerName="nova-api-log" containerID="cri-o://ac41cac3e8aa088d4a24a51d2af71fb2a17b3b8ba1cdb4e868128f884132244e" gracePeriod=30 Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.979056 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="fd9e4394-01fa-485a-9e85-05b2178445e8" containerName="nova-api-api" containerID="cri-o://a0ecd98ed58aec4504de27538119d64f6110d3754aa6003370bad164ec160264" gracePeriod=30 Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.985405 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="fd9e4394-01fa-485a-9e85-05b2178445e8" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": EOF" Nov 24 13:35:45 crc kubenswrapper[4790]: I1124 13:35:45.985473 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="fd9e4394-01fa-485a-9e85-05b2178445e8" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": EOF" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.012991 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.013061 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.024102 4790 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aabbe361-2ef8-444e-acf2-fe48abecf27b-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.024134 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aabbe361-2ef8-444e-acf2-fe48abecf27b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.024147 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aabbe361-2ef8-444e-acf2-fe48abecf27b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.052173 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.140934 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-f6lwg" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.226934 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6803c7f5-eace-4b73-8ed7-39609fc9334a-config-data\") pod \"6803c7f5-eace-4b73-8ed7-39609fc9334a\" (UID: \"6803c7f5-eace-4b73-8ed7-39609fc9334a\") " Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.227173 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkglg\" (UniqueName: \"kubernetes.io/projected/6803c7f5-eace-4b73-8ed7-39609fc9334a-kube-api-access-pkglg\") pod \"6803c7f5-eace-4b73-8ed7-39609fc9334a\" (UID: \"6803c7f5-eace-4b73-8ed7-39609fc9334a\") " Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.227220 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6803c7f5-eace-4b73-8ed7-39609fc9334a-combined-ca-bundle\") pod \"6803c7f5-eace-4b73-8ed7-39609fc9334a\" (UID: \"6803c7f5-eace-4b73-8ed7-39609fc9334a\") " Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.227250 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6803c7f5-eace-4b73-8ed7-39609fc9334a-scripts\") pod \"6803c7f5-eace-4b73-8ed7-39609fc9334a\" (UID: \"6803c7f5-eace-4b73-8ed7-39609fc9334a\") " Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.232854 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6803c7f5-eace-4b73-8ed7-39609fc9334a-kube-api-access-pkglg" (OuterVolumeSpecName: "kube-api-access-pkglg") pod "6803c7f5-eace-4b73-8ed7-39609fc9334a" (UID: "6803c7f5-eace-4b73-8ed7-39609fc9334a"). InnerVolumeSpecName "kube-api-access-pkglg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.233847 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6803c7f5-eace-4b73-8ed7-39609fc9334a-scripts" (OuterVolumeSpecName: "scripts") pod "6803c7f5-eace-4b73-8ed7-39609fc9334a" (UID: "6803c7f5-eace-4b73-8ed7-39609fc9334a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.268106 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6803c7f5-eace-4b73-8ed7-39609fc9334a-config-data" (OuterVolumeSpecName: "config-data") pod "6803c7f5-eace-4b73-8ed7-39609fc9334a" (UID: "6803c7f5-eace-4b73-8ed7-39609fc9334a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.274403 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.297063 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6803c7f5-eace-4b73-8ed7-39609fc9334a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6803c7f5-eace-4b73-8ed7-39609fc9334a" (UID: "6803c7f5-eace-4b73-8ed7-39609fc9334a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.330005 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkglg\" (UniqueName: \"kubernetes.io/projected/6803c7f5-eace-4b73-8ed7-39609fc9334a-kube-api-access-pkglg\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.330481 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6803c7f5-eace-4b73-8ed7-39609fc9334a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.330494 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6803c7f5-eace-4b73-8ed7-39609fc9334a-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.330504 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6803c7f5-eace-4b73-8ed7-39609fc9334a-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.677156 4790 generic.go:334] "Generic (PLEG): container finished" podID="fd9e4394-01fa-485a-9e85-05b2178445e8" containerID="ac41cac3e8aa088d4a24a51d2af71fb2a17b3b8ba1cdb4e868128f884132244e" exitCode=143 Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.677217 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fd9e4394-01fa-485a-9e85-05b2178445e8","Type":"ContainerDied","Data":"ac41cac3e8aa088d4a24a51d2af71fb2a17b3b8ba1cdb4e868128f884132244e"} Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.679664 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-snppf" event={"ID":"aabbe361-2ef8-444e-acf2-fe48abecf27b","Type":"ContainerDied","Data":"6450b3b94bcfe1c5cb8992047032bb4da107a91cd1b01885dd410144995f21c8"} Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.679709 4790 scope.go:117] "RemoveContainer" containerID="6bee98f4451efc6d196589802f29d5f46905beb31c259a7f4c4ee20b26895450" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.679832 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-snppf" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.681840 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-f6lwg" event={"ID":"6803c7f5-eace-4b73-8ed7-39609fc9334a","Type":"ContainerDied","Data":"a35414d917640009d30748ea0c5f39c81f4319468ae27a6fd2dab106f7c071b1"} Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.681893 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a35414d917640009d30748ea0c5f39c81f4319468ae27a6fd2dab106f7c071b1" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.681991 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="adc9b1fa-6906-429f-8238-32798c6f563a" containerName="nova-metadata-log" containerID="cri-o://841f6a28de463438bab63ae94767350083795ba72c0122aa68ad0bbb8d5fea31" gracePeriod=30 Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.682014 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-f6lwg" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.682060 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="adc9b1fa-6906-429f-8238-32798c6f563a" containerName="nova-metadata-metadata" containerID="cri-o://247fde78a48a936a512702ebf27b9f4fbd05111a078a61ed16aa6dfb332af6ce" gracePeriod=30 Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.720183 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-snppf"] Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.721410 4790 scope.go:117] "RemoveContainer" containerID="2d446b0816edf4925a4d9a1f9987de79a211c49b58ca1683a329e8584d8616cb" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.729060 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-snppf"] Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.771801 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 13:35:46 crc kubenswrapper[4790]: E1124 13:35:46.772303 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fddd7928-5ec3-4605-94fb-a663c7f410e3" containerName="nova-manage" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.772381 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="fddd7928-5ec3-4605-94fb-a663c7f410e3" containerName="nova-manage" Nov 24 13:35:46 crc kubenswrapper[4790]: E1124 13:35:46.772412 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aabbe361-2ef8-444e-acf2-fe48abecf27b" containerName="init" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.772422 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="aabbe361-2ef8-444e-acf2-fe48abecf27b" containerName="init" Nov 24 13:35:46 crc kubenswrapper[4790]: E1124 13:35:46.772437 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6803c7f5-eace-4b73-8ed7-39609fc9334a" containerName="nova-cell1-conductor-db-sync" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.772446 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="6803c7f5-eace-4b73-8ed7-39609fc9334a" containerName="nova-cell1-conductor-db-sync" Nov 24 13:35:46 crc kubenswrapper[4790]: E1124 13:35:46.772469 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aabbe361-2ef8-444e-acf2-fe48abecf27b" containerName="dnsmasq-dns" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.772477 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="aabbe361-2ef8-444e-acf2-fe48abecf27b" containerName="dnsmasq-dns" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.772715 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="6803c7f5-eace-4b73-8ed7-39609fc9334a" containerName="nova-cell1-conductor-db-sync" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.772740 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="aabbe361-2ef8-444e-acf2-fe48abecf27b" containerName="dnsmasq-dns" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.772751 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="fddd7928-5ec3-4605-94fb-a663c7f410e3" containerName="nova-manage" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.773583 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.779488 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.831815 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.839292 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39d594c5-aae1-4cd7-a7de-9de435ad83b8-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"39d594c5-aae1-4cd7-a7de-9de435ad83b8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.839333 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p524t\" (UniqueName: \"kubernetes.io/projected/39d594c5-aae1-4cd7-a7de-9de435ad83b8-kube-api-access-p524t\") pod \"nova-cell1-conductor-0\" (UID: \"39d594c5-aae1-4cd7-a7de-9de435ad83b8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.839466 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39d594c5-aae1-4cd7-a7de-9de435ad83b8-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"39d594c5-aae1-4cd7-a7de-9de435ad83b8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.941665 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39d594c5-aae1-4cd7-a7de-9de435ad83b8-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"39d594c5-aae1-4cd7-a7de-9de435ad83b8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.941785 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39d594c5-aae1-4cd7-a7de-9de435ad83b8-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"39d594c5-aae1-4cd7-a7de-9de435ad83b8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.941822 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p524t\" (UniqueName: \"kubernetes.io/projected/39d594c5-aae1-4cd7-a7de-9de435ad83b8-kube-api-access-p524t\") pod \"nova-cell1-conductor-0\" (UID: \"39d594c5-aae1-4cd7-a7de-9de435ad83b8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.946112 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39d594c5-aae1-4cd7-a7de-9de435ad83b8-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"39d594c5-aae1-4cd7-a7de-9de435ad83b8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.948098 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39d594c5-aae1-4cd7-a7de-9de435ad83b8-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"39d594c5-aae1-4cd7-a7de-9de435ad83b8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 13:35:46 crc kubenswrapper[4790]: I1124 13:35:46.959695 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p524t\" (UniqueName: \"kubernetes.io/projected/39d594c5-aae1-4cd7-a7de-9de435ad83b8-kube-api-access-p524t\") pod \"nova-cell1-conductor-0\" (UID: \"39d594c5-aae1-4cd7-a7de-9de435ad83b8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.160501 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.266119 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.266358 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="d9726b87-983d-46fc-8880-7e19cbe1e35d" containerName="kube-state-metrics" containerID="cri-o://9d976a5e89a22f94794138f6f993ae7529618b4e787e03ded2687af1b33877b2" gracePeriod=30 Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.489050 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.566973 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adc9b1fa-6906-429f-8238-32798c6f563a-config-data\") pod \"adc9b1fa-6906-429f-8238-32798c6f563a\" (UID: \"adc9b1fa-6906-429f-8238-32798c6f563a\") " Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.567028 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adc9b1fa-6906-429f-8238-32798c6f563a-combined-ca-bundle\") pod \"adc9b1fa-6906-429f-8238-32798c6f563a\" (UID: \"adc9b1fa-6906-429f-8238-32798c6f563a\") " Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.567125 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tngwk\" (UniqueName: \"kubernetes.io/projected/adc9b1fa-6906-429f-8238-32798c6f563a-kube-api-access-tngwk\") pod \"adc9b1fa-6906-429f-8238-32798c6f563a\" (UID: \"adc9b1fa-6906-429f-8238-32798c6f563a\") " Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.567203 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adc9b1fa-6906-429f-8238-32798c6f563a-logs\") pod \"adc9b1fa-6906-429f-8238-32798c6f563a\" (UID: \"adc9b1fa-6906-429f-8238-32798c6f563a\") " Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.567258 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/adc9b1fa-6906-429f-8238-32798c6f563a-nova-metadata-tls-certs\") pod \"adc9b1fa-6906-429f-8238-32798c6f563a\" (UID: \"adc9b1fa-6906-429f-8238-32798c6f563a\") " Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.567641 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adc9b1fa-6906-429f-8238-32798c6f563a-logs" (OuterVolumeSpecName: "logs") pod "adc9b1fa-6906-429f-8238-32798c6f563a" (UID: "adc9b1fa-6906-429f-8238-32798c6f563a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.572810 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adc9b1fa-6906-429f-8238-32798c6f563a-kube-api-access-tngwk" (OuterVolumeSpecName: "kube-api-access-tngwk") pod "adc9b1fa-6906-429f-8238-32798c6f563a" (UID: "adc9b1fa-6906-429f-8238-32798c6f563a"). InnerVolumeSpecName "kube-api-access-tngwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.604863 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adc9b1fa-6906-429f-8238-32798c6f563a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "adc9b1fa-6906-429f-8238-32798c6f563a" (UID: "adc9b1fa-6906-429f-8238-32798c6f563a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.609865 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adc9b1fa-6906-429f-8238-32798c6f563a-config-data" (OuterVolumeSpecName: "config-data") pod "adc9b1fa-6906-429f-8238-32798c6f563a" (UID: "adc9b1fa-6906-429f-8238-32798c6f563a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.638987 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adc9b1fa-6906-429f-8238-32798c6f563a-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "adc9b1fa-6906-429f-8238-32798c6f563a" (UID: "adc9b1fa-6906-429f-8238-32798c6f563a"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.674639 4790 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/adc9b1fa-6906-429f-8238-32798c6f563a-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.675176 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adc9b1fa-6906-429f-8238-32798c6f563a-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.675211 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adc9b1fa-6906-429f-8238-32798c6f563a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.675222 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tngwk\" (UniqueName: \"kubernetes.io/projected/adc9b1fa-6906-429f-8238-32798c6f563a-kube-api-access-tngwk\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.675237 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adc9b1fa-6906-429f-8238-32798c6f563a-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.697574 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.714521 4790 generic.go:334] "Generic (PLEG): container finished" podID="d9726b87-983d-46fc-8880-7e19cbe1e35d" containerID="9d976a5e89a22f94794138f6f993ae7529618b4e787e03ded2687af1b33877b2" exitCode=2 Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.714569 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d9726b87-983d-46fc-8880-7e19cbe1e35d","Type":"ContainerDied","Data":"9d976a5e89a22f94794138f6f993ae7529618b4e787e03ded2687af1b33877b2"} Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.716979 4790 generic.go:334] "Generic (PLEG): container finished" podID="adc9b1fa-6906-429f-8238-32798c6f563a" containerID="247fde78a48a936a512702ebf27b9f4fbd05111a078a61ed16aa6dfb332af6ce" exitCode=0 Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.716999 4790 generic.go:334] "Generic (PLEG): container finished" podID="adc9b1fa-6906-429f-8238-32798c6f563a" containerID="841f6a28de463438bab63ae94767350083795ba72c0122aa68ad0bbb8d5fea31" exitCode=143 Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.717026 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"adc9b1fa-6906-429f-8238-32798c6f563a","Type":"ContainerDied","Data":"247fde78a48a936a512702ebf27b9f4fbd05111a078a61ed16aa6dfb332af6ce"} Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.717041 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"adc9b1fa-6906-429f-8238-32798c6f563a","Type":"ContainerDied","Data":"841f6a28de463438bab63ae94767350083795ba72c0122aa68ad0bbb8d5fea31"} Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.717051 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"adc9b1fa-6906-429f-8238-32798c6f563a","Type":"ContainerDied","Data":"8f950e5cbd0e7faca115a9a2f61ec2a33ed9d1e104a01467e44b6523a7c4368a"} Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.717071 4790 scope.go:117] "RemoveContainer" containerID="247fde78a48a936a512702ebf27b9f4fbd05111a078a61ed16aa6dfb332af6ce" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.717203 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.726617 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="7e988296-b952-43b2-bad8-39c2916bcecd" containerName="nova-scheduler-scheduler" containerID="cri-o://33636b9e3d7522fa5b65628a38a9b55c3049d7876c4ab7273d5e0270bede393e" gracePeriod=30 Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.727020 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"39d594c5-aae1-4cd7-a7de-9de435ad83b8","Type":"ContainerStarted","Data":"846877fb35e003c0d6457b8ae7c84f49884f7e6761533a1bfde34d0a491f0b51"} Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.740798 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.764263 4790 scope.go:117] "RemoveContainer" containerID="841f6a28de463438bab63ae94767350083795ba72c0122aa68ad0bbb8d5fea31" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.764405 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.776575 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8th6r\" (UniqueName: \"kubernetes.io/projected/d9726b87-983d-46fc-8880-7e19cbe1e35d-kube-api-access-8th6r\") pod \"d9726b87-983d-46fc-8880-7e19cbe1e35d\" (UID: \"d9726b87-983d-46fc-8880-7e19cbe1e35d\") " Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.787928 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9726b87-983d-46fc-8880-7e19cbe1e35d-kube-api-access-8th6r" (OuterVolumeSpecName: "kube-api-access-8th6r") pod "d9726b87-983d-46fc-8880-7e19cbe1e35d" (UID: "d9726b87-983d-46fc-8880-7e19cbe1e35d"). InnerVolumeSpecName "kube-api-access-8th6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.803673 4790 scope.go:117] "RemoveContainer" containerID="247fde78a48a936a512702ebf27b9f4fbd05111a078a61ed16aa6dfb332af6ce" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.803718 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:35:47 crc kubenswrapper[4790]: E1124 13:35:47.805203 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"247fde78a48a936a512702ebf27b9f4fbd05111a078a61ed16aa6dfb332af6ce\": container with ID starting with 247fde78a48a936a512702ebf27b9f4fbd05111a078a61ed16aa6dfb332af6ce not found: ID does not exist" containerID="247fde78a48a936a512702ebf27b9f4fbd05111a078a61ed16aa6dfb332af6ce" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.805271 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"247fde78a48a936a512702ebf27b9f4fbd05111a078a61ed16aa6dfb332af6ce"} err="failed to get container status \"247fde78a48a936a512702ebf27b9f4fbd05111a078a61ed16aa6dfb332af6ce\": rpc error: code = NotFound desc = could not find container \"247fde78a48a936a512702ebf27b9f4fbd05111a078a61ed16aa6dfb332af6ce\": container with ID starting with 247fde78a48a936a512702ebf27b9f4fbd05111a078a61ed16aa6dfb332af6ce not found: ID does not exist" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.805326 4790 scope.go:117] "RemoveContainer" containerID="841f6a28de463438bab63ae94767350083795ba72c0122aa68ad0bbb8d5fea31" Nov 24 13:35:47 crc kubenswrapper[4790]: E1124 13:35:47.810466 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"841f6a28de463438bab63ae94767350083795ba72c0122aa68ad0bbb8d5fea31\": container with ID starting with 841f6a28de463438bab63ae94767350083795ba72c0122aa68ad0bbb8d5fea31 not found: ID does not exist" containerID="841f6a28de463438bab63ae94767350083795ba72c0122aa68ad0bbb8d5fea31" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.810624 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"841f6a28de463438bab63ae94767350083795ba72c0122aa68ad0bbb8d5fea31"} err="failed to get container status \"841f6a28de463438bab63ae94767350083795ba72c0122aa68ad0bbb8d5fea31\": rpc error: code = NotFound desc = could not find container \"841f6a28de463438bab63ae94767350083795ba72c0122aa68ad0bbb8d5fea31\": container with ID starting with 841f6a28de463438bab63ae94767350083795ba72c0122aa68ad0bbb8d5fea31 not found: ID does not exist" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.810660 4790 scope.go:117] "RemoveContainer" containerID="247fde78a48a936a512702ebf27b9f4fbd05111a078a61ed16aa6dfb332af6ce" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.812350 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"247fde78a48a936a512702ebf27b9f4fbd05111a078a61ed16aa6dfb332af6ce"} err="failed to get container status \"247fde78a48a936a512702ebf27b9f4fbd05111a078a61ed16aa6dfb332af6ce\": rpc error: code = NotFound desc = could not find container \"247fde78a48a936a512702ebf27b9f4fbd05111a078a61ed16aa6dfb332af6ce\": container with ID starting with 247fde78a48a936a512702ebf27b9f4fbd05111a078a61ed16aa6dfb332af6ce not found: ID does not exist" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.812410 4790 scope.go:117] "RemoveContainer" containerID="841f6a28de463438bab63ae94767350083795ba72c0122aa68ad0bbb8d5fea31" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.815515 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"841f6a28de463438bab63ae94767350083795ba72c0122aa68ad0bbb8d5fea31"} err="failed to get container status \"841f6a28de463438bab63ae94767350083795ba72c0122aa68ad0bbb8d5fea31\": rpc error: code = NotFound desc = could not find container \"841f6a28de463438bab63ae94767350083795ba72c0122aa68ad0bbb8d5fea31\": container with ID starting with 841f6a28de463438bab63ae94767350083795ba72c0122aa68ad0bbb8d5fea31 not found: ID does not exist" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.845420 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:35:47 crc kubenswrapper[4790]: E1124 13:35:47.845950 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9726b87-983d-46fc-8880-7e19cbe1e35d" containerName="kube-state-metrics" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.845973 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9726b87-983d-46fc-8880-7e19cbe1e35d" containerName="kube-state-metrics" Nov 24 13:35:47 crc kubenswrapper[4790]: E1124 13:35:47.846003 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adc9b1fa-6906-429f-8238-32798c6f563a" containerName="nova-metadata-metadata" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.846012 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="adc9b1fa-6906-429f-8238-32798c6f563a" containerName="nova-metadata-metadata" Nov 24 13:35:47 crc kubenswrapper[4790]: E1124 13:35:47.846030 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adc9b1fa-6906-429f-8238-32798c6f563a" containerName="nova-metadata-log" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.846040 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="adc9b1fa-6906-429f-8238-32798c6f563a" containerName="nova-metadata-log" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.846586 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9726b87-983d-46fc-8880-7e19cbe1e35d" containerName="kube-state-metrics" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.846615 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="adc9b1fa-6906-429f-8238-32798c6f563a" containerName="nova-metadata-log" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.846632 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="adc9b1fa-6906-429f-8238-32798c6f563a" containerName="nova-metadata-metadata" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.848248 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.851382 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.851610 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.857773 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.878448 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65369c6a-c495-49d7-829f-e9adb21172ff-config-data\") pod \"nova-metadata-0\" (UID: \"65369c6a-c495-49d7-829f-e9adb21172ff\") " pod="openstack/nova-metadata-0" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.878565 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxrbx\" (UniqueName: \"kubernetes.io/projected/65369c6a-c495-49d7-829f-e9adb21172ff-kube-api-access-zxrbx\") pod \"nova-metadata-0\" (UID: \"65369c6a-c495-49d7-829f-e9adb21172ff\") " pod="openstack/nova-metadata-0" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.878588 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65369c6a-c495-49d7-829f-e9adb21172ff-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"65369c6a-c495-49d7-829f-e9adb21172ff\") " pod="openstack/nova-metadata-0" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.878766 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/65369c6a-c495-49d7-829f-e9adb21172ff-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"65369c6a-c495-49d7-829f-e9adb21172ff\") " pod="openstack/nova-metadata-0" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.878844 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65369c6a-c495-49d7-829f-e9adb21172ff-logs\") pod \"nova-metadata-0\" (UID: \"65369c6a-c495-49d7-829f-e9adb21172ff\") " pod="openstack/nova-metadata-0" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.879260 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8th6r\" (UniqueName: \"kubernetes.io/projected/d9726b87-983d-46fc-8880-7e19cbe1e35d-kube-api-access-8th6r\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.980939 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65369c6a-c495-49d7-829f-e9adb21172ff-config-data\") pod \"nova-metadata-0\" (UID: \"65369c6a-c495-49d7-829f-e9adb21172ff\") " pod="openstack/nova-metadata-0" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.981041 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxrbx\" (UniqueName: \"kubernetes.io/projected/65369c6a-c495-49d7-829f-e9adb21172ff-kube-api-access-zxrbx\") pod \"nova-metadata-0\" (UID: \"65369c6a-c495-49d7-829f-e9adb21172ff\") " pod="openstack/nova-metadata-0" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.981063 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65369c6a-c495-49d7-829f-e9adb21172ff-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"65369c6a-c495-49d7-829f-e9adb21172ff\") " pod="openstack/nova-metadata-0" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.981105 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/65369c6a-c495-49d7-829f-e9adb21172ff-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"65369c6a-c495-49d7-829f-e9adb21172ff\") " pod="openstack/nova-metadata-0" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.981130 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65369c6a-c495-49d7-829f-e9adb21172ff-logs\") pod \"nova-metadata-0\" (UID: \"65369c6a-c495-49d7-829f-e9adb21172ff\") " pod="openstack/nova-metadata-0" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.981481 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65369c6a-c495-49d7-829f-e9adb21172ff-logs\") pod \"nova-metadata-0\" (UID: \"65369c6a-c495-49d7-829f-e9adb21172ff\") " pod="openstack/nova-metadata-0" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.989966 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65369c6a-c495-49d7-829f-e9adb21172ff-config-data\") pod \"nova-metadata-0\" (UID: \"65369c6a-c495-49d7-829f-e9adb21172ff\") " pod="openstack/nova-metadata-0" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.991679 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65369c6a-c495-49d7-829f-e9adb21172ff-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"65369c6a-c495-49d7-829f-e9adb21172ff\") " pod="openstack/nova-metadata-0" Nov 24 13:35:47 crc kubenswrapper[4790]: I1124 13:35:47.994758 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/65369c6a-c495-49d7-829f-e9adb21172ff-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"65369c6a-c495-49d7-829f-e9adb21172ff\") " pod="openstack/nova-metadata-0" Nov 24 13:35:48 crc kubenswrapper[4790]: I1124 13:35:48.003923 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxrbx\" (UniqueName: \"kubernetes.io/projected/65369c6a-c495-49d7-829f-e9adb21172ff-kube-api-access-zxrbx\") pod \"nova-metadata-0\" (UID: \"65369c6a-c495-49d7-829f-e9adb21172ff\") " pod="openstack/nova-metadata-0" Nov 24 13:35:48 crc kubenswrapper[4790]: I1124 13:35:48.174507 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 13:35:48 crc kubenswrapper[4790]: I1124 13:35:48.335628 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aabbe361-2ef8-444e-acf2-fe48abecf27b" path="/var/lib/kubelet/pods/aabbe361-2ef8-444e-acf2-fe48abecf27b/volumes" Nov 24 13:35:48 crc kubenswrapper[4790]: I1124 13:35:48.336695 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adc9b1fa-6906-429f-8238-32798c6f563a" path="/var/lib/kubelet/pods/adc9b1fa-6906-429f-8238-32798c6f563a/volumes" Nov 24 13:35:48 crc kubenswrapper[4790]: I1124 13:35:48.624524 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:35:48 crc kubenswrapper[4790]: I1124 13:35:48.737456 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"65369c6a-c495-49d7-829f-e9adb21172ff","Type":"ContainerStarted","Data":"9e5d5f8346a57c976fc559e02ab3f537626fbf4fd16eb9df72b80146699585bd"} Nov 24 13:35:48 crc kubenswrapper[4790]: I1124 13:35:48.749491 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d9726b87-983d-46fc-8880-7e19cbe1e35d","Type":"ContainerDied","Data":"5ef77fdaab4bbd04d6baec365f079e0bb59ea2d4e9c12b47ab434f3c80a17f74"} Nov 24 13:35:48 crc kubenswrapper[4790]: I1124 13:35:48.749547 4790 scope.go:117] "RemoveContainer" containerID="9d976a5e89a22f94794138f6f993ae7529618b4e787e03ded2687af1b33877b2" Nov 24 13:35:48 crc kubenswrapper[4790]: I1124 13:35:48.749503 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 13:35:48 crc kubenswrapper[4790]: I1124 13:35:48.758534 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"39d594c5-aae1-4cd7-a7de-9de435ad83b8","Type":"ContainerStarted","Data":"eb8a4e1ecddd7fb54e90dd1a525962762b3153943ebd954bb7544913710d2ec2"} Nov 24 13:35:48 crc kubenswrapper[4790]: I1124 13:35:48.758776 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 24 13:35:48 crc kubenswrapper[4790]: I1124 13:35:48.773494 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 13:35:48 crc kubenswrapper[4790]: I1124 13:35:48.784108 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 13:35:48 crc kubenswrapper[4790]: I1124 13:35:48.793960 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 13:35:48 crc kubenswrapper[4790]: I1124 13:35:48.796255 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 13:35:48 crc kubenswrapper[4790]: I1124 13:35:48.798580 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 24 13:35:48 crc kubenswrapper[4790]: I1124 13:35:48.799748 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 24 13:35:48 crc kubenswrapper[4790]: I1124 13:35:48.800631 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.800610487 podStartE2EDuration="2.800610487s" podCreationTimestamp="2025-11-24 13:35:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:35:48.782874241 +0000 UTC m=+1397.162767923" watchObservedRunningTime="2025-11-24 13:35:48.800610487 +0000 UTC m=+1397.180504149" Nov 24 13:35:48 crc kubenswrapper[4790]: I1124 13:35:48.831357 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 13:35:48 crc kubenswrapper[4790]: I1124 13:35:48.896168 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/23704906-0f33-4a4e-8bbb-a796a9a1733f-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"23704906-0f33-4a4e-8bbb-a796a9a1733f\") " pod="openstack/kube-state-metrics-0" Nov 24 13:35:48 crc kubenswrapper[4790]: I1124 13:35:48.896221 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6bbx\" (UniqueName: \"kubernetes.io/projected/23704906-0f33-4a4e-8bbb-a796a9a1733f-kube-api-access-t6bbx\") pod \"kube-state-metrics-0\" (UID: \"23704906-0f33-4a4e-8bbb-a796a9a1733f\") " pod="openstack/kube-state-metrics-0" Nov 24 13:35:48 crc kubenswrapper[4790]: I1124 13:35:48.896244 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/23704906-0f33-4a4e-8bbb-a796a9a1733f-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"23704906-0f33-4a4e-8bbb-a796a9a1733f\") " pod="openstack/kube-state-metrics-0" Nov 24 13:35:48 crc kubenswrapper[4790]: I1124 13:35:48.896384 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23704906-0f33-4a4e-8bbb-a796a9a1733f-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"23704906-0f33-4a4e-8bbb-a796a9a1733f\") " pod="openstack/kube-state-metrics-0" Nov 24 13:35:48 crc kubenswrapper[4790]: I1124 13:35:48.998225 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23704906-0f33-4a4e-8bbb-a796a9a1733f-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"23704906-0f33-4a4e-8bbb-a796a9a1733f\") " pod="openstack/kube-state-metrics-0" Nov 24 13:35:48 crc kubenswrapper[4790]: I1124 13:35:48.998363 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/23704906-0f33-4a4e-8bbb-a796a9a1733f-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"23704906-0f33-4a4e-8bbb-a796a9a1733f\") " pod="openstack/kube-state-metrics-0" Nov 24 13:35:48 crc kubenswrapper[4790]: I1124 13:35:48.998397 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6bbx\" (UniqueName: \"kubernetes.io/projected/23704906-0f33-4a4e-8bbb-a796a9a1733f-kube-api-access-t6bbx\") pod \"kube-state-metrics-0\" (UID: \"23704906-0f33-4a4e-8bbb-a796a9a1733f\") " pod="openstack/kube-state-metrics-0" Nov 24 13:35:48 crc kubenswrapper[4790]: I1124 13:35:48.998423 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/23704906-0f33-4a4e-8bbb-a796a9a1733f-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"23704906-0f33-4a4e-8bbb-a796a9a1733f\") " pod="openstack/kube-state-metrics-0" Nov 24 13:35:49 crc kubenswrapper[4790]: I1124 13:35:49.002084 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/23704906-0f33-4a4e-8bbb-a796a9a1733f-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"23704906-0f33-4a4e-8bbb-a796a9a1733f\") " pod="openstack/kube-state-metrics-0" Nov 24 13:35:49 crc kubenswrapper[4790]: I1124 13:35:49.003397 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/23704906-0f33-4a4e-8bbb-a796a9a1733f-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"23704906-0f33-4a4e-8bbb-a796a9a1733f\") " pod="openstack/kube-state-metrics-0" Nov 24 13:35:49 crc kubenswrapper[4790]: I1124 13:35:49.004077 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23704906-0f33-4a4e-8bbb-a796a9a1733f-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"23704906-0f33-4a4e-8bbb-a796a9a1733f\") " pod="openstack/kube-state-metrics-0" Nov 24 13:35:49 crc kubenswrapper[4790]: I1124 13:35:49.017542 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6bbx\" (UniqueName: \"kubernetes.io/projected/23704906-0f33-4a4e-8bbb-a796a9a1733f-kube-api-access-t6bbx\") pod \"kube-state-metrics-0\" (UID: \"23704906-0f33-4a4e-8bbb-a796a9a1733f\") " pod="openstack/kube-state-metrics-0" Nov 24 13:35:49 crc kubenswrapper[4790]: I1124 13:35:49.117377 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 13:35:49 crc kubenswrapper[4790]: I1124 13:35:49.474322 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:35:49 crc kubenswrapper[4790]: I1124 13:35:49.475008 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="237e7166-6c17-46d8-a18b-2af671d75e28" containerName="sg-core" containerID="cri-o://04275059fed1ba9eabf4afa369dbc5147006448caf80244a63fecf325ccc57e7" gracePeriod=30 Nov 24 13:35:49 crc kubenswrapper[4790]: I1124 13:35:49.475021 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="237e7166-6c17-46d8-a18b-2af671d75e28" containerName="proxy-httpd" containerID="cri-o://243b98c6ff9675ec9019b09fe6a4ee32b15708475572e6be2dd633baf9614988" gracePeriod=30 Nov 24 13:35:49 crc kubenswrapper[4790]: I1124 13:35:49.475028 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="237e7166-6c17-46d8-a18b-2af671d75e28" containerName="ceilometer-notification-agent" containerID="cri-o://da4959e122210172c5153b9713ea054958931f764b05982716558bf2d9239f5f" gracePeriod=30 Nov 24 13:35:49 crc kubenswrapper[4790]: I1124 13:35:49.475199 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="237e7166-6c17-46d8-a18b-2af671d75e28" containerName="ceilometer-central-agent" containerID="cri-o://6b70cf6fea320d4bb5acf67752d705bc8f4e8b0eea3d825bd9b466ef671d7cee" gracePeriod=30 Nov 24 13:35:49 crc kubenswrapper[4790]: I1124 13:35:49.620063 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 13:35:49 crc kubenswrapper[4790]: W1124 13:35:49.623322 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23704906_0f33_4a4e_8bbb_a796a9a1733f.slice/crio-69e6c7f78b3c6faf70cdd89bc87e230d3e05f3f269931940d6787a110523daf7 WatchSource:0}: Error finding container 69e6c7f78b3c6faf70cdd89bc87e230d3e05f3f269931940d6787a110523daf7: Status 404 returned error can't find the container with id 69e6c7f78b3c6faf70cdd89bc87e230d3e05f3f269931940d6787a110523daf7 Nov 24 13:35:49 crc kubenswrapper[4790]: I1124 13:35:49.769366 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"65369c6a-c495-49d7-829f-e9adb21172ff","Type":"ContainerStarted","Data":"d5cc8dc52fd100b95d8a516cbdf245be0d1ffd7c60cfbbaee80ea00d834358ca"} Nov 24 13:35:49 crc kubenswrapper[4790]: I1124 13:35:49.769423 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"65369c6a-c495-49d7-829f-e9adb21172ff","Type":"ContainerStarted","Data":"28ce7a4d4c4dc879f7be8c9eb8f16679b7296eea8d6956705834a8e39f388e31"} Nov 24 13:35:49 crc kubenswrapper[4790]: I1124 13:35:49.772783 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"23704906-0f33-4a4e-8bbb-a796a9a1733f","Type":"ContainerStarted","Data":"69e6c7f78b3c6faf70cdd89bc87e230d3e05f3f269931940d6787a110523daf7"} Nov 24 13:35:49 crc kubenswrapper[4790]: I1124 13:35:49.777607 4790 generic.go:334] "Generic (PLEG): container finished" podID="237e7166-6c17-46d8-a18b-2af671d75e28" containerID="243b98c6ff9675ec9019b09fe6a4ee32b15708475572e6be2dd633baf9614988" exitCode=0 Nov 24 13:35:49 crc kubenswrapper[4790]: I1124 13:35:49.777640 4790 generic.go:334] "Generic (PLEG): container finished" podID="237e7166-6c17-46d8-a18b-2af671d75e28" containerID="04275059fed1ba9eabf4afa369dbc5147006448caf80244a63fecf325ccc57e7" exitCode=2 Nov 24 13:35:49 crc kubenswrapper[4790]: I1124 13:35:49.777650 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"237e7166-6c17-46d8-a18b-2af671d75e28","Type":"ContainerDied","Data":"243b98c6ff9675ec9019b09fe6a4ee32b15708475572e6be2dd633baf9614988"} Nov 24 13:35:49 crc kubenswrapper[4790]: I1124 13:35:49.777707 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"237e7166-6c17-46d8-a18b-2af671d75e28","Type":"ContainerDied","Data":"04275059fed1ba9eabf4afa369dbc5147006448caf80244a63fecf325ccc57e7"} Nov 24 13:35:49 crc kubenswrapper[4790]: I1124 13:35:49.791443 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.791422508 podStartE2EDuration="2.791422508s" podCreationTimestamp="2025-11-24 13:35:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:35:49.788235436 +0000 UTC m=+1398.168129099" watchObservedRunningTime="2025-11-24 13:35:49.791422508 +0000 UTC m=+1398.171316170" Nov 24 13:35:50 crc kubenswrapper[4790]: E1124 13:35:50.211718 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 33636b9e3d7522fa5b65628a38a9b55c3049d7876c4ab7273d5e0270bede393e is running failed: container process not found" containerID="33636b9e3d7522fa5b65628a38a9b55c3049d7876c4ab7273d5e0270bede393e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 13:35:50 crc kubenswrapper[4790]: E1124 13:35:50.213703 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 33636b9e3d7522fa5b65628a38a9b55c3049d7876c4ab7273d5e0270bede393e is running failed: container process not found" containerID="33636b9e3d7522fa5b65628a38a9b55c3049d7876c4ab7273d5e0270bede393e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 13:35:50 crc kubenswrapper[4790]: E1124 13:35:50.214169 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 33636b9e3d7522fa5b65628a38a9b55c3049d7876c4ab7273d5e0270bede393e is running failed: container process not found" containerID="33636b9e3d7522fa5b65628a38a9b55c3049d7876c4ab7273d5e0270bede393e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 13:35:50 crc kubenswrapper[4790]: E1124 13:35:50.214216 4790 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 33636b9e3d7522fa5b65628a38a9b55c3049d7876c4ab7273d5e0270bede393e is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="7e988296-b952-43b2-bad8-39c2916bcecd" containerName="nova-scheduler-scheduler" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.331477 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9726b87-983d-46fc-8880-7e19cbe1e35d" path="/var/lib/kubelet/pods/d9726b87-983d-46fc-8880-7e19cbe1e35d/volumes" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.544960 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.633796 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/237e7166-6c17-46d8-a18b-2af671d75e28-sg-core-conf-yaml\") pod \"237e7166-6c17-46d8-a18b-2af671d75e28\" (UID: \"237e7166-6c17-46d8-a18b-2af671d75e28\") " Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.634205 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/237e7166-6c17-46d8-a18b-2af671d75e28-scripts\") pod \"237e7166-6c17-46d8-a18b-2af671d75e28\" (UID: \"237e7166-6c17-46d8-a18b-2af671d75e28\") " Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.634218 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.635248 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/237e7166-6c17-46d8-a18b-2af671d75e28-log-httpd\") pod \"237e7166-6c17-46d8-a18b-2af671d75e28\" (UID: \"237e7166-6c17-46d8-a18b-2af671d75e28\") " Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.635288 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/237e7166-6c17-46d8-a18b-2af671d75e28-run-httpd\") pod \"237e7166-6c17-46d8-a18b-2af671d75e28\" (UID: \"237e7166-6c17-46d8-a18b-2af671d75e28\") " Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.635337 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/237e7166-6c17-46d8-a18b-2af671d75e28-combined-ca-bundle\") pod \"237e7166-6c17-46d8-a18b-2af671d75e28\" (UID: \"237e7166-6c17-46d8-a18b-2af671d75e28\") " Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.635359 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pz2k4\" (UniqueName: \"kubernetes.io/projected/237e7166-6c17-46d8-a18b-2af671d75e28-kube-api-access-pz2k4\") pod \"237e7166-6c17-46d8-a18b-2af671d75e28\" (UID: \"237e7166-6c17-46d8-a18b-2af671d75e28\") " Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.635374 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/237e7166-6c17-46d8-a18b-2af671d75e28-config-data\") pod \"237e7166-6c17-46d8-a18b-2af671d75e28\" (UID: \"237e7166-6c17-46d8-a18b-2af671d75e28\") " Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.636550 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/237e7166-6c17-46d8-a18b-2af671d75e28-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "237e7166-6c17-46d8-a18b-2af671d75e28" (UID: "237e7166-6c17-46d8-a18b-2af671d75e28"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.636899 4790 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/237e7166-6c17-46d8-a18b-2af671d75e28-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.636994 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/237e7166-6c17-46d8-a18b-2af671d75e28-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "237e7166-6c17-46d8-a18b-2af671d75e28" (UID: "237e7166-6c17-46d8-a18b-2af671d75e28"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.639768 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/237e7166-6c17-46d8-a18b-2af671d75e28-scripts" (OuterVolumeSpecName: "scripts") pod "237e7166-6c17-46d8-a18b-2af671d75e28" (UID: "237e7166-6c17-46d8-a18b-2af671d75e28"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.641155 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/237e7166-6c17-46d8-a18b-2af671d75e28-kube-api-access-pz2k4" (OuterVolumeSpecName: "kube-api-access-pz2k4") pod "237e7166-6c17-46d8-a18b-2af671d75e28" (UID: "237e7166-6c17-46d8-a18b-2af671d75e28"). InnerVolumeSpecName "kube-api-access-pz2k4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.673753 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/237e7166-6c17-46d8-a18b-2af671d75e28-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "237e7166-6c17-46d8-a18b-2af671d75e28" (UID: "237e7166-6c17-46d8-a18b-2af671d75e28"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.763012 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e988296-b952-43b2-bad8-39c2916bcecd-combined-ca-bundle\") pod \"7e988296-b952-43b2-bad8-39c2916bcecd\" (UID: \"7e988296-b952-43b2-bad8-39c2916bcecd\") " Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.763179 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e988296-b952-43b2-bad8-39c2916bcecd-config-data\") pod \"7e988296-b952-43b2-bad8-39c2916bcecd\" (UID: \"7e988296-b952-43b2-bad8-39c2916bcecd\") " Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.763266 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhk95\" (UniqueName: \"kubernetes.io/projected/7e988296-b952-43b2-bad8-39c2916bcecd-kube-api-access-jhk95\") pod \"7e988296-b952-43b2-bad8-39c2916bcecd\" (UID: \"7e988296-b952-43b2-bad8-39c2916bcecd\") " Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.763636 4790 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/237e7166-6c17-46d8-a18b-2af671d75e28-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.763661 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/237e7166-6c17-46d8-a18b-2af671d75e28-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.763672 4790 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/237e7166-6c17-46d8-a18b-2af671d75e28-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.763681 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pz2k4\" (UniqueName: \"kubernetes.io/projected/237e7166-6c17-46d8-a18b-2af671d75e28-kube-api-access-pz2k4\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.766716 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e988296-b952-43b2-bad8-39c2916bcecd-kube-api-access-jhk95" (OuterVolumeSpecName: "kube-api-access-jhk95") pod "7e988296-b952-43b2-bad8-39c2916bcecd" (UID: "7e988296-b952-43b2-bad8-39c2916bcecd"). InnerVolumeSpecName "kube-api-access-jhk95". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.831106 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e988296-b952-43b2-bad8-39c2916bcecd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7e988296-b952-43b2-bad8-39c2916bcecd" (UID: "7e988296-b952-43b2-bad8-39c2916bcecd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.831222 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e988296-b952-43b2-bad8-39c2916bcecd-config-data" (OuterVolumeSpecName: "config-data") pod "7e988296-b952-43b2-bad8-39c2916bcecd" (UID: "7e988296-b952-43b2-bad8-39c2916bcecd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.843100 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/237e7166-6c17-46d8-a18b-2af671d75e28-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "237e7166-6c17-46d8-a18b-2af671d75e28" (UID: "237e7166-6c17-46d8-a18b-2af671d75e28"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.858127 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"23704906-0f33-4a4e-8bbb-a796a9a1733f","Type":"ContainerStarted","Data":"7eb4e19d335efc1382a2f1934fe4914e9bce012cc0d8c8dd38049967fe85a33d"} Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.859362 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.865318 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhk95\" (UniqueName: \"kubernetes.io/projected/7e988296-b952-43b2-bad8-39c2916bcecd-kube-api-access-jhk95\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.865356 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/237e7166-6c17-46d8-a18b-2af671d75e28-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.865364 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e988296-b952-43b2-bad8-39c2916bcecd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.865374 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e988296-b952-43b2-bad8-39c2916bcecd-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.870754 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/237e7166-6c17-46d8-a18b-2af671d75e28-config-data" (OuterVolumeSpecName: "config-data") pod "237e7166-6c17-46d8-a18b-2af671d75e28" (UID: "237e7166-6c17-46d8-a18b-2af671d75e28"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.871200 4790 generic.go:334] "Generic (PLEG): container finished" podID="7e988296-b952-43b2-bad8-39c2916bcecd" containerID="33636b9e3d7522fa5b65628a38a9b55c3049d7876c4ab7273d5e0270bede393e" exitCode=0 Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.871269 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7e988296-b952-43b2-bad8-39c2916bcecd","Type":"ContainerDied","Data":"33636b9e3d7522fa5b65628a38a9b55c3049d7876c4ab7273d5e0270bede393e"} Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.871287 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.871309 4790 scope.go:117] "RemoveContainer" containerID="33636b9e3d7522fa5b65628a38a9b55c3049d7876c4ab7273d5e0270bede393e" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.871298 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7e988296-b952-43b2-bad8-39c2916bcecd","Type":"ContainerDied","Data":"f22bd6efe39521b9e1992013ea3771042237c9a20af4ad476a5e69d3c0a5f2e2"} Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.880706 4790 generic.go:334] "Generic (PLEG): container finished" podID="237e7166-6c17-46d8-a18b-2af671d75e28" containerID="da4959e122210172c5153b9713ea054958931f764b05982716558bf2d9239f5f" exitCode=0 Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.880734 4790 generic.go:334] "Generic (PLEG): container finished" podID="237e7166-6c17-46d8-a18b-2af671d75e28" containerID="6b70cf6fea320d4bb5acf67752d705bc8f4e8b0eea3d825bd9b466ef671d7cee" exitCode=0 Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.881488 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.892965 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"237e7166-6c17-46d8-a18b-2af671d75e28","Type":"ContainerDied","Data":"da4959e122210172c5153b9713ea054958931f764b05982716558bf2d9239f5f"} Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.893026 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"237e7166-6c17-46d8-a18b-2af671d75e28","Type":"ContainerDied","Data":"6b70cf6fea320d4bb5acf67752d705bc8f4e8b0eea3d825bd9b466ef671d7cee"} Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.893042 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"237e7166-6c17-46d8-a18b-2af671d75e28","Type":"ContainerDied","Data":"4def04435b8fa9bcf11de020fdd00930a2050f00a84542747dad5156b9bea006"} Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.895661 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.319774915 podStartE2EDuration="2.895636369s" podCreationTimestamp="2025-11-24 13:35:48 +0000 UTC" firstStartedPulling="2025-11-24 13:35:49.627680499 +0000 UTC m=+1398.007574161" lastFinishedPulling="2025-11-24 13:35:50.203541953 +0000 UTC m=+1398.583435615" observedRunningTime="2025-11-24 13:35:50.893371644 +0000 UTC m=+1399.273265316" watchObservedRunningTime="2025-11-24 13:35:50.895636369 +0000 UTC m=+1399.275530031" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.905912 4790 scope.go:117] "RemoveContainer" containerID="33636b9e3d7522fa5b65628a38a9b55c3049d7876c4ab7273d5e0270bede393e" Nov 24 13:35:50 crc kubenswrapper[4790]: E1124 13:35:50.907300 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33636b9e3d7522fa5b65628a38a9b55c3049d7876c4ab7273d5e0270bede393e\": container with ID starting with 33636b9e3d7522fa5b65628a38a9b55c3049d7876c4ab7273d5e0270bede393e not found: ID does not exist" containerID="33636b9e3d7522fa5b65628a38a9b55c3049d7876c4ab7273d5e0270bede393e" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.907349 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33636b9e3d7522fa5b65628a38a9b55c3049d7876c4ab7273d5e0270bede393e"} err="failed to get container status \"33636b9e3d7522fa5b65628a38a9b55c3049d7876c4ab7273d5e0270bede393e\": rpc error: code = NotFound desc = could not find container \"33636b9e3d7522fa5b65628a38a9b55c3049d7876c4ab7273d5e0270bede393e\": container with ID starting with 33636b9e3d7522fa5b65628a38a9b55c3049d7876c4ab7273d5e0270bede393e not found: ID does not exist" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.907373 4790 scope.go:117] "RemoveContainer" containerID="243b98c6ff9675ec9019b09fe6a4ee32b15708475572e6be2dd633baf9614988" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.933171 4790 scope.go:117] "RemoveContainer" containerID="04275059fed1ba9eabf4afa369dbc5147006448caf80244a63fecf325ccc57e7" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.940393 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.951564 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.959138 4790 scope.go:117] "RemoveContainer" containerID="da4959e122210172c5153b9713ea054958931f764b05982716558bf2d9239f5f" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.966837 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/237e7166-6c17-46d8-a18b-2af671d75e28-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.984955 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:35:50 crc kubenswrapper[4790]: I1124 13:35:50.998718 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.008214 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 13:35:51 crc kubenswrapper[4790]: E1124 13:35:51.008777 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="237e7166-6c17-46d8-a18b-2af671d75e28" containerName="ceilometer-central-agent" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.008846 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="237e7166-6c17-46d8-a18b-2af671d75e28" containerName="ceilometer-central-agent" Nov 24 13:35:51 crc kubenswrapper[4790]: E1124 13:35:51.009148 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e988296-b952-43b2-bad8-39c2916bcecd" containerName="nova-scheduler-scheduler" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.009266 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e988296-b952-43b2-bad8-39c2916bcecd" containerName="nova-scheduler-scheduler" Nov 24 13:35:51 crc kubenswrapper[4790]: E1124 13:35:51.009330 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="237e7166-6c17-46d8-a18b-2af671d75e28" containerName="ceilometer-notification-agent" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.009383 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="237e7166-6c17-46d8-a18b-2af671d75e28" containerName="ceilometer-notification-agent" Nov 24 13:35:51 crc kubenswrapper[4790]: E1124 13:35:51.009452 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="237e7166-6c17-46d8-a18b-2af671d75e28" containerName="sg-core" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.009505 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="237e7166-6c17-46d8-a18b-2af671d75e28" containerName="sg-core" Nov 24 13:35:51 crc kubenswrapper[4790]: E1124 13:35:51.009584 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="237e7166-6c17-46d8-a18b-2af671d75e28" containerName="proxy-httpd" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.009636 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="237e7166-6c17-46d8-a18b-2af671d75e28" containerName="proxy-httpd" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.009852 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="237e7166-6c17-46d8-a18b-2af671d75e28" containerName="ceilometer-notification-agent" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.009942 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="237e7166-6c17-46d8-a18b-2af671d75e28" containerName="ceilometer-central-agent" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.010010 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="237e7166-6c17-46d8-a18b-2af671d75e28" containerName="proxy-httpd" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.010078 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e988296-b952-43b2-bad8-39c2916bcecd" containerName="nova-scheduler-scheduler" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.010147 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="237e7166-6c17-46d8-a18b-2af671d75e28" containerName="sg-core" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.010909 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.014119 4790 scope.go:117] "RemoveContainer" containerID="6b70cf6fea320d4bb5acf67752d705bc8f4e8b0eea3d825bd9b466ef671d7cee" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.016241 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.019179 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.027788 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.031666 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.035983 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.036308 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.038061 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.044102 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.069543 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f2d4bc2-9482-4790-beee-c221236609f6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3f2d4bc2-9482-4790-beee-c221236609f6\") " pod="openstack/ceilometer-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.069826 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3f2d4bc2-9482-4790-beee-c221236609f6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3f2d4bc2-9482-4790-beee-c221236609f6\") " pod="openstack/ceilometer-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.069969 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c864a2b1-1693-4200-be70-a140b1526bac-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c864a2b1-1693-4200-be70-a140b1526bac\") " pod="openstack/nova-scheduler-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.070079 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f2d4bc2-9482-4790-beee-c221236609f6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3f2d4bc2-9482-4790-beee-c221236609f6\") " pod="openstack/ceilometer-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.070168 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f2d4bc2-9482-4790-beee-c221236609f6-config-data\") pod \"ceilometer-0\" (UID: \"3f2d4bc2-9482-4790-beee-c221236609f6\") " pod="openstack/ceilometer-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.070263 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f2d4bc2-9482-4790-beee-c221236609f6-scripts\") pod \"ceilometer-0\" (UID: \"3f2d4bc2-9482-4790-beee-c221236609f6\") " pod="openstack/ceilometer-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.070397 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c864a2b1-1693-4200-be70-a140b1526bac-config-data\") pod \"nova-scheduler-0\" (UID: \"c864a2b1-1693-4200-be70-a140b1526bac\") " pod="openstack/nova-scheduler-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.070525 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f2d4bc2-9482-4790-beee-c221236609f6-run-httpd\") pod \"ceilometer-0\" (UID: \"3f2d4bc2-9482-4790-beee-c221236609f6\") " pod="openstack/ceilometer-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.070653 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6shng\" (UniqueName: \"kubernetes.io/projected/c864a2b1-1693-4200-be70-a140b1526bac-kube-api-access-6shng\") pod \"nova-scheduler-0\" (UID: \"c864a2b1-1693-4200-be70-a140b1526bac\") " pod="openstack/nova-scheduler-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.070800 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f2d4bc2-9482-4790-beee-c221236609f6-log-httpd\") pod \"ceilometer-0\" (UID: \"3f2d4bc2-9482-4790-beee-c221236609f6\") " pod="openstack/ceilometer-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.071005 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbnww\" (UniqueName: \"kubernetes.io/projected/3f2d4bc2-9482-4790-beee-c221236609f6-kube-api-access-bbnww\") pod \"ceilometer-0\" (UID: \"3f2d4bc2-9482-4790-beee-c221236609f6\") " pod="openstack/ceilometer-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.108005 4790 scope.go:117] "RemoveContainer" containerID="243b98c6ff9675ec9019b09fe6a4ee32b15708475572e6be2dd633baf9614988" Nov 24 13:35:51 crc kubenswrapper[4790]: E1124 13:35:51.108343 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"243b98c6ff9675ec9019b09fe6a4ee32b15708475572e6be2dd633baf9614988\": container with ID starting with 243b98c6ff9675ec9019b09fe6a4ee32b15708475572e6be2dd633baf9614988 not found: ID does not exist" containerID="243b98c6ff9675ec9019b09fe6a4ee32b15708475572e6be2dd633baf9614988" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.108374 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"243b98c6ff9675ec9019b09fe6a4ee32b15708475572e6be2dd633baf9614988"} err="failed to get container status \"243b98c6ff9675ec9019b09fe6a4ee32b15708475572e6be2dd633baf9614988\": rpc error: code = NotFound desc = could not find container \"243b98c6ff9675ec9019b09fe6a4ee32b15708475572e6be2dd633baf9614988\": container with ID starting with 243b98c6ff9675ec9019b09fe6a4ee32b15708475572e6be2dd633baf9614988 not found: ID does not exist" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.108400 4790 scope.go:117] "RemoveContainer" containerID="04275059fed1ba9eabf4afa369dbc5147006448caf80244a63fecf325ccc57e7" Nov 24 13:35:51 crc kubenswrapper[4790]: E1124 13:35:51.108801 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04275059fed1ba9eabf4afa369dbc5147006448caf80244a63fecf325ccc57e7\": container with ID starting with 04275059fed1ba9eabf4afa369dbc5147006448caf80244a63fecf325ccc57e7 not found: ID does not exist" containerID="04275059fed1ba9eabf4afa369dbc5147006448caf80244a63fecf325ccc57e7" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.108821 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04275059fed1ba9eabf4afa369dbc5147006448caf80244a63fecf325ccc57e7"} err="failed to get container status \"04275059fed1ba9eabf4afa369dbc5147006448caf80244a63fecf325ccc57e7\": rpc error: code = NotFound desc = could not find container \"04275059fed1ba9eabf4afa369dbc5147006448caf80244a63fecf325ccc57e7\": container with ID starting with 04275059fed1ba9eabf4afa369dbc5147006448caf80244a63fecf325ccc57e7 not found: ID does not exist" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.108834 4790 scope.go:117] "RemoveContainer" containerID="da4959e122210172c5153b9713ea054958931f764b05982716558bf2d9239f5f" Nov 24 13:35:51 crc kubenswrapper[4790]: E1124 13:35:51.109241 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da4959e122210172c5153b9713ea054958931f764b05982716558bf2d9239f5f\": container with ID starting with da4959e122210172c5153b9713ea054958931f764b05982716558bf2d9239f5f not found: ID does not exist" containerID="da4959e122210172c5153b9713ea054958931f764b05982716558bf2d9239f5f" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.109283 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da4959e122210172c5153b9713ea054958931f764b05982716558bf2d9239f5f"} err="failed to get container status \"da4959e122210172c5153b9713ea054958931f764b05982716558bf2d9239f5f\": rpc error: code = NotFound desc = could not find container \"da4959e122210172c5153b9713ea054958931f764b05982716558bf2d9239f5f\": container with ID starting with da4959e122210172c5153b9713ea054958931f764b05982716558bf2d9239f5f not found: ID does not exist" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.109310 4790 scope.go:117] "RemoveContainer" containerID="6b70cf6fea320d4bb5acf67752d705bc8f4e8b0eea3d825bd9b466ef671d7cee" Nov 24 13:35:51 crc kubenswrapper[4790]: E1124 13:35:51.109554 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b70cf6fea320d4bb5acf67752d705bc8f4e8b0eea3d825bd9b466ef671d7cee\": container with ID starting with 6b70cf6fea320d4bb5acf67752d705bc8f4e8b0eea3d825bd9b466ef671d7cee not found: ID does not exist" containerID="6b70cf6fea320d4bb5acf67752d705bc8f4e8b0eea3d825bd9b466ef671d7cee" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.109578 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b70cf6fea320d4bb5acf67752d705bc8f4e8b0eea3d825bd9b466ef671d7cee"} err="failed to get container status \"6b70cf6fea320d4bb5acf67752d705bc8f4e8b0eea3d825bd9b466ef671d7cee\": rpc error: code = NotFound desc = could not find container \"6b70cf6fea320d4bb5acf67752d705bc8f4e8b0eea3d825bd9b466ef671d7cee\": container with ID starting with 6b70cf6fea320d4bb5acf67752d705bc8f4e8b0eea3d825bd9b466ef671d7cee not found: ID does not exist" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.109593 4790 scope.go:117] "RemoveContainer" containerID="243b98c6ff9675ec9019b09fe6a4ee32b15708475572e6be2dd633baf9614988" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.109831 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"243b98c6ff9675ec9019b09fe6a4ee32b15708475572e6be2dd633baf9614988"} err="failed to get container status \"243b98c6ff9675ec9019b09fe6a4ee32b15708475572e6be2dd633baf9614988\": rpc error: code = NotFound desc = could not find container \"243b98c6ff9675ec9019b09fe6a4ee32b15708475572e6be2dd633baf9614988\": container with ID starting with 243b98c6ff9675ec9019b09fe6a4ee32b15708475572e6be2dd633baf9614988 not found: ID does not exist" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.109942 4790 scope.go:117] "RemoveContainer" containerID="04275059fed1ba9eabf4afa369dbc5147006448caf80244a63fecf325ccc57e7" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.110236 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04275059fed1ba9eabf4afa369dbc5147006448caf80244a63fecf325ccc57e7"} err="failed to get container status \"04275059fed1ba9eabf4afa369dbc5147006448caf80244a63fecf325ccc57e7\": rpc error: code = NotFound desc = could not find container \"04275059fed1ba9eabf4afa369dbc5147006448caf80244a63fecf325ccc57e7\": container with ID starting with 04275059fed1ba9eabf4afa369dbc5147006448caf80244a63fecf325ccc57e7 not found: ID does not exist" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.110256 4790 scope.go:117] "RemoveContainer" containerID="da4959e122210172c5153b9713ea054958931f764b05982716558bf2d9239f5f" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.111865 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da4959e122210172c5153b9713ea054958931f764b05982716558bf2d9239f5f"} err="failed to get container status \"da4959e122210172c5153b9713ea054958931f764b05982716558bf2d9239f5f\": rpc error: code = NotFound desc = could not find container \"da4959e122210172c5153b9713ea054958931f764b05982716558bf2d9239f5f\": container with ID starting with da4959e122210172c5153b9713ea054958931f764b05982716558bf2d9239f5f not found: ID does not exist" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.111907 4790 scope.go:117] "RemoveContainer" containerID="6b70cf6fea320d4bb5acf67752d705bc8f4e8b0eea3d825bd9b466ef671d7cee" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.112138 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b70cf6fea320d4bb5acf67752d705bc8f4e8b0eea3d825bd9b466ef671d7cee"} err="failed to get container status \"6b70cf6fea320d4bb5acf67752d705bc8f4e8b0eea3d825bd9b466ef671d7cee\": rpc error: code = NotFound desc = could not find container \"6b70cf6fea320d4bb5acf67752d705bc8f4e8b0eea3d825bd9b466ef671d7cee\": container with ID starting with 6b70cf6fea320d4bb5acf67752d705bc8f4e8b0eea3d825bd9b466ef671d7cee not found: ID does not exist" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.173119 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f2d4bc2-9482-4790-beee-c221236609f6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3f2d4bc2-9482-4790-beee-c221236609f6\") " pod="openstack/ceilometer-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.173172 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3f2d4bc2-9482-4790-beee-c221236609f6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3f2d4bc2-9482-4790-beee-c221236609f6\") " pod="openstack/ceilometer-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.173203 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c864a2b1-1693-4200-be70-a140b1526bac-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c864a2b1-1693-4200-be70-a140b1526bac\") " pod="openstack/nova-scheduler-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.173239 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f2d4bc2-9482-4790-beee-c221236609f6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3f2d4bc2-9482-4790-beee-c221236609f6\") " pod="openstack/ceilometer-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.173258 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f2d4bc2-9482-4790-beee-c221236609f6-config-data\") pod \"ceilometer-0\" (UID: \"3f2d4bc2-9482-4790-beee-c221236609f6\") " pod="openstack/ceilometer-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.173280 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f2d4bc2-9482-4790-beee-c221236609f6-scripts\") pod \"ceilometer-0\" (UID: \"3f2d4bc2-9482-4790-beee-c221236609f6\") " pod="openstack/ceilometer-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.173307 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c864a2b1-1693-4200-be70-a140b1526bac-config-data\") pod \"nova-scheduler-0\" (UID: \"c864a2b1-1693-4200-be70-a140b1526bac\") " pod="openstack/nova-scheduler-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.173337 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f2d4bc2-9482-4790-beee-c221236609f6-run-httpd\") pod \"ceilometer-0\" (UID: \"3f2d4bc2-9482-4790-beee-c221236609f6\") " pod="openstack/ceilometer-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.173356 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6shng\" (UniqueName: \"kubernetes.io/projected/c864a2b1-1693-4200-be70-a140b1526bac-kube-api-access-6shng\") pod \"nova-scheduler-0\" (UID: \"c864a2b1-1693-4200-be70-a140b1526bac\") " pod="openstack/nova-scheduler-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.173393 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f2d4bc2-9482-4790-beee-c221236609f6-log-httpd\") pod \"ceilometer-0\" (UID: \"3f2d4bc2-9482-4790-beee-c221236609f6\") " pod="openstack/ceilometer-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.173439 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbnww\" (UniqueName: \"kubernetes.io/projected/3f2d4bc2-9482-4790-beee-c221236609f6-kube-api-access-bbnww\") pod \"ceilometer-0\" (UID: \"3f2d4bc2-9482-4790-beee-c221236609f6\") " pod="openstack/ceilometer-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.174523 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f2d4bc2-9482-4790-beee-c221236609f6-run-httpd\") pod \"ceilometer-0\" (UID: \"3f2d4bc2-9482-4790-beee-c221236609f6\") " pod="openstack/ceilometer-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.177681 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3f2d4bc2-9482-4790-beee-c221236609f6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3f2d4bc2-9482-4790-beee-c221236609f6\") " pod="openstack/ceilometer-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.177769 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f2d4bc2-9482-4790-beee-c221236609f6-scripts\") pod \"ceilometer-0\" (UID: \"3f2d4bc2-9482-4790-beee-c221236609f6\") " pod="openstack/ceilometer-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.178136 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f2d4bc2-9482-4790-beee-c221236609f6-config-data\") pod \"ceilometer-0\" (UID: \"3f2d4bc2-9482-4790-beee-c221236609f6\") " pod="openstack/ceilometer-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.178240 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f2d4bc2-9482-4790-beee-c221236609f6-log-httpd\") pod \"ceilometer-0\" (UID: \"3f2d4bc2-9482-4790-beee-c221236609f6\") " pod="openstack/ceilometer-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.178740 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c864a2b1-1693-4200-be70-a140b1526bac-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c864a2b1-1693-4200-be70-a140b1526bac\") " pod="openstack/nova-scheduler-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.178971 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f2d4bc2-9482-4790-beee-c221236609f6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3f2d4bc2-9482-4790-beee-c221236609f6\") " pod="openstack/ceilometer-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.180829 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f2d4bc2-9482-4790-beee-c221236609f6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3f2d4bc2-9482-4790-beee-c221236609f6\") " pod="openstack/ceilometer-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.191182 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c864a2b1-1693-4200-be70-a140b1526bac-config-data\") pod \"nova-scheduler-0\" (UID: \"c864a2b1-1693-4200-be70-a140b1526bac\") " pod="openstack/nova-scheduler-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.193691 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6shng\" (UniqueName: \"kubernetes.io/projected/c864a2b1-1693-4200-be70-a140b1526bac-kube-api-access-6shng\") pod \"nova-scheduler-0\" (UID: \"c864a2b1-1693-4200-be70-a140b1526bac\") " pod="openstack/nova-scheduler-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.197464 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbnww\" (UniqueName: \"kubernetes.io/projected/3f2d4bc2-9482-4790-beee-c221236609f6-kube-api-access-bbnww\") pod \"ceilometer-0\" (UID: \"3f2d4bc2-9482-4790-beee-c221236609f6\") " pod="openstack/ceilometer-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.413431 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.421293 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.870757 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:35:51 crc kubenswrapper[4790]: W1124 13:35:51.874466 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f2d4bc2_9482_4790_beee_c221236609f6.slice/crio-5b48cd6aa385fc09325c0d7e099958b8e93d18ecdbf32a51eb696d87d5d620e0 WatchSource:0}: Error finding container 5b48cd6aa385fc09325c0d7e099958b8e93d18ecdbf32a51eb696d87d5d620e0: Status 404 returned error can't find the container with id 5b48cd6aa385fc09325c0d7e099958b8e93d18ecdbf32a51eb696d87d5d620e0 Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.894323 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f2d4bc2-9482-4790-beee-c221236609f6","Type":"ContainerStarted","Data":"5b48cd6aa385fc09325c0d7e099958b8e93d18ecdbf32a51eb696d87d5d620e0"} Nov 24 13:35:51 crc kubenswrapper[4790]: I1124 13:35:51.953321 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 13:35:52 crc kubenswrapper[4790]: I1124 13:35:52.188072 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 24 13:35:52 crc kubenswrapper[4790]: I1124 13:35:52.329849 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="237e7166-6c17-46d8-a18b-2af671d75e28" path="/var/lib/kubelet/pods/237e7166-6c17-46d8-a18b-2af671d75e28/volumes" Nov 24 13:35:52 crc kubenswrapper[4790]: I1124 13:35:52.332625 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e988296-b952-43b2-bad8-39c2916bcecd" path="/var/lib/kubelet/pods/7e988296-b952-43b2-bad8-39c2916bcecd/volumes" Nov 24 13:35:52 crc kubenswrapper[4790]: I1124 13:35:52.898510 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 13:35:52 crc kubenswrapper[4790]: I1124 13:35:52.911945 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c864a2b1-1693-4200-be70-a140b1526bac","Type":"ContainerStarted","Data":"285bb94beb2c560eb1cadca29c368e78f28a1d7b266ed21f459cb58d61a3f93b"} Nov 24 13:35:52 crc kubenswrapper[4790]: I1124 13:35:52.911985 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c864a2b1-1693-4200-be70-a140b1526bac","Type":"ContainerStarted","Data":"b9c285f81ae78d4cda166049a582c17eef8cb14d241437f53ce18dd1d2ab72fb"} Nov 24 13:35:52 crc kubenswrapper[4790]: I1124 13:35:52.913596 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f2d4bc2-9482-4790-beee-c221236609f6","Type":"ContainerStarted","Data":"7d2e4dbd94cd4c5d426c877165c167e175b4d892096b798b2a694a7fc6873761"} Nov 24 13:35:52 crc kubenswrapper[4790]: I1124 13:35:52.917166 4790 generic.go:334] "Generic (PLEG): container finished" podID="fd9e4394-01fa-485a-9e85-05b2178445e8" containerID="a0ecd98ed58aec4504de27538119d64f6110d3754aa6003370bad164ec160264" exitCode=0 Nov 24 13:35:52 crc kubenswrapper[4790]: I1124 13:35:52.917790 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 13:35:52 crc kubenswrapper[4790]: I1124 13:35:52.917963 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fd9e4394-01fa-485a-9e85-05b2178445e8","Type":"ContainerDied","Data":"a0ecd98ed58aec4504de27538119d64f6110d3754aa6003370bad164ec160264"} Nov 24 13:35:52 crc kubenswrapper[4790]: I1124 13:35:52.917985 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fd9e4394-01fa-485a-9e85-05b2178445e8","Type":"ContainerDied","Data":"881171b3621bf54b14036d2b0347475c6c3360777aa49b39971c14ff9178cbb0"} Nov 24 13:35:52 crc kubenswrapper[4790]: I1124 13:35:52.918000 4790 scope.go:117] "RemoveContainer" containerID="a0ecd98ed58aec4504de27538119d64f6110d3754aa6003370bad164ec160264" Nov 24 13:35:52 crc kubenswrapper[4790]: I1124 13:35:52.944354 4790 scope.go:117] "RemoveContainer" containerID="ac41cac3e8aa088d4a24a51d2af71fb2a17b3b8ba1cdb4e868128f884132244e" Nov 24 13:35:52 crc kubenswrapper[4790]: I1124 13:35:52.951696 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.951674987 podStartE2EDuration="2.951674987s" podCreationTimestamp="2025-11-24 13:35:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:35:52.949369331 +0000 UTC m=+1401.329262993" watchObservedRunningTime="2025-11-24 13:35:52.951674987 +0000 UTC m=+1401.331568649" Nov 24 13:35:52 crc kubenswrapper[4790]: I1124 13:35:52.966793 4790 scope.go:117] "RemoveContainer" containerID="a0ecd98ed58aec4504de27538119d64f6110d3754aa6003370bad164ec160264" Nov 24 13:35:52 crc kubenswrapper[4790]: E1124 13:35:52.968124 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0ecd98ed58aec4504de27538119d64f6110d3754aa6003370bad164ec160264\": container with ID starting with a0ecd98ed58aec4504de27538119d64f6110d3754aa6003370bad164ec160264 not found: ID does not exist" containerID="a0ecd98ed58aec4504de27538119d64f6110d3754aa6003370bad164ec160264" Nov 24 13:35:52 crc kubenswrapper[4790]: I1124 13:35:52.968151 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0ecd98ed58aec4504de27538119d64f6110d3754aa6003370bad164ec160264"} err="failed to get container status \"a0ecd98ed58aec4504de27538119d64f6110d3754aa6003370bad164ec160264\": rpc error: code = NotFound desc = could not find container \"a0ecd98ed58aec4504de27538119d64f6110d3754aa6003370bad164ec160264\": container with ID starting with a0ecd98ed58aec4504de27538119d64f6110d3754aa6003370bad164ec160264 not found: ID does not exist" Nov 24 13:35:52 crc kubenswrapper[4790]: I1124 13:35:52.968170 4790 scope.go:117] "RemoveContainer" containerID="ac41cac3e8aa088d4a24a51d2af71fb2a17b3b8ba1cdb4e868128f884132244e" Nov 24 13:35:52 crc kubenswrapper[4790]: E1124 13:35:52.968504 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac41cac3e8aa088d4a24a51d2af71fb2a17b3b8ba1cdb4e868128f884132244e\": container with ID starting with ac41cac3e8aa088d4a24a51d2af71fb2a17b3b8ba1cdb4e868128f884132244e not found: ID does not exist" containerID="ac41cac3e8aa088d4a24a51d2af71fb2a17b3b8ba1cdb4e868128f884132244e" Nov 24 13:35:52 crc kubenswrapper[4790]: I1124 13:35:52.968539 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac41cac3e8aa088d4a24a51d2af71fb2a17b3b8ba1cdb4e868128f884132244e"} err="failed to get container status \"ac41cac3e8aa088d4a24a51d2af71fb2a17b3b8ba1cdb4e868128f884132244e\": rpc error: code = NotFound desc = could not find container \"ac41cac3e8aa088d4a24a51d2af71fb2a17b3b8ba1cdb4e868128f884132244e\": container with ID starting with ac41cac3e8aa088d4a24a51d2af71fb2a17b3b8ba1cdb4e868128f884132244e not found: ID does not exist" Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.008596 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd9e4394-01fa-485a-9e85-05b2178445e8-logs\") pod \"fd9e4394-01fa-485a-9e85-05b2178445e8\" (UID: \"fd9e4394-01fa-485a-9e85-05b2178445e8\") " Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.009026 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd9e4394-01fa-485a-9e85-05b2178445e8-combined-ca-bundle\") pod \"fd9e4394-01fa-485a-9e85-05b2178445e8\" (UID: \"fd9e4394-01fa-485a-9e85-05b2178445e8\") " Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.009822 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd9e4394-01fa-485a-9e85-05b2178445e8-config-data\") pod \"fd9e4394-01fa-485a-9e85-05b2178445e8\" (UID: \"fd9e4394-01fa-485a-9e85-05b2178445e8\") " Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.009225 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd9e4394-01fa-485a-9e85-05b2178445e8-logs" (OuterVolumeSpecName: "logs") pod "fd9e4394-01fa-485a-9e85-05b2178445e8" (UID: "fd9e4394-01fa-485a-9e85-05b2178445e8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.010350 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hw7cx\" (UniqueName: \"kubernetes.io/projected/fd9e4394-01fa-485a-9e85-05b2178445e8-kube-api-access-hw7cx\") pod \"fd9e4394-01fa-485a-9e85-05b2178445e8\" (UID: \"fd9e4394-01fa-485a-9e85-05b2178445e8\") " Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.011095 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd9e4394-01fa-485a-9e85-05b2178445e8-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.015090 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd9e4394-01fa-485a-9e85-05b2178445e8-kube-api-access-hw7cx" (OuterVolumeSpecName: "kube-api-access-hw7cx") pod "fd9e4394-01fa-485a-9e85-05b2178445e8" (UID: "fd9e4394-01fa-485a-9e85-05b2178445e8"). InnerVolumeSpecName "kube-api-access-hw7cx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.035650 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd9e4394-01fa-485a-9e85-05b2178445e8-config-data" (OuterVolumeSpecName: "config-data") pod "fd9e4394-01fa-485a-9e85-05b2178445e8" (UID: "fd9e4394-01fa-485a-9e85-05b2178445e8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.036550 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd9e4394-01fa-485a-9e85-05b2178445e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fd9e4394-01fa-485a-9e85-05b2178445e8" (UID: "fd9e4394-01fa-485a-9e85-05b2178445e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.112874 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hw7cx\" (UniqueName: \"kubernetes.io/projected/fd9e4394-01fa-485a-9e85-05b2178445e8-kube-api-access-hw7cx\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.112922 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd9e4394-01fa-485a-9e85-05b2178445e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.112937 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd9e4394-01fa-485a-9e85-05b2178445e8-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.175486 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.177022 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.264037 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.278753 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.290089 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 13:35:53 crc kubenswrapper[4790]: E1124 13:35:53.290477 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd9e4394-01fa-485a-9e85-05b2178445e8" containerName="nova-api-api" Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.290496 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd9e4394-01fa-485a-9e85-05b2178445e8" containerName="nova-api-api" Nov 24 13:35:53 crc kubenswrapper[4790]: E1124 13:35:53.290554 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd9e4394-01fa-485a-9e85-05b2178445e8" containerName="nova-api-log" Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.290562 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd9e4394-01fa-485a-9e85-05b2178445e8" containerName="nova-api-log" Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.290742 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd9e4394-01fa-485a-9e85-05b2178445e8" containerName="nova-api-api" Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.290768 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd9e4394-01fa-485a-9e85-05b2178445e8" containerName="nova-api-log" Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.291982 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.293709 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.298919 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.417918 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd50520e-ec7e-48da-bb07-27c3cc3b7a1e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cd50520e-ec7e-48da-bb07-27c3cc3b7a1e\") " pod="openstack/nova-api-0" Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.417985 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd50520e-ec7e-48da-bb07-27c3cc3b7a1e-config-data\") pod \"nova-api-0\" (UID: \"cd50520e-ec7e-48da-bb07-27c3cc3b7a1e\") " pod="openstack/nova-api-0" Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.418040 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd50520e-ec7e-48da-bb07-27c3cc3b7a1e-logs\") pod \"nova-api-0\" (UID: \"cd50520e-ec7e-48da-bb07-27c3cc3b7a1e\") " pod="openstack/nova-api-0" Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.418086 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pc8cz\" (UniqueName: \"kubernetes.io/projected/cd50520e-ec7e-48da-bb07-27c3cc3b7a1e-kube-api-access-pc8cz\") pod \"nova-api-0\" (UID: \"cd50520e-ec7e-48da-bb07-27c3cc3b7a1e\") " pod="openstack/nova-api-0" Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.519505 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd50520e-ec7e-48da-bb07-27c3cc3b7a1e-config-data\") pod \"nova-api-0\" (UID: \"cd50520e-ec7e-48da-bb07-27c3cc3b7a1e\") " pod="openstack/nova-api-0" Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.519571 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd50520e-ec7e-48da-bb07-27c3cc3b7a1e-logs\") pod \"nova-api-0\" (UID: \"cd50520e-ec7e-48da-bb07-27c3cc3b7a1e\") " pod="openstack/nova-api-0" Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.519599 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pc8cz\" (UniqueName: \"kubernetes.io/projected/cd50520e-ec7e-48da-bb07-27c3cc3b7a1e-kube-api-access-pc8cz\") pod \"nova-api-0\" (UID: \"cd50520e-ec7e-48da-bb07-27c3cc3b7a1e\") " pod="openstack/nova-api-0" Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.519723 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd50520e-ec7e-48da-bb07-27c3cc3b7a1e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cd50520e-ec7e-48da-bb07-27c3cc3b7a1e\") " pod="openstack/nova-api-0" Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.520382 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd50520e-ec7e-48da-bb07-27c3cc3b7a1e-logs\") pod \"nova-api-0\" (UID: \"cd50520e-ec7e-48da-bb07-27c3cc3b7a1e\") " pod="openstack/nova-api-0" Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.526601 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd50520e-ec7e-48da-bb07-27c3cc3b7a1e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cd50520e-ec7e-48da-bb07-27c3cc3b7a1e\") " pod="openstack/nova-api-0" Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.526638 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd50520e-ec7e-48da-bb07-27c3cc3b7a1e-config-data\") pod \"nova-api-0\" (UID: \"cd50520e-ec7e-48da-bb07-27c3cc3b7a1e\") " pod="openstack/nova-api-0" Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.540370 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pc8cz\" (UniqueName: \"kubernetes.io/projected/cd50520e-ec7e-48da-bb07-27c3cc3b7a1e-kube-api-access-pc8cz\") pod \"nova-api-0\" (UID: \"cd50520e-ec7e-48da-bb07-27c3cc3b7a1e\") " pod="openstack/nova-api-0" Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.607746 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 13:35:53 crc kubenswrapper[4790]: I1124 13:35:53.934381 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f2d4bc2-9482-4790-beee-c221236609f6","Type":"ContainerStarted","Data":"86597fa39ea0ecee6911f4fc85f8531e905d1efe57b06fa776d6f51272ddf530"} Nov 24 13:35:54 crc kubenswrapper[4790]: W1124 13:35:54.053278 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd50520e_ec7e_48da_bb07_27c3cc3b7a1e.slice/crio-6ae5f0616e5d0d587636caa46f15a059dc11c0a0f4463df491907bc54aca2f75 WatchSource:0}: Error finding container 6ae5f0616e5d0d587636caa46f15a059dc11c0a0f4463df491907bc54aca2f75: Status 404 returned error can't find the container with id 6ae5f0616e5d0d587636caa46f15a059dc11c0a0f4463df491907bc54aca2f75 Nov 24 13:35:54 crc kubenswrapper[4790]: I1124 13:35:54.056930 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:35:54 crc kubenswrapper[4790]: I1124 13:35:54.350722 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd9e4394-01fa-485a-9e85-05b2178445e8" path="/var/lib/kubelet/pods/fd9e4394-01fa-485a-9e85-05b2178445e8/volumes" Nov 24 13:35:54 crc kubenswrapper[4790]: I1124 13:35:54.946532 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f2d4bc2-9482-4790-beee-c221236609f6","Type":"ContainerStarted","Data":"29318ee81f035c8fbbce11db915e3bd719dba88643c8a572632f173b6508e5b6"} Nov 24 13:35:54 crc kubenswrapper[4790]: I1124 13:35:54.948247 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cd50520e-ec7e-48da-bb07-27c3cc3b7a1e","Type":"ContainerStarted","Data":"6275194871d5daca61dffd87f7b3b078bf95e6b3d5437279f936212a99985c4e"} Nov 24 13:35:54 crc kubenswrapper[4790]: I1124 13:35:54.948300 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cd50520e-ec7e-48da-bb07-27c3cc3b7a1e","Type":"ContainerStarted","Data":"8d4d871f41ae1bd375fb1d586afe5870f6c6704be950bb394f58cb336951aee6"} Nov 24 13:35:54 crc kubenswrapper[4790]: I1124 13:35:54.948315 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cd50520e-ec7e-48da-bb07-27c3cc3b7a1e","Type":"ContainerStarted","Data":"6ae5f0616e5d0d587636caa46f15a059dc11c0a0f4463df491907bc54aca2f75"} Nov 24 13:35:54 crc kubenswrapper[4790]: I1124 13:35:54.977226 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.9772080939999999 podStartE2EDuration="1.977208094s" podCreationTimestamp="2025-11-24 13:35:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:35:54.967793185 +0000 UTC m=+1403.347686847" watchObservedRunningTime="2025-11-24 13:35:54.977208094 +0000 UTC m=+1403.357101756" Nov 24 13:35:56 crc kubenswrapper[4790]: I1124 13:35:56.414686 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 13:35:56 crc kubenswrapper[4790]: I1124 13:35:56.969420 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f2d4bc2-9482-4790-beee-c221236609f6","Type":"ContainerStarted","Data":"7d502a90c4cd6e3fcfbf940d7dcf81f76a64e192cd1d6335bd6c3fb5b94549ec"} Nov 24 13:35:56 crc kubenswrapper[4790]: I1124 13:35:56.971369 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 13:35:57 crc kubenswrapper[4790]: I1124 13:35:57.011390 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.069546945 podStartE2EDuration="7.011367896s" podCreationTimestamp="2025-11-24 13:35:50 +0000 UTC" firstStartedPulling="2025-11-24 13:35:51.87592961 +0000 UTC m=+1400.255823272" lastFinishedPulling="2025-11-24 13:35:55.817750561 +0000 UTC m=+1404.197644223" observedRunningTime="2025-11-24 13:35:56.999654421 +0000 UTC m=+1405.379548113" watchObservedRunningTime="2025-11-24 13:35:57.011367896 +0000 UTC m=+1405.391261558" Nov 24 13:35:58 crc kubenswrapper[4790]: I1124 13:35:58.174925 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 13:35:58 crc kubenswrapper[4790]: I1124 13:35:58.175314 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 13:35:59 crc kubenswrapper[4790]: I1124 13:35:59.128146 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 24 13:35:59 crc kubenswrapper[4790]: I1124 13:35:59.199145 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="65369c6a-c495-49d7-829f-e9adb21172ff" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.189:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 13:35:59 crc kubenswrapper[4790]: I1124 13:35:59.199145 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="65369c6a-c495-49d7-829f-e9adb21172ff" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.189:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 13:36:01 crc kubenswrapper[4790]: I1124 13:36:01.088153 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nkvgt"] Nov 24 13:36:01 crc kubenswrapper[4790]: I1124 13:36:01.090408 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nkvgt" Nov 24 13:36:01 crc kubenswrapper[4790]: I1124 13:36:01.098237 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nkvgt"] Nov 24 13:36:01 crc kubenswrapper[4790]: I1124 13:36:01.155371 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/099c81d1-2be3-4e10-b733-65e66912b605-catalog-content\") pod \"redhat-operators-nkvgt\" (UID: \"099c81d1-2be3-4e10-b733-65e66912b605\") " pod="openshift-marketplace/redhat-operators-nkvgt" Nov 24 13:36:01 crc kubenswrapper[4790]: I1124 13:36:01.155431 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zb5f\" (UniqueName: \"kubernetes.io/projected/099c81d1-2be3-4e10-b733-65e66912b605-kube-api-access-5zb5f\") pod \"redhat-operators-nkvgt\" (UID: \"099c81d1-2be3-4e10-b733-65e66912b605\") " pod="openshift-marketplace/redhat-operators-nkvgt" Nov 24 13:36:01 crc kubenswrapper[4790]: I1124 13:36:01.155506 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/099c81d1-2be3-4e10-b733-65e66912b605-utilities\") pod \"redhat-operators-nkvgt\" (UID: \"099c81d1-2be3-4e10-b733-65e66912b605\") " pod="openshift-marketplace/redhat-operators-nkvgt" Nov 24 13:36:01 crc kubenswrapper[4790]: I1124 13:36:01.257072 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/099c81d1-2be3-4e10-b733-65e66912b605-utilities\") pod \"redhat-operators-nkvgt\" (UID: \"099c81d1-2be3-4e10-b733-65e66912b605\") " pod="openshift-marketplace/redhat-operators-nkvgt" Nov 24 13:36:01 crc kubenswrapper[4790]: I1124 13:36:01.257194 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/099c81d1-2be3-4e10-b733-65e66912b605-catalog-content\") pod \"redhat-operators-nkvgt\" (UID: \"099c81d1-2be3-4e10-b733-65e66912b605\") " pod="openshift-marketplace/redhat-operators-nkvgt" Nov 24 13:36:01 crc kubenswrapper[4790]: I1124 13:36:01.257219 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zb5f\" (UniqueName: \"kubernetes.io/projected/099c81d1-2be3-4e10-b733-65e66912b605-kube-api-access-5zb5f\") pod \"redhat-operators-nkvgt\" (UID: \"099c81d1-2be3-4e10-b733-65e66912b605\") " pod="openshift-marketplace/redhat-operators-nkvgt" Nov 24 13:36:01 crc kubenswrapper[4790]: I1124 13:36:01.257999 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/099c81d1-2be3-4e10-b733-65e66912b605-utilities\") pod \"redhat-operators-nkvgt\" (UID: \"099c81d1-2be3-4e10-b733-65e66912b605\") " pod="openshift-marketplace/redhat-operators-nkvgt" Nov 24 13:36:01 crc kubenswrapper[4790]: I1124 13:36:01.258208 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/099c81d1-2be3-4e10-b733-65e66912b605-catalog-content\") pod \"redhat-operators-nkvgt\" (UID: \"099c81d1-2be3-4e10-b733-65e66912b605\") " pod="openshift-marketplace/redhat-operators-nkvgt" Nov 24 13:36:01 crc kubenswrapper[4790]: I1124 13:36:01.276835 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zb5f\" (UniqueName: \"kubernetes.io/projected/099c81d1-2be3-4e10-b733-65e66912b605-kube-api-access-5zb5f\") pod \"redhat-operators-nkvgt\" (UID: \"099c81d1-2be3-4e10-b733-65e66912b605\") " pod="openshift-marketplace/redhat-operators-nkvgt" Nov 24 13:36:01 crc kubenswrapper[4790]: I1124 13:36:01.414889 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 13:36:01 crc kubenswrapper[4790]: I1124 13:36:01.419759 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nkvgt" Nov 24 13:36:01 crc kubenswrapper[4790]: I1124 13:36:01.444223 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 13:36:01 crc kubenswrapper[4790]: I1124 13:36:01.876585 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nkvgt"] Nov 24 13:36:02 crc kubenswrapper[4790]: I1124 13:36:02.022620 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nkvgt" event={"ID":"099c81d1-2be3-4e10-b733-65e66912b605","Type":"ContainerStarted","Data":"602948df3da2e9a0dc56c8375a5d5ea011cc76e7e12e256d80d5d768be413041"} Nov 24 13:36:02 crc kubenswrapper[4790]: I1124 13:36:02.062229 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 13:36:03 crc kubenswrapper[4790]: I1124 13:36:03.033088 4790 generic.go:334] "Generic (PLEG): container finished" podID="099c81d1-2be3-4e10-b733-65e66912b605" containerID="71ef147cadaf1a19d06f10591436329c47702dc6721878e0208bc04ca4c4e395" exitCode=0 Nov 24 13:36:03 crc kubenswrapper[4790]: I1124 13:36:03.033200 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nkvgt" event={"ID":"099c81d1-2be3-4e10-b733-65e66912b605","Type":"ContainerDied","Data":"71ef147cadaf1a19d06f10591436329c47702dc6721878e0208bc04ca4c4e395"} Nov 24 13:36:03 crc kubenswrapper[4790]: I1124 13:36:03.035431 4790 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 13:36:03 crc kubenswrapper[4790]: I1124 13:36:03.608646 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 13:36:03 crc kubenswrapper[4790]: I1124 13:36:03.608987 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 13:36:04 crc kubenswrapper[4790]: I1124 13:36:04.690199 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="cd50520e-ec7e-48da-bb07-27c3cc3b7a1e" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 13:36:04 crc kubenswrapper[4790]: I1124 13:36:04.690236 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="cd50520e-ec7e-48da-bb07-27c3cc3b7a1e" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 13:36:05 crc kubenswrapper[4790]: I1124 13:36:05.056215 4790 generic.go:334] "Generic (PLEG): container finished" podID="099c81d1-2be3-4e10-b733-65e66912b605" containerID="86fe297d369579aab3fd9785a4ca060b16eafb20531318ab7a41138922921292" exitCode=0 Nov 24 13:36:05 crc kubenswrapper[4790]: I1124 13:36:05.056281 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nkvgt" event={"ID":"099c81d1-2be3-4e10-b733-65e66912b605","Type":"ContainerDied","Data":"86fe297d369579aab3fd9785a4ca060b16eafb20531318ab7a41138922921292"} Nov 24 13:36:08 crc kubenswrapper[4790]: I1124 13:36:08.083496 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nkvgt" event={"ID":"099c81d1-2be3-4e10-b733-65e66912b605","Type":"ContainerStarted","Data":"88abd8704b089a541505f154340ecb0756ce273f1b8e926a5a9549fcd8886f64"} Nov 24 13:36:08 crc kubenswrapper[4790]: I1124 13:36:08.104946 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nkvgt" podStartSLOduration=2.583393762 podStartE2EDuration="7.104921705s" podCreationTimestamp="2025-11-24 13:36:01 +0000 UTC" firstStartedPulling="2025-11-24 13:36:03.035170837 +0000 UTC m=+1411.415064499" lastFinishedPulling="2025-11-24 13:36:07.55669878 +0000 UTC m=+1415.936592442" observedRunningTime="2025-11-24 13:36:08.098069689 +0000 UTC m=+1416.477963351" watchObservedRunningTime="2025-11-24 13:36:08.104921705 +0000 UTC m=+1416.484815367" Nov 24 13:36:08 crc kubenswrapper[4790]: I1124 13:36:08.184924 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 13:36:08 crc kubenswrapper[4790]: I1124 13:36:08.186560 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 13:36:08 crc kubenswrapper[4790]: I1124 13:36:08.191508 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 13:36:09 crc kubenswrapper[4790]: I1124 13:36:09.099964 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 13:36:10 crc kubenswrapper[4790]: I1124 13:36:10.104923 4790 generic.go:334] "Generic (PLEG): container finished" podID="5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0" containerID="aba5dac078aa2c293348629db3cab9f01c0d490999bb82d7e064e7ea72ffe77c" exitCode=137 Nov 24 13:36:10 crc kubenswrapper[4790]: I1124 13:36:10.104962 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0","Type":"ContainerDied","Data":"aba5dac078aa2c293348629db3cab9f01c0d490999bb82d7e064e7ea72ffe77c"} Nov 24 13:36:10 crc kubenswrapper[4790]: I1124 13:36:10.274479 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:10 crc kubenswrapper[4790]: I1124 13:36:10.324615 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0-combined-ca-bundle\") pod \"5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0\" (UID: \"5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0\") " Nov 24 13:36:10 crc kubenswrapper[4790]: I1124 13:36:10.324736 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0-config-data\") pod \"5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0\" (UID: \"5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0\") " Nov 24 13:36:10 crc kubenswrapper[4790]: I1124 13:36:10.324822 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4krqq\" (UniqueName: \"kubernetes.io/projected/5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0-kube-api-access-4krqq\") pod \"5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0\" (UID: \"5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0\") " Nov 24 13:36:10 crc kubenswrapper[4790]: I1124 13:36:10.335182 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0-kube-api-access-4krqq" (OuterVolumeSpecName: "kube-api-access-4krqq") pod "5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0" (UID: "5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0"). InnerVolumeSpecName "kube-api-access-4krqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:36:10 crc kubenswrapper[4790]: I1124 13:36:10.357335 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0" (UID: "5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:10 crc kubenswrapper[4790]: I1124 13:36:10.358206 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0-config-data" (OuterVolumeSpecName: "config-data") pod "5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0" (UID: "5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:10 crc kubenswrapper[4790]: I1124 13:36:10.426465 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:10 crc kubenswrapper[4790]: I1124 13:36:10.426501 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:10 crc kubenswrapper[4790]: I1124 13:36:10.426511 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4krqq\" (UniqueName: \"kubernetes.io/projected/5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0-kube-api-access-4krqq\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:11 crc kubenswrapper[4790]: I1124 13:36:11.113176 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:11 crc kubenswrapper[4790]: I1124 13:36:11.113180 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0","Type":"ContainerDied","Data":"dab33147fa8de28ae5aa9c9af8db4f22d54ae18131601532ffcfb4240b37cbca"} Nov 24 13:36:11 crc kubenswrapper[4790]: I1124 13:36:11.113239 4790 scope.go:117] "RemoveContainer" containerID="aba5dac078aa2c293348629db3cab9f01c0d490999bb82d7e064e7ea72ffe77c" Nov 24 13:36:11 crc kubenswrapper[4790]: I1124 13:36:11.147315 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 13:36:11 crc kubenswrapper[4790]: I1124 13:36:11.155115 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 13:36:11 crc kubenswrapper[4790]: I1124 13:36:11.174985 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 13:36:11 crc kubenswrapper[4790]: E1124 13:36:11.176117 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 13:36:11 crc kubenswrapper[4790]: I1124 13:36:11.176140 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 13:36:11 crc kubenswrapper[4790]: I1124 13:36:11.179791 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 13:36:11 crc kubenswrapper[4790]: I1124 13:36:11.180603 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:11 crc kubenswrapper[4790]: I1124 13:36:11.182746 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 24 13:36:11 crc kubenswrapper[4790]: I1124 13:36:11.182985 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 24 13:36:11 crc kubenswrapper[4790]: I1124 13:36:11.183669 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 24 13:36:11 crc kubenswrapper[4790]: I1124 13:36:11.189743 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 13:36:11 crc kubenswrapper[4790]: I1124 13:36:11.245843 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/698a7113-712a-4940-aef9-aabd791b021d-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"698a7113-712a-4940-aef9-aabd791b021d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:11 crc kubenswrapper[4790]: I1124 13:36:11.245946 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/698a7113-712a-4940-aef9-aabd791b021d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"698a7113-712a-4940-aef9-aabd791b021d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:11 crc kubenswrapper[4790]: I1124 13:36:11.246077 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csgm4\" (UniqueName: \"kubernetes.io/projected/698a7113-712a-4940-aef9-aabd791b021d-kube-api-access-csgm4\") pod \"nova-cell1-novncproxy-0\" (UID: \"698a7113-712a-4940-aef9-aabd791b021d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:11 crc kubenswrapper[4790]: I1124 13:36:11.246198 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/698a7113-712a-4940-aef9-aabd791b021d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"698a7113-712a-4940-aef9-aabd791b021d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:11 crc kubenswrapper[4790]: I1124 13:36:11.246505 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/698a7113-712a-4940-aef9-aabd791b021d-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"698a7113-712a-4940-aef9-aabd791b021d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:11 crc kubenswrapper[4790]: I1124 13:36:11.347950 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/698a7113-712a-4940-aef9-aabd791b021d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"698a7113-712a-4940-aef9-aabd791b021d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:11 crc kubenswrapper[4790]: I1124 13:36:11.348022 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csgm4\" (UniqueName: \"kubernetes.io/projected/698a7113-712a-4940-aef9-aabd791b021d-kube-api-access-csgm4\") pod \"nova-cell1-novncproxy-0\" (UID: \"698a7113-712a-4940-aef9-aabd791b021d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:11 crc kubenswrapper[4790]: I1124 13:36:11.348075 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/698a7113-712a-4940-aef9-aabd791b021d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"698a7113-712a-4940-aef9-aabd791b021d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:11 crc kubenswrapper[4790]: I1124 13:36:11.348174 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/698a7113-712a-4940-aef9-aabd791b021d-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"698a7113-712a-4940-aef9-aabd791b021d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:11 crc kubenswrapper[4790]: I1124 13:36:11.348221 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/698a7113-712a-4940-aef9-aabd791b021d-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"698a7113-712a-4940-aef9-aabd791b021d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:11 crc kubenswrapper[4790]: I1124 13:36:11.352332 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/698a7113-712a-4940-aef9-aabd791b021d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"698a7113-712a-4940-aef9-aabd791b021d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:11 crc kubenswrapper[4790]: I1124 13:36:11.352826 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/698a7113-712a-4940-aef9-aabd791b021d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"698a7113-712a-4940-aef9-aabd791b021d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:11 crc kubenswrapper[4790]: I1124 13:36:11.352938 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/698a7113-712a-4940-aef9-aabd791b021d-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"698a7113-712a-4940-aef9-aabd791b021d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:11 crc kubenswrapper[4790]: I1124 13:36:11.352998 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/698a7113-712a-4940-aef9-aabd791b021d-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"698a7113-712a-4940-aef9-aabd791b021d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:11 crc kubenswrapper[4790]: I1124 13:36:11.369535 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csgm4\" (UniqueName: \"kubernetes.io/projected/698a7113-712a-4940-aef9-aabd791b021d-kube-api-access-csgm4\") pod \"nova-cell1-novncproxy-0\" (UID: \"698a7113-712a-4940-aef9-aabd791b021d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:11 crc kubenswrapper[4790]: I1124 13:36:11.420425 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nkvgt" Nov 24 13:36:11 crc kubenswrapper[4790]: I1124 13:36:11.420473 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nkvgt" Nov 24 13:36:11 crc kubenswrapper[4790]: I1124 13:36:11.501615 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:11 crc kubenswrapper[4790]: I1124 13:36:11.954378 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 13:36:11 crc kubenswrapper[4790]: W1124 13:36:11.958101 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod698a7113_712a_4940_aef9_aabd791b021d.slice/crio-3c27269dde176c00ab82bdbd7fd94332ddf4c94c67a1bae0d5a1cc1b7f27c4ca WatchSource:0}: Error finding container 3c27269dde176c00ab82bdbd7fd94332ddf4c94c67a1bae0d5a1cc1b7f27c4ca: Status 404 returned error can't find the container with id 3c27269dde176c00ab82bdbd7fd94332ddf4c94c67a1bae0d5a1cc1b7f27c4ca Nov 24 13:36:12 crc kubenswrapper[4790]: I1124 13:36:12.127738 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"698a7113-712a-4940-aef9-aabd791b021d","Type":"ContainerStarted","Data":"3c27269dde176c00ab82bdbd7fd94332ddf4c94c67a1bae0d5a1cc1b7f27c4ca"} Nov 24 13:36:12 crc kubenswrapper[4790]: I1124 13:36:12.324539 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0" path="/var/lib/kubelet/pods/5fa5cbcf-526b-4e11-b7aa-aa532bf7e6b0/volumes" Nov 24 13:36:12 crc kubenswrapper[4790]: I1124 13:36:12.468668 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nkvgt" podUID="099c81d1-2be3-4e10-b733-65e66912b605" containerName="registry-server" probeResult="failure" output=< Nov 24 13:36:12 crc kubenswrapper[4790]: timeout: failed to connect service ":50051" within 1s Nov 24 13:36:12 crc kubenswrapper[4790]: > Nov 24 13:36:13 crc kubenswrapper[4790]: I1124 13:36:13.136585 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"698a7113-712a-4940-aef9-aabd791b021d","Type":"ContainerStarted","Data":"a3bbb077cbcc6509c4a0bd5685f5bed2599daa1c8f9b2790688ae3b5133c5b0f"} Nov 24 13:36:13 crc kubenswrapper[4790]: I1124 13:36:13.158343 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.158305197 podStartE2EDuration="2.158305197s" podCreationTimestamp="2025-11-24 13:36:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:36:13.155386484 +0000 UTC m=+1421.535280146" watchObservedRunningTime="2025-11-24 13:36:13.158305197 +0000 UTC m=+1421.538198859" Nov 24 13:36:13 crc kubenswrapper[4790]: I1124 13:36:13.614206 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 13:36:13 crc kubenswrapper[4790]: I1124 13:36:13.614589 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 13:36:13 crc kubenswrapper[4790]: I1124 13:36:13.615843 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 13:36:13 crc kubenswrapper[4790]: I1124 13:36:13.615902 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 13:36:13 crc kubenswrapper[4790]: I1124 13:36:13.625535 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 13:36:13 crc kubenswrapper[4790]: I1124 13:36:13.625589 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 13:36:13 crc kubenswrapper[4790]: I1124 13:36:13.830389 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-48rv7"] Nov 24 13:36:13 crc kubenswrapper[4790]: I1124 13:36:13.832507 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-48rv7" Nov 24 13:36:13 crc kubenswrapper[4790]: I1124 13:36:13.868320 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-48rv7"] Nov 24 13:36:13 crc kubenswrapper[4790]: I1124 13:36:13.907548 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c428221a-e989-4844-894c-de96782508a3-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-48rv7\" (UID: \"c428221a-e989-4844-894c-de96782508a3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-48rv7" Nov 24 13:36:13 crc kubenswrapper[4790]: I1124 13:36:13.907597 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c428221a-e989-4844-894c-de96782508a3-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-48rv7\" (UID: \"c428221a-e989-4844-894c-de96782508a3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-48rv7" Nov 24 13:36:13 crc kubenswrapper[4790]: I1124 13:36:13.907718 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c428221a-e989-4844-894c-de96782508a3-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-48rv7\" (UID: \"c428221a-e989-4844-894c-de96782508a3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-48rv7" Nov 24 13:36:13 crc kubenswrapper[4790]: I1124 13:36:13.907749 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c428221a-e989-4844-894c-de96782508a3-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-48rv7\" (UID: \"c428221a-e989-4844-894c-de96782508a3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-48rv7" Nov 24 13:36:13 crc kubenswrapper[4790]: I1124 13:36:13.907770 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfbd6\" (UniqueName: \"kubernetes.io/projected/c428221a-e989-4844-894c-de96782508a3-kube-api-access-wfbd6\") pod \"dnsmasq-dns-cd5cbd7b9-48rv7\" (UID: \"c428221a-e989-4844-894c-de96782508a3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-48rv7" Nov 24 13:36:13 crc kubenswrapper[4790]: I1124 13:36:13.907811 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c428221a-e989-4844-894c-de96782508a3-config\") pod \"dnsmasq-dns-cd5cbd7b9-48rv7\" (UID: \"c428221a-e989-4844-894c-de96782508a3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-48rv7" Nov 24 13:36:13 crc kubenswrapper[4790]: I1124 13:36:13.939113 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:36:13 crc kubenswrapper[4790]: I1124 13:36:13.939179 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:36:14 crc kubenswrapper[4790]: I1124 13:36:14.009548 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c428221a-e989-4844-894c-de96782508a3-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-48rv7\" (UID: \"c428221a-e989-4844-894c-de96782508a3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-48rv7" Nov 24 13:36:14 crc kubenswrapper[4790]: I1124 13:36:14.009604 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c428221a-e989-4844-894c-de96782508a3-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-48rv7\" (UID: \"c428221a-e989-4844-894c-de96782508a3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-48rv7" Nov 24 13:36:14 crc kubenswrapper[4790]: I1124 13:36:14.009630 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfbd6\" (UniqueName: \"kubernetes.io/projected/c428221a-e989-4844-894c-de96782508a3-kube-api-access-wfbd6\") pod \"dnsmasq-dns-cd5cbd7b9-48rv7\" (UID: \"c428221a-e989-4844-894c-de96782508a3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-48rv7" Nov 24 13:36:14 crc kubenswrapper[4790]: I1124 13:36:14.009677 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c428221a-e989-4844-894c-de96782508a3-config\") pod \"dnsmasq-dns-cd5cbd7b9-48rv7\" (UID: \"c428221a-e989-4844-894c-de96782508a3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-48rv7" Nov 24 13:36:14 crc kubenswrapper[4790]: I1124 13:36:14.009712 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c428221a-e989-4844-894c-de96782508a3-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-48rv7\" (UID: \"c428221a-e989-4844-894c-de96782508a3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-48rv7" Nov 24 13:36:14 crc kubenswrapper[4790]: I1124 13:36:14.009866 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c428221a-e989-4844-894c-de96782508a3-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-48rv7\" (UID: \"c428221a-e989-4844-894c-de96782508a3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-48rv7" Nov 24 13:36:14 crc kubenswrapper[4790]: I1124 13:36:14.010473 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c428221a-e989-4844-894c-de96782508a3-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-48rv7\" (UID: \"c428221a-e989-4844-894c-de96782508a3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-48rv7" Nov 24 13:36:14 crc kubenswrapper[4790]: I1124 13:36:14.011399 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c428221a-e989-4844-894c-de96782508a3-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-48rv7\" (UID: \"c428221a-e989-4844-894c-de96782508a3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-48rv7" Nov 24 13:36:14 crc kubenswrapper[4790]: I1124 13:36:14.011434 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c428221a-e989-4844-894c-de96782508a3-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-48rv7\" (UID: \"c428221a-e989-4844-894c-de96782508a3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-48rv7" Nov 24 13:36:14 crc kubenswrapper[4790]: I1124 13:36:14.012085 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c428221a-e989-4844-894c-de96782508a3-config\") pod \"dnsmasq-dns-cd5cbd7b9-48rv7\" (UID: \"c428221a-e989-4844-894c-de96782508a3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-48rv7" Nov 24 13:36:14 crc kubenswrapper[4790]: I1124 13:36:14.012246 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c428221a-e989-4844-894c-de96782508a3-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-48rv7\" (UID: \"c428221a-e989-4844-894c-de96782508a3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-48rv7" Nov 24 13:36:14 crc kubenswrapper[4790]: I1124 13:36:14.031489 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfbd6\" (UniqueName: \"kubernetes.io/projected/c428221a-e989-4844-894c-de96782508a3-kube-api-access-wfbd6\") pod \"dnsmasq-dns-cd5cbd7b9-48rv7\" (UID: \"c428221a-e989-4844-894c-de96782508a3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-48rv7" Nov 24 13:36:14 crc kubenswrapper[4790]: I1124 13:36:14.154700 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-48rv7" Nov 24 13:36:14 crc kubenswrapper[4790]: I1124 13:36:14.660153 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-48rv7"] Nov 24 13:36:14 crc kubenswrapper[4790]: W1124 13:36:14.678007 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc428221a_e989_4844_894c_de96782508a3.slice/crio-a9e629ddc77758865107cad9072d8ed2001b03905c40bf6f5631cb19a96b4cd2 WatchSource:0}: Error finding container a9e629ddc77758865107cad9072d8ed2001b03905c40bf6f5631cb19a96b4cd2: Status 404 returned error can't find the container with id a9e629ddc77758865107cad9072d8ed2001b03905c40bf6f5631cb19a96b4cd2 Nov 24 13:36:15 crc kubenswrapper[4790]: I1124 13:36:15.154030 4790 generic.go:334] "Generic (PLEG): container finished" podID="c428221a-e989-4844-894c-de96782508a3" containerID="fc25723ed92be3d0368cbe18e81e746be59930af5dbf78c90256e70ecf2cbca4" exitCode=0 Nov 24 13:36:15 crc kubenswrapper[4790]: I1124 13:36:15.154085 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-48rv7" event={"ID":"c428221a-e989-4844-894c-de96782508a3","Type":"ContainerDied","Data":"fc25723ed92be3d0368cbe18e81e746be59930af5dbf78c90256e70ecf2cbca4"} Nov 24 13:36:15 crc kubenswrapper[4790]: I1124 13:36:15.154419 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-48rv7" event={"ID":"c428221a-e989-4844-894c-de96782508a3","Type":"ContainerStarted","Data":"a9e629ddc77758865107cad9072d8ed2001b03905c40bf6f5631cb19a96b4cd2"} Nov 24 13:36:15 crc kubenswrapper[4790]: I1124 13:36:15.801449 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:36:15 crc kubenswrapper[4790]: I1124 13:36:15.801754 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3f2d4bc2-9482-4790-beee-c221236609f6" containerName="ceilometer-central-agent" containerID="cri-o://7d2e4dbd94cd4c5d426c877165c167e175b4d892096b798b2a694a7fc6873761" gracePeriod=30 Nov 24 13:36:15 crc kubenswrapper[4790]: I1124 13:36:15.801902 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3f2d4bc2-9482-4790-beee-c221236609f6" containerName="proxy-httpd" containerID="cri-o://7d502a90c4cd6e3fcfbf940d7dcf81f76a64e192cd1d6335bd6c3fb5b94549ec" gracePeriod=30 Nov 24 13:36:15 crc kubenswrapper[4790]: I1124 13:36:15.801943 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3f2d4bc2-9482-4790-beee-c221236609f6" containerName="sg-core" containerID="cri-o://29318ee81f035c8fbbce11db915e3bd719dba88643c8a572632f173b6508e5b6" gracePeriod=30 Nov 24 13:36:15 crc kubenswrapper[4790]: I1124 13:36:15.801970 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3f2d4bc2-9482-4790-beee-c221236609f6" containerName="ceilometer-notification-agent" containerID="cri-o://86597fa39ea0ecee6911f4fc85f8531e905d1efe57b06fa776d6f51272ddf530" gracePeriod=30 Nov 24 13:36:15 crc kubenswrapper[4790]: I1124 13:36:15.823722 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="3f2d4bc2-9482-4790-beee-c221236609f6" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.192:3000/\": EOF" Nov 24 13:36:16 crc kubenswrapper[4790]: I1124 13:36:16.165429 4790 generic.go:334] "Generic (PLEG): container finished" podID="3f2d4bc2-9482-4790-beee-c221236609f6" containerID="7d502a90c4cd6e3fcfbf940d7dcf81f76a64e192cd1d6335bd6c3fb5b94549ec" exitCode=0 Nov 24 13:36:16 crc kubenswrapper[4790]: I1124 13:36:16.165462 4790 generic.go:334] "Generic (PLEG): container finished" podID="3f2d4bc2-9482-4790-beee-c221236609f6" containerID="29318ee81f035c8fbbce11db915e3bd719dba88643c8a572632f173b6508e5b6" exitCode=2 Nov 24 13:36:16 crc kubenswrapper[4790]: I1124 13:36:16.165515 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f2d4bc2-9482-4790-beee-c221236609f6","Type":"ContainerDied","Data":"7d502a90c4cd6e3fcfbf940d7dcf81f76a64e192cd1d6335bd6c3fb5b94549ec"} Nov 24 13:36:16 crc kubenswrapper[4790]: I1124 13:36:16.165545 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f2d4bc2-9482-4790-beee-c221236609f6","Type":"ContainerDied","Data":"29318ee81f035c8fbbce11db915e3bd719dba88643c8a572632f173b6508e5b6"} Nov 24 13:36:16 crc kubenswrapper[4790]: I1124 13:36:16.167711 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-48rv7" event={"ID":"c428221a-e989-4844-894c-de96782508a3","Type":"ContainerStarted","Data":"31d1bc29353608425771effac466332709c34ebd3c734237989fcb2339d732f9"} Nov 24 13:36:16 crc kubenswrapper[4790]: I1124 13:36:16.168125 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cd5cbd7b9-48rv7" Nov 24 13:36:16 crc kubenswrapper[4790]: I1124 13:36:16.206148 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cd5cbd7b9-48rv7" podStartSLOduration=3.206130014 podStartE2EDuration="3.206130014s" podCreationTimestamp="2025-11-24 13:36:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:36:16.195934583 +0000 UTC m=+1424.575828245" watchObservedRunningTime="2025-11-24 13:36:16.206130014 +0000 UTC m=+1424.586023686" Nov 24 13:36:16 crc kubenswrapper[4790]: I1124 13:36:16.501813 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:16 crc kubenswrapper[4790]: I1124 13:36:16.966489 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:36:16 crc kubenswrapper[4790]: I1124 13:36:16.966701 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="cd50520e-ec7e-48da-bb07-27c3cc3b7a1e" containerName="nova-api-log" containerID="cri-o://8d4d871f41ae1bd375fb1d586afe5870f6c6704be950bb394f58cb336951aee6" gracePeriod=30 Nov 24 13:36:16 crc kubenswrapper[4790]: I1124 13:36:16.966804 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="cd50520e-ec7e-48da-bb07-27c3cc3b7a1e" containerName="nova-api-api" containerID="cri-o://6275194871d5daca61dffd87f7b3b078bf95e6b3d5437279f936212a99985c4e" gracePeriod=30 Nov 24 13:36:17 crc kubenswrapper[4790]: I1124 13:36:17.180457 4790 generic.go:334] "Generic (PLEG): container finished" podID="3f2d4bc2-9482-4790-beee-c221236609f6" containerID="7d2e4dbd94cd4c5d426c877165c167e175b4d892096b798b2a694a7fc6873761" exitCode=0 Nov 24 13:36:17 crc kubenswrapper[4790]: I1124 13:36:17.180857 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f2d4bc2-9482-4790-beee-c221236609f6","Type":"ContainerDied","Data":"7d2e4dbd94cd4c5d426c877165c167e175b4d892096b798b2a694a7fc6873761"} Nov 24 13:36:17 crc kubenswrapper[4790]: I1124 13:36:17.182526 4790 generic.go:334] "Generic (PLEG): container finished" podID="cd50520e-ec7e-48da-bb07-27c3cc3b7a1e" containerID="8d4d871f41ae1bd375fb1d586afe5870f6c6704be950bb394f58cb336951aee6" exitCode=143 Nov 24 13:36:17 crc kubenswrapper[4790]: I1124 13:36:17.183419 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cd50520e-ec7e-48da-bb07-27c3cc3b7a1e","Type":"ContainerDied","Data":"8d4d871f41ae1bd375fb1d586afe5870f6c6704be950bb394f58cb336951aee6"} Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.220170 4790 generic.go:334] "Generic (PLEG): container finished" podID="3f2d4bc2-9482-4790-beee-c221236609f6" containerID="86597fa39ea0ecee6911f4fc85f8531e905d1efe57b06fa776d6f51272ddf530" exitCode=0 Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.220500 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f2d4bc2-9482-4790-beee-c221236609f6","Type":"ContainerDied","Data":"86597fa39ea0ecee6911f4fc85f8531e905d1efe57b06fa776d6f51272ddf530"} Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.222365 4790 generic.go:334] "Generic (PLEG): container finished" podID="cd50520e-ec7e-48da-bb07-27c3cc3b7a1e" containerID="6275194871d5daca61dffd87f7b3b078bf95e6b3d5437279f936212a99985c4e" exitCode=0 Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.222408 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cd50520e-ec7e-48da-bb07-27c3cc3b7a1e","Type":"ContainerDied","Data":"6275194871d5daca61dffd87f7b3b078bf95e6b3d5437279f936212a99985c4e"} Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.396229 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.431943 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f2d4bc2-9482-4790-beee-c221236609f6-config-data\") pod \"3f2d4bc2-9482-4790-beee-c221236609f6\" (UID: \"3f2d4bc2-9482-4790-beee-c221236609f6\") " Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.432020 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbnww\" (UniqueName: \"kubernetes.io/projected/3f2d4bc2-9482-4790-beee-c221236609f6-kube-api-access-bbnww\") pod \"3f2d4bc2-9482-4790-beee-c221236609f6\" (UID: \"3f2d4bc2-9482-4790-beee-c221236609f6\") " Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.432063 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f2d4bc2-9482-4790-beee-c221236609f6-scripts\") pod \"3f2d4bc2-9482-4790-beee-c221236609f6\" (UID: \"3f2d4bc2-9482-4790-beee-c221236609f6\") " Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.432147 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f2d4bc2-9482-4790-beee-c221236609f6-combined-ca-bundle\") pod \"3f2d4bc2-9482-4790-beee-c221236609f6\" (UID: \"3f2d4bc2-9482-4790-beee-c221236609f6\") " Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.432172 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3f2d4bc2-9482-4790-beee-c221236609f6-sg-core-conf-yaml\") pod \"3f2d4bc2-9482-4790-beee-c221236609f6\" (UID: \"3f2d4bc2-9482-4790-beee-c221236609f6\") " Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.432279 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f2d4bc2-9482-4790-beee-c221236609f6-ceilometer-tls-certs\") pod \"3f2d4bc2-9482-4790-beee-c221236609f6\" (UID: \"3f2d4bc2-9482-4790-beee-c221236609f6\") " Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.432634 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f2d4bc2-9482-4790-beee-c221236609f6-log-httpd\") pod \"3f2d4bc2-9482-4790-beee-c221236609f6\" (UID: \"3f2d4bc2-9482-4790-beee-c221236609f6\") " Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.432713 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f2d4bc2-9482-4790-beee-c221236609f6-run-httpd\") pod \"3f2d4bc2-9482-4790-beee-c221236609f6\" (UID: \"3f2d4bc2-9482-4790-beee-c221236609f6\") " Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.433142 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f2d4bc2-9482-4790-beee-c221236609f6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3f2d4bc2-9482-4790-beee-c221236609f6" (UID: "3f2d4bc2-9482-4790-beee-c221236609f6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.452110 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f2d4bc2-9482-4790-beee-c221236609f6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3f2d4bc2-9482-4790-beee-c221236609f6" (UID: "3f2d4bc2-9482-4790-beee-c221236609f6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.453331 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f2d4bc2-9482-4790-beee-c221236609f6-kube-api-access-bbnww" (OuterVolumeSpecName: "kube-api-access-bbnww") pod "3f2d4bc2-9482-4790-beee-c221236609f6" (UID: "3f2d4bc2-9482-4790-beee-c221236609f6"). InnerVolumeSpecName "kube-api-access-bbnww". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.453524 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f2d4bc2-9482-4790-beee-c221236609f6-scripts" (OuterVolumeSpecName: "scripts") pod "3f2d4bc2-9482-4790-beee-c221236609f6" (UID: "3f2d4bc2-9482-4790-beee-c221236609f6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.463961 4790 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f2d4bc2-9482-4790-beee-c221236609f6-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.528026 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f2d4bc2-9482-4790-beee-c221236609f6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3f2d4bc2-9482-4790-beee-c221236609f6" (UID: "3f2d4bc2-9482-4790-beee-c221236609f6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.543977 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.546180 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f2d4bc2-9482-4790-beee-c221236609f6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3f2d4bc2-9482-4790-beee-c221236609f6" (UID: "3f2d4bc2-9482-4790-beee-c221236609f6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.557330 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f2d4bc2-9482-4790-beee-c221236609f6-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "3f2d4bc2-9482-4790-beee-c221236609f6" (UID: "3f2d4bc2-9482-4790-beee-c221236609f6"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.568435 4790 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f2d4bc2-9482-4790-beee-c221236609f6-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.568470 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbnww\" (UniqueName: \"kubernetes.io/projected/3f2d4bc2-9482-4790-beee-c221236609f6-kube-api-access-bbnww\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.568480 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f2d4bc2-9482-4790-beee-c221236609f6-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.568490 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f2d4bc2-9482-4790-beee-c221236609f6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.568498 4790 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3f2d4bc2-9482-4790-beee-c221236609f6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.568507 4790 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f2d4bc2-9482-4790-beee-c221236609f6-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.604428 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f2d4bc2-9482-4790-beee-c221236609f6-config-data" (OuterVolumeSpecName: "config-data") pod "3f2d4bc2-9482-4790-beee-c221236609f6" (UID: "3f2d4bc2-9482-4790-beee-c221236609f6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.669917 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pc8cz\" (UniqueName: \"kubernetes.io/projected/cd50520e-ec7e-48da-bb07-27c3cc3b7a1e-kube-api-access-pc8cz\") pod \"cd50520e-ec7e-48da-bb07-27c3cc3b7a1e\" (UID: \"cd50520e-ec7e-48da-bb07-27c3cc3b7a1e\") " Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.669999 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd50520e-ec7e-48da-bb07-27c3cc3b7a1e-combined-ca-bundle\") pod \"cd50520e-ec7e-48da-bb07-27c3cc3b7a1e\" (UID: \"cd50520e-ec7e-48da-bb07-27c3cc3b7a1e\") " Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.670045 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd50520e-ec7e-48da-bb07-27c3cc3b7a1e-config-data\") pod \"cd50520e-ec7e-48da-bb07-27c3cc3b7a1e\" (UID: \"cd50520e-ec7e-48da-bb07-27c3cc3b7a1e\") " Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.670186 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd50520e-ec7e-48da-bb07-27c3cc3b7a1e-logs\") pod \"cd50520e-ec7e-48da-bb07-27c3cc3b7a1e\" (UID: \"cd50520e-ec7e-48da-bb07-27c3cc3b7a1e\") " Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.670628 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f2d4bc2-9482-4790-beee-c221236609f6-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.671101 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd50520e-ec7e-48da-bb07-27c3cc3b7a1e-logs" (OuterVolumeSpecName: "logs") pod "cd50520e-ec7e-48da-bb07-27c3cc3b7a1e" (UID: "cd50520e-ec7e-48da-bb07-27c3cc3b7a1e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.675148 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd50520e-ec7e-48da-bb07-27c3cc3b7a1e-kube-api-access-pc8cz" (OuterVolumeSpecName: "kube-api-access-pc8cz") pod "cd50520e-ec7e-48da-bb07-27c3cc3b7a1e" (UID: "cd50520e-ec7e-48da-bb07-27c3cc3b7a1e"). InnerVolumeSpecName "kube-api-access-pc8cz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.700632 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd50520e-ec7e-48da-bb07-27c3cc3b7a1e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cd50520e-ec7e-48da-bb07-27c3cc3b7a1e" (UID: "cd50520e-ec7e-48da-bb07-27c3cc3b7a1e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.710544 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd50520e-ec7e-48da-bb07-27c3cc3b7a1e-config-data" (OuterVolumeSpecName: "config-data") pod "cd50520e-ec7e-48da-bb07-27c3cc3b7a1e" (UID: "cd50520e-ec7e-48da-bb07-27c3cc3b7a1e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.772127 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd50520e-ec7e-48da-bb07-27c3cc3b7a1e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.772168 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd50520e-ec7e-48da-bb07-27c3cc3b7a1e-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.772178 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd50520e-ec7e-48da-bb07-27c3cc3b7a1e-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:20 crc kubenswrapper[4790]: I1124 13:36:20.772188 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pc8cz\" (UniqueName: \"kubernetes.io/projected/cd50520e-ec7e-48da-bb07-27c3cc3b7a1e-kube-api-access-pc8cz\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.233291 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f2d4bc2-9482-4790-beee-c221236609f6","Type":"ContainerDied","Data":"5b48cd6aa385fc09325c0d7e099958b8e93d18ecdbf32a51eb696d87d5d620e0"} Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.233372 4790 scope.go:117] "RemoveContainer" containerID="7d502a90c4cd6e3fcfbf940d7dcf81f76a64e192cd1d6335bd6c3fb5b94549ec" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.234856 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.236351 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cd50520e-ec7e-48da-bb07-27c3cc3b7a1e","Type":"ContainerDied","Data":"6ae5f0616e5d0d587636caa46f15a059dc11c0a0f4463df491907bc54aca2f75"} Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.236433 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.268321 4790 scope.go:117] "RemoveContainer" containerID="29318ee81f035c8fbbce11db915e3bd719dba88643c8a572632f173b6508e5b6" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.277248 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.289105 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.303413 4790 scope.go:117] "RemoveContainer" containerID="86597fa39ea0ecee6911f4fc85f8531e905d1efe57b06fa776d6f51272ddf530" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.304418 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.314260 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.321745 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 13:36:21 crc kubenswrapper[4790]: E1124 13:36:21.322262 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f2d4bc2-9482-4790-beee-c221236609f6" containerName="proxy-httpd" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.322287 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f2d4bc2-9482-4790-beee-c221236609f6" containerName="proxy-httpd" Nov 24 13:36:21 crc kubenswrapper[4790]: E1124 13:36:21.322338 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f2d4bc2-9482-4790-beee-c221236609f6" containerName="ceilometer-notification-agent" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.322349 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f2d4bc2-9482-4790-beee-c221236609f6" containerName="ceilometer-notification-agent" Nov 24 13:36:21 crc kubenswrapper[4790]: E1124 13:36:21.322365 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd50520e-ec7e-48da-bb07-27c3cc3b7a1e" containerName="nova-api-log" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.322373 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd50520e-ec7e-48da-bb07-27c3cc3b7a1e" containerName="nova-api-log" Nov 24 13:36:21 crc kubenswrapper[4790]: E1124 13:36:21.322383 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd50520e-ec7e-48da-bb07-27c3cc3b7a1e" containerName="nova-api-api" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.322392 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd50520e-ec7e-48da-bb07-27c3cc3b7a1e" containerName="nova-api-api" Nov 24 13:36:21 crc kubenswrapper[4790]: E1124 13:36:21.322404 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f2d4bc2-9482-4790-beee-c221236609f6" containerName="sg-core" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.322411 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f2d4bc2-9482-4790-beee-c221236609f6" containerName="sg-core" Nov 24 13:36:21 crc kubenswrapper[4790]: E1124 13:36:21.322424 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f2d4bc2-9482-4790-beee-c221236609f6" containerName="ceilometer-central-agent" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.322431 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f2d4bc2-9482-4790-beee-c221236609f6" containerName="ceilometer-central-agent" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.322642 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f2d4bc2-9482-4790-beee-c221236609f6" containerName="ceilometer-central-agent" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.322673 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f2d4bc2-9482-4790-beee-c221236609f6" containerName="proxy-httpd" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.322693 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd50520e-ec7e-48da-bb07-27c3cc3b7a1e" containerName="nova-api-api" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.322708 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f2d4bc2-9482-4790-beee-c221236609f6" containerName="sg-core" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.322721 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f2d4bc2-9482-4790-beee-c221236609f6" containerName="ceilometer-notification-agent" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.322733 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd50520e-ec7e-48da-bb07-27c3cc3b7a1e" containerName="nova-api-log" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.324086 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.331041 4790 scope.go:117] "RemoveContainer" containerID="7d2e4dbd94cd4c5d426c877165c167e175b4d892096b798b2a694a7fc6873761" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.331437 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.331645 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.331798 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.332821 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.335180 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.337402 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.337605 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.337805 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.339136 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.352101 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.366056 4790 scope.go:117] "RemoveContainer" containerID="6275194871d5daca61dffd87f7b3b078bf95e6b3d5437279f936212a99985c4e" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.383991 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7761cabb-a3e9-4b72-a562-10c066271fe0-config-data\") pod \"nova-api-0\" (UID: \"7761cabb-a3e9-4b72-a562-10c066271fe0\") " pod="openstack/nova-api-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.384033 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7761cabb-a3e9-4b72-a562-10c066271fe0-logs\") pod \"nova-api-0\" (UID: \"7761cabb-a3e9-4b72-a562-10c066271fe0\") " pod="openstack/nova-api-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.384293 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7761cabb-a3e9-4b72-a562-10c066271fe0-public-tls-certs\") pod \"nova-api-0\" (UID: \"7761cabb-a3e9-4b72-a562-10c066271fe0\") " pod="openstack/nova-api-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.384425 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7761cabb-a3e9-4b72-a562-10c066271fe0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7761cabb-a3e9-4b72-a562-10c066271fe0\") " pod="openstack/nova-api-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.384462 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rbrr\" (UniqueName: \"kubernetes.io/projected/7761cabb-a3e9-4b72-a562-10c066271fe0-kube-api-access-5rbrr\") pod \"nova-api-0\" (UID: \"7761cabb-a3e9-4b72-a562-10c066271fe0\") " pod="openstack/nova-api-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.384696 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7761cabb-a3e9-4b72-a562-10c066271fe0-internal-tls-certs\") pod \"nova-api-0\" (UID: \"7761cabb-a3e9-4b72-a562-10c066271fe0\") " pod="openstack/nova-api-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.403013 4790 scope.go:117] "RemoveContainer" containerID="8d4d871f41ae1bd375fb1d586afe5870f6c6704be950bb394f58cb336951aee6" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.487844 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9d8z\" (UniqueName: \"kubernetes.io/projected/3fc4314a-6a31-49d1-8652-f4de6da97916-kube-api-access-m9d8z\") pod \"ceilometer-0\" (UID: \"3fc4314a-6a31-49d1-8652-f4de6da97916\") " pod="openstack/ceilometer-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.488146 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3fc4314a-6a31-49d1-8652-f4de6da97916-scripts\") pod \"ceilometer-0\" (UID: \"3fc4314a-6a31-49d1-8652-f4de6da97916\") " pod="openstack/ceilometer-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.488188 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7761cabb-a3e9-4b72-a562-10c066271fe0-config-data\") pod \"nova-api-0\" (UID: \"7761cabb-a3e9-4b72-a562-10c066271fe0\") " pod="openstack/nova-api-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.488207 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7761cabb-a3e9-4b72-a562-10c066271fe0-logs\") pod \"nova-api-0\" (UID: \"7761cabb-a3e9-4b72-a562-10c066271fe0\") " pod="openstack/nova-api-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.488404 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7761cabb-a3e9-4b72-a562-10c066271fe0-public-tls-certs\") pod \"nova-api-0\" (UID: \"7761cabb-a3e9-4b72-a562-10c066271fe0\") " pod="openstack/nova-api-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.488464 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fc4314a-6a31-49d1-8652-f4de6da97916-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3fc4314a-6a31-49d1-8652-f4de6da97916\") " pod="openstack/ceilometer-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.488489 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fc4314a-6a31-49d1-8652-f4de6da97916-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3fc4314a-6a31-49d1-8652-f4de6da97916\") " pod="openstack/ceilometer-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.488544 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3fc4314a-6a31-49d1-8652-f4de6da97916-log-httpd\") pod \"ceilometer-0\" (UID: \"3fc4314a-6a31-49d1-8652-f4de6da97916\") " pod="openstack/ceilometer-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.488586 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7761cabb-a3e9-4b72-a562-10c066271fe0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7761cabb-a3e9-4b72-a562-10c066271fe0\") " pod="openstack/nova-api-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.488637 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7761cabb-a3e9-4b72-a562-10c066271fe0-logs\") pod \"nova-api-0\" (UID: \"7761cabb-a3e9-4b72-a562-10c066271fe0\") " pod="openstack/nova-api-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.488737 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rbrr\" (UniqueName: \"kubernetes.io/projected/7761cabb-a3e9-4b72-a562-10c066271fe0-kube-api-access-5rbrr\") pod \"nova-api-0\" (UID: \"7761cabb-a3e9-4b72-a562-10c066271fe0\") " pod="openstack/nova-api-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.488869 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3fc4314a-6a31-49d1-8652-f4de6da97916-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3fc4314a-6a31-49d1-8652-f4de6da97916\") " pod="openstack/ceilometer-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.488969 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7761cabb-a3e9-4b72-a562-10c066271fe0-internal-tls-certs\") pod \"nova-api-0\" (UID: \"7761cabb-a3e9-4b72-a562-10c066271fe0\") " pod="openstack/nova-api-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.489082 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fc4314a-6a31-49d1-8652-f4de6da97916-config-data\") pod \"ceilometer-0\" (UID: \"3fc4314a-6a31-49d1-8652-f4de6da97916\") " pod="openstack/ceilometer-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.489118 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3fc4314a-6a31-49d1-8652-f4de6da97916-run-httpd\") pod \"ceilometer-0\" (UID: \"3fc4314a-6a31-49d1-8652-f4de6da97916\") " pod="openstack/ceilometer-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.492484 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7761cabb-a3e9-4b72-a562-10c066271fe0-config-data\") pod \"nova-api-0\" (UID: \"7761cabb-a3e9-4b72-a562-10c066271fe0\") " pod="openstack/nova-api-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.493436 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7761cabb-a3e9-4b72-a562-10c066271fe0-internal-tls-certs\") pod \"nova-api-0\" (UID: \"7761cabb-a3e9-4b72-a562-10c066271fe0\") " pod="openstack/nova-api-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.494858 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7761cabb-a3e9-4b72-a562-10c066271fe0-public-tls-certs\") pod \"nova-api-0\" (UID: \"7761cabb-a3e9-4b72-a562-10c066271fe0\") " pod="openstack/nova-api-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.496210 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7761cabb-a3e9-4b72-a562-10c066271fe0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7761cabb-a3e9-4b72-a562-10c066271fe0\") " pod="openstack/nova-api-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.502870 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.510336 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rbrr\" (UniqueName: \"kubernetes.io/projected/7761cabb-a3e9-4b72-a562-10c066271fe0-kube-api-access-5rbrr\") pod \"nova-api-0\" (UID: \"7761cabb-a3e9-4b72-a562-10c066271fe0\") " pod="openstack/nova-api-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.526054 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.591197 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3fc4314a-6a31-49d1-8652-f4de6da97916-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3fc4314a-6a31-49d1-8652-f4de6da97916\") " pod="openstack/ceilometer-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.591306 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fc4314a-6a31-49d1-8652-f4de6da97916-config-data\") pod \"ceilometer-0\" (UID: \"3fc4314a-6a31-49d1-8652-f4de6da97916\") " pod="openstack/ceilometer-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.591331 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3fc4314a-6a31-49d1-8652-f4de6da97916-run-httpd\") pod \"ceilometer-0\" (UID: \"3fc4314a-6a31-49d1-8652-f4de6da97916\") " pod="openstack/ceilometer-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.591360 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9d8z\" (UniqueName: \"kubernetes.io/projected/3fc4314a-6a31-49d1-8652-f4de6da97916-kube-api-access-m9d8z\") pod \"ceilometer-0\" (UID: \"3fc4314a-6a31-49d1-8652-f4de6da97916\") " pod="openstack/ceilometer-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.591382 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3fc4314a-6a31-49d1-8652-f4de6da97916-scripts\") pod \"ceilometer-0\" (UID: \"3fc4314a-6a31-49d1-8652-f4de6da97916\") " pod="openstack/ceilometer-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.591469 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fc4314a-6a31-49d1-8652-f4de6da97916-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3fc4314a-6a31-49d1-8652-f4de6da97916\") " pod="openstack/ceilometer-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.591486 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fc4314a-6a31-49d1-8652-f4de6da97916-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3fc4314a-6a31-49d1-8652-f4de6da97916\") " pod="openstack/ceilometer-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.591518 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3fc4314a-6a31-49d1-8652-f4de6da97916-log-httpd\") pod \"ceilometer-0\" (UID: \"3fc4314a-6a31-49d1-8652-f4de6da97916\") " pod="openstack/ceilometer-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.591941 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3fc4314a-6a31-49d1-8652-f4de6da97916-log-httpd\") pod \"ceilometer-0\" (UID: \"3fc4314a-6a31-49d1-8652-f4de6da97916\") " pod="openstack/ceilometer-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.594130 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3fc4314a-6a31-49d1-8652-f4de6da97916-run-httpd\") pod \"ceilometer-0\" (UID: \"3fc4314a-6a31-49d1-8652-f4de6da97916\") " pod="openstack/ceilometer-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.595844 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3fc4314a-6a31-49d1-8652-f4de6da97916-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3fc4314a-6a31-49d1-8652-f4de6da97916\") " pod="openstack/ceilometer-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.596705 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fc4314a-6a31-49d1-8652-f4de6da97916-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3fc4314a-6a31-49d1-8652-f4de6da97916\") " pod="openstack/ceilometer-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.597101 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fc4314a-6a31-49d1-8652-f4de6da97916-config-data\") pod \"ceilometer-0\" (UID: \"3fc4314a-6a31-49d1-8652-f4de6da97916\") " pod="openstack/ceilometer-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.597159 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3fc4314a-6a31-49d1-8652-f4de6da97916-scripts\") pod \"ceilometer-0\" (UID: \"3fc4314a-6a31-49d1-8652-f4de6da97916\") " pod="openstack/ceilometer-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.597378 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fc4314a-6a31-49d1-8652-f4de6da97916-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3fc4314a-6a31-49d1-8652-f4de6da97916\") " pod="openstack/ceilometer-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.614748 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9d8z\" (UniqueName: \"kubernetes.io/projected/3fc4314a-6a31-49d1-8652-f4de6da97916-kube-api-access-m9d8z\") pod \"ceilometer-0\" (UID: \"3fc4314a-6a31-49d1-8652-f4de6da97916\") " pod="openstack/ceilometer-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.663860 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 13:36:21 crc kubenswrapper[4790]: I1124 13:36:21.676697 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:36:22 crc kubenswrapper[4790]: I1124 13:36:22.119556 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:36:22 crc kubenswrapper[4790]: W1124 13:36:22.121145 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3fc4314a_6a31_49d1_8652_f4de6da97916.slice/crio-e8a1a03fbff93f503eaa9aa5d709b529d2dbba23cfdbd560e8362b99eb1a00ea WatchSource:0}: Error finding container e8a1a03fbff93f503eaa9aa5d709b529d2dbba23cfdbd560e8362b99eb1a00ea: Status 404 returned error can't find the container with id e8a1a03fbff93f503eaa9aa5d709b529d2dbba23cfdbd560e8362b99eb1a00ea Nov 24 13:36:22 crc kubenswrapper[4790]: I1124 13:36:22.200270 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:36:22 crc kubenswrapper[4790]: W1124 13:36:22.204110 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7761cabb_a3e9_4b72_a562_10c066271fe0.slice/crio-556b986f67dfb66817f056f99b2b2d2f6aa25342ba76083de031757f163731ee WatchSource:0}: Error finding container 556b986f67dfb66817f056f99b2b2d2f6aa25342ba76083de031757f163731ee: Status 404 returned error can't find the container with id 556b986f67dfb66817f056f99b2b2d2f6aa25342ba76083de031757f163731ee Nov 24 13:36:22 crc kubenswrapper[4790]: I1124 13:36:22.250569 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3fc4314a-6a31-49d1-8652-f4de6da97916","Type":"ContainerStarted","Data":"e8a1a03fbff93f503eaa9aa5d709b529d2dbba23cfdbd560e8362b99eb1a00ea"} Nov 24 13:36:22 crc kubenswrapper[4790]: I1124 13:36:22.252306 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7761cabb-a3e9-4b72-a562-10c066271fe0","Type":"ContainerStarted","Data":"556b986f67dfb66817f056f99b2b2d2f6aa25342ba76083de031757f163731ee"} Nov 24 13:36:22 crc kubenswrapper[4790]: I1124 13:36:22.275155 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:22 crc kubenswrapper[4790]: I1124 13:36:22.363273 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f2d4bc2-9482-4790-beee-c221236609f6" path="/var/lib/kubelet/pods/3f2d4bc2-9482-4790-beee-c221236609f6/volumes" Nov 24 13:36:22 crc kubenswrapper[4790]: I1124 13:36:22.364234 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd50520e-ec7e-48da-bb07-27c3cc3b7a1e" path="/var/lib/kubelet/pods/cd50520e-ec7e-48da-bb07-27c3cc3b7a1e/volumes" Nov 24 13:36:22 crc kubenswrapper[4790]: I1124 13:36:22.479964 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nkvgt" podUID="099c81d1-2be3-4e10-b733-65e66912b605" containerName="registry-server" probeResult="failure" output=< Nov 24 13:36:22 crc kubenswrapper[4790]: timeout: failed to connect service ":50051" within 1s Nov 24 13:36:22 crc kubenswrapper[4790]: > Nov 24 13:36:22 crc kubenswrapper[4790]: I1124 13:36:22.555470 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-ckhz8"] Nov 24 13:36:22 crc kubenswrapper[4790]: I1124 13:36:22.556721 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-ckhz8" Nov 24 13:36:22 crc kubenswrapper[4790]: I1124 13:36:22.560557 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 24 13:36:22 crc kubenswrapper[4790]: I1124 13:36:22.562181 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 24 13:36:22 crc kubenswrapper[4790]: I1124 13:36:22.588646 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-ckhz8"] Nov 24 13:36:22 crc kubenswrapper[4790]: I1124 13:36:22.610109 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5d01a68-e29a-44b3-93e6-f35b91673300-config-data\") pod \"nova-cell1-cell-mapping-ckhz8\" (UID: \"d5d01a68-e29a-44b3-93e6-f35b91673300\") " pod="openstack/nova-cell1-cell-mapping-ckhz8" Nov 24 13:36:22 crc kubenswrapper[4790]: I1124 13:36:22.610189 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5d01a68-e29a-44b3-93e6-f35b91673300-scripts\") pod \"nova-cell1-cell-mapping-ckhz8\" (UID: \"d5d01a68-e29a-44b3-93e6-f35b91673300\") " pod="openstack/nova-cell1-cell-mapping-ckhz8" Nov 24 13:36:22 crc kubenswrapper[4790]: I1124 13:36:22.610222 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdwzm\" (UniqueName: \"kubernetes.io/projected/d5d01a68-e29a-44b3-93e6-f35b91673300-kube-api-access-hdwzm\") pod \"nova-cell1-cell-mapping-ckhz8\" (UID: \"d5d01a68-e29a-44b3-93e6-f35b91673300\") " pod="openstack/nova-cell1-cell-mapping-ckhz8" Nov 24 13:36:22 crc kubenswrapper[4790]: I1124 13:36:22.610273 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5d01a68-e29a-44b3-93e6-f35b91673300-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-ckhz8\" (UID: \"d5d01a68-e29a-44b3-93e6-f35b91673300\") " pod="openstack/nova-cell1-cell-mapping-ckhz8" Nov 24 13:36:22 crc kubenswrapper[4790]: I1124 13:36:22.712902 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdwzm\" (UniqueName: \"kubernetes.io/projected/d5d01a68-e29a-44b3-93e6-f35b91673300-kube-api-access-hdwzm\") pod \"nova-cell1-cell-mapping-ckhz8\" (UID: \"d5d01a68-e29a-44b3-93e6-f35b91673300\") " pod="openstack/nova-cell1-cell-mapping-ckhz8" Nov 24 13:36:22 crc kubenswrapper[4790]: I1124 13:36:22.712978 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5d01a68-e29a-44b3-93e6-f35b91673300-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-ckhz8\" (UID: \"d5d01a68-e29a-44b3-93e6-f35b91673300\") " pod="openstack/nova-cell1-cell-mapping-ckhz8" Nov 24 13:36:22 crc kubenswrapper[4790]: I1124 13:36:22.713071 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5d01a68-e29a-44b3-93e6-f35b91673300-config-data\") pod \"nova-cell1-cell-mapping-ckhz8\" (UID: \"d5d01a68-e29a-44b3-93e6-f35b91673300\") " pod="openstack/nova-cell1-cell-mapping-ckhz8" Nov 24 13:36:22 crc kubenswrapper[4790]: I1124 13:36:22.714094 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5d01a68-e29a-44b3-93e6-f35b91673300-scripts\") pod \"nova-cell1-cell-mapping-ckhz8\" (UID: \"d5d01a68-e29a-44b3-93e6-f35b91673300\") " pod="openstack/nova-cell1-cell-mapping-ckhz8" Nov 24 13:36:22 crc kubenswrapper[4790]: I1124 13:36:22.724848 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5d01a68-e29a-44b3-93e6-f35b91673300-scripts\") pod \"nova-cell1-cell-mapping-ckhz8\" (UID: \"d5d01a68-e29a-44b3-93e6-f35b91673300\") " pod="openstack/nova-cell1-cell-mapping-ckhz8" Nov 24 13:36:22 crc kubenswrapper[4790]: I1124 13:36:22.724991 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5d01a68-e29a-44b3-93e6-f35b91673300-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-ckhz8\" (UID: \"d5d01a68-e29a-44b3-93e6-f35b91673300\") " pod="openstack/nova-cell1-cell-mapping-ckhz8" Nov 24 13:36:22 crc kubenswrapper[4790]: I1124 13:36:22.725453 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5d01a68-e29a-44b3-93e6-f35b91673300-config-data\") pod \"nova-cell1-cell-mapping-ckhz8\" (UID: \"d5d01a68-e29a-44b3-93e6-f35b91673300\") " pod="openstack/nova-cell1-cell-mapping-ckhz8" Nov 24 13:36:22 crc kubenswrapper[4790]: I1124 13:36:22.736476 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdwzm\" (UniqueName: \"kubernetes.io/projected/d5d01a68-e29a-44b3-93e6-f35b91673300-kube-api-access-hdwzm\") pod \"nova-cell1-cell-mapping-ckhz8\" (UID: \"d5d01a68-e29a-44b3-93e6-f35b91673300\") " pod="openstack/nova-cell1-cell-mapping-ckhz8" Nov 24 13:36:22 crc kubenswrapper[4790]: I1124 13:36:22.888752 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-ckhz8" Nov 24 13:36:23 crc kubenswrapper[4790]: I1124 13:36:23.262593 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3fc4314a-6a31-49d1-8652-f4de6da97916","Type":"ContainerStarted","Data":"63bc861e59ed27e8b2dec969395af49861e48d187c85af757c5022b25349826a"} Nov 24 13:36:23 crc kubenswrapper[4790]: I1124 13:36:23.265621 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7761cabb-a3e9-4b72-a562-10c066271fe0","Type":"ContainerStarted","Data":"bf087df426faea9131fd7a339ec6ef0fa4f85c11747061f4d0231a6ff8d3aa23"} Nov 24 13:36:23 crc kubenswrapper[4790]: I1124 13:36:23.265700 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7761cabb-a3e9-4b72-a562-10c066271fe0","Type":"ContainerStarted","Data":"cfc9c08d243310fe7960670de9b3e2278cc8874acd78a44359837cb3dcdc86b0"} Nov 24 13:36:23 crc kubenswrapper[4790]: I1124 13:36:23.303331 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.303308844 podStartE2EDuration="2.303308844s" podCreationTimestamp="2025-11-24 13:36:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:36:23.285632159 +0000 UTC m=+1431.665525831" watchObservedRunningTime="2025-11-24 13:36:23.303308844 +0000 UTC m=+1431.683202506" Nov 24 13:36:23 crc kubenswrapper[4790]: I1124 13:36:23.347902 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-ckhz8"] Nov 24 13:36:23 crc kubenswrapper[4790]: W1124 13:36:23.350226 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd5d01a68_e29a_44b3_93e6_f35b91673300.slice/crio-d26f5fff446e51e69c6d4583c5b52b41372bc5e95e8504385d56b3b761dfa9a8 WatchSource:0}: Error finding container d26f5fff446e51e69c6d4583c5b52b41372bc5e95e8504385d56b3b761dfa9a8: Status 404 returned error can't find the container with id d26f5fff446e51e69c6d4583c5b52b41372bc5e95e8504385d56b3b761dfa9a8 Nov 24 13:36:24 crc kubenswrapper[4790]: I1124 13:36:24.156031 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cd5cbd7b9-48rv7" Nov 24 13:36:24 crc kubenswrapper[4790]: I1124 13:36:24.226955 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-skvpz"] Nov 24 13:36:24 crc kubenswrapper[4790]: I1124 13:36:24.227520 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bccf8f775-skvpz" podUID="6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412" containerName="dnsmasq-dns" containerID="cri-o://56bf93c28a64fc47955b7be4f3cb20d93bfc8f764601c7c707b0215042bde17d" gracePeriod=10 Nov 24 13:36:24 crc kubenswrapper[4790]: I1124 13:36:24.275707 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-ckhz8" event={"ID":"d5d01a68-e29a-44b3-93e6-f35b91673300","Type":"ContainerStarted","Data":"2461d3602b2db4efea3d420eb4e138e451c6048fe390dc3d4b479c8e5a584628"} Nov 24 13:36:24 crc kubenswrapper[4790]: I1124 13:36:24.275750 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-ckhz8" event={"ID":"d5d01a68-e29a-44b3-93e6-f35b91673300","Type":"ContainerStarted","Data":"d26f5fff446e51e69c6d4583c5b52b41372bc5e95e8504385d56b3b761dfa9a8"} Nov 24 13:36:24 crc kubenswrapper[4790]: I1124 13:36:24.281366 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3fc4314a-6a31-49d1-8652-f4de6da97916","Type":"ContainerStarted","Data":"b5e492786f34b12c4fccd95cb12360cbb4386005a0414f30c8e1be783d5252e0"} Nov 24 13:36:24 crc kubenswrapper[4790]: I1124 13:36:24.281403 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3fc4314a-6a31-49d1-8652-f4de6da97916","Type":"ContainerStarted","Data":"593dfb8aab769d099082292784d20c9dd7fc610d3ac9c0659d6fd1635968dfa7"} Nov 24 13:36:24 crc kubenswrapper[4790]: I1124 13:36:24.300893 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-ckhz8" podStartSLOduration=2.300857248 podStartE2EDuration="2.300857248s" podCreationTimestamp="2025-11-24 13:36:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:36:24.292198461 +0000 UTC m=+1432.672092123" watchObservedRunningTime="2025-11-24 13:36:24.300857248 +0000 UTC m=+1432.680750920" Nov 24 13:36:24 crc kubenswrapper[4790]: I1124 13:36:24.707158 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-skvpz" Nov 24 13:36:24 crc kubenswrapper[4790]: I1124 13:36:24.786118 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-dns-svc\") pod \"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412\" (UID: \"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412\") " Nov 24 13:36:24 crc kubenswrapper[4790]: I1124 13:36:24.786185 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-config\") pod \"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412\" (UID: \"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412\") " Nov 24 13:36:24 crc kubenswrapper[4790]: I1124 13:36:24.786237 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-ovsdbserver-sb\") pod \"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412\" (UID: \"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412\") " Nov 24 13:36:24 crc kubenswrapper[4790]: I1124 13:36:24.786283 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-ovsdbserver-nb\") pod \"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412\" (UID: \"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412\") " Nov 24 13:36:24 crc kubenswrapper[4790]: I1124 13:36:24.786367 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-dns-swift-storage-0\") pod \"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412\" (UID: \"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412\") " Nov 24 13:36:24 crc kubenswrapper[4790]: I1124 13:36:24.787038 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzk5l\" (UniqueName: \"kubernetes.io/projected/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-kube-api-access-gzk5l\") pod \"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412\" (UID: \"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412\") " Nov 24 13:36:24 crc kubenswrapper[4790]: I1124 13:36:24.791198 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-kube-api-access-gzk5l" (OuterVolumeSpecName: "kube-api-access-gzk5l") pod "6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412" (UID: "6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412"). InnerVolumeSpecName "kube-api-access-gzk5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:36:24 crc kubenswrapper[4790]: I1124 13:36:24.846792 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412" (UID: "6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:36:24 crc kubenswrapper[4790]: I1124 13:36:24.851820 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412" (UID: "6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:36:24 crc kubenswrapper[4790]: I1124 13:36:24.854277 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-config" (OuterVolumeSpecName: "config") pod "6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412" (UID: "6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:36:24 crc kubenswrapper[4790]: I1124 13:36:24.862616 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412" (UID: "6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:36:24 crc kubenswrapper[4790]: I1124 13:36:24.869487 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412" (UID: "6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:36:24 crc kubenswrapper[4790]: I1124 13:36:24.888506 4790 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:24 crc kubenswrapper[4790]: I1124 13:36:24.888537 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzk5l\" (UniqueName: \"kubernetes.io/projected/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-kube-api-access-gzk5l\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:24 crc kubenswrapper[4790]: I1124 13:36:24.888548 4790 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:24 crc kubenswrapper[4790]: I1124 13:36:24.888556 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:24 crc kubenswrapper[4790]: I1124 13:36:24.888565 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:24 crc kubenswrapper[4790]: I1124 13:36:24.888574 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:25 crc kubenswrapper[4790]: I1124 13:36:25.291270 4790 generic.go:334] "Generic (PLEG): container finished" podID="6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412" containerID="56bf93c28a64fc47955b7be4f3cb20d93bfc8f764601c7c707b0215042bde17d" exitCode=0 Nov 24 13:36:25 crc kubenswrapper[4790]: I1124 13:36:25.291317 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-skvpz" event={"ID":"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412","Type":"ContainerDied","Data":"56bf93c28a64fc47955b7be4f3cb20d93bfc8f764601c7c707b0215042bde17d"} Nov 24 13:36:25 crc kubenswrapper[4790]: I1124 13:36:25.291342 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-skvpz" Nov 24 13:36:25 crc kubenswrapper[4790]: I1124 13:36:25.291362 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-skvpz" event={"ID":"6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412","Type":"ContainerDied","Data":"0d68d75d9bcf475988f9548ceb92947f84704648eeb7fb199bb0afbcefa283a3"} Nov 24 13:36:25 crc kubenswrapper[4790]: I1124 13:36:25.291384 4790 scope.go:117] "RemoveContainer" containerID="56bf93c28a64fc47955b7be4f3cb20d93bfc8f764601c7c707b0215042bde17d" Nov 24 13:36:25 crc kubenswrapper[4790]: I1124 13:36:25.319863 4790 scope.go:117] "RemoveContainer" containerID="1239012cbfa2d3cd8a5ff484e5663a42f10626d77202e5759db80d62f1dc224f" Nov 24 13:36:25 crc kubenswrapper[4790]: I1124 13:36:25.338240 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-skvpz"] Nov 24 13:36:25 crc kubenswrapper[4790]: I1124 13:36:25.345932 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-skvpz"] Nov 24 13:36:25 crc kubenswrapper[4790]: I1124 13:36:25.353909 4790 scope.go:117] "RemoveContainer" containerID="56bf93c28a64fc47955b7be4f3cb20d93bfc8f764601c7c707b0215042bde17d" Nov 24 13:36:25 crc kubenswrapper[4790]: E1124 13:36:25.354311 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56bf93c28a64fc47955b7be4f3cb20d93bfc8f764601c7c707b0215042bde17d\": container with ID starting with 56bf93c28a64fc47955b7be4f3cb20d93bfc8f764601c7c707b0215042bde17d not found: ID does not exist" containerID="56bf93c28a64fc47955b7be4f3cb20d93bfc8f764601c7c707b0215042bde17d" Nov 24 13:36:25 crc kubenswrapper[4790]: I1124 13:36:25.354345 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56bf93c28a64fc47955b7be4f3cb20d93bfc8f764601c7c707b0215042bde17d"} err="failed to get container status \"56bf93c28a64fc47955b7be4f3cb20d93bfc8f764601c7c707b0215042bde17d\": rpc error: code = NotFound desc = could not find container \"56bf93c28a64fc47955b7be4f3cb20d93bfc8f764601c7c707b0215042bde17d\": container with ID starting with 56bf93c28a64fc47955b7be4f3cb20d93bfc8f764601c7c707b0215042bde17d not found: ID does not exist" Nov 24 13:36:25 crc kubenswrapper[4790]: I1124 13:36:25.354372 4790 scope.go:117] "RemoveContainer" containerID="1239012cbfa2d3cd8a5ff484e5663a42f10626d77202e5759db80d62f1dc224f" Nov 24 13:36:25 crc kubenswrapper[4790]: E1124 13:36:25.354591 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1239012cbfa2d3cd8a5ff484e5663a42f10626d77202e5759db80d62f1dc224f\": container with ID starting with 1239012cbfa2d3cd8a5ff484e5663a42f10626d77202e5759db80d62f1dc224f not found: ID does not exist" containerID="1239012cbfa2d3cd8a5ff484e5663a42f10626d77202e5759db80d62f1dc224f" Nov 24 13:36:25 crc kubenswrapper[4790]: I1124 13:36:25.354615 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1239012cbfa2d3cd8a5ff484e5663a42f10626d77202e5759db80d62f1dc224f"} err="failed to get container status \"1239012cbfa2d3cd8a5ff484e5663a42f10626d77202e5759db80d62f1dc224f\": rpc error: code = NotFound desc = could not find container \"1239012cbfa2d3cd8a5ff484e5663a42f10626d77202e5759db80d62f1dc224f\": container with ID starting with 1239012cbfa2d3cd8a5ff484e5663a42f10626d77202e5759db80d62f1dc224f not found: ID does not exist" Nov 24 13:36:26 crc kubenswrapper[4790]: I1124 13:36:26.308776 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3fc4314a-6a31-49d1-8652-f4de6da97916","Type":"ContainerStarted","Data":"3bdd8dbb30e64b7458fddef45148be02e9280fe559bf6018d442a5bf9eb8dada"} Nov 24 13:36:26 crc kubenswrapper[4790]: I1124 13:36:26.309171 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 13:36:26 crc kubenswrapper[4790]: I1124 13:36:26.334220 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412" path="/var/lib/kubelet/pods/6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412/volumes" Nov 24 13:36:26 crc kubenswrapper[4790]: I1124 13:36:26.334404 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.984302309 podStartE2EDuration="5.334382762s" podCreationTimestamp="2025-11-24 13:36:21 +0000 UTC" firstStartedPulling="2025-11-24 13:36:22.123158793 +0000 UTC m=+1430.503052455" lastFinishedPulling="2025-11-24 13:36:25.473239246 +0000 UTC m=+1433.853132908" observedRunningTime="2025-11-24 13:36:26.328660669 +0000 UTC m=+1434.708554321" watchObservedRunningTime="2025-11-24 13:36:26.334382762 +0000 UTC m=+1434.714276424" Nov 24 13:36:29 crc kubenswrapper[4790]: I1124 13:36:29.335834 4790 generic.go:334] "Generic (PLEG): container finished" podID="d5d01a68-e29a-44b3-93e6-f35b91673300" containerID="2461d3602b2db4efea3d420eb4e138e451c6048fe390dc3d4b479c8e5a584628" exitCode=0 Nov 24 13:36:29 crc kubenswrapper[4790]: I1124 13:36:29.335910 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-ckhz8" event={"ID":"d5d01a68-e29a-44b3-93e6-f35b91673300","Type":"ContainerDied","Data":"2461d3602b2db4efea3d420eb4e138e451c6048fe390dc3d4b479c8e5a584628"} Nov 24 13:36:30 crc kubenswrapper[4790]: I1124 13:36:30.733285 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-ckhz8" Nov 24 13:36:30 crc kubenswrapper[4790]: I1124 13:36:30.801870 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdwzm\" (UniqueName: \"kubernetes.io/projected/d5d01a68-e29a-44b3-93e6-f35b91673300-kube-api-access-hdwzm\") pod \"d5d01a68-e29a-44b3-93e6-f35b91673300\" (UID: \"d5d01a68-e29a-44b3-93e6-f35b91673300\") " Nov 24 13:36:30 crc kubenswrapper[4790]: I1124 13:36:30.802029 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5d01a68-e29a-44b3-93e6-f35b91673300-scripts\") pod \"d5d01a68-e29a-44b3-93e6-f35b91673300\" (UID: \"d5d01a68-e29a-44b3-93e6-f35b91673300\") " Nov 24 13:36:30 crc kubenswrapper[4790]: I1124 13:36:30.802054 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5d01a68-e29a-44b3-93e6-f35b91673300-combined-ca-bundle\") pod \"d5d01a68-e29a-44b3-93e6-f35b91673300\" (UID: \"d5d01a68-e29a-44b3-93e6-f35b91673300\") " Nov 24 13:36:30 crc kubenswrapper[4790]: I1124 13:36:30.802218 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5d01a68-e29a-44b3-93e6-f35b91673300-config-data\") pod \"d5d01a68-e29a-44b3-93e6-f35b91673300\" (UID: \"d5d01a68-e29a-44b3-93e6-f35b91673300\") " Nov 24 13:36:30 crc kubenswrapper[4790]: I1124 13:36:30.807594 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5d01a68-e29a-44b3-93e6-f35b91673300-kube-api-access-hdwzm" (OuterVolumeSpecName: "kube-api-access-hdwzm") pod "d5d01a68-e29a-44b3-93e6-f35b91673300" (UID: "d5d01a68-e29a-44b3-93e6-f35b91673300"). InnerVolumeSpecName "kube-api-access-hdwzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:36:30 crc kubenswrapper[4790]: I1124 13:36:30.807841 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5d01a68-e29a-44b3-93e6-f35b91673300-scripts" (OuterVolumeSpecName: "scripts") pod "d5d01a68-e29a-44b3-93e6-f35b91673300" (UID: "d5d01a68-e29a-44b3-93e6-f35b91673300"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:30 crc kubenswrapper[4790]: E1124 13:36:30.840658 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d5d01a68-e29a-44b3-93e6-f35b91673300-combined-ca-bundle podName:d5d01a68-e29a-44b3-93e6-f35b91673300 nodeName:}" failed. No retries permitted until 2025-11-24 13:36:31.340633611 +0000 UTC m=+1439.720527273 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "combined-ca-bundle" (UniqueName: "kubernetes.io/secret/d5d01a68-e29a-44b3-93e6-f35b91673300-combined-ca-bundle") pod "d5d01a68-e29a-44b3-93e6-f35b91673300" (UID: "d5d01a68-e29a-44b3-93e6-f35b91673300") : error deleting /var/lib/kubelet/pods/d5d01a68-e29a-44b3-93e6-f35b91673300/volume-subpaths: remove /var/lib/kubelet/pods/d5d01a68-e29a-44b3-93e6-f35b91673300/volume-subpaths: no such file or directory Nov 24 13:36:30 crc kubenswrapper[4790]: I1124 13:36:30.843609 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5d01a68-e29a-44b3-93e6-f35b91673300-config-data" (OuterVolumeSpecName: "config-data") pod "d5d01a68-e29a-44b3-93e6-f35b91673300" (UID: "d5d01a68-e29a-44b3-93e6-f35b91673300"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:30 crc kubenswrapper[4790]: I1124 13:36:30.904221 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdwzm\" (UniqueName: \"kubernetes.io/projected/d5d01a68-e29a-44b3-93e6-f35b91673300-kube-api-access-hdwzm\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:30 crc kubenswrapper[4790]: I1124 13:36:30.904257 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5d01a68-e29a-44b3-93e6-f35b91673300-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:30 crc kubenswrapper[4790]: I1124 13:36:30.904266 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5d01a68-e29a-44b3-93e6-f35b91673300-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:31 crc kubenswrapper[4790]: I1124 13:36:31.380335 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-ckhz8" event={"ID":"d5d01a68-e29a-44b3-93e6-f35b91673300","Type":"ContainerDied","Data":"d26f5fff446e51e69c6d4583c5b52b41372bc5e95e8504385d56b3b761dfa9a8"} Nov 24 13:36:31 crc kubenswrapper[4790]: I1124 13:36:31.380675 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d26f5fff446e51e69c6d4583c5b52b41372bc5e95e8504385d56b3b761dfa9a8" Nov 24 13:36:31 crc kubenswrapper[4790]: I1124 13:36:31.380377 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-ckhz8" Nov 24 13:36:31 crc kubenswrapper[4790]: I1124 13:36:31.413049 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5d01a68-e29a-44b3-93e6-f35b91673300-combined-ca-bundle\") pod \"d5d01a68-e29a-44b3-93e6-f35b91673300\" (UID: \"d5d01a68-e29a-44b3-93e6-f35b91673300\") " Nov 24 13:36:31 crc kubenswrapper[4790]: I1124 13:36:31.418626 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5d01a68-e29a-44b3-93e6-f35b91673300-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d5d01a68-e29a-44b3-93e6-f35b91673300" (UID: "d5d01a68-e29a-44b3-93e6-f35b91673300"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:31 crc kubenswrapper[4790]: I1124 13:36:31.508064 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:36:31 crc kubenswrapper[4790]: I1124 13:36:31.508390 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7761cabb-a3e9-4b72-a562-10c066271fe0" containerName="nova-api-log" containerID="cri-o://cfc9c08d243310fe7960670de9b3e2278cc8874acd78a44359837cb3dcdc86b0" gracePeriod=30 Nov 24 13:36:31 crc kubenswrapper[4790]: I1124 13:36:31.508451 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7761cabb-a3e9-4b72-a562-10c066271fe0" containerName="nova-api-api" containerID="cri-o://bf087df426faea9131fd7a339ec6ef0fa4f85c11747061f4d0231a6ff8d3aa23" gracePeriod=30 Nov 24 13:36:31 crc kubenswrapper[4790]: I1124 13:36:31.516961 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5d01a68-e29a-44b3-93e6-f35b91673300-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:31 crc kubenswrapper[4790]: I1124 13:36:31.521652 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 13:36:31 crc kubenswrapper[4790]: I1124 13:36:31.521933 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="c864a2b1-1693-4200-be70-a140b1526bac" containerName="nova-scheduler-scheduler" containerID="cri-o://285bb94beb2c560eb1cadca29c368e78f28a1d7b266ed21f459cb58d61a3f93b" gracePeriod=30 Nov 24 13:36:31 crc kubenswrapper[4790]: I1124 13:36:31.555002 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:36:31 crc kubenswrapper[4790]: I1124 13:36:31.555261 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="65369c6a-c495-49d7-829f-e9adb21172ff" containerName="nova-metadata-log" containerID="cri-o://28ce7a4d4c4dc879f7be8c9eb8f16679b7296eea8d6956705834a8e39f388e31" gracePeriod=30 Nov 24 13:36:31 crc kubenswrapper[4790]: I1124 13:36:31.555749 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="65369c6a-c495-49d7-829f-e9adb21172ff" containerName="nova-metadata-metadata" containerID="cri-o://d5cc8dc52fd100b95d8a516cbdf245be0d1ffd7c60cfbbaee80ea00d834358ca" gracePeriod=30 Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.180429 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.229015 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7761cabb-a3e9-4b72-a562-10c066271fe0-config-data\") pod \"7761cabb-a3e9-4b72-a562-10c066271fe0\" (UID: \"7761cabb-a3e9-4b72-a562-10c066271fe0\") " Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.229146 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7761cabb-a3e9-4b72-a562-10c066271fe0-logs\") pod \"7761cabb-a3e9-4b72-a562-10c066271fe0\" (UID: \"7761cabb-a3e9-4b72-a562-10c066271fe0\") " Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.229620 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7761cabb-a3e9-4b72-a562-10c066271fe0-logs" (OuterVolumeSpecName: "logs") pod "7761cabb-a3e9-4b72-a562-10c066271fe0" (UID: "7761cabb-a3e9-4b72-a562-10c066271fe0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.229741 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rbrr\" (UniqueName: \"kubernetes.io/projected/7761cabb-a3e9-4b72-a562-10c066271fe0-kube-api-access-5rbrr\") pod \"7761cabb-a3e9-4b72-a562-10c066271fe0\" (UID: \"7761cabb-a3e9-4b72-a562-10c066271fe0\") " Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.230207 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7761cabb-a3e9-4b72-a562-10c066271fe0-internal-tls-certs\") pod \"7761cabb-a3e9-4b72-a562-10c066271fe0\" (UID: \"7761cabb-a3e9-4b72-a562-10c066271fe0\") " Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.230239 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7761cabb-a3e9-4b72-a562-10c066271fe0-combined-ca-bundle\") pod \"7761cabb-a3e9-4b72-a562-10c066271fe0\" (UID: \"7761cabb-a3e9-4b72-a562-10c066271fe0\") " Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.230275 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7761cabb-a3e9-4b72-a562-10c066271fe0-public-tls-certs\") pod \"7761cabb-a3e9-4b72-a562-10c066271fe0\" (UID: \"7761cabb-a3e9-4b72-a562-10c066271fe0\") " Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.230804 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7761cabb-a3e9-4b72-a562-10c066271fe0-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.234677 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7761cabb-a3e9-4b72-a562-10c066271fe0-kube-api-access-5rbrr" (OuterVolumeSpecName: "kube-api-access-5rbrr") pod "7761cabb-a3e9-4b72-a562-10c066271fe0" (UID: "7761cabb-a3e9-4b72-a562-10c066271fe0"). InnerVolumeSpecName "kube-api-access-5rbrr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.259203 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7761cabb-a3e9-4b72-a562-10c066271fe0-config-data" (OuterVolumeSpecName: "config-data") pod "7761cabb-a3e9-4b72-a562-10c066271fe0" (UID: "7761cabb-a3e9-4b72-a562-10c066271fe0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.259197 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7761cabb-a3e9-4b72-a562-10c066271fe0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7761cabb-a3e9-4b72-a562-10c066271fe0" (UID: "7761cabb-a3e9-4b72-a562-10c066271fe0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.285687 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7761cabb-a3e9-4b72-a562-10c066271fe0-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "7761cabb-a3e9-4b72-a562-10c066271fe0" (UID: "7761cabb-a3e9-4b72-a562-10c066271fe0"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.297481 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7761cabb-a3e9-4b72-a562-10c066271fe0-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "7761cabb-a3e9-4b72-a562-10c066271fe0" (UID: "7761cabb-a3e9-4b72-a562-10c066271fe0"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.333094 4790 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7761cabb-a3e9-4b72-a562-10c066271fe0-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.333128 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7761cabb-a3e9-4b72-a562-10c066271fe0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.333140 4790 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7761cabb-a3e9-4b72-a562-10c066271fe0-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.333152 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7761cabb-a3e9-4b72-a562-10c066271fe0-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.333162 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rbrr\" (UniqueName: \"kubernetes.io/projected/7761cabb-a3e9-4b72-a562-10c066271fe0-kube-api-access-5rbrr\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.389950 4790 generic.go:334] "Generic (PLEG): container finished" podID="65369c6a-c495-49d7-829f-e9adb21172ff" containerID="28ce7a4d4c4dc879f7be8c9eb8f16679b7296eea8d6956705834a8e39f388e31" exitCode=143 Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.390009 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"65369c6a-c495-49d7-829f-e9adb21172ff","Type":"ContainerDied","Data":"28ce7a4d4c4dc879f7be8c9eb8f16679b7296eea8d6956705834a8e39f388e31"} Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.391842 4790 generic.go:334] "Generic (PLEG): container finished" podID="7761cabb-a3e9-4b72-a562-10c066271fe0" containerID="bf087df426faea9131fd7a339ec6ef0fa4f85c11747061f4d0231a6ff8d3aa23" exitCode=0 Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.391899 4790 generic.go:334] "Generic (PLEG): container finished" podID="7761cabb-a3e9-4b72-a562-10c066271fe0" containerID="cfc9c08d243310fe7960670de9b3e2278cc8874acd78a44359837cb3dcdc86b0" exitCode=143 Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.391923 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7761cabb-a3e9-4b72-a562-10c066271fe0","Type":"ContainerDied","Data":"bf087df426faea9131fd7a339ec6ef0fa4f85c11747061f4d0231a6ff8d3aa23"} Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.391935 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.391952 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7761cabb-a3e9-4b72-a562-10c066271fe0","Type":"ContainerDied","Data":"cfc9c08d243310fe7960670de9b3e2278cc8874acd78a44359837cb3dcdc86b0"} Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.391962 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7761cabb-a3e9-4b72-a562-10c066271fe0","Type":"ContainerDied","Data":"556b986f67dfb66817f056f99b2b2d2f6aa25342ba76083de031757f163731ee"} Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.391977 4790 scope.go:117] "RemoveContainer" containerID="bf087df426faea9131fd7a339ec6ef0fa4f85c11747061f4d0231a6ff8d3aa23" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.418576 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.419058 4790 scope.go:117] "RemoveContainer" containerID="cfc9c08d243310fe7960670de9b3e2278cc8874acd78a44359837cb3dcdc86b0" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.426133 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.445105 4790 scope.go:117] "RemoveContainer" containerID="bf087df426faea9131fd7a339ec6ef0fa4f85c11747061f4d0231a6ff8d3aa23" Nov 24 13:36:32 crc kubenswrapper[4790]: E1124 13:36:32.445495 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf087df426faea9131fd7a339ec6ef0fa4f85c11747061f4d0231a6ff8d3aa23\": container with ID starting with bf087df426faea9131fd7a339ec6ef0fa4f85c11747061f4d0231a6ff8d3aa23 not found: ID does not exist" containerID="bf087df426faea9131fd7a339ec6ef0fa4f85c11747061f4d0231a6ff8d3aa23" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.445524 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf087df426faea9131fd7a339ec6ef0fa4f85c11747061f4d0231a6ff8d3aa23"} err="failed to get container status \"bf087df426faea9131fd7a339ec6ef0fa4f85c11747061f4d0231a6ff8d3aa23\": rpc error: code = NotFound desc = could not find container \"bf087df426faea9131fd7a339ec6ef0fa4f85c11747061f4d0231a6ff8d3aa23\": container with ID starting with bf087df426faea9131fd7a339ec6ef0fa4f85c11747061f4d0231a6ff8d3aa23 not found: ID does not exist" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.445547 4790 scope.go:117] "RemoveContainer" containerID="cfc9c08d243310fe7960670de9b3e2278cc8874acd78a44359837cb3dcdc86b0" Nov 24 13:36:32 crc kubenswrapper[4790]: E1124 13:36:32.445876 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfc9c08d243310fe7960670de9b3e2278cc8874acd78a44359837cb3dcdc86b0\": container with ID starting with cfc9c08d243310fe7960670de9b3e2278cc8874acd78a44359837cb3dcdc86b0 not found: ID does not exist" containerID="cfc9c08d243310fe7960670de9b3e2278cc8874acd78a44359837cb3dcdc86b0" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.445910 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfc9c08d243310fe7960670de9b3e2278cc8874acd78a44359837cb3dcdc86b0"} err="failed to get container status \"cfc9c08d243310fe7960670de9b3e2278cc8874acd78a44359837cb3dcdc86b0\": rpc error: code = NotFound desc = could not find container \"cfc9c08d243310fe7960670de9b3e2278cc8874acd78a44359837cb3dcdc86b0\": container with ID starting with cfc9c08d243310fe7960670de9b3e2278cc8874acd78a44359837cb3dcdc86b0 not found: ID does not exist" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.445925 4790 scope.go:117] "RemoveContainer" containerID="bf087df426faea9131fd7a339ec6ef0fa4f85c11747061f4d0231a6ff8d3aa23" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.446160 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf087df426faea9131fd7a339ec6ef0fa4f85c11747061f4d0231a6ff8d3aa23"} err="failed to get container status \"bf087df426faea9131fd7a339ec6ef0fa4f85c11747061f4d0231a6ff8d3aa23\": rpc error: code = NotFound desc = could not find container \"bf087df426faea9131fd7a339ec6ef0fa4f85c11747061f4d0231a6ff8d3aa23\": container with ID starting with bf087df426faea9131fd7a339ec6ef0fa4f85c11747061f4d0231a6ff8d3aa23 not found: ID does not exist" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.446192 4790 scope.go:117] "RemoveContainer" containerID="cfc9c08d243310fe7960670de9b3e2278cc8874acd78a44359837cb3dcdc86b0" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.446467 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfc9c08d243310fe7960670de9b3e2278cc8874acd78a44359837cb3dcdc86b0"} err="failed to get container status \"cfc9c08d243310fe7960670de9b3e2278cc8874acd78a44359837cb3dcdc86b0\": rpc error: code = NotFound desc = could not find container \"cfc9c08d243310fe7960670de9b3e2278cc8874acd78a44359837cb3dcdc86b0\": container with ID starting with cfc9c08d243310fe7960670de9b3e2278cc8874acd78a44359837cb3dcdc86b0 not found: ID does not exist" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.448718 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 13:36:32 crc kubenswrapper[4790]: E1124 13:36:32.449146 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7761cabb-a3e9-4b72-a562-10c066271fe0" containerName="nova-api-api" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.449164 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="7761cabb-a3e9-4b72-a562-10c066271fe0" containerName="nova-api-api" Nov 24 13:36:32 crc kubenswrapper[4790]: E1124 13:36:32.449174 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412" containerName="init" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.449181 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412" containerName="init" Nov 24 13:36:32 crc kubenswrapper[4790]: E1124 13:36:32.449205 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5d01a68-e29a-44b3-93e6-f35b91673300" containerName="nova-manage" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.449212 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5d01a68-e29a-44b3-93e6-f35b91673300" containerName="nova-manage" Nov 24 13:36:32 crc kubenswrapper[4790]: E1124 13:36:32.449228 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412" containerName="dnsmasq-dns" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.449233 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412" containerName="dnsmasq-dns" Nov 24 13:36:32 crc kubenswrapper[4790]: E1124 13:36:32.449255 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7761cabb-a3e9-4b72-a562-10c066271fe0" containerName="nova-api-log" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.449261 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="7761cabb-a3e9-4b72-a562-10c066271fe0" containerName="nova-api-log" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.449411 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b922e9a-ffc7-43ed-ad5c-d5d7f2ac3412" containerName="dnsmasq-dns" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.449426 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="7761cabb-a3e9-4b72-a562-10c066271fe0" containerName="nova-api-log" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.449437 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5d01a68-e29a-44b3-93e6-f35b91673300" containerName="nova-manage" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.449456 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="7761cabb-a3e9-4b72-a562-10c066271fe0" containerName="nova-api-api" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.450439 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.455069 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.455265 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.459242 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.469948 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.479305 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nkvgt" podUID="099c81d1-2be3-4e10-b733-65e66912b605" containerName="registry-server" probeResult="failure" output=< Nov 24 13:36:32 crc kubenswrapper[4790]: timeout: failed to connect service ":50051" within 1s Nov 24 13:36:32 crc kubenswrapper[4790]: > Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.541857 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c786f8b4-6946-45c1-8cde-00e56bf899ed-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c786f8b4-6946-45c1-8cde-00e56bf899ed\") " pod="openstack/nova-api-0" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.541974 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c786f8b4-6946-45c1-8cde-00e56bf899ed-config-data\") pod \"nova-api-0\" (UID: \"c786f8b4-6946-45c1-8cde-00e56bf899ed\") " pod="openstack/nova-api-0" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.541999 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6gxb\" (UniqueName: \"kubernetes.io/projected/c786f8b4-6946-45c1-8cde-00e56bf899ed-kube-api-access-m6gxb\") pod \"nova-api-0\" (UID: \"c786f8b4-6946-45c1-8cde-00e56bf899ed\") " pod="openstack/nova-api-0" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.542050 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c786f8b4-6946-45c1-8cde-00e56bf899ed-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c786f8b4-6946-45c1-8cde-00e56bf899ed\") " pod="openstack/nova-api-0" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.542103 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c786f8b4-6946-45c1-8cde-00e56bf899ed-public-tls-certs\") pod \"nova-api-0\" (UID: \"c786f8b4-6946-45c1-8cde-00e56bf899ed\") " pod="openstack/nova-api-0" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.542158 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c786f8b4-6946-45c1-8cde-00e56bf899ed-logs\") pod \"nova-api-0\" (UID: \"c786f8b4-6946-45c1-8cde-00e56bf899ed\") " pod="openstack/nova-api-0" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.643528 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c786f8b4-6946-45c1-8cde-00e56bf899ed-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c786f8b4-6946-45c1-8cde-00e56bf899ed\") " pod="openstack/nova-api-0" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.643599 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c786f8b4-6946-45c1-8cde-00e56bf899ed-config-data\") pod \"nova-api-0\" (UID: \"c786f8b4-6946-45c1-8cde-00e56bf899ed\") " pod="openstack/nova-api-0" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.643624 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6gxb\" (UniqueName: \"kubernetes.io/projected/c786f8b4-6946-45c1-8cde-00e56bf899ed-kube-api-access-m6gxb\") pod \"nova-api-0\" (UID: \"c786f8b4-6946-45c1-8cde-00e56bf899ed\") " pod="openstack/nova-api-0" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.643642 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c786f8b4-6946-45c1-8cde-00e56bf899ed-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c786f8b4-6946-45c1-8cde-00e56bf899ed\") " pod="openstack/nova-api-0" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.643689 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c786f8b4-6946-45c1-8cde-00e56bf899ed-public-tls-certs\") pod \"nova-api-0\" (UID: \"c786f8b4-6946-45c1-8cde-00e56bf899ed\") " pod="openstack/nova-api-0" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.643729 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c786f8b4-6946-45c1-8cde-00e56bf899ed-logs\") pod \"nova-api-0\" (UID: \"c786f8b4-6946-45c1-8cde-00e56bf899ed\") " pod="openstack/nova-api-0" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.644101 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c786f8b4-6946-45c1-8cde-00e56bf899ed-logs\") pod \"nova-api-0\" (UID: \"c786f8b4-6946-45c1-8cde-00e56bf899ed\") " pod="openstack/nova-api-0" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.647238 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c786f8b4-6946-45c1-8cde-00e56bf899ed-config-data\") pod \"nova-api-0\" (UID: \"c786f8b4-6946-45c1-8cde-00e56bf899ed\") " pod="openstack/nova-api-0" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.647444 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c786f8b4-6946-45c1-8cde-00e56bf899ed-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c786f8b4-6946-45c1-8cde-00e56bf899ed\") " pod="openstack/nova-api-0" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.647865 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c786f8b4-6946-45c1-8cde-00e56bf899ed-public-tls-certs\") pod \"nova-api-0\" (UID: \"c786f8b4-6946-45c1-8cde-00e56bf899ed\") " pod="openstack/nova-api-0" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.649743 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c786f8b4-6946-45c1-8cde-00e56bf899ed-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c786f8b4-6946-45c1-8cde-00e56bf899ed\") " pod="openstack/nova-api-0" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.660649 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6gxb\" (UniqueName: \"kubernetes.io/projected/c786f8b4-6946-45c1-8cde-00e56bf899ed-kube-api-access-m6gxb\") pod \"nova-api-0\" (UID: \"c786f8b4-6946-45c1-8cde-00e56bf899ed\") " pod="openstack/nova-api-0" Nov 24 13:36:32 crc kubenswrapper[4790]: I1124 13:36:32.768091 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 13:36:33 crc kubenswrapper[4790]: I1124 13:36:33.196399 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:36:33 crc kubenswrapper[4790]: I1124 13:36:33.401822 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c786f8b4-6946-45c1-8cde-00e56bf899ed","Type":"ContainerStarted","Data":"7e1e61d9e8e5a5c4687b8cc095ff2bfb83f045d184dc2b3cab025c89bd6960a1"} Nov 24 13:36:33 crc kubenswrapper[4790]: I1124 13:36:33.401870 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c786f8b4-6946-45c1-8cde-00e56bf899ed","Type":"ContainerStarted","Data":"967fad38c66d7248a13455d3a6db225ef5fd1da2a231cbb2ae3462876c679f42"} Nov 24 13:36:34 crc kubenswrapper[4790]: I1124 13:36:34.324254 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7761cabb-a3e9-4b72-a562-10c066271fe0" path="/var/lib/kubelet/pods/7761cabb-a3e9-4b72-a562-10c066271fe0/volumes" Nov 24 13:36:34 crc kubenswrapper[4790]: I1124 13:36:34.415688 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c786f8b4-6946-45c1-8cde-00e56bf899ed","Type":"ContainerStarted","Data":"ddb81a6c66eee6e319c0646f64dd78e67d54faaf3b3d4b87630354653dbf2c4c"} Nov 24 13:36:34 crc kubenswrapper[4790]: I1124 13:36:34.443945 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.4439234499999998 podStartE2EDuration="2.44392345s" podCreationTimestamp="2025-11-24 13:36:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:36:34.438911906 +0000 UTC m=+1442.818805608" watchObservedRunningTime="2025-11-24 13:36:34.44392345 +0000 UTC m=+1442.823817112" Nov 24 13:36:34 crc kubenswrapper[4790]: I1124 13:36:34.676365 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="65369c6a-c495-49d7-829f-e9adb21172ff" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.189:8775/\": read tcp 10.217.0.2:48992->10.217.0.189:8775: read: connection reset by peer" Nov 24 13:36:34 crc kubenswrapper[4790]: I1124 13:36:34.676763 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="65369c6a-c495-49d7-829f-e9adb21172ff" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.189:8775/\": read tcp 10.217.0.2:48994->10.217.0.189:8775: read: connection reset by peer" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.148832 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.296469 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65369c6a-c495-49d7-829f-e9adb21172ff-combined-ca-bundle\") pod \"65369c6a-c495-49d7-829f-e9adb21172ff\" (UID: \"65369c6a-c495-49d7-829f-e9adb21172ff\") " Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.296587 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65369c6a-c495-49d7-829f-e9adb21172ff-logs\") pod \"65369c6a-c495-49d7-829f-e9adb21172ff\" (UID: \"65369c6a-c495-49d7-829f-e9adb21172ff\") " Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.296677 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/65369c6a-c495-49d7-829f-e9adb21172ff-nova-metadata-tls-certs\") pod \"65369c6a-c495-49d7-829f-e9adb21172ff\" (UID: \"65369c6a-c495-49d7-829f-e9adb21172ff\") " Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.296722 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxrbx\" (UniqueName: \"kubernetes.io/projected/65369c6a-c495-49d7-829f-e9adb21172ff-kube-api-access-zxrbx\") pod \"65369c6a-c495-49d7-829f-e9adb21172ff\" (UID: \"65369c6a-c495-49d7-829f-e9adb21172ff\") " Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.296828 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65369c6a-c495-49d7-829f-e9adb21172ff-config-data\") pod \"65369c6a-c495-49d7-829f-e9adb21172ff\" (UID: \"65369c6a-c495-49d7-829f-e9adb21172ff\") " Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.297391 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65369c6a-c495-49d7-829f-e9adb21172ff-logs" (OuterVolumeSpecName: "logs") pod "65369c6a-c495-49d7-829f-e9adb21172ff" (UID: "65369c6a-c495-49d7-829f-e9adb21172ff"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.308169 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65369c6a-c495-49d7-829f-e9adb21172ff-kube-api-access-zxrbx" (OuterVolumeSpecName: "kube-api-access-zxrbx") pod "65369c6a-c495-49d7-829f-e9adb21172ff" (UID: "65369c6a-c495-49d7-829f-e9adb21172ff"). InnerVolumeSpecName "kube-api-access-zxrbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.329093 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65369c6a-c495-49d7-829f-e9adb21172ff-config-data" (OuterVolumeSpecName: "config-data") pod "65369c6a-c495-49d7-829f-e9adb21172ff" (UID: "65369c6a-c495-49d7-829f-e9adb21172ff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.330712 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65369c6a-c495-49d7-829f-e9adb21172ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "65369c6a-c495-49d7-829f-e9adb21172ff" (UID: "65369c6a-c495-49d7-829f-e9adb21172ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.353165 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65369c6a-c495-49d7-829f-e9adb21172ff-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "65369c6a-c495-49d7-829f-e9adb21172ff" (UID: "65369c6a-c495-49d7-829f-e9adb21172ff"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.398972 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65369c6a-c495-49d7-829f-e9adb21172ff-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.399347 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65369c6a-c495-49d7-829f-e9adb21172ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.399359 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65369c6a-c495-49d7-829f-e9adb21172ff-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.399367 4790 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/65369c6a-c495-49d7-829f-e9adb21172ff-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.399378 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxrbx\" (UniqueName: \"kubernetes.io/projected/65369c6a-c495-49d7-829f-e9adb21172ff-kube-api-access-zxrbx\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.428058 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.428012 4790 generic.go:334] "Generic (PLEG): container finished" podID="65369c6a-c495-49d7-829f-e9adb21172ff" containerID="d5cc8dc52fd100b95d8a516cbdf245be0d1ffd7c60cfbbaee80ea00d834358ca" exitCode=0 Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.428436 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"65369c6a-c495-49d7-829f-e9adb21172ff","Type":"ContainerDied","Data":"d5cc8dc52fd100b95d8a516cbdf245be0d1ffd7c60cfbbaee80ea00d834358ca"} Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.428648 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"65369c6a-c495-49d7-829f-e9adb21172ff","Type":"ContainerDied","Data":"9e5d5f8346a57c976fc559e02ab3f537626fbf4fd16eb9df72b80146699585bd"} Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.428682 4790 scope.go:117] "RemoveContainer" containerID="d5cc8dc52fd100b95d8a516cbdf245be0d1ffd7c60cfbbaee80ea00d834358ca" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.451609 4790 scope.go:117] "RemoveContainer" containerID="28ce7a4d4c4dc879f7be8c9eb8f16679b7296eea8d6956705834a8e39f388e31" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.524477 4790 scope.go:117] "RemoveContainer" containerID="d5cc8dc52fd100b95d8a516cbdf245be0d1ffd7c60cfbbaee80ea00d834358ca" Nov 24 13:36:35 crc kubenswrapper[4790]: E1124 13:36:35.524969 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5cc8dc52fd100b95d8a516cbdf245be0d1ffd7c60cfbbaee80ea00d834358ca\": container with ID starting with d5cc8dc52fd100b95d8a516cbdf245be0d1ffd7c60cfbbaee80ea00d834358ca not found: ID does not exist" containerID="d5cc8dc52fd100b95d8a516cbdf245be0d1ffd7c60cfbbaee80ea00d834358ca" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.525011 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5cc8dc52fd100b95d8a516cbdf245be0d1ffd7c60cfbbaee80ea00d834358ca"} err="failed to get container status \"d5cc8dc52fd100b95d8a516cbdf245be0d1ffd7c60cfbbaee80ea00d834358ca\": rpc error: code = NotFound desc = could not find container \"d5cc8dc52fd100b95d8a516cbdf245be0d1ffd7c60cfbbaee80ea00d834358ca\": container with ID starting with d5cc8dc52fd100b95d8a516cbdf245be0d1ffd7c60cfbbaee80ea00d834358ca not found: ID does not exist" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.525034 4790 scope.go:117] "RemoveContainer" containerID="28ce7a4d4c4dc879f7be8c9eb8f16679b7296eea8d6956705834a8e39f388e31" Nov 24 13:36:35 crc kubenswrapper[4790]: E1124 13:36:35.525608 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28ce7a4d4c4dc879f7be8c9eb8f16679b7296eea8d6956705834a8e39f388e31\": container with ID starting with 28ce7a4d4c4dc879f7be8c9eb8f16679b7296eea8d6956705834a8e39f388e31 not found: ID does not exist" containerID="28ce7a4d4c4dc879f7be8c9eb8f16679b7296eea8d6956705834a8e39f388e31" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.525648 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28ce7a4d4c4dc879f7be8c9eb8f16679b7296eea8d6956705834a8e39f388e31"} err="failed to get container status \"28ce7a4d4c4dc879f7be8c9eb8f16679b7296eea8d6956705834a8e39f388e31\": rpc error: code = NotFound desc = could not find container \"28ce7a4d4c4dc879f7be8c9eb8f16679b7296eea8d6956705834a8e39f388e31\": container with ID starting with 28ce7a4d4c4dc879f7be8c9eb8f16679b7296eea8d6956705834a8e39f388e31 not found: ID does not exist" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.540126 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.550351 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.569283 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:36:35 crc kubenswrapper[4790]: E1124 13:36:35.569778 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65369c6a-c495-49d7-829f-e9adb21172ff" containerName="nova-metadata-metadata" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.569806 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="65369c6a-c495-49d7-829f-e9adb21172ff" containerName="nova-metadata-metadata" Nov 24 13:36:35 crc kubenswrapper[4790]: E1124 13:36:35.569868 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65369c6a-c495-49d7-829f-e9adb21172ff" containerName="nova-metadata-log" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.569891 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="65369c6a-c495-49d7-829f-e9adb21172ff" containerName="nova-metadata-log" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.570104 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="65369c6a-c495-49d7-829f-e9adb21172ff" containerName="nova-metadata-metadata" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.570162 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="65369c6a-c495-49d7-829f-e9adb21172ff" containerName="nova-metadata-log" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.571631 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.573978 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.574225 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.597053 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.704225 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dchb\" (UniqueName: \"kubernetes.io/projected/81757ef9-7330-44bb-a35e-8e6f4c137c0e-kube-api-access-4dchb\") pod \"nova-metadata-0\" (UID: \"81757ef9-7330-44bb-a35e-8e6f4c137c0e\") " pod="openstack/nova-metadata-0" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.704286 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81757ef9-7330-44bb-a35e-8e6f4c137c0e-logs\") pod \"nova-metadata-0\" (UID: \"81757ef9-7330-44bb-a35e-8e6f4c137c0e\") " pod="openstack/nova-metadata-0" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.704478 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81757ef9-7330-44bb-a35e-8e6f4c137c0e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"81757ef9-7330-44bb-a35e-8e6f4c137c0e\") " pod="openstack/nova-metadata-0" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.704738 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/81757ef9-7330-44bb-a35e-8e6f4c137c0e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"81757ef9-7330-44bb-a35e-8e6f4c137c0e\") " pod="openstack/nova-metadata-0" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.704790 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81757ef9-7330-44bb-a35e-8e6f4c137c0e-config-data\") pod \"nova-metadata-0\" (UID: \"81757ef9-7330-44bb-a35e-8e6f4c137c0e\") " pod="openstack/nova-metadata-0" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.806144 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81757ef9-7330-44bb-a35e-8e6f4c137c0e-config-data\") pod \"nova-metadata-0\" (UID: \"81757ef9-7330-44bb-a35e-8e6f4c137c0e\") " pod="openstack/nova-metadata-0" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.806233 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dchb\" (UniqueName: \"kubernetes.io/projected/81757ef9-7330-44bb-a35e-8e6f4c137c0e-kube-api-access-4dchb\") pod \"nova-metadata-0\" (UID: \"81757ef9-7330-44bb-a35e-8e6f4c137c0e\") " pod="openstack/nova-metadata-0" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.806255 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81757ef9-7330-44bb-a35e-8e6f4c137c0e-logs\") pod \"nova-metadata-0\" (UID: \"81757ef9-7330-44bb-a35e-8e6f4c137c0e\") " pod="openstack/nova-metadata-0" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.806598 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81757ef9-7330-44bb-a35e-8e6f4c137c0e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"81757ef9-7330-44bb-a35e-8e6f4c137c0e\") " pod="openstack/nova-metadata-0" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.806696 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81757ef9-7330-44bb-a35e-8e6f4c137c0e-logs\") pod \"nova-metadata-0\" (UID: \"81757ef9-7330-44bb-a35e-8e6f4c137c0e\") " pod="openstack/nova-metadata-0" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.806999 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/81757ef9-7330-44bb-a35e-8e6f4c137c0e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"81757ef9-7330-44bb-a35e-8e6f4c137c0e\") " pod="openstack/nova-metadata-0" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.811325 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/81757ef9-7330-44bb-a35e-8e6f4c137c0e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"81757ef9-7330-44bb-a35e-8e6f4c137c0e\") " pod="openstack/nova-metadata-0" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.811840 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81757ef9-7330-44bb-a35e-8e6f4c137c0e-config-data\") pod \"nova-metadata-0\" (UID: \"81757ef9-7330-44bb-a35e-8e6f4c137c0e\") " pod="openstack/nova-metadata-0" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.811892 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81757ef9-7330-44bb-a35e-8e6f4c137c0e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"81757ef9-7330-44bb-a35e-8e6f4c137c0e\") " pod="openstack/nova-metadata-0" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.823440 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dchb\" (UniqueName: \"kubernetes.io/projected/81757ef9-7330-44bb-a35e-8e6f4c137c0e-kube-api-access-4dchb\") pod \"nova-metadata-0\" (UID: \"81757ef9-7330-44bb-a35e-8e6f4c137c0e\") " pod="openstack/nova-metadata-0" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.890237 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 13:36:35 crc kubenswrapper[4790]: I1124 13:36:35.922423 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.010313 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c864a2b1-1693-4200-be70-a140b1526bac-config-data\") pod \"c864a2b1-1693-4200-be70-a140b1526bac\" (UID: \"c864a2b1-1693-4200-be70-a140b1526bac\") " Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.010482 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c864a2b1-1693-4200-be70-a140b1526bac-combined-ca-bundle\") pod \"c864a2b1-1693-4200-be70-a140b1526bac\" (UID: \"c864a2b1-1693-4200-be70-a140b1526bac\") " Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.010547 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6shng\" (UniqueName: \"kubernetes.io/projected/c864a2b1-1693-4200-be70-a140b1526bac-kube-api-access-6shng\") pod \"c864a2b1-1693-4200-be70-a140b1526bac\" (UID: \"c864a2b1-1693-4200-be70-a140b1526bac\") " Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.015637 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c864a2b1-1693-4200-be70-a140b1526bac-kube-api-access-6shng" (OuterVolumeSpecName: "kube-api-access-6shng") pod "c864a2b1-1693-4200-be70-a140b1526bac" (UID: "c864a2b1-1693-4200-be70-a140b1526bac"). InnerVolumeSpecName "kube-api-access-6shng". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.041021 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c864a2b1-1693-4200-be70-a140b1526bac-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c864a2b1-1693-4200-be70-a140b1526bac" (UID: "c864a2b1-1693-4200-be70-a140b1526bac"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.041323 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c864a2b1-1693-4200-be70-a140b1526bac-config-data" (OuterVolumeSpecName: "config-data") pod "c864a2b1-1693-4200-be70-a140b1526bac" (UID: "c864a2b1-1693-4200-be70-a140b1526bac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.113573 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c864a2b1-1693-4200-be70-a140b1526bac-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.113609 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c864a2b1-1693-4200-be70-a140b1526bac-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.113645 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6shng\" (UniqueName: \"kubernetes.io/projected/c864a2b1-1693-4200-be70-a140b1526bac-kube-api-access-6shng\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.340560 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65369c6a-c495-49d7-829f-e9adb21172ff" path="/var/lib/kubelet/pods/65369c6a-c495-49d7-829f-e9adb21172ff/volumes" Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.441850 4790 generic.go:334] "Generic (PLEG): container finished" podID="c864a2b1-1693-4200-be70-a140b1526bac" containerID="285bb94beb2c560eb1cadca29c368e78f28a1d7b266ed21f459cb58d61a3f93b" exitCode=0 Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.441953 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.441953 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c864a2b1-1693-4200-be70-a140b1526bac","Type":"ContainerDied","Data":"285bb94beb2c560eb1cadca29c368e78f28a1d7b266ed21f459cb58d61a3f93b"} Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.442088 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c864a2b1-1693-4200-be70-a140b1526bac","Type":"ContainerDied","Data":"b9c285f81ae78d4cda166049a582c17eef8cb14d241437f53ce18dd1d2ab72fb"} Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.442111 4790 scope.go:117] "RemoveContainer" containerID="285bb94beb2c560eb1cadca29c368e78f28a1d7b266ed21f459cb58d61a3f93b" Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.483308 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.484641 4790 scope.go:117] "RemoveContainer" containerID="285bb94beb2c560eb1cadca29c368e78f28a1d7b266ed21f459cb58d61a3f93b" Nov 24 13:36:36 crc kubenswrapper[4790]: E1124 13:36:36.485126 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"285bb94beb2c560eb1cadca29c368e78f28a1d7b266ed21f459cb58d61a3f93b\": container with ID starting with 285bb94beb2c560eb1cadca29c368e78f28a1d7b266ed21f459cb58d61a3f93b not found: ID does not exist" containerID="285bb94beb2c560eb1cadca29c368e78f28a1d7b266ed21f459cb58d61a3f93b" Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.485157 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"285bb94beb2c560eb1cadca29c368e78f28a1d7b266ed21f459cb58d61a3f93b"} err="failed to get container status \"285bb94beb2c560eb1cadca29c368e78f28a1d7b266ed21f459cb58d61a3f93b\": rpc error: code = NotFound desc = could not find container \"285bb94beb2c560eb1cadca29c368e78f28a1d7b266ed21f459cb58d61a3f93b\": container with ID starting with 285bb94beb2c560eb1cadca29c368e78f28a1d7b266ed21f459cb58d61a3f93b not found: ID does not exist" Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.497747 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.507242 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 13:36:36 crc kubenswrapper[4790]: E1124 13:36:36.507708 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c864a2b1-1693-4200-be70-a140b1526bac" containerName="nova-scheduler-scheduler" Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.507724 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="c864a2b1-1693-4200-be70-a140b1526bac" containerName="nova-scheduler-scheduler" Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.508035 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="c864a2b1-1693-4200-be70-a140b1526bac" containerName="nova-scheduler-scheduler" Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.508771 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.511163 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.516684 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.629774 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de6f5df1-eb57-493b-9d89-5e710c6f4304-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"de6f5df1-eb57-493b-9d89-5e710c6f4304\") " pod="openstack/nova-scheduler-0" Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.629971 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de6f5df1-eb57-493b-9d89-5e710c6f4304-config-data\") pod \"nova-scheduler-0\" (UID: \"de6f5df1-eb57-493b-9d89-5e710c6f4304\") " pod="openstack/nova-scheduler-0" Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.630242 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6p57k\" (UniqueName: \"kubernetes.io/projected/de6f5df1-eb57-493b-9d89-5e710c6f4304-kube-api-access-6p57k\") pod \"nova-scheduler-0\" (UID: \"de6f5df1-eb57-493b-9d89-5e710c6f4304\") " pod="openstack/nova-scheduler-0" Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.732051 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6p57k\" (UniqueName: \"kubernetes.io/projected/de6f5df1-eb57-493b-9d89-5e710c6f4304-kube-api-access-6p57k\") pod \"nova-scheduler-0\" (UID: \"de6f5df1-eb57-493b-9d89-5e710c6f4304\") " pod="openstack/nova-scheduler-0" Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.732176 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de6f5df1-eb57-493b-9d89-5e710c6f4304-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"de6f5df1-eb57-493b-9d89-5e710c6f4304\") " pod="openstack/nova-scheduler-0" Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.732230 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de6f5df1-eb57-493b-9d89-5e710c6f4304-config-data\") pod \"nova-scheduler-0\" (UID: \"de6f5df1-eb57-493b-9d89-5e710c6f4304\") " pod="openstack/nova-scheduler-0" Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.736452 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de6f5df1-eb57-493b-9d89-5e710c6f4304-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"de6f5df1-eb57-493b-9d89-5e710c6f4304\") " pod="openstack/nova-scheduler-0" Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.736799 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de6f5df1-eb57-493b-9d89-5e710c6f4304-config-data\") pod \"nova-scheduler-0\" (UID: \"de6f5df1-eb57-493b-9d89-5e710c6f4304\") " pod="openstack/nova-scheduler-0" Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.748260 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6p57k\" (UniqueName: \"kubernetes.io/projected/de6f5df1-eb57-493b-9d89-5e710c6f4304-kube-api-access-6p57k\") pod \"nova-scheduler-0\" (UID: \"de6f5df1-eb57-493b-9d89-5e710c6f4304\") " pod="openstack/nova-scheduler-0" Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.823996 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 13:36:36 crc kubenswrapper[4790]: I1124 13:36:36.928406 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:36:36 crc kubenswrapper[4790]: W1124 13:36:36.933584 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81757ef9_7330_44bb_a35e_8e6f4c137c0e.slice/crio-dd52f7be4dc7b7ab84f9f1447bde4fa92633674da2d16d7841751f839f7e2b13 WatchSource:0}: Error finding container dd52f7be4dc7b7ab84f9f1447bde4fa92633674da2d16d7841751f839f7e2b13: Status 404 returned error can't find the container with id dd52f7be4dc7b7ab84f9f1447bde4fa92633674da2d16d7841751f839f7e2b13 Nov 24 13:36:37 crc kubenswrapper[4790]: W1124 13:36:37.252755 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde6f5df1_eb57_493b_9d89_5e710c6f4304.slice/crio-42f4b21d70f156369f559c7e6ef7a64596f9aba627a9c1ab988a7e9aee79943c WatchSource:0}: Error finding container 42f4b21d70f156369f559c7e6ef7a64596f9aba627a9c1ab988a7e9aee79943c: Status 404 returned error can't find the container with id 42f4b21d70f156369f559c7e6ef7a64596f9aba627a9c1ab988a7e9aee79943c Nov 24 13:36:37 crc kubenswrapper[4790]: I1124 13:36:37.253473 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 13:36:37 crc kubenswrapper[4790]: I1124 13:36:37.453741 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"de6f5df1-eb57-493b-9d89-5e710c6f4304","Type":"ContainerStarted","Data":"6b8183ab06a7a4bd08c674c5788831aa9879a328f22b998c507e78d8d453b26d"} Nov 24 13:36:37 crc kubenswrapper[4790]: I1124 13:36:37.453782 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"de6f5df1-eb57-493b-9d89-5e710c6f4304","Type":"ContainerStarted","Data":"42f4b21d70f156369f559c7e6ef7a64596f9aba627a9c1ab988a7e9aee79943c"} Nov 24 13:36:37 crc kubenswrapper[4790]: I1124 13:36:37.456603 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"81757ef9-7330-44bb-a35e-8e6f4c137c0e","Type":"ContainerStarted","Data":"6485471927d5a223e611cbd164262141a1a0f94b13c44e9d6ea257f4db80ed9a"} Nov 24 13:36:37 crc kubenswrapper[4790]: I1124 13:36:37.456629 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"81757ef9-7330-44bb-a35e-8e6f4c137c0e","Type":"ContainerStarted","Data":"6b1d4a6d406564bd94038ede12823e6303e9c099f716a3ed7065653038dff83e"} Nov 24 13:36:37 crc kubenswrapper[4790]: I1124 13:36:37.456639 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"81757ef9-7330-44bb-a35e-8e6f4c137c0e","Type":"ContainerStarted","Data":"dd52f7be4dc7b7ab84f9f1447bde4fa92633674da2d16d7841751f839f7e2b13"} Nov 24 13:36:37 crc kubenswrapper[4790]: I1124 13:36:37.477190 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.47717289 podStartE2EDuration="1.47717289s" podCreationTimestamp="2025-11-24 13:36:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:36:37.469318095 +0000 UTC m=+1445.849211757" watchObservedRunningTime="2025-11-24 13:36:37.47717289 +0000 UTC m=+1445.857066552" Nov 24 13:36:37 crc kubenswrapper[4790]: I1124 13:36:37.491508 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.491483139 podStartE2EDuration="2.491483139s" podCreationTimestamp="2025-11-24 13:36:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:36:37.486933169 +0000 UTC m=+1445.866826841" watchObservedRunningTime="2025-11-24 13:36:37.491483139 +0000 UTC m=+1445.871376811" Nov 24 13:36:38 crc kubenswrapper[4790]: I1124 13:36:38.329322 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c864a2b1-1693-4200-be70-a140b1526bac" path="/var/lib/kubelet/pods/c864a2b1-1693-4200-be70-a140b1526bac/volumes" Nov 24 13:36:40 crc kubenswrapper[4790]: I1124 13:36:40.890530 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 13:36:40 crc kubenswrapper[4790]: I1124 13:36:40.891123 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 13:36:41 crc kubenswrapper[4790]: I1124 13:36:41.469220 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nkvgt" Nov 24 13:36:41 crc kubenswrapper[4790]: I1124 13:36:41.518073 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nkvgt" Nov 24 13:36:41 crc kubenswrapper[4790]: I1124 13:36:41.704103 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nkvgt"] Nov 24 13:36:41 crc kubenswrapper[4790]: I1124 13:36:41.824421 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 13:36:42 crc kubenswrapper[4790]: I1124 13:36:42.510727 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nkvgt" podUID="099c81d1-2be3-4e10-b733-65e66912b605" containerName="registry-server" containerID="cri-o://88abd8704b089a541505f154340ecb0756ce273f1b8e926a5a9549fcd8886f64" gracePeriod=2 Nov 24 13:36:42 crc kubenswrapper[4790]: I1124 13:36:42.768579 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 13:36:42 crc kubenswrapper[4790]: I1124 13:36:42.768906 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 13:36:43 crc kubenswrapper[4790]: I1124 13:36:43.012334 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nkvgt" Nov 24 13:36:43 crc kubenswrapper[4790]: I1124 13:36:43.146422 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/099c81d1-2be3-4e10-b733-65e66912b605-catalog-content\") pod \"099c81d1-2be3-4e10-b733-65e66912b605\" (UID: \"099c81d1-2be3-4e10-b733-65e66912b605\") " Nov 24 13:36:43 crc kubenswrapper[4790]: I1124 13:36:43.146508 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zb5f\" (UniqueName: \"kubernetes.io/projected/099c81d1-2be3-4e10-b733-65e66912b605-kube-api-access-5zb5f\") pod \"099c81d1-2be3-4e10-b733-65e66912b605\" (UID: \"099c81d1-2be3-4e10-b733-65e66912b605\") " Nov 24 13:36:43 crc kubenswrapper[4790]: I1124 13:36:43.146705 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/099c81d1-2be3-4e10-b733-65e66912b605-utilities\") pod \"099c81d1-2be3-4e10-b733-65e66912b605\" (UID: \"099c81d1-2be3-4e10-b733-65e66912b605\") " Nov 24 13:36:43 crc kubenswrapper[4790]: I1124 13:36:43.147556 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/099c81d1-2be3-4e10-b733-65e66912b605-utilities" (OuterVolumeSpecName: "utilities") pod "099c81d1-2be3-4e10-b733-65e66912b605" (UID: "099c81d1-2be3-4e10-b733-65e66912b605"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:36:43 crc kubenswrapper[4790]: I1124 13:36:43.153497 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/099c81d1-2be3-4e10-b733-65e66912b605-kube-api-access-5zb5f" (OuterVolumeSpecName: "kube-api-access-5zb5f") pod "099c81d1-2be3-4e10-b733-65e66912b605" (UID: "099c81d1-2be3-4e10-b733-65e66912b605"). InnerVolumeSpecName "kube-api-access-5zb5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:36:43 crc kubenswrapper[4790]: I1124 13:36:43.230499 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/099c81d1-2be3-4e10-b733-65e66912b605-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "099c81d1-2be3-4e10-b733-65e66912b605" (UID: "099c81d1-2be3-4e10-b733-65e66912b605"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:36:43 crc kubenswrapper[4790]: I1124 13:36:43.248717 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/099c81d1-2be3-4e10-b733-65e66912b605-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:43 crc kubenswrapper[4790]: I1124 13:36:43.248742 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zb5f\" (UniqueName: \"kubernetes.io/projected/099c81d1-2be3-4e10-b733-65e66912b605-kube-api-access-5zb5f\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:43 crc kubenswrapper[4790]: I1124 13:36:43.248751 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/099c81d1-2be3-4e10-b733-65e66912b605-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:43 crc kubenswrapper[4790]: I1124 13:36:43.522941 4790 generic.go:334] "Generic (PLEG): container finished" podID="099c81d1-2be3-4e10-b733-65e66912b605" containerID="88abd8704b089a541505f154340ecb0756ce273f1b8e926a5a9549fcd8886f64" exitCode=0 Nov 24 13:36:43 crc kubenswrapper[4790]: I1124 13:36:43.522995 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nkvgt" Nov 24 13:36:43 crc kubenswrapper[4790]: I1124 13:36:43.523009 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nkvgt" event={"ID":"099c81d1-2be3-4e10-b733-65e66912b605","Type":"ContainerDied","Data":"88abd8704b089a541505f154340ecb0756ce273f1b8e926a5a9549fcd8886f64"} Nov 24 13:36:43 crc kubenswrapper[4790]: I1124 13:36:43.524648 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nkvgt" event={"ID":"099c81d1-2be3-4e10-b733-65e66912b605","Type":"ContainerDied","Data":"602948df3da2e9a0dc56c8375a5d5ea011cc76e7e12e256d80d5d768be413041"} Nov 24 13:36:43 crc kubenswrapper[4790]: I1124 13:36:43.524675 4790 scope.go:117] "RemoveContainer" containerID="88abd8704b089a541505f154340ecb0756ce273f1b8e926a5a9549fcd8886f64" Nov 24 13:36:43 crc kubenswrapper[4790]: I1124 13:36:43.545894 4790 scope.go:117] "RemoveContainer" containerID="86fe297d369579aab3fd9785a4ca060b16eafb20531318ab7a41138922921292" Nov 24 13:36:43 crc kubenswrapper[4790]: I1124 13:36:43.568293 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nkvgt"] Nov 24 13:36:43 crc kubenswrapper[4790]: I1124 13:36:43.577449 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nkvgt"] Nov 24 13:36:43 crc kubenswrapper[4790]: I1124 13:36:43.621092 4790 scope.go:117] "RemoveContainer" containerID="71ef147cadaf1a19d06f10591436329c47702dc6721878e0208bc04ca4c4e395" Nov 24 13:36:43 crc kubenswrapper[4790]: I1124 13:36:43.647029 4790 scope.go:117] "RemoveContainer" containerID="88abd8704b089a541505f154340ecb0756ce273f1b8e926a5a9549fcd8886f64" Nov 24 13:36:43 crc kubenswrapper[4790]: E1124 13:36:43.647573 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88abd8704b089a541505f154340ecb0756ce273f1b8e926a5a9549fcd8886f64\": container with ID starting with 88abd8704b089a541505f154340ecb0756ce273f1b8e926a5a9549fcd8886f64 not found: ID does not exist" containerID="88abd8704b089a541505f154340ecb0756ce273f1b8e926a5a9549fcd8886f64" Nov 24 13:36:43 crc kubenswrapper[4790]: I1124 13:36:43.647612 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88abd8704b089a541505f154340ecb0756ce273f1b8e926a5a9549fcd8886f64"} err="failed to get container status \"88abd8704b089a541505f154340ecb0756ce273f1b8e926a5a9549fcd8886f64\": rpc error: code = NotFound desc = could not find container \"88abd8704b089a541505f154340ecb0756ce273f1b8e926a5a9549fcd8886f64\": container with ID starting with 88abd8704b089a541505f154340ecb0756ce273f1b8e926a5a9549fcd8886f64 not found: ID does not exist" Nov 24 13:36:43 crc kubenswrapper[4790]: I1124 13:36:43.647639 4790 scope.go:117] "RemoveContainer" containerID="86fe297d369579aab3fd9785a4ca060b16eafb20531318ab7a41138922921292" Nov 24 13:36:43 crc kubenswrapper[4790]: E1124 13:36:43.647975 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86fe297d369579aab3fd9785a4ca060b16eafb20531318ab7a41138922921292\": container with ID starting with 86fe297d369579aab3fd9785a4ca060b16eafb20531318ab7a41138922921292 not found: ID does not exist" containerID="86fe297d369579aab3fd9785a4ca060b16eafb20531318ab7a41138922921292" Nov 24 13:36:43 crc kubenswrapper[4790]: I1124 13:36:43.648028 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86fe297d369579aab3fd9785a4ca060b16eafb20531318ab7a41138922921292"} err="failed to get container status \"86fe297d369579aab3fd9785a4ca060b16eafb20531318ab7a41138922921292\": rpc error: code = NotFound desc = could not find container \"86fe297d369579aab3fd9785a4ca060b16eafb20531318ab7a41138922921292\": container with ID starting with 86fe297d369579aab3fd9785a4ca060b16eafb20531318ab7a41138922921292 not found: ID does not exist" Nov 24 13:36:43 crc kubenswrapper[4790]: I1124 13:36:43.648054 4790 scope.go:117] "RemoveContainer" containerID="71ef147cadaf1a19d06f10591436329c47702dc6721878e0208bc04ca4c4e395" Nov 24 13:36:43 crc kubenswrapper[4790]: E1124 13:36:43.648354 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71ef147cadaf1a19d06f10591436329c47702dc6721878e0208bc04ca4c4e395\": container with ID starting with 71ef147cadaf1a19d06f10591436329c47702dc6721878e0208bc04ca4c4e395 not found: ID does not exist" containerID="71ef147cadaf1a19d06f10591436329c47702dc6721878e0208bc04ca4c4e395" Nov 24 13:36:43 crc kubenswrapper[4790]: I1124 13:36:43.648396 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71ef147cadaf1a19d06f10591436329c47702dc6721878e0208bc04ca4c4e395"} err="failed to get container status \"71ef147cadaf1a19d06f10591436329c47702dc6721878e0208bc04ca4c4e395\": rpc error: code = NotFound desc = could not find container \"71ef147cadaf1a19d06f10591436329c47702dc6721878e0208bc04ca4c4e395\": container with ID starting with 71ef147cadaf1a19d06f10591436329c47702dc6721878e0208bc04ca4c4e395 not found: ID does not exist" Nov 24 13:36:43 crc kubenswrapper[4790]: I1124 13:36:43.783072 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c786f8b4-6946-45c1-8cde-00e56bf899ed" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.200:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 13:36:43 crc kubenswrapper[4790]: I1124 13:36:43.783082 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c786f8b4-6946-45c1-8cde-00e56bf899ed" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.200:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 13:36:43 crc kubenswrapper[4790]: I1124 13:36:43.938598 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:36:43 crc kubenswrapper[4790]: I1124 13:36:43.938653 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:36:44 crc kubenswrapper[4790]: I1124 13:36:44.330160 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="099c81d1-2be3-4e10-b733-65e66912b605" path="/var/lib/kubelet/pods/099c81d1-2be3-4e10-b733-65e66912b605/volumes" Nov 24 13:36:45 crc kubenswrapper[4790]: I1124 13:36:45.890568 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 13:36:45 crc kubenswrapper[4790]: I1124 13:36:45.890925 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 13:36:46 crc kubenswrapper[4790]: I1124 13:36:46.824778 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 13:36:46 crc kubenswrapper[4790]: I1124 13:36:46.860557 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 13:36:46 crc kubenswrapper[4790]: I1124 13:36:46.898153 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="81757ef9-7330-44bb-a35e-8e6f4c137c0e" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 13:36:46 crc kubenswrapper[4790]: I1124 13:36:46.906057 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="81757ef9-7330-44bb-a35e-8e6f4c137c0e" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 13:36:47 crc kubenswrapper[4790]: I1124 13:36:47.607683 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 13:36:49 crc kubenswrapper[4790]: I1124 13:36:49.682341 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-f455q"] Nov 24 13:36:49 crc kubenswrapper[4790]: E1124 13:36:49.683326 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="099c81d1-2be3-4e10-b733-65e66912b605" containerName="extract-utilities" Nov 24 13:36:49 crc kubenswrapper[4790]: I1124 13:36:49.683341 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="099c81d1-2be3-4e10-b733-65e66912b605" containerName="extract-utilities" Nov 24 13:36:49 crc kubenswrapper[4790]: E1124 13:36:49.683356 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="099c81d1-2be3-4e10-b733-65e66912b605" containerName="extract-content" Nov 24 13:36:49 crc kubenswrapper[4790]: I1124 13:36:49.683363 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="099c81d1-2be3-4e10-b733-65e66912b605" containerName="extract-content" Nov 24 13:36:49 crc kubenswrapper[4790]: E1124 13:36:49.683394 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="099c81d1-2be3-4e10-b733-65e66912b605" containerName="registry-server" Nov 24 13:36:49 crc kubenswrapper[4790]: I1124 13:36:49.683400 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="099c81d1-2be3-4e10-b733-65e66912b605" containerName="registry-server" Nov 24 13:36:49 crc kubenswrapper[4790]: I1124 13:36:49.683608 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="099c81d1-2be3-4e10-b733-65e66912b605" containerName="registry-server" Nov 24 13:36:49 crc kubenswrapper[4790]: I1124 13:36:49.685149 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f455q" Nov 24 13:36:49 crc kubenswrapper[4790]: I1124 13:36:49.693291 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f455q"] Nov 24 13:36:49 crc kubenswrapper[4790]: I1124 13:36:49.793286 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndkmg\" (UniqueName: \"kubernetes.io/projected/ce775c5d-0636-476c-8f6f-938b7b864729-kube-api-access-ndkmg\") pod \"community-operators-f455q\" (UID: \"ce775c5d-0636-476c-8f6f-938b7b864729\") " pod="openshift-marketplace/community-operators-f455q" Nov 24 13:36:49 crc kubenswrapper[4790]: I1124 13:36:49.794203 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce775c5d-0636-476c-8f6f-938b7b864729-catalog-content\") pod \"community-operators-f455q\" (UID: \"ce775c5d-0636-476c-8f6f-938b7b864729\") " pod="openshift-marketplace/community-operators-f455q" Nov 24 13:36:49 crc kubenswrapper[4790]: I1124 13:36:49.794328 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce775c5d-0636-476c-8f6f-938b7b864729-utilities\") pod \"community-operators-f455q\" (UID: \"ce775c5d-0636-476c-8f6f-938b7b864729\") " pod="openshift-marketplace/community-operators-f455q" Nov 24 13:36:49 crc kubenswrapper[4790]: I1124 13:36:49.895950 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce775c5d-0636-476c-8f6f-938b7b864729-utilities\") pod \"community-operators-f455q\" (UID: \"ce775c5d-0636-476c-8f6f-938b7b864729\") " pod="openshift-marketplace/community-operators-f455q" Nov 24 13:36:49 crc kubenswrapper[4790]: I1124 13:36:49.896189 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndkmg\" (UniqueName: \"kubernetes.io/projected/ce775c5d-0636-476c-8f6f-938b7b864729-kube-api-access-ndkmg\") pod \"community-operators-f455q\" (UID: \"ce775c5d-0636-476c-8f6f-938b7b864729\") " pod="openshift-marketplace/community-operators-f455q" Nov 24 13:36:49 crc kubenswrapper[4790]: I1124 13:36:49.896216 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce775c5d-0636-476c-8f6f-938b7b864729-catalog-content\") pod \"community-operators-f455q\" (UID: \"ce775c5d-0636-476c-8f6f-938b7b864729\") " pod="openshift-marketplace/community-operators-f455q" Nov 24 13:36:49 crc kubenswrapper[4790]: I1124 13:36:49.896683 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce775c5d-0636-476c-8f6f-938b7b864729-utilities\") pod \"community-operators-f455q\" (UID: \"ce775c5d-0636-476c-8f6f-938b7b864729\") " pod="openshift-marketplace/community-operators-f455q" Nov 24 13:36:49 crc kubenswrapper[4790]: I1124 13:36:49.896718 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce775c5d-0636-476c-8f6f-938b7b864729-catalog-content\") pod \"community-operators-f455q\" (UID: \"ce775c5d-0636-476c-8f6f-938b7b864729\") " pod="openshift-marketplace/community-operators-f455q" Nov 24 13:36:49 crc kubenswrapper[4790]: I1124 13:36:49.918099 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndkmg\" (UniqueName: \"kubernetes.io/projected/ce775c5d-0636-476c-8f6f-938b7b864729-kube-api-access-ndkmg\") pod \"community-operators-f455q\" (UID: \"ce775c5d-0636-476c-8f6f-938b7b864729\") " pod="openshift-marketplace/community-operators-f455q" Nov 24 13:36:50 crc kubenswrapper[4790]: I1124 13:36:50.014642 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f455q" Nov 24 13:36:50 crc kubenswrapper[4790]: I1124 13:36:50.509766 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f455q"] Nov 24 13:36:50 crc kubenswrapper[4790]: W1124 13:36:50.510505 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce775c5d_0636_476c_8f6f_938b7b864729.slice/crio-6780db93f1d68e768ff86d6fff6cdac452006bf3e3e4496349e66c745ea4a75c WatchSource:0}: Error finding container 6780db93f1d68e768ff86d6fff6cdac452006bf3e3e4496349e66c745ea4a75c: Status 404 returned error can't find the container with id 6780db93f1d68e768ff86d6fff6cdac452006bf3e3e4496349e66c745ea4a75c Nov 24 13:36:50 crc kubenswrapper[4790]: I1124 13:36:50.601626 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f455q" event={"ID":"ce775c5d-0636-476c-8f6f-938b7b864729","Type":"ContainerStarted","Data":"6780db93f1d68e768ff86d6fff6cdac452006bf3e3e4496349e66c745ea4a75c"} Nov 24 13:36:51 crc kubenswrapper[4790]: I1124 13:36:51.626513 4790 generic.go:334] "Generic (PLEG): container finished" podID="ce775c5d-0636-476c-8f6f-938b7b864729" containerID="bc39982243cba0449a5052e0735ed482ccb8aa0bdf2a0b7f866b880b0c76716a" exitCode=0 Nov 24 13:36:51 crc kubenswrapper[4790]: I1124 13:36:51.626556 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f455q" event={"ID":"ce775c5d-0636-476c-8f6f-938b7b864729","Type":"ContainerDied","Data":"bc39982243cba0449a5052e0735ed482ccb8aa0bdf2a0b7f866b880b0c76716a"} Nov 24 13:36:51 crc kubenswrapper[4790]: I1124 13:36:51.711285 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 13:36:52 crc kubenswrapper[4790]: I1124 13:36:52.776188 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 13:36:52 crc kubenswrapper[4790]: I1124 13:36:52.777130 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 13:36:52 crc kubenswrapper[4790]: I1124 13:36:52.777415 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 13:36:52 crc kubenswrapper[4790]: I1124 13:36:52.784540 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 13:36:53 crc kubenswrapper[4790]: I1124 13:36:53.644446 4790 generic.go:334] "Generic (PLEG): container finished" podID="ce775c5d-0636-476c-8f6f-938b7b864729" containerID="01df8dde17854c668b8d9428adcace8f324086d0cd64bf4249ea0a226f3444b2" exitCode=0 Nov 24 13:36:53 crc kubenswrapper[4790]: I1124 13:36:53.644502 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f455q" event={"ID":"ce775c5d-0636-476c-8f6f-938b7b864729","Type":"ContainerDied","Data":"01df8dde17854c668b8d9428adcace8f324086d0cd64bf4249ea0a226f3444b2"} Nov 24 13:36:53 crc kubenswrapper[4790]: I1124 13:36:53.645269 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 13:36:53 crc kubenswrapper[4790]: I1124 13:36:53.653328 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 13:36:55 crc kubenswrapper[4790]: I1124 13:36:55.665402 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f455q" event={"ID":"ce775c5d-0636-476c-8f6f-938b7b864729","Type":"ContainerStarted","Data":"6da24a9eb1a22e378be65a1e456969f40b9c9412b8704836f080961ecd88f313"} Nov 24 13:36:55 crc kubenswrapper[4790]: I1124 13:36:55.695465 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-f455q" podStartSLOduration=3.169947771 podStartE2EDuration="6.695435696s" podCreationTimestamp="2025-11-24 13:36:49 +0000 UTC" firstStartedPulling="2025-11-24 13:36:51.628127029 +0000 UTC m=+1460.008020691" lastFinishedPulling="2025-11-24 13:36:55.153614954 +0000 UTC m=+1463.533508616" observedRunningTime="2025-11-24 13:36:55.685484552 +0000 UTC m=+1464.065378224" watchObservedRunningTime="2025-11-24 13:36:55.695435696 +0000 UTC m=+1464.075329358" Nov 24 13:36:55 crc kubenswrapper[4790]: I1124 13:36:55.896873 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 13:36:55 crc kubenswrapper[4790]: I1124 13:36:55.897543 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 13:36:55 crc kubenswrapper[4790]: I1124 13:36:55.909433 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 13:36:56 crc kubenswrapper[4790]: I1124 13:36:56.679530 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 13:37:00 crc kubenswrapper[4790]: I1124 13:37:00.015369 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-f455q" Nov 24 13:37:00 crc kubenswrapper[4790]: I1124 13:37:00.015808 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-f455q" Nov 24 13:37:00 crc kubenswrapper[4790]: I1124 13:37:00.064065 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-f455q" Nov 24 13:37:00 crc kubenswrapper[4790]: I1124 13:37:00.786216 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-f455q" Nov 24 13:37:00 crc kubenswrapper[4790]: I1124 13:37:00.840974 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f455q"] Nov 24 13:37:02 crc kubenswrapper[4790]: I1124 13:37:02.739350 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-f455q" podUID="ce775c5d-0636-476c-8f6f-938b7b864729" containerName="registry-server" containerID="cri-o://6da24a9eb1a22e378be65a1e456969f40b9c9412b8704836f080961ecd88f313" gracePeriod=2 Nov 24 13:37:03 crc kubenswrapper[4790]: I1124 13:37:03.169302 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f455q" Nov 24 13:37:03 crc kubenswrapper[4790]: I1124 13:37:03.241646 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce775c5d-0636-476c-8f6f-938b7b864729-catalog-content\") pod \"ce775c5d-0636-476c-8f6f-938b7b864729\" (UID: \"ce775c5d-0636-476c-8f6f-938b7b864729\") " Nov 24 13:37:03 crc kubenswrapper[4790]: I1124 13:37:03.241855 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ndkmg\" (UniqueName: \"kubernetes.io/projected/ce775c5d-0636-476c-8f6f-938b7b864729-kube-api-access-ndkmg\") pod \"ce775c5d-0636-476c-8f6f-938b7b864729\" (UID: \"ce775c5d-0636-476c-8f6f-938b7b864729\") " Nov 24 13:37:03 crc kubenswrapper[4790]: I1124 13:37:03.241922 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce775c5d-0636-476c-8f6f-938b7b864729-utilities\") pod \"ce775c5d-0636-476c-8f6f-938b7b864729\" (UID: \"ce775c5d-0636-476c-8f6f-938b7b864729\") " Nov 24 13:37:03 crc kubenswrapper[4790]: I1124 13:37:03.243027 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce775c5d-0636-476c-8f6f-938b7b864729-utilities" (OuterVolumeSpecName: "utilities") pod "ce775c5d-0636-476c-8f6f-938b7b864729" (UID: "ce775c5d-0636-476c-8f6f-938b7b864729"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:03 crc kubenswrapper[4790]: I1124 13:37:03.252263 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce775c5d-0636-476c-8f6f-938b7b864729-kube-api-access-ndkmg" (OuterVolumeSpecName: "kube-api-access-ndkmg") pod "ce775c5d-0636-476c-8f6f-938b7b864729" (UID: "ce775c5d-0636-476c-8f6f-938b7b864729"). InnerVolumeSpecName "kube-api-access-ndkmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:03 crc kubenswrapper[4790]: I1124 13:37:03.315580 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce775c5d-0636-476c-8f6f-938b7b864729-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ce775c5d-0636-476c-8f6f-938b7b864729" (UID: "ce775c5d-0636-476c-8f6f-938b7b864729"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:03 crc kubenswrapper[4790]: I1124 13:37:03.344021 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ndkmg\" (UniqueName: \"kubernetes.io/projected/ce775c5d-0636-476c-8f6f-938b7b864729-kube-api-access-ndkmg\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:03 crc kubenswrapper[4790]: I1124 13:37:03.344067 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce775c5d-0636-476c-8f6f-938b7b864729-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:03 crc kubenswrapper[4790]: I1124 13:37:03.344083 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce775c5d-0636-476c-8f6f-938b7b864729-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:03 crc kubenswrapper[4790]: I1124 13:37:03.755324 4790 generic.go:334] "Generic (PLEG): container finished" podID="ce775c5d-0636-476c-8f6f-938b7b864729" containerID="6da24a9eb1a22e378be65a1e456969f40b9c9412b8704836f080961ecd88f313" exitCode=0 Nov 24 13:37:03 crc kubenswrapper[4790]: I1124 13:37:03.755390 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f455q" event={"ID":"ce775c5d-0636-476c-8f6f-938b7b864729","Type":"ContainerDied","Data":"6da24a9eb1a22e378be65a1e456969f40b9c9412b8704836f080961ecd88f313"} Nov 24 13:37:03 crc kubenswrapper[4790]: I1124 13:37:03.755405 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f455q" Nov 24 13:37:03 crc kubenswrapper[4790]: I1124 13:37:03.755441 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f455q" event={"ID":"ce775c5d-0636-476c-8f6f-938b7b864729","Type":"ContainerDied","Data":"6780db93f1d68e768ff86d6fff6cdac452006bf3e3e4496349e66c745ea4a75c"} Nov 24 13:37:03 crc kubenswrapper[4790]: I1124 13:37:03.755465 4790 scope.go:117] "RemoveContainer" containerID="6da24a9eb1a22e378be65a1e456969f40b9c9412b8704836f080961ecd88f313" Nov 24 13:37:03 crc kubenswrapper[4790]: I1124 13:37:03.788138 4790 scope.go:117] "RemoveContainer" containerID="01df8dde17854c668b8d9428adcace8f324086d0cd64bf4249ea0a226f3444b2" Nov 24 13:37:03 crc kubenswrapper[4790]: I1124 13:37:03.792911 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f455q"] Nov 24 13:37:03 crc kubenswrapper[4790]: I1124 13:37:03.802540 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-f455q"] Nov 24 13:37:03 crc kubenswrapper[4790]: I1124 13:37:03.807452 4790 scope.go:117] "RemoveContainer" containerID="bc39982243cba0449a5052e0735ed482ccb8aa0bdf2a0b7f866b880b0c76716a" Nov 24 13:37:03 crc kubenswrapper[4790]: I1124 13:37:03.848461 4790 scope.go:117] "RemoveContainer" containerID="6da24a9eb1a22e378be65a1e456969f40b9c9412b8704836f080961ecd88f313" Nov 24 13:37:03 crc kubenswrapper[4790]: E1124 13:37:03.849015 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6da24a9eb1a22e378be65a1e456969f40b9c9412b8704836f080961ecd88f313\": container with ID starting with 6da24a9eb1a22e378be65a1e456969f40b9c9412b8704836f080961ecd88f313 not found: ID does not exist" containerID="6da24a9eb1a22e378be65a1e456969f40b9c9412b8704836f080961ecd88f313" Nov 24 13:37:03 crc kubenswrapper[4790]: I1124 13:37:03.849245 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6da24a9eb1a22e378be65a1e456969f40b9c9412b8704836f080961ecd88f313"} err="failed to get container status \"6da24a9eb1a22e378be65a1e456969f40b9c9412b8704836f080961ecd88f313\": rpc error: code = NotFound desc = could not find container \"6da24a9eb1a22e378be65a1e456969f40b9c9412b8704836f080961ecd88f313\": container with ID starting with 6da24a9eb1a22e378be65a1e456969f40b9c9412b8704836f080961ecd88f313 not found: ID does not exist" Nov 24 13:37:03 crc kubenswrapper[4790]: I1124 13:37:03.849334 4790 scope.go:117] "RemoveContainer" containerID="01df8dde17854c668b8d9428adcace8f324086d0cd64bf4249ea0a226f3444b2" Nov 24 13:37:03 crc kubenswrapper[4790]: E1124 13:37:03.849741 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01df8dde17854c668b8d9428adcace8f324086d0cd64bf4249ea0a226f3444b2\": container with ID starting with 01df8dde17854c668b8d9428adcace8f324086d0cd64bf4249ea0a226f3444b2 not found: ID does not exist" containerID="01df8dde17854c668b8d9428adcace8f324086d0cd64bf4249ea0a226f3444b2" Nov 24 13:37:03 crc kubenswrapper[4790]: I1124 13:37:03.849763 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01df8dde17854c668b8d9428adcace8f324086d0cd64bf4249ea0a226f3444b2"} err="failed to get container status \"01df8dde17854c668b8d9428adcace8f324086d0cd64bf4249ea0a226f3444b2\": rpc error: code = NotFound desc = could not find container \"01df8dde17854c668b8d9428adcace8f324086d0cd64bf4249ea0a226f3444b2\": container with ID starting with 01df8dde17854c668b8d9428adcace8f324086d0cd64bf4249ea0a226f3444b2 not found: ID does not exist" Nov 24 13:37:03 crc kubenswrapper[4790]: I1124 13:37:03.849781 4790 scope.go:117] "RemoveContainer" containerID="bc39982243cba0449a5052e0735ed482ccb8aa0bdf2a0b7f866b880b0c76716a" Nov 24 13:37:03 crc kubenswrapper[4790]: E1124 13:37:03.850062 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc39982243cba0449a5052e0735ed482ccb8aa0bdf2a0b7f866b880b0c76716a\": container with ID starting with bc39982243cba0449a5052e0735ed482ccb8aa0bdf2a0b7f866b880b0c76716a not found: ID does not exist" containerID="bc39982243cba0449a5052e0735ed482ccb8aa0bdf2a0b7f866b880b0c76716a" Nov 24 13:37:03 crc kubenswrapper[4790]: I1124 13:37:03.850082 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc39982243cba0449a5052e0735ed482ccb8aa0bdf2a0b7f866b880b0c76716a"} err="failed to get container status \"bc39982243cba0449a5052e0735ed482ccb8aa0bdf2a0b7f866b880b0c76716a\": rpc error: code = NotFound desc = could not find container \"bc39982243cba0449a5052e0735ed482ccb8aa0bdf2a0b7f866b880b0c76716a\": container with ID starting with bc39982243cba0449a5052e0735ed482ccb8aa0bdf2a0b7f866b880b0c76716a not found: ID does not exist" Nov 24 13:37:04 crc kubenswrapper[4790]: I1124 13:37:04.324673 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce775c5d-0636-476c-8f6f-938b7b864729" path="/var/lib/kubelet/pods/ce775c5d-0636-476c-8f6f-938b7b864729/volumes" Nov 24 13:37:06 crc kubenswrapper[4790]: I1124 13:37:06.685023 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bh9qg"] Nov 24 13:37:06 crc kubenswrapper[4790]: E1124 13:37:06.685803 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce775c5d-0636-476c-8f6f-938b7b864729" containerName="extract-content" Nov 24 13:37:06 crc kubenswrapper[4790]: I1124 13:37:06.685822 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce775c5d-0636-476c-8f6f-938b7b864729" containerName="extract-content" Nov 24 13:37:06 crc kubenswrapper[4790]: E1124 13:37:06.685849 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce775c5d-0636-476c-8f6f-938b7b864729" containerName="extract-utilities" Nov 24 13:37:06 crc kubenswrapper[4790]: I1124 13:37:06.685859 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce775c5d-0636-476c-8f6f-938b7b864729" containerName="extract-utilities" Nov 24 13:37:06 crc kubenswrapper[4790]: E1124 13:37:06.685873 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce775c5d-0636-476c-8f6f-938b7b864729" containerName="registry-server" Nov 24 13:37:06 crc kubenswrapper[4790]: I1124 13:37:06.685906 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce775c5d-0636-476c-8f6f-938b7b864729" containerName="registry-server" Nov 24 13:37:06 crc kubenswrapper[4790]: I1124 13:37:06.686191 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce775c5d-0636-476c-8f6f-938b7b864729" containerName="registry-server" Nov 24 13:37:06 crc kubenswrapper[4790]: I1124 13:37:06.687906 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bh9qg" Nov 24 13:37:06 crc kubenswrapper[4790]: I1124 13:37:06.701686 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bh9qg"] Nov 24 13:37:06 crc kubenswrapper[4790]: I1124 13:37:06.703863 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb72fd7a-89c4-4a8a-a7b0-c758322a08b7-utilities\") pod \"redhat-marketplace-bh9qg\" (UID: \"bb72fd7a-89c4-4a8a-a7b0-c758322a08b7\") " pod="openshift-marketplace/redhat-marketplace-bh9qg" Nov 24 13:37:06 crc kubenswrapper[4790]: I1124 13:37:06.703949 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb72fd7a-89c4-4a8a-a7b0-c758322a08b7-catalog-content\") pod \"redhat-marketplace-bh9qg\" (UID: \"bb72fd7a-89c4-4a8a-a7b0-c758322a08b7\") " pod="openshift-marketplace/redhat-marketplace-bh9qg" Nov 24 13:37:06 crc kubenswrapper[4790]: I1124 13:37:06.704129 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrwvv\" (UniqueName: \"kubernetes.io/projected/bb72fd7a-89c4-4a8a-a7b0-c758322a08b7-kube-api-access-xrwvv\") pod \"redhat-marketplace-bh9qg\" (UID: \"bb72fd7a-89c4-4a8a-a7b0-c758322a08b7\") " pod="openshift-marketplace/redhat-marketplace-bh9qg" Nov 24 13:37:06 crc kubenswrapper[4790]: I1124 13:37:06.805267 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrwvv\" (UniqueName: \"kubernetes.io/projected/bb72fd7a-89c4-4a8a-a7b0-c758322a08b7-kube-api-access-xrwvv\") pod \"redhat-marketplace-bh9qg\" (UID: \"bb72fd7a-89c4-4a8a-a7b0-c758322a08b7\") " pod="openshift-marketplace/redhat-marketplace-bh9qg" Nov 24 13:37:06 crc kubenswrapper[4790]: I1124 13:37:06.805331 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb72fd7a-89c4-4a8a-a7b0-c758322a08b7-utilities\") pod \"redhat-marketplace-bh9qg\" (UID: \"bb72fd7a-89c4-4a8a-a7b0-c758322a08b7\") " pod="openshift-marketplace/redhat-marketplace-bh9qg" Nov 24 13:37:06 crc kubenswrapper[4790]: I1124 13:37:06.805360 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb72fd7a-89c4-4a8a-a7b0-c758322a08b7-catalog-content\") pod \"redhat-marketplace-bh9qg\" (UID: \"bb72fd7a-89c4-4a8a-a7b0-c758322a08b7\") " pod="openshift-marketplace/redhat-marketplace-bh9qg" Nov 24 13:37:06 crc kubenswrapper[4790]: I1124 13:37:06.805778 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb72fd7a-89c4-4a8a-a7b0-c758322a08b7-catalog-content\") pod \"redhat-marketplace-bh9qg\" (UID: \"bb72fd7a-89c4-4a8a-a7b0-c758322a08b7\") " pod="openshift-marketplace/redhat-marketplace-bh9qg" Nov 24 13:37:06 crc kubenswrapper[4790]: I1124 13:37:06.805856 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb72fd7a-89c4-4a8a-a7b0-c758322a08b7-utilities\") pod \"redhat-marketplace-bh9qg\" (UID: \"bb72fd7a-89c4-4a8a-a7b0-c758322a08b7\") " pod="openshift-marketplace/redhat-marketplace-bh9qg" Nov 24 13:37:06 crc kubenswrapper[4790]: I1124 13:37:06.824341 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrwvv\" (UniqueName: \"kubernetes.io/projected/bb72fd7a-89c4-4a8a-a7b0-c758322a08b7-kube-api-access-xrwvv\") pod \"redhat-marketplace-bh9qg\" (UID: \"bb72fd7a-89c4-4a8a-a7b0-c758322a08b7\") " pod="openshift-marketplace/redhat-marketplace-bh9qg" Nov 24 13:37:07 crc kubenswrapper[4790]: I1124 13:37:07.011561 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bh9qg" Nov 24 13:37:07 crc kubenswrapper[4790]: I1124 13:37:07.480597 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bh9qg"] Nov 24 13:37:07 crc kubenswrapper[4790]: I1124 13:37:07.791958 4790 generic.go:334] "Generic (PLEG): container finished" podID="bb72fd7a-89c4-4a8a-a7b0-c758322a08b7" containerID="a0bab3581d08313eb5fb228fef8c8dc47ab3fb05a7b2805a913fce1b4d13047e" exitCode=0 Nov 24 13:37:07 crc kubenswrapper[4790]: I1124 13:37:07.792053 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bh9qg" event={"ID":"bb72fd7a-89c4-4a8a-a7b0-c758322a08b7","Type":"ContainerDied","Data":"a0bab3581d08313eb5fb228fef8c8dc47ab3fb05a7b2805a913fce1b4d13047e"} Nov 24 13:37:07 crc kubenswrapper[4790]: I1124 13:37:07.792294 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bh9qg" event={"ID":"bb72fd7a-89c4-4a8a-a7b0-c758322a08b7","Type":"ContainerStarted","Data":"64380177ab7771b2d7b3c1ad694c3450d1fb45313e2b57caf0b4716126ab5d21"} Nov 24 13:37:08 crc kubenswrapper[4790]: I1124 13:37:08.802225 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bh9qg" event={"ID":"bb72fd7a-89c4-4a8a-a7b0-c758322a08b7","Type":"ContainerStarted","Data":"212a63776bee7e4aa0a7c001871fb866bdc261b22ca9ead26adc668341a14cd7"} Nov 24 13:37:09 crc kubenswrapper[4790]: I1124 13:37:09.814521 4790 generic.go:334] "Generic (PLEG): container finished" podID="bb72fd7a-89c4-4a8a-a7b0-c758322a08b7" containerID="212a63776bee7e4aa0a7c001871fb866bdc261b22ca9ead26adc668341a14cd7" exitCode=0 Nov 24 13:37:09 crc kubenswrapper[4790]: I1124 13:37:09.814575 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bh9qg" event={"ID":"bb72fd7a-89c4-4a8a-a7b0-c758322a08b7","Type":"ContainerDied","Data":"212a63776bee7e4aa0a7c001871fb866bdc261b22ca9ead26adc668341a14cd7"} Nov 24 13:37:10 crc kubenswrapper[4790]: I1124 13:37:10.824158 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bh9qg" event={"ID":"bb72fd7a-89c4-4a8a-a7b0-c758322a08b7","Type":"ContainerStarted","Data":"56961087bbae3f0602d7c24580f19c5fa8eba1503fda903df84c3561d57e0b05"} Nov 24 13:37:10 crc kubenswrapper[4790]: I1124 13:37:10.847986 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bh9qg" podStartSLOduration=2.44868932 podStartE2EDuration="4.847968806s" podCreationTimestamp="2025-11-24 13:37:06 +0000 UTC" firstStartedPulling="2025-11-24 13:37:07.794543028 +0000 UTC m=+1476.174436690" lastFinishedPulling="2025-11-24 13:37:10.193822514 +0000 UTC m=+1478.573716176" observedRunningTime="2025-11-24 13:37:10.842237172 +0000 UTC m=+1479.222130844" watchObservedRunningTime="2025-11-24 13:37:10.847968806 +0000 UTC m=+1479.227862468" Nov 24 13:37:13 crc kubenswrapper[4790]: I1124 13:37:13.938649 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:37:13 crc kubenswrapper[4790]: I1124 13:37:13.939299 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:37:13 crc kubenswrapper[4790]: I1124 13:37:13.939348 4790 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 13:37:13 crc kubenswrapper[4790]: I1124 13:37:13.940103 4790 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8544370c5579779a1c61d01762c06ad3cb058bc1368085e6b5750e300ddbf449"} pod="openshift-machine-config-operator/machine-config-daemon-xz49t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 13:37:13 crc kubenswrapper[4790]: I1124 13:37:13.940174 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" containerID="cri-o://8544370c5579779a1c61d01762c06ad3cb058bc1368085e6b5750e300ddbf449" gracePeriod=600 Nov 24 13:37:14 crc kubenswrapper[4790]: I1124 13:37:14.866382 4790 generic.go:334] "Generic (PLEG): container finished" podID="0d73b133-48f0-455f-8f6a-742e633f631a" containerID="8544370c5579779a1c61d01762c06ad3cb058bc1368085e6b5750e300ddbf449" exitCode=0 Nov 24 13:37:14 crc kubenswrapper[4790]: I1124 13:37:14.866455 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerDied","Data":"8544370c5579779a1c61d01762c06ad3cb058bc1368085e6b5750e300ddbf449"} Nov 24 13:37:14 crc kubenswrapper[4790]: I1124 13:37:14.866710 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"0e39e707a2686ae4ea3e31dae8bfe7595a875cadfa529bb73dd02722c8084b9e"} Nov 24 13:37:14 crc kubenswrapper[4790]: I1124 13:37:14.866729 4790 scope.go:117] "RemoveContainer" containerID="56e28eeaa0ab287a217413e69cba64d88c1ede93557f55581b2a1331d38d609a" Nov 24 13:37:16 crc kubenswrapper[4790]: I1124 13:37:16.280629 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 24 13:37:16 crc kubenswrapper[4790]: I1124 13:37:16.281284 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c" containerName="openstackclient" containerID="cri-o://0981452a5c31b7ac7be6aca3f35f652875423629b90d147c20e6f8b123ad006d" gracePeriod=2 Nov 24 13:37:16 crc kubenswrapper[4790]: I1124 13:37:16.357437 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 24 13:37:16 crc kubenswrapper[4790]: I1124 13:37:16.696954 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-28cwj"] Nov 24 13:37:16 crc kubenswrapper[4790]: I1124 13:37:16.736947 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-rkxr6"] Nov 24 13:37:16 crc kubenswrapper[4790]: I1124 13:37:16.773952 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 13:37:16 crc kubenswrapper[4790]: I1124 13:37:16.795105 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement0fc6-account-delete-rl7sl"] Nov 24 13:37:16 crc kubenswrapper[4790]: E1124 13:37:16.795541 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c" containerName="openstackclient" Nov 24 13:37:16 crc kubenswrapper[4790]: I1124 13:37:16.795581 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c" containerName="openstackclient" Nov 24 13:37:16 crc kubenswrapper[4790]: I1124 13:37:16.795792 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c" containerName="openstackclient" Nov 24 13:37:16 crc kubenswrapper[4790]: I1124 13:37:16.796452 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement0fc6-account-delete-rl7sl" Nov 24 13:37:16 crc kubenswrapper[4790]: I1124 13:37:16.833172 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement0fc6-account-delete-rl7sl"] Nov 24 13:37:16 crc kubenswrapper[4790]: I1124 13:37:16.901943 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 13:37:16 crc kubenswrapper[4790]: I1124 13:37:16.902213 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="a3fa97ac-7189-4828-8b95-664e852dea8e" containerName="ovn-northd" containerID="cri-o://c71cf899ea7e4a2eff4caa8290abac4a40d0072f283f44d449955d488101621e" gracePeriod=30 Nov 24 13:37:16 crc kubenswrapper[4790]: I1124 13:37:16.902596 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="a3fa97ac-7189-4828-8b95-664e852dea8e" containerName="openstack-network-exporter" containerID="cri-o://889043e225f3f968efc7dfcf48a4fcb3822a1e4d8cf106f1af9b36a1cda57399" gracePeriod=30 Nov 24 13:37:16 crc kubenswrapper[4790]: I1124 13:37:16.906048 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51552477-67be-4c73-ae48-aa131945caea-operator-scripts\") pod \"placement0fc6-account-delete-rl7sl\" (UID: \"51552477-67be-4c73-ae48-aa131945caea\") " pod="openstack/placement0fc6-account-delete-rl7sl" Nov 24 13:37:16 crc kubenswrapper[4790]: I1124 13:37:16.906276 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcxrv\" (UniqueName: \"kubernetes.io/projected/51552477-67be-4c73-ae48-aa131945caea-kube-api-access-hcxrv\") pod \"placement0fc6-account-delete-rl7sl\" (UID: \"51552477-67be-4c73-ae48-aa131945caea\") " pod="openstack/placement0fc6-account-delete-rl7sl" Nov 24 13:37:16 crc kubenswrapper[4790]: E1124 13:37:16.907410 4790 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 24 13:37:16 crc kubenswrapper[4790]: E1124 13:37:16.907537 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/85e04b92-ae17-4eb6-a736-395c5d44f563-config-data podName:85e04b92-ae17-4eb6-a736-395c5d44f563 nodeName:}" failed. No retries permitted until 2025-11-24 13:37:17.407494197 +0000 UTC m=+1485.787387859 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/85e04b92-ae17-4eb6-a736-395c5d44f563-config-data") pod "rabbitmq-cell1-server-0" (UID: "85e04b92-ae17-4eb6-a736-395c5d44f563") : configmap "rabbitmq-cell1-config-data" not found Nov 24 13:37:16 crc kubenswrapper[4790]: I1124 13:37:16.917724 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbicana018-account-delete-6z4kv"] Nov 24 13:37:16 crc kubenswrapper[4790]: I1124 13:37:16.918989 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbicana018-account-delete-6z4kv" Nov 24 13:37:16 crc kubenswrapper[4790]: I1124 13:37:16.927737 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-q4sb2"] Nov 24 13:37:16 crc kubenswrapper[4790]: I1124 13:37:16.952325 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-q4sb2"] Nov 24 13:37:16 crc kubenswrapper[4790]: I1124 13:37:16.967034 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbicana018-account-delete-6z4kv"] Nov 24 13:37:16 crc kubenswrapper[4790]: I1124 13:37:16.990662 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-z8gjd"] Nov 24 13:37:16 crc kubenswrapper[4790]: I1124 13:37:16.990920 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-z8gjd" podUID="a9c5a116-6438-47b5-8462-207eb656b553" containerName="openstack-network-exporter" containerID="cri-o://c6902baf408c3891cff01a43b41f9f037b6b29b99555835a2a635032bcee8c85" gracePeriod=30 Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.007065 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinderc394-account-delete-gnvr4"] Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.008716 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinderc394-account-delete-gnvr4" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.013193 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bh9qg" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.014293 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bh9qg" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.019691 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51552477-67be-4c73-ae48-aa131945caea-operator-scripts\") pod \"placement0fc6-account-delete-rl7sl\" (UID: \"51552477-67be-4c73-ae48-aa131945caea\") " pod="openstack/placement0fc6-account-delete-rl7sl" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.019867 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mw5sg\" (UniqueName: \"kubernetes.io/projected/7f52457e-46da-4edd-8d76-dc4a28f02f0a-kube-api-access-mw5sg\") pod \"barbicana018-account-delete-6z4kv\" (UID: \"7f52457e-46da-4edd-8d76-dc4a28f02f0a\") " pod="openstack/barbicana018-account-delete-6z4kv" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.019995 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcxrv\" (UniqueName: \"kubernetes.io/projected/51552477-67be-4c73-ae48-aa131945caea-kube-api-access-hcxrv\") pod \"placement0fc6-account-delete-rl7sl\" (UID: \"51552477-67be-4c73-ae48-aa131945caea\") " pod="openstack/placement0fc6-account-delete-rl7sl" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.020098 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f52457e-46da-4edd-8d76-dc4a28f02f0a-operator-scripts\") pod \"barbicana018-account-delete-6z4kv\" (UID: \"7f52457e-46da-4edd-8d76-dc4a28f02f0a\") " pod="openstack/barbicana018-account-delete-6z4kv" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.020547 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51552477-67be-4c73-ae48-aa131945caea-operator-scripts\") pod \"placement0fc6-account-delete-rl7sl\" (UID: \"51552477-67be-4c73-ae48-aa131945caea\") " pod="openstack/placement0fc6-account-delete-rl7sl" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.023424 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinderc394-account-delete-gnvr4"] Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.062990 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcxrv\" (UniqueName: \"kubernetes.io/projected/51552477-67be-4c73-ae48-aa131945caea-kube-api-access-hcxrv\") pod \"placement0fc6-account-delete-rl7sl\" (UID: \"51552477-67be-4c73-ae48-aa131945caea\") " pod="openstack/placement0fc6-account-delete-rl7sl" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.067582 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-mbm86"] Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.081288 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-mbm86"] Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.122578 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bh9qg" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.127397 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pw9pq\" (UniqueName: \"kubernetes.io/projected/8095e4aa-7290-441d-ba67-79e48c23b8a6-kube-api-access-pw9pq\") pod \"cinderc394-account-delete-gnvr4\" (UID: \"8095e4aa-7290-441d-ba67-79e48c23b8a6\") " pod="openstack/cinderc394-account-delete-gnvr4" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.127478 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f52457e-46da-4edd-8d76-dc4a28f02f0a-operator-scripts\") pod \"barbicana018-account-delete-6z4kv\" (UID: \"7f52457e-46da-4edd-8d76-dc4a28f02f0a\") " pod="openstack/barbicana018-account-delete-6z4kv" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.127572 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8095e4aa-7290-441d-ba67-79e48c23b8a6-operator-scripts\") pod \"cinderc394-account-delete-gnvr4\" (UID: \"8095e4aa-7290-441d-ba67-79e48c23b8a6\") " pod="openstack/cinderc394-account-delete-gnvr4" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.128731 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mw5sg\" (UniqueName: \"kubernetes.io/projected/7f52457e-46da-4edd-8d76-dc4a28f02f0a-kube-api-access-mw5sg\") pod \"barbicana018-account-delete-6z4kv\" (UID: \"7f52457e-46da-4edd-8d76-dc4a28f02f0a\") " pod="openstack/barbicana018-account-delete-6z4kv" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.134795 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement0fc6-account-delete-rl7sl" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.147037 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f52457e-46da-4edd-8d76-dc4a28f02f0a-operator-scripts\") pod \"barbicana018-account-delete-6z4kv\" (UID: \"7f52457e-46da-4edd-8d76-dc4a28f02f0a\") " pod="openstack/barbicana018-account-delete-6z4kv" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.165994 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance388c-account-delete-8hfh2"] Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.167185 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance388c-account-delete-8hfh2" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.206967 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance388c-account-delete-8hfh2"] Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.210708 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mw5sg\" (UniqueName: \"kubernetes.io/projected/7f52457e-46da-4edd-8d76-dc4a28f02f0a-kube-api-access-mw5sg\") pod \"barbicana018-account-delete-6z4kv\" (UID: \"7f52457e-46da-4edd-8d76-dc4a28f02f0a\") " pod="openstack/barbicana018-account-delete-6z4kv" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.217353 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutronf6b9-account-delete-tdltj"] Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.220318 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutronf6b9-account-delete-tdltj" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.234415 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8095e4aa-7290-441d-ba67-79e48c23b8a6-operator-scripts\") pod \"cinderc394-account-delete-gnvr4\" (UID: \"8095e4aa-7290-441d-ba67-79e48c23b8a6\") " pod="openstack/cinderc394-account-delete-gnvr4" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.234583 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pw9pq\" (UniqueName: \"kubernetes.io/projected/8095e4aa-7290-441d-ba67-79e48c23b8a6-kube-api-access-pw9pq\") pod \"cinderc394-account-delete-gnvr4\" (UID: \"8095e4aa-7290-441d-ba67-79e48c23b8a6\") " pod="openstack/cinderc394-account-delete-gnvr4" Nov 24 13:37:17 crc kubenswrapper[4790]: E1124 13:37:17.268439 4790 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err="command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: " execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-rkxr6" message="Exiting ovn-controller (1) " Nov 24 13:37:17 crc kubenswrapper[4790]: E1124 13:37:17.268484 4790 kuberuntime_container.go:691] "PreStop hook failed" err="command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: " pod="openstack/ovn-controller-rkxr6" podUID="c8fa811d-92cf-46a2-a9fd-539130369eb1" containerName="ovn-controller" containerID="cri-o://3e36b6d6b9a369ac071cdab1b1d854f1f347521f301b072cb644c63a4b575d0c" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.268522 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-rkxr6" podUID="c8fa811d-92cf-46a2-a9fd-539130369eb1" containerName="ovn-controller" containerID="cri-o://3e36b6d6b9a369ac071cdab1b1d854f1f347521f301b072cb644c63a4b575d0c" gracePeriod=30 Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.271170 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbicana018-account-delete-6z4kv" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.299075 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pw9pq\" (UniqueName: \"kubernetes.io/projected/8095e4aa-7290-441d-ba67-79e48c23b8a6-kube-api-access-pw9pq\") pod \"cinderc394-account-delete-gnvr4\" (UID: \"8095e4aa-7290-441d-ba67-79e48c23b8a6\") " pod="openstack/cinderc394-account-delete-gnvr4" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.314896 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8095e4aa-7290-441d-ba67-79e48c23b8a6-operator-scripts\") pod \"cinderc394-account-delete-gnvr4\" (UID: \"8095e4aa-7290-441d-ba67-79e48c23b8a6\") " pod="openstack/cinderc394-account-delete-gnvr4" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.316743 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutronf6b9-account-delete-tdltj"] Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.361555 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qv5lh\" (UniqueName: \"kubernetes.io/projected/f76f637a-192d-4827-92b3-68e0fc348629-kube-api-access-qv5lh\") pod \"glance388c-account-delete-8hfh2\" (UID: \"f76f637a-192d-4827-92b3-68e0fc348629\") " pod="openstack/glance388c-account-delete-8hfh2" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.361615 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8psfq\" (UniqueName: \"kubernetes.io/projected/cc8592b6-2527-449c-94a2-bf40f8ea4b81-kube-api-access-8psfq\") pod \"neutronf6b9-account-delete-tdltj\" (UID: \"cc8592b6-2527-449c-94a2-bf40f8ea4b81\") " pod="openstack/neutronf6b9-account-delete-tdltj" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.362279 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f76f637a-192d-4827-92b3-68e0fc348629-operator-scripts\") pod \"glance388c-account-delete-8hfh2\" (UID: \"f76f637a-192d-4827-92b3-68e0fc348629\") " pod="openstack/glance388c-account-delete-8hfh2" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.362371 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc8592b6-2527-449c-94a2-bf40f8ea4b81-operator-scripts\") pod \"neutronf6b9-account-delete-tdltj\" (UID: \"cc8592b6-2527-449c-94a2-bf40f8ea4b81\") " pod="openstack/neutronf6b9-account-delete-tdltj" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.373571 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-8dvn5"] Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.395936 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-8dvn5"] Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.418651 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinderc394-account-delete-gnvr4" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.463658 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f76f637a-192d-4827-92b3-68e0fc348629-operator-scripts\") pod \"glance388c-account-delete-8hfh2\" (UID: \"f76f637a-192d-4827-92b3-68e0fc348629\") " pod="openstack/glance388c-account-delete-8hfh2" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.464014 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc8592b6-2527-449c-94a2-bf40f8ea4b81-operator-scripts\") pod \"neutronf6b9-account-delete-tdltj\" (UID: \"cc8592b6-2527-449c-94a2-bf40f8ea4b81\") " pod="openstack/neutronf6b9-account-delete-tdltj" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.464061 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qv5lh\" (UniqueName: \"kubernetes.io/projected/f76f637a-192d-4827-92b3-68e0fc348629-kube-api-access-qv5lh\") pod \"glance388c-account-delete-8hfh2\" (UID: \"f76f637a-192d-4827-92b3-68e0fc348629\") " pod="openstack/glance388c-account-delete-8hfh2" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.464093 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8psfq\" (UniqueName: \"kubernetes.io/projected/cc8592b6-2527-449c-94a2-bf40f8ea4b81-kube-api-access-8psfq\") pod \"neutronf6b9-account-delete-tdltj\" (UID: \"cc8592b6-2527-449c-94a2-bf40f8ea4b81\") " pod="openstack/neutronf6b9-account-delete-tdltj" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.465118 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f76f637a-192d-4827-92b3-68e0fc348629-operator-scripts\") pod \"glance388c-account-delete-8hfh2\" (UID: \"f76f637a-192d-4827-92b3-68e0fc348629\") " pod="openstack/glance388c-account-delete-8hfh2" Nov 24 13:37:17 crc kubenswrapper[4790]: E1124 13:37:17.465208 4790 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 24 13:37:17 crc kubenswrapper[4790]: E1124 13:37:17.465264 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/85e04b92-ae17-4eb6-a736-395c5d44f563-config-data podName:85e04b92-ae17-4eb6-a736-395c5d44f563 nodeName:}" failed. No retries permitted until 2025-11-24 13:37:18.465244974 +0000 UTC m=+1486.845138686 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/85e04b92-ae17-4eb6-a736-395c5d44f563-config-data") pod "rabbitmq-cell1-server-0" (UID: "85e04b92-ae17-4eb6-a736-395c5d44f563") : configmap "rabbitmq-cell1-config-data" not found Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.482461 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc8592b6-2527-449c-94a2-bf40f8ea4b81-operator-scripts\") pod \"neutronf6b9-account-delete-tdltj\" (UID: \"cc8592b6-2527-449c-94a2-bf40f8ea4b81\") " pod="openstack/neutronf6b9-account-delete-tdltj" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.483070 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.483525 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="23e9fd30-8462-4b44-b9ae-11e3a367c814" containerName="openstack-network-exporter" containerID="cri-o://72c394197db4eb126b656feee3931bef19393201c83eb96112678edb0aa5ebd7" gracePeriod=300 Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.524057 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qv5lh\" (UniqueName: \"kubernetes.io/projected/f76f637a-192d-4827-92b3-68e0fc348629-kube-api-access-qv5lh\") pod \"glance388c-account-delete-8hfh2\" (UID: \"f76f637a-192d-4827-92b3-68e0fc348629\") " pod="openstack/glance388c-account-delete-8hfh2" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.524675 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8psfq\" (UniqueName: \"kubernetes.io/projected/cc8592b6-2527-449c-94a2-bf40f8ea4b81-kube-api-access-8psfq\") pod \"neutronf6b9-account-delete-tdltj\" (UID: \"cc8592b6-2527-449c-94a2-bf40f8ea4b81\") " pod="openstack/neutronf6b9-account-delete-tdltj" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.544972 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-chcfj"] Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.586292 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-chcfj"] Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.672776 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell1964b-account-delete-xhklw"] Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.674046 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell1964b-account-delete-xhklw" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.677003 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="23e9fd30-8462-4b44-b9ae-11e3a367c814" containerName="ovsdbserver-sb" containerID="cri-o://995eda8d4390f50082b69ba23ac398c85f53e9011ae0253de271653fcc9abd18" gracePeriod=300 Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.698944 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell1964b-account-delete-xhklw"] Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.714231 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novaapid63e-account-delete-k96dh"] Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.720837 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapid63e-account-delete-k96dh" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.730917 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapid63e-account-delete-k96dh"] Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.752978 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.764009 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.764339 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance388c-account-delete-8hfh2" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.764641 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="08baac2a-300a-498f-8d0c-d1f261e6c8fc" containerName="openstack-network-exporter" containerID="cri-o://500eaeafad5bdd0b7c31ca6862cf936a5fd98d2a297af4a1a569abfdeafe28a7" gracePeriod=300 Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.768032 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutronf6b9-account-delete-tdltj" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.769899 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szr6v\" (UniqueName: \"kubernetes.io/projected/fef75227-ffd5-4c97-87c7-e05fe899eb25-kube-api-access-szr6v\") pod \"novaapid63e-account-delete-k96dh\" (UID: \"fef75227-ffd5-4c97-87c7-e05fe899eb25\") " pod="openstack/novaapid63e-account-delete-k96dh" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.769998 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fef75227-ffd5-4c97-87c7-e05fe899eb25-operator-scripts\") pod \"novaapid63e-account-delete-k96dh\" (UID: \"fef75227-ffd5-4c97-87c7-e05fe899eb25\") " pod="openstack/novaapid63e-account-delete-k96dh" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.770027 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfk9j\" (UniqueName: \"kubernetes.io/projected/20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3-kube-api-access-qfk9j\") pod \"novacell1964b-account-delete-xhklw\" (UID: \"20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3\") " pod="openstack/novacell1964b-account-delete-xhklw" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.770044 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3-operator-scripts\") pod \"novacell1964b-account-delete-xhklw\" (UID: \"20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3\") " pod="openstack/novacell1964b-account-delete-xhklw" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.815915 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a3fa97ac-7189-4828-8b95-664e852dea8e/ovn-northd/0.log" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.815985 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.826856 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-7957b6846-f8n7f"] Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.827129 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-7957b6846-f8n7f" podUID="1b507c50-0776-4d50-ad41-10f26f25fd5f" containerName="placement-log" containerID="cri-o://96fb30b3c69ce48bcf01e52ed956c73a35d60273c9db96fac4575e7dc8dd1004" gracePeriod=30 Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.827326 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-7957b6846-f8n7f" podUID="1b507c50-0776-4d50-ad41-10f26f25fd5f" containerName="placement-api" containerID="cri-o://2f816c8e5f36c0c446e29e140b1f74aeb5df0880612029daa8ef51dfaee75bee" gracePeriod=30 Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.848302 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell00c43-account-delete-dlpjf"] Nov 24 13:37:17 crc kubenswrapper[4790]: E1124 13:37:17.848702 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3fa97ac-7189-4828-8b95-664e852dea8e" containerName="ovn-northd" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.848715 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3fa97ac-7189-4828-8b95-664e852dea8e" containerName="ovn-northd" Nov 24 13:37:17 crc kubenswrapper[4790]: E1124 13:37:17.848730 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3fa97ac-7189-4828-8b95-664e852dea8e" containerName="openstack-network-exporter" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.848735 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3fa97ac-7189-4828-8b95-664e852dea8e" containerName="openstack-network-exporter" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.848954 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3fa97ac-7189-4828-8b95-664e852dea8e" containerName="openstack-network-exporter" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.848971 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3fa97ac-7189-4828-8b95-664e852dea8e" containerName="ovn-northd" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.849557 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell00c43-account-delete-dlpjf" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.859015 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell00c43-account-delete-dlpjf"] Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.874683 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3fa97ac-7189-4828-8b95-664e852dea8e-combined-ca-bundle\") pod \"a3fa97ac-7189-4828-8b95-664e852dea8e\" (UID: \"a3fa97ac-7189-4828-8b95-664e852dea8e\") " Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.874837 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3fa97ac-7189-4828-8b95-664e852dea8e-metrics-certs-tls-certs\") pod \"a3fa97ac-7189-4828-8b95-664e852dea8e\" (UID: \"a3fa97ac-7189-4828-8b95-664e852dea8e\") " Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.874892 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96dm4\" (UniqueName: \"kubernetes.io/projected/a3fa97ac-7189-4828-8b95-664e852dea8e-kube-api-access-96dm4\") pod \"a3fa97ac-7189-4828-8b95-664e852dea8e\" (UID: \"a3fa97ac-7189-4828-8b95-664e852dea8e\") " Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.874941 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3fa97ac-7189-4828-8b95-664e852dea8e-config\") pod \"a3fa97ac-7189-4828-8b95-664e852dea8e\" (UID: \"a3fa97ac-7189-4828-8b95-664e852dea8e\") " Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.874998 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3fa97ac-7189-4828-8b95-664e852dea8e-ovn-northd-tls-certs\") pod \"a3fa97ac-7189-4828-8b95-664e852dea8e\" (UID: \"a3fa97ac-7189-4828-8b95-664e852dea8e\") " Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.875030 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a3fa97ac-7189-4828-8b95-664e852dea8e-scripts\") pod \"a3fa97ac-7189-4828-8b95-664e852dea8e\" (UID: \"a3fa97ac-7189-4828-8b95-664e852dea8e\") " Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.875111 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a3fa97ac-7189-4828-8b95-664e852dea8e-ovn-rundir\") pod \"a3fa97ac-7189-4828-8b95-664e852dea8e\" (UID: \"a3fa97ac-7189-4828-8b95-664e852dea8e\") " Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.875383 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fef75227-ffd5-4c97-87c7-e05fe899eb25-operator-scripts\") pod \"novaapid63e-account-delete-k96dh\" (UID: \"fef75227-ffd5-4c97-87c7-e05fe899eb25\") " pod="openstack/novaapid63e-account-delete-k96dh" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.875435 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfk9j\" (UniqueName: \"kubernetes.io/projected/20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3-kube-api-access-qfk9j\") pod \"novacell1964b-account-delete-xhklw\" (UID: \"20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3\") " pod="openstack/novacell1964b-account-delete-xhklw" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.875457 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3-operator-scripts\") pod \"novacell1964b-account-delete-xhklw\" (UID: \"20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3\") " pod="openstack/novacell1964b-account-delete-xhklw" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.875821 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szr6v\" (UniqueName: \"kubernetes.io/projected/fef75227-ffd5-4c97-87c7-e05fe899eb25-kube-api-access-szr6v\") pod \"novaapid63e-account-delete-k96dh\" (UID: \"fef75227-ffd5-4c97-87c7-e05fe899eb25\") " pod="openstack/novaapid63e-account-delete-k96dh" Nov 24 13:37:17 crc kubenswrapper[4790]: E1124 13:37:17.879984 4790 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 24 13:37:17 crc kubenswrapper[4790]: E1124 13:37:17.880104 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/39508987-8655-46ed-861b-e1448652ddf6-config-data podName:39508987-8655-46ed-861b-e1448652ddf6 nodeName:}" failed. No retries permitted until 2025-11-24 13:37:18.380081467 +0000 UTC m=+1486.759975129 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/39508987-8655-46ed-861b-e1448652ddf6-config-data") pod "rabbitmq-server-0" (UID: "39508987-8655-46ed-861b-e1448652ddf6") : configmap "rabbitmq-config-data" not found Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.880288 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3fa97ac-7189-4828-8b95-664e852dea8e-config" (OuterVolumeSpecName: "config") pod "a3fa97ac-7189-4828-8b95-664e852dea8e" (UID: "a3fa97ac-7189-4828-8b95-664e852dea8e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:17 crc kubenswrapper[4790]: E1124 13:37:17.881566 4790 configmap.go:193] Couldn't get configMap openstack/openstack-cell1-scripts: configmap "openstack-cell1-scripts" not found Nov 24 13:37:17 crc kubenswrapper[4790]: E1124 13:37:17.881703 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3-operator-scripts podName:20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3 nodeName:}" failed. No retries permitted until 2025-11-24 13:37:18.381652542 +0000 UTC m=+1486.761546274 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3-operator-scripts") pod "novacell1964b-account-delete-xhklw" (UID: "20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3") : configmap "openstack-cell1-scripts" not found Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.882685 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3fa97ac-7189-4828-8b95-664e852dea8e-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "a3fa97ac-7189-4828-8b95-664e852dea8e" (UID: "a3fa97ac-7189-4828-8b95-664e852dea8e"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.883483 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fef75227-ffd5-4c97-87c7-e05fe899eb25-operator-scripts\") pod \"novaapid63e-account-delete-k96dh\" (UID: \"fef75227-ffd5-4c97-87c7-e05fe899eb25\") " pod="openstack/novaapid63e-account-delete-k96dh" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.884053 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="08baac2a-300a-498f-8d0c-d1f261e6c8fc" containerName="ovsdbserver-nb" containerID="cri-o://c70103120e0d00167ed2d2d0209229ea06b056321fe8833b38fe79ae3cb43775" gracePeriod=300 Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.884576 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3fa97ac-7189-4828-8b95-664e852dea8e-scripts" (OuterVolumeSpecName: "scripts") pod "a3fa97ac-7189-4828-8b95-664e852dea8e" (UID: "a3fa97ac-7189-4828-8b95-664e852dea8e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:17 crc kubenswrapper[4790]: E1124 13:37:17.894178 4790 projected.go:194] Error preparing data for projected volume kube-api-access-qfk9j for pod openstack/novacell1964b-account-delete-xhklw: failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Nov 24 13:37:17 crc kubenswrapper[4790]: E1124 13:37:17.894254 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3-kube-api-access-qfk9j podName:20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3 nodeName:}" failed. No retries permitted until 2025-11-24 13:37:18.394221601 +0000 UTC m=+1486.774115253 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-qfk9j" (UniqueName: "kubernetes.io/projected/20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3-kube-api-access-qfk9j") pod "novacell1964b-account-delete-xhklw" (UID: "20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3") : failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.894444 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3fa97ac-7189-4828-8b95-664e852dea8e-kube-api-access-96dm4" (OuterVolumeSpecName: "kube-api-access-96dm4") pod "a3fa97ac-7189-4828-8b95-664e852dea8e" (UID: "a3fa97ac-7189-4828-8b95-664e852dea8e"). InnerVolumeSpecName "kube-api-access-96dm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.937498 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szr6v\" (UniqueName: \"kubernetes.io/projected/fef75227-ffd5-4c97-87c7-e05fe899eb25-kube-api-access-szr6v\") pod \"novaapid63e-account-delete-k96dh\" (UID: \"fef75227-ffd5-4c97-87c7-e05fe899eb25\") " pod="openstack/novaapid63e-account-delete-k96dh" Nov 24 13:37:17 crc kubenswrapper[4790]: I1124 13:37:17.981580 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-p5msf"] Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.002727 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-p5msf"] Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.009230 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbb22\" (UniqueName: \"kubernetes.io/projected/dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008-kube-api-access-mbb22\") pod \"novacell00c43-account-delete-dlpjf\" (UID: \"dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008\") " pod="openstack/novacell00c43-account-delete-dlpjf" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.009280 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008-operator-scripts\") pod \"novacell00c43-account-delete-dlpjf\" (UID: \"dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008\") " pod="openstack/novacell00c43-account-delete-dlpjf" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.009345 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96dm4\" (UniqueName: \"kubernetes.io/projected/a3fa97ac-7189-4828-8b95-664e852dea8e-kube-api-access-96dm4\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.009357 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3fa97ac-7189-4828-8b95-664e852dea8e-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.009366 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a3fa97ac-7189-4828-8b95-664e852dea8e-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.009375 4790 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a3fa97ac-7189-4828-8b95-664e852dea8e-ovn-rundir\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.029637 4790 generic.go:334] "Generic (PLEG): container finished" podID="08baac2a-300a-498f-8d0c-d1f261e6c8fc" containerID="500eaeafad5bdd0b7c31ca6862cf936a5fd98d2a297af4a1a569abfdeafe28a7" exitCode=2 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.029757 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"08baac2a-300a-498f-8d0c-d1f261e6c8fc","Type":"ContainerDied","Data":"500eaeafad5bdd0b7c31ca6862cf936a5fd98d2a297af4a1a569abfdeafe28a7"} Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.059828 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-48flt"] Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.064505 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3fa97ac-7189-4828-8b95-664e852dea8e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a3fa97ac-7189-4828-8b95-664e852dea8e" (UID: "a3fa97ac-7189-4828-8b95-664e852dea8e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.080150 4790 generic.go:334] "Generic (PLEG): container finished" podID="c8fa811d-92cf-46a2-a9fd-539130369eb1" containerID="3e36b6d6b9a369ac071cdab1b1d854f1f347521f301b072cb644c63a4b575d0c" exitCode=0 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.080427 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rkxr6" event={"ID":"c8fa811d-92cf-46a2-a9fd-539130369eb1","Type":"ContainerDied","Data":"3e36b6d6b9a369ac071cdab1b1d854f1f347521f301b072cb644c63a4b575d0c"} Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.104295 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-48flt"] Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.111148 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008-operator-scripts\") pod \"novacell00c43-account-delete-dlpjf\" (UID: \"dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008\") " pod="openstack/novacell00c43-account-delete-dlpjf" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.117872 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbb22\" (UniqueName: \"kubernetes.io/projected/dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008-kube-api-access-mbb22\") pod \"novacell00c43-account-delete-dlpjf\" (UID: \"dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008\") " pod="openstack/novacell00c43-account-delete-dlpjf" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.112286 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008-operator-scripts\") pod \"novacell00c43-account-delete-dlpjf\" (UID: \"dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008\") " pod="openstack/novacell00c43-account-delete-dlpjf" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.113931 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-48rv7"] Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.138597 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3fa97ac-7189-4828-8b95-664e852dea8e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.139786 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a3fa97ac-7189-4828-8b95-664e852dea8e/ovn-northd/0.log" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.139831 4790 generic.go:334] "Generic (PLEG): container finished" podID="a3fa97ac-7189-4828-8b95-664e852dea8e" containerID="889043e225f3f968efc7dfcf48a4fcb3822a1e4d8cf106f1af9b36a1cda57399" exitCode=2 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.139848 4790 generic.go:334] "Generic (PLEG): container finished" podID="a3fa97ac-7189-4828-8b95-664e852dea8e" containerID="c71cf899ea7e4a2eff4caa8290abac4a40d0072f283f44d449955d488101621e" exitCode=143 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.139914 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a3fa97ac-7189-4828-8b95-664e852dea8e","Type":"ContainerDied","Data":"889043e225f3f968efc7dfcf48a4fcb3822a1e4d8cf106f1af9b36a1cda57399"} Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.139940 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a3fa97ac-7189-4828-8b95-664e852dea8e","Type":"ContainerDied","Data":"c71cf899ea7e4a2eff4caa8290abac4a40d0072f283f44d449955d488101621e"} Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.139950 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a3fa97ac-7189-4828-8b95-664e852dea8e","Type":"ContainerDied","Data":"e10229356b9c9fecc124d25b09f6663da5e7a6d25dcf71f7791d4f45c918d2f0"} Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.139966 4790 scope.go:117] "RemoveContainer" containerID="889043e225f3f968efc7dfcf48a4fcb3822a1e4d8cf106f1af9b36a1cda57399" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.140036 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cd5cbd7b9-48rv7" podUID="c428221a-e989-4844-894c-de96782508a3" containerName="dnsmasq-dns" containerID="cri-o://31d1bc29353608425771effac466332709c34ebd3c734237989fcb2339d732f9" gracePeriod=10 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.140091 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.140762 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.141173 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="account-server" containerID="cri-o://23963383dd944af8fb81fedb5a24119440fcae6b836f04def348d26a2568c09f" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.141229 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="swift-recon-cron" containerID="cri-o://b7e87dedf7cf3279e0cd55e6ed5885652aac1f5a97cb5db5d608e23cb07f7dc4" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.141263 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="rsync" containerID="cri-o://7d0ce12b3e15290c2d364a5d9b9eb84a5315993172cd4dded95d480621cad447" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.141299 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="object-expirer" containerID="cri-o://c2b3e00524e13f39df13207ebcb11976c8f2a411cb2cf3f7922cf052a42ee301" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.141340 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="object-updater" containerID="cri-o://578285e8c4f233014516fdb85665a982ed0526d96886404924f292cd08a273f2" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.141371 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="object-auditor" containerID="cri-o://b471731f94d67b5b7e6123cfb1a2a5c179f4f9ec0d47b37c19690321409b4711" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.141411 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="object-replicator" containerID="cri-o://d23adafa420c3790e972ee180913c8f5a9453ac2fffd584b5675a60e8c5a418a" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.141440 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="object-server" containerID="cri-o://ca07858ec58622d34fa209b9cd36987ee9d21a975afb78c55e5eebc7646c3b1a" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.141468 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="container-updater" containerID="cri-o://09f43e9aef5856b296a5ef62f8bc35fdd45d245130733c53d7e35377992a954b" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.141497 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="container-auditor" containerID="cri-o://c08ee946a46469851d49ebb63d97976087646f7f74bc00d1aba900aeb1b780eb" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.141528 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="container-replicator" containerID="cri-o://891362e93b7800953eed0ac6306309542589bcd92b30aa5ef910fdb82394fd3d" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.141555 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="container-server" containerID="cri-o://3122135b69c368fd946107ad82977792d44cf6ae8bfae58942ad6057773fa295" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.141583 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="account-reaper" containerID="cri-o://a9637198b1cf74e71f1558bec0b1a2c24a29c81f04cc33568637f3f6a30f9759" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.141611 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="account-auditor" containerID="cri-o://b3904f2398af315163d3ca972a5d587c4d9aa8034f444bc2b60c9b44aa285d9d" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.141641 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="account-replicator" containerID="cri-o://54b741f22ad4d511172f550e71786b130b052e565c75124d9ad50b4066de4deb" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.150179 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbb22\" (UniqueName: \"kubernetes.io/projected/dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008-kube-api-access-mbb22\") pod \"novacell00c43-account-delete-dlpjf\" (UID: \"dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008\") " pod="openstack/novacell00c43-account-delete-dlpjf" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.160687 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3fa97ac-7189-4828-8b95-664e852dea8e-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "a3fa97ac-7189-4828-8b95-664e852dea8e" (UID: "a3fa97ac-7189-4828-8b95-664e852dea8e"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.160825 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3fa97ac-7189-4828-8b95-664e852dea8e-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "a3fa97ac-7189-4828-8b95-664e852dea8e" (UID: "a3fa97ac-7189-4828-8b95-664e852dea8e"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.168609 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.168905 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="d0fa96c6-35fb-4e66-a7dc-31f6fa589132" containerName="cinder-scheduler" containerID="cri-o://92ffb21f402a267f3a3f49d256f427879130f4670ece784d82a260628c06ed0d" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.169037 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="d0fa96c6-35fb-4e66-a7dc-31f6fa589132" containerName="probe" containerID="cri-o://fb2d4795a70584cebf1c4bedf0675e9e61a8fae6696ec9e5217fb4161288c248" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.192316 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapid63e-account-delete-k96dh" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.201783 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-gjs8m"] Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.249989 4790 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3fa97ac-7189-4828-8b95-664e852dea8e-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.250033 4790 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3fa97ac-7189-4828-8b95-664e852dea8e-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.262654 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell00c43-account-delete-dlpjf" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.272060 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-z8gjd_a9c5a116-6438-47b5-8462-207eb656b553/openstack-network-exporter/0.log" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.272130 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-z8gjd" event={"ID":"a9c5a116-6438-47b5-8462-207eb656b553","Type":"ContainerDied","Data":"c6902baf408c3891cff01a43b41f9f037b6b29b99555835a2a635032bcee8c85"} Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.272107 4790 generic.go:334] "Generic (PLEG): container finished" podID="a9c5a116-6438-47b5-8462-207eb656b553" containerID="c6902baf408c3891cff01a43b41f9f037b6b29b99555835a2a635032bcee8c85" exitCode=2 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.298606 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-gjs8m"] Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.332531 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-28cwj" podUID="4d6d5d88-adda-4679-80de-0df0e0847bcb" containerName="ovs-vswitchd" containerID="cri-o://46182cf796e680922c3d2c23de93f19051dee361d1797948d997345bb2652287" gracePeriod=29 Nov 24 13:37:18 crc kubenswrapper[4790]: E1124 13:37:18.332670 4790 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Nov 24 13:37:18 crc kubenswrapper[4790]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Nov 24 13:37:18 crc kubenswrapper[4790]: + source /usr/local/bin/container-scripts/functions Nov 24 13:37:18 crc kubenswrapper[4790]: ++ OVNBridge=br-int Nov 24 13:37:18 crc kubenswrapper[4790]: ++ OVNRemote=tcp:localhost:6642 Nov 24 13:37:18 crc kubenswrapper[4790]: ++ OVNEncapType=geneve Nov 24 13:37:18 crc kubenswrapper[4790]: ++ OVNAvailabilityZones= Nov 24 13:37:18 crc kubenswrapper[4790]: ++ EnableChassisAsGateway=true Nov 24 13:37:18 crc kubenswrapper[4790]: ++ PhysicalNetworks= Nov 24 13:37:18 crc kubenswrapper[4790]: ++ OVNHostName= Nov 24 13:37:18 crc kubenswrapper[4790]: ++ DB_FILE=/etc/openvswitch/conf.db Nov 24 13:37:18 crc kubenswrapper[4790]: ++ ovs_dir=/var/lib/openvswitch Nov 24 13:37:18 crc kubenswrapper[4790]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Nov 24 13:37:18 crc kubenswrapper[4790]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Nov 24 13:37:18 crc kubenswrapper[4790]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 24 13:37:18 crc kubenswrapper[4790]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 24 13:37:18 crc kubenswrapper[4790]: + sleep 0.5 Nov 24 13:37:18 crc kubenswrapper[4790]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 24 13:37:18 crc kubenswrapper[4790]: + sleep 0.5 Nov 24 13:37:18 crc kubenswrapper[4790]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 24 13:37:18 crc kubenswrapper[4790]: + cleanup_ovsdb_server_semaphore Nov 24 13:37:18 crc kubenswrapper[4790]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 24 13:37:18 crc kubenswrapper[4790]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Nov 24 13:37:18 crc kubenswrapper[4790]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-28cwj" message=< Nov 24 13:37:18 crc kubenswrapper[4790]: Exiting ovsdb-server (5) [ OK ] Nov 24 13:37:18 crc kubenswrapper[4790]: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Nov 24 13:37:18 crc kubenswrapper[4790]: + source /usr/local/bin/container-scripts/functions Nov 24 13:37:18 crc kubenswrapper[4790]: ++ OVNBridge=br-int Nov 24 13:37:18 crc kubenswrapper[4790]: ++ OVNRemote=tcp:localhost:6642 Nov 24 13:37:18 crc kubenswrapper[4790]: ++ OVNEncapType=geneve Nov 24 13:37:18 crc kubenswrapper[4790]: ++ OVNAvailabilityZones= Nov 24 13:37:18 crc kubenswrapper[4790]: ++ EnableChassisAsGateway=true Nov 24 13:37:18 crc kubenswrapper[4790]: ++ PhysicalNetworks= Nov 24 13:37:18 crc kubenswrapper[4790]: ++ OVNHostName= Nov 24 13:37:18 crc kubenswrapper[4790]: ++ DB_FILE=/etc/openvswitch/conf.db Nov 24 13:37:18 crc kubenswrapper[4790]: ++ ovs_dir=/var/lib/openvswitch Nov 24 13:37:18 crc kubenswrapper[4790]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Nov 24 13:37:18 crc kubenswrapper[4790]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Nov 24 13:37:18 crc kubenswrapper[4790]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 24 13:37:18 crc kubenswrapper[4790]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 24 13:37:18 crc kubenswrapper[4790]: + sleep 0.5 Nov 24 13:37:18 crc kubenswrapper[4790]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 24 13:37:18 crc kubenswrapper[4790]: + sleep 0.5 Nov 24 13:37:18 crc kubenswrapper[4790]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 24 13:37:18 crc kubenswrapper[4790]: + cleanup_ovsdb_server_semaphore Nov 24 13:37:18 crc kubenswrapper[4790]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 24 13:37:18 crc kubenswrapper[4790]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Nov 24 13:37:18 crc kubenswrapper[4790]: > Nov 24 13:37:18 crc kubenswrapper[4790]: E1124 13:37:18.332698 4790 kuberuntime_container.go:691] "PreStop hook failed" err=< Nov 24 13:37:18 crc kubenswrapper[4790]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Nov 24 13:37:18 crc kubenswrapper[4790]: + source /usr/local/bin/container-scripts/functions Nov 24 13:37:18 crc kubenswrapper[4790]: ++ OVNBridge=br-int Nov 24 13:37:18 crc kubenswrapper[4790]: ++ OVNRemote=tcp:localhost:6642 Nov 24 13:37:18 crc kubenswrapper[4790]: ++ OVNEncapType=geneve Nov 24 13:37:18 crc kubenswrapper[4790]: ++ OVNAvailabilityZones= Nov 24 13:37:18 crc kubenswrapper[4790]: ++ EnableChassisAsGateway=true Nov 24 13:37:18 crc kubenswrapper[4790]: ++ PhysicalNetworks= Nov 24 13:37:18 crc kubenswrapper[4790]: ++ OVNHostName= Nov 24 13:37:18 crc kubenswrapper[4790]: ++ DB_FILE=/etc/openvswitch/conf.db Nov 24 13:37:18 crc kubenswrapper[4790]: ++ ovs_dir=/var/lib/openvswitch Nov 24 13:37:18 crc kubenswrapper[4790]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Nov 24 13:37:18 crc kubenswrapper[4790]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Nov 24 13:37:18 crc kubenswrapper[4790]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 24 13:37:18 crc kubenswrapper[4790]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 24 13:37:18 crc kubenswrapper[4790]: + sleep 0.5 Nov 24 13:37:18 crc kubenswrapper[4790]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 24 13:37:18 crc kubenswrapper[4790]: + sleep 0.5 Nov 24 13:37:18 crc kubenswrapper[4790]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 24 13:37:18 crc kubenswrapper[4790]: + cleanup_ovsdb_server_semaphore Nov 24 13:37:18 crc kubenswrapper[4790]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 24 13:37:18 crc kubenswrapper[4790]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Nov 24 13:37:18 crc kubenswrapper[4790]: > pod="openstack/ovn-controller-ovs-28cwj" podUID="4d6d5d88-adda-4679-80de-0df0e0847bcb" containerName="ovsdb-server" containerID="cri-o://e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.332733 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-28cwj" podUID="4d6d5d88-adda-4679-80de-0df0e0847bcb" containerName="ovsdb-server" containerID="cri-o://e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3" gracePeriod=29 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.352353 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_23e9fd30-8462-4b44-b9ae-11e3a367c814/ovsdbserver-sb/0.log" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.352411 4790 generic.go:334] "Generic (PLEG): container finished" podID="23e9fd30-8462-4b44-b9ae-11e3a367c814" containerID="72c394197db4eb126b656feee3931bef19393201c83eb96112678edb0aa5ebd7" exitCode=2 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.352458 4790 generic.go:334] "Generic (PLEG): container finished" podID="23e9fd30-8462-4b44-b9ae-11e3a367c814" containerID="995eda8d4390f50082b69ba23ac398c85f53e9011ae0253de271653fcc9abd18" exitCode=143 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.379502 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e5246df-ca01-43e6-bcd9-c8d9b6b78279" path="/var/lib/kubelet/pods/3e5246df-ca01-43e6-bcd9-c8d9b6b78279/volumes" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.380507 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f5d721b-e52b-490d-89f3-4a5edaf403ec" path="/var/lib/kubelet/pods/6f5d721b-e52b-490d-89f3-4a5edaf403ec/volumes" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.381109 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84a50ea0-c597-42e2-b4cd-7e628a517ddf" path="/var/lib/kubelet/pods/84a50ea0-c597-42e2-b4cd-7e628a517ddf/volumes" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.421564 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87e08198-d2d0-4608-87f3-02d832fe7c55" path="/var/lib/kubelet/pods/87e08198-d2d0-4608-87f3-02d832fe7c55/volumes" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.425137 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c49c82dc-b231-4f2f-8553-323fcb24f67f" path="/var/lib/kubelet/pods/c49c82dc-b231-4f2f-8553-323fcb24f67f/volumes" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.431645 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca484cf0-4ab0-4581-bb65-b1701db63df3" path="/var/lib/kubelet/pods/ca484cf0-4ab0-4581-bb65-b1701db63df3/volumes" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.437458 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fddd7928-5ec3-4605-94fb-a663c7f410e3" path="/var/lib/kubelet/pods/fddd7928-5ec3-4605-94fb-a663c7f410e3/volumes" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.444839 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"23e9fd30-8462-4b44-b9ae-11e3a367c814","Type":"ContainerDied","Data":"72c394197db4eb126b656feee3931bef19393201c83eb96112678edb0aa5ebd7"} Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.444900 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"23e9fd30-8462-4b44-b9ae-11e3a367c814","Type":"ContainerDied","Data":"995eda8d4390f50082b69ba23ac398c85f53e9011ae0253de271653fcc9abd18"} Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.444919 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-ckhz8"] Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.444944 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-ckhz8"] Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.499712 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfk9j\" (UniqueName: \"kubernetes.io/projected/20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3-kube-api-access-qfk9j\") pod \"novacell1964b-account-delete-xhklw\" (UID: \"20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3\") " pod="openstack/novacell1964b-account-delete-xhklw" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.500062 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3-operator-scripts\") pod \"novacell1964b-account-delete-xhklw\" (UID: \"20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3\") " pod="openstack/novacell1964b-account-delete-xhklw" Nov 24 13:37:18 crc kubenswrapper[4790]: E1124 13:37:18.501528 4790 configmap.go:193] Couldn't get configMap openstack/openstack-cell1-scripts: configmap "openstack-cell1-scripts" not found Nov 24 13:37:18 crc kubenswrapper[4790]: E1124 13:37:18.504934 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3-operator-scripts podName:20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3 nodeName:}" failed. No retries permitted until 2025-11-24 13:37:19.50490619 +0000 UTC m=+1487.884799852 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3-operator-scripts") pod "novacell1964b-account-delete-xhklw" (UID: "20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3") : configmap "openstack-cell1-scripts" not found Nov 24 13:37:18 crc kubenswrapper[4790]: E1124 13:37:18.507048 4790 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 24 13:37:18 crc kubenswrapper[4790]: E1124 13:37:18.507692 4790 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 24 13:37:18 crc kubenswrapper[4790]: E1124 13:37:18.508674 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/85e04b92-ae17-4eb6-a736-395c5d44f563-config-data podName:85e04b92-ae17-4eb6-a736-395c5d44f563 nodeName:}" failed. No retries permitted until 2025-11-24 13:37:20.507112363 +0000 UTC m=+1488.887006025 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/85e04b92-ae17-4eb6-a736-395c5d44f563-config-data") pod "rabbitmq-cell1-server-0" (UID: "85e04b92-ae17-4eb6-a736-395c5d44f563") : configmap "rabbitmq-cell1-config-data" not found Nov 24 13:37:18 crc kubenswrapper[4790]: E1124 13:37:18.508799 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/39508987-8655-46ed-861b-e1448652ddf6-config-data podName:39508987-8655-46ed-861b-e1448652ddf6 nodeName:}" failed. No retries permitted until 2025-11-24 13:37:19.50876716 +0000 UTC m=+1487.888660822 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/39508987-8655-46ed-861b-e1448652ddf6-config-data") pod "rabbitmq-server-0" (UID: "39508987-8655-46ed-861b-e1448652ddf6") : configmap "rabbitmq-config-data" not found Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.509794 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bh9qg" Nov 24 13:37:18 crc kubenswrapper[4790]: E1124 13:37:18.513863 4790 projected.go:194] Error preparing data for projected volume kube-api-access-qfk9j for pod openstack/novacell1964b-account-delete-xhklw: failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Nov 24 13:37:18 crc kubenswrapper[4790]: E1124 13:37:18.514247 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3-kube-api-access-qfk9j podName:20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3 nodeName:}" failed. No retries permitted until 2025-11-24 13:37:19.514225616 +0000 UTC m=+1487.894119278 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-qfk9j" (UniqueName: "kubernetes.io/projected/20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3-kube-api-access-qfk9j") pod "novacell1964b-account-delete-xhklw" (UID: "20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3") : failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.548207 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.548836 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="e6d696a7-d618-4416-b499-aae08088b079" containerName="cinder-api-log" containerID="cri-o://7d8bdf173f30451f767a3b85fb31741d1552f46cecc70197217878b14326622c" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.549099 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="e6d696a7-d618-4416-b499-aae08088b079" containerName="cinder-api" containerID="cri-o://002e858954a60800481c9819989042b1ad82aae0f1762f13550da43a5f262b74" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: E1124 13:37:18.579155 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3e36b6d6b9a369ac071cdab1b1d854f1f347521f301b072cb644c63a4b575d0c is running failed: container process not found" containerID="3e36b6d6b9a369ac071cdab1b1d854f1f347521f301b072cb644c63a4b575d0c" cmd=["/usr/local/bin/container-scripts/ovn_controller_readiness.sh"] Nov 24 13:37:18 crc kubenswrapper[4790]: E1124 13:37:18.587057 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3e36b6d6b9a369ac071cdab1b1d854f1f347521f301b072cb644c63a4b575d0c is running failed: container process not found" containerID="3e36b6d6b9a369ac071cdab1b1d854f1f347521f301b072cb644c63a4b575d0c" cmd=["/usr/local/bin/container-scripts/ovn_controller_readiness.sh"] Nov 24 13:37:18 crc kubenswrapper[4790]: E1124 13:37:18.587245 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3 is running failed: container process not found" containerID="e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 13:37:18 crc kubenswrapper[4790]: E1124 13:37:18.603443 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3 is running failed: container process not found" containerID="e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 13:37:18 crc kubenswrapper[4790]: E1124 13:37:18.603568 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3e36b6d6b9a369ac071cdab1b1d854f1f347521f301b072cb644c63a4b575d0c is running failed: container process not found" containerID="3e36b6d6b9a369ac071cdab1b1d854f1f347521f301b072cb644c63a4b575d0c" cmd=["/usr/local/bin/container-scripts/ovn_controller_readiness.sh"] Nov 24 13:37:18 crc kubenswrapper[4790]: E1124 13:37:18.603594 4790 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3e36b6d6b9a369ac071cdab1b1d854f1f347521f301b072cb644c63a4b575d0c is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-rkxr6" podUID="c8fa811d-92cf-46a2-a9fd-539130369eb1" containerName="ovn-controller" Nov 24 13:37:18 crc kubenswrapper[4790]: E1124 13:37:18.609073 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3 is running failed: container process not found" containerID="e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 13:37:18 crc kubenswrapper[4790]: E1124 13:37:18.609130 4790 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-28cwj" podUID="4d6d5d88-adda-4679-80de-0df0e0847bcb" containerName="ovsdb-server" Nov 24 13:37:18 crc kubenswrapper[4790]: E1124 13:37:18.609180 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="46182cf796e680922c3d2c23de93f19051dee361d1797948d997345bb2652287" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 13:37:18 crc kubenswrapper[4790]: E1124 13:37:18.635932 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="46182cf796e680922c3d2c23de93f19051dee361d1797948d997345bb2652287" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 13:37:18 crc kubenswrapper[4790]: E1124 13:37:18.644938 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="46182cf796e680922c3d2c23de93f19051dee361d1797948d997345bb2652287" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 13:37:18 crc kubenswrapper[4790]: E1124 13:37:18.644993 4790 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-28cwj" podUID="4d6d5d88-adda-4679-80de-0df0e0847bcb" containerName="ovs-vswitchd" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.648230 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-z8gjd_a9c5a116-6438-47b5-8462-207eb656b553/openstack-network-exporter/0.log" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.648294 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-z8gjd" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.656267 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.656475 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="8cbd6e74-a758-4bbf-81eb-acffdfa56955" containerName="glance-log" containerID="cri-o://1d279a19d65b0962ffcf048273f881a8a609e8c6c381d7ee412f985feb7393fb" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.656578 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="8cbd6e74-a758-4bbf-81eb-acffdfa56955" containerName="glance-httpd" containerID="cri-o://8b87dbc9b7ac42a8486a73feacf6d51b270b4dd6af8a81539a4ea922eb8a0e1b" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.659681 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rkxr6" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.681768 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-78cd478f4f-bqm94"] Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.682327 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-78cd478f4f-bqm94" podUID="4d4e45de-9768-40e9-9c75-6abba8b38559" containerName="neutron-api" containerID="cri-o://deee7a82aed113ecc740dcca448cfcb53642b959cd494942c6732db6590cb066" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.682476 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-78cd478f4f-bqm94" podUID="4d4e45de-9768-40e9-9c75-6abba8b38559" containerName="neutron-httpd" containerID="cri-o://cef1cb69d1e5ecd83dedc3637e51913646334d098991071163c9a6574695cd3e" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.702188 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.702426 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="8f8fe1f4-7b63-455c-97d1-d3c78826c3c4" containerName="glance-log" containerID="cri-o://3e9b86f86ff0501669479b2ac2723096fea52e20d1dedb6c13ee82635ab6f1aa" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.702566 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="8f8fe1f4-7b63-455c-97d1-d3c78826c3c4" containerName="glance-httpd" containerID="cri-o://77d68f10a4550d5451c7aeeccb710c9bb4cd97f4057890444850ad09cff25f6b" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: E1124 13:37:18.706496 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 995eda8d4390f50082b69ba23ac398c85f53e9011ae0253de271653fcc9abd18 is running failed: container process not found" containerID="995eda8d4390f50082b69ba23ac398c85f53e9011ae0253de271653fcc9abd18" cmd=["/usr/bin/pidof","ovsdb-server"] Nov 24 13:37:18 crc kubenswrapper[4790]: E1124 13:37:18.710313 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 995eda8d4390f50082b69ba23ac398c85f53e9011ae0253de271653fcc9abd18 is running failed: container process not found" containerID="995eda8d4390f50082b69ba23ac398c85f53e9011ae0253de271653fcc9abd18" cmd=["/usr/bin/pidof","ovsdb-server"] Nov 24 13:37:18 crc kubenswrapper[4790]: E1124 13:37:18.714076 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 995eda8d4390f50082b69ba23ac398c85f53e9011ae0253de271653fcc9abd18 is running failed: container process not found" containerID="995eda8d4390f50082b69ba23ac398c85f53e9011ae0253de271653fcc9abd18" cmd=["/usr/bin/pidof","ovsdb-server"] Nov 24 13:37:18 crc kubenswrapper[4790]: E1124 13:37:18.714135 4790 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 995eda8d4390f50082b69ba23ac398c85f53e9011ae0253de271653fcc9abd18 is running failed: container process not found" probeType="Readiness" pod="openstack/ovsdbserver-sb-0" podUID="23e9fd30-8462-4b44-b9ae-11e3a367c814" containerName="ovsdbserver-sb" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.721959 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c8fa811d-92cf-46a2-a9fd-539130369eb1-var-run-ovn\") pod \"c8fa811d-92cf-46a2-a9fd-539130369eb1\" (UID: \"c8fa811d-92cf-46a2-a9fd-539130369eb1\") " Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.722013 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9c5a116-6438-47b5-8462-207eb656b553-metrics-certs-tls-certs\") pod \"a9c5a116-6438-47b5-8462-207eb656b553\" (UID: \"a9c5a116-6438-47b5-8462-207eb656b553\") " Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.722033 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8fa811d-92cf-46a2-a9fd-539130369eb1-combined-ca-bundle\") pod \"c8fa811d-92cf-46a2-a9fd-539130369eb1\" (UID: \"c8fa811d-92cf-46a2-a9fd-539130369eb1\") " Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.722047 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9c5a116-6438-47b5-8462-207eb656b553-combined-ca-bundle\") pod \"a9c5a116-6438-47b5-8462-207eb656b553\" (UID: \"a9c5a116-6438-47b5-8462-207eb656b553\") " Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.722081 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c8fa811d-92cf-46a2-a9fd-539130369eb1-var-run\") pod \"c8fa811d-92cf-46a2-a9fd-539130369eb1\" (UID: \"c8fa811d-92cf-46a2-a9fd-539130369eb1\") " Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.722096 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c8fa811d-92cf-46a2-a9fd-539130369eb1-var-log-ovn\") pod \"c8fa811d-92cf-46a2-a9fd-539130369eb1\" (UID: \"c8fa811d-92cf-46a2-a9fd-539130369eb1\") " Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.722126 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a9c5a116-6438-47b5-8462-207eb656b553-ovn-rundir\") pod \"a9c5a116-6438-47b5-8462-207eb656b553\" (UID: \"a9c5a116-6438-47b5-8462-207eb656b553\") " Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.722146 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9c5a116-6438-47b5-8462-207eb656b553-config\") pod \"a9c5a116-6438-47b5-8462-207eb656b553\" (UID: \"a9c5a116-6438-47b5-8462-207eb656b553\") " Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.722166 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kbtgd\" (UniqueName: \"kubernetes.io/projected/a9c5a116-6438-47b5-8462-207eb656b553-kube-api-access-kbtgd\") pod \"a9c5a116-6438-47b5-8462-207eb656b553\" (UID: \"a9c5a116-6438-47b5-8462-207eb656b553\") " Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.722186 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a9c5a116-6438-47b5-8462-207eb656b553-ovs-rundir\") pod \"a9c5a116-6438-47b5-8462-207eb656b553\" (UID: \"a9c5a116-6438-47b5-8462-207eb656b553\") " Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.722205 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wwpr\" (UniqueName: \"kubernetes.io/projected/c8fa811d-92cf-46a2-a9fd-539130369eb1-kube-api-access-2wwpr\") pod \"c8fa811d-92cf-46a2-a9fd-539130369eb1\" (UID: \"c8fa811d-92cf-46a2-a9fd-539130369eb1\") " Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.722320 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c8fa811d-92cf-46a2-a9fd-539130369eb1-scripts\") pod \"c8fa811d-92cf-46a2-a9fd-539130369eb1\" (UID: \"c8fa811d-92cf-46a2-a9fd-539130369eb1\") " Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.722384 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8fa811d-92cf-46a2-a9fd-539130369eb1-ovn-controller-tls-certs\") pod \"c8fa811d-92cf-46a2-a9fd-539130369eb1\" (UID: \"c8fa811d-92cf-46a2-a9fd-539130369eb1\") " Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.724656 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8fa811d-92cf-46a2-a9fd-539130369eb1-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "c8fa811d-92cf-46a2-a9fd-539130369eb1" (UID: "c8fa811d-92cf-46a2-a9fd-539130369eb1"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.724725 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8fa811d-92cf-46a2-a9fd-539130369eb1-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "c8fa811d-92cf-46a2-a9fd-539130369eb1" (UID: "c8fa811d-92cf-46a2-a9fd-539130369eb1"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.730260 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9c5a116-6438-47b5-8462-207eb656b553-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "a9c5a116-6438-47b5-8462-207eb656b553" (UID: "a9c5a116-6438-47b5-8462-207eb656b553"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.731003 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9c5a116-6438-47b5-8462-207eb656b553-config" (OuterVolumeSpecName: "config") pod "a9c5a116-6438-47b5-8462-207eb656b553" (UID: "a9c5a116-6438-47b5-8462-207eb656b553"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.735436 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8fa811d-92cf-46a2-a9fd-539130369eb1-scripts" (OuterVolumeSpecName: "scripts") pod "c8fa811d-92cf-46a2-a9fd-539130369eb1" (UID: "c8fa811d-92cf-46a2-a9fd-539130369eb1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.735512 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9c5a116-6438-47b5-8462-207eb656b553-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "a9c5a116-6438-47b5-8462-207eb656b553" (UID: "a9c5a116-6438-47b5-8462-207eb656b553"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.735553 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-845f9d57f-2s48z"] Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.735801 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-845f9d57f-2s48z" podUID="3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c" containerName="barbican-worker-log" containerID="cri-o://9191be68775994f407e5cef4ebbca1e4aed9cfffa519ea7b28b2742579b483ad" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.735995 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-845f9d57f-2s48z" podUID="3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c" containerName="barbican-worker" containerID="cri-o://32dece0aef9f5c5f700993f3c33e9dc9b09465378d0d7abd138bd90e7d5c6319" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.736206 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8fa811d-92cf-46a2-a9fd-539130369eb1-var-run" (OuterVolumeSpecName: "var-run") pod "c8fa811d-92cf-46a2-a9fd-539130369eb1" (UID: "c8fa811d-92cf-46a2-a9fd-539130369eb1"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.739904 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8fa811d-92cf-46a2-a9fd-539130369eb1-kube-api-access-2wwpr" (OuterVolumeSpecName: "kube-api-access-2wwpr") pod "c8fa811d-92cf-46a2-a9fd-539130369eb1" (UID: "c8fa811d-92cf-46a2-a9fd-539130369eb1"). InnerVolumeSpecName "kube-api-access-2wwpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.742602 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9c5a116-6438-47b5-8462-207eb656b553-kube-api-access-kbtgd" (OuterVolumeSpecName: "kube-api-access-kbtgd") pod "a9c5a116-6438-47b5-8462-207eb656b553" (UID: "a9c5a116-6438-47b5-8462-207eb656b553"). InnerVolumeSpecName "kube-api-access-kbtgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.782815 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-58b5bcbcb-b6kbh"] Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.783121 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-58b5bcbcb-b6kbh" podUID="a2ee8af7-26fb-4652-be37-594db62f1146" containerName="barbican-keystone-listener-log" containerID="cri-o://2b38eafa7ed5a3895c1dd40a15a595d2c6fed37ac22fb3628de8e8a277d110ed" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.783336 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-58b5bcbcb-b6kbh" podUID="a2ee8af7-26fb-4652-be37-594db62f1146" containerName="barbican-keystone-listener" containerID="cri-o://ef998419eb802553e1ad108205b17dff1eada33d519b61e4493aea07e0efefc0" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.801185 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7d69688466-r2bq9"] Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.801416 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7d69688466-r2bq9" podUID="8311dfa1-6e87-493d-92e0-4eb69c792afe" containerName="barbican-api-log" containerID="cri-o://5a96182e01bd62e46435ce7bc8d7492b1b623ec41db4e9fa84ec0792133db77d" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.803907 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7d69688466-r2bq9" podUID="8311dfa1-6e87-493d-92e0-4eb69c792afe" containerName="barbican-api" containerID="cri-o://d07411834060975e29533a55785dd1993dfe65d5e3ce92d8eb62b08cba9dc0d1" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.811343 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8fa811d-92cf-46a2-a9fd-539130369eb1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c8fa811d-92cf-46a2-a9fd-539130369eb1" (UID: "c8fa811d-92cf-46a2-a9fd-539130369eb1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.833823 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.838251 4790 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c8fa811d-92cf-46a2-a9fd-539130369eb1-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.838298 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8fa811d-92cf-46a2-a9fd-539130369eb1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.838316 4790 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c8fa811d-92cf-46a2-a9fd-539130369eb1-var-run\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.838327 4790 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c8fa811d-92cf-46a2-a9fd-539130369eb1-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.838338 4790 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a9c5a116-6438-47b5-8462-207eb656b553-ovn-rundir\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.838352 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9c5a116-6438-47b5-8462-207eb656b553-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.838365 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kbtgd\" (UniqueName: \"kubernetes.io/projected/a9c5a116-6438-47b5-8462-207eb656b553-kube-api-access-kbtgd\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.838377 4790 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a9c5a116-6438-47b5-8462-207eb656b553-ovs-rundir\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.838388 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wwpr\" (UniqueName: \"kubernetes.io/projected/c8fa811d-92cf-46a2-a9fd-539130369eb1-kube-api-access-2wwpr\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.838398 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c8fa811d-92cf-46a2-a9fd-539130369eb1-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.840791 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9c5a116-6438-47b5-8462-207eb656b553-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a9c5a116-6438-47b5-8462-207eb656b553" (UID: "a9c5a116-6438-47b5-8462-207eb656b553"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.858501 4790 scope.go:117] "RemoveContainer" containerID="c71cf899ea7e4a2eff4caa8290abac4a40d0072f283f44d449955d488101621e" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.861956 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.862225 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c786f8b4-6946-45c1-8cde-00e56bf899ed" containerName="nova-api-log" containerID="cri-o://7e1e61d9e8e5a5c4687b8cc095ff2bfb83f045d184dc2b3cab025c89bd6960a1" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.862722 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c786f8b4-6946-45c1-8cde-00e56bf899ed" containerName="nova-api-api" containerID="cri-o://ddb81a6c66eee6e319c0646f64dd78e67d54faaf3b3d4b87630354653dbf2c4c" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.874200 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.874540 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="81757ef9-7330-44bb-a35e-8e6f4c137c0e" containerName="nova-metadata-log" containerID="cri-o://6b1d4a6d406564bd94038ede12823e6303e9c099f716a3ed7065653038dff83e" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.874770 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="81757ef9-7330-44bb-a35e-8e6f4c137c0e" containerName="nova-metadata-metadata" containerID="cri-o://6485471927d5a223e611cbd164262141a1a0f94b13c44e9d6ea257f4db80ed9a" gracePeriod=30 Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.887346 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.914639 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-rncp4"] Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.934256 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell1964b-account-delete-xhklw"] Nov 24 13:37:18 crc kubenswrapper[4790]: E1124 13:37:18.934674 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-qfk9j operator-scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/novacell1964b-account-delete-xhklw" podUID="20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.941679 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9c5a116-6438-47b5-8462-207eb656b553-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.980419 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-rncp4"] Nov 24 13:37:18 crc kubenswrapper[4790]: I1124 13:37:18.993217 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-964b-account-create-f25wh"] Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.006021 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-964b-account-create-f25wh"] Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.019826 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.020182 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="698a7113-712a-4940-aef9-aabd791b021d" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://a3bbb077cbcc6509c4a0bd5685f5bed2599daa1c8f9b2790688ae3b5133c5b0f" gracePeriod=30 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.023703 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.044221 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_23e9fd30-8462-4b44-b9ae-11e3a367c814/ovsdbserver-sb/0.log" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.044284 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.050155 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bh9qg"] Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.073150 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.077427 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="85e04b92-ae17-4eb6-a736-395c5d44f563" containerName="rabbitmq" containerID="cri-o://8c7b1f8ad061938b73c9c94a4ef4bfe3303a40fe30294f1a92f9a97da7d69a3e" gracePeriod=604800 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.083508 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.105008 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8fa811d-92cf-46a2-a9fd-539130369eb1-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "c8fa811d-92cf-46a2-a9fd-539130369eb1" (UID: "c8fa811d-92cf-46a2-a9fd-539130369eb1"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.106559 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-f6lwg"] Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.112946 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-f6lwg"] Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.113180 4790 scope.go:117] "RemoveContainer" containerID="889043e225f3f968efc7dfcf48a4fcb3822a1e4d8cf106f1af9b36a1cda57399" Nov 24 13:37:19 crc kubenswrapper[4790]: E1124 13:37:19.114148 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"889043e225f3f968efc7dfcf48a4fcb3822a1e4d8cf106f1af9b36a1cda57399\": container with ID starting with 889043e225f3f968efc7dfcf48a4fcb3822a1e4d8cf106f1af9b36a1cda57399 not found: ID does not exist" containerID="889043e225f3f968efc7dfcf48a4fcb3822a1e4d8cf106f1af9b36a1cda57399" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.114894 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"889043e225f3f968efc7dfcf48a4fcb3822a1e4d8cf106f1af9b36a1cda57399"} err="failed to get container status \"889043e225f3f968efc7dfcf48a4fcb3822a1e4d8cf106f1af9b36a1cda57399\": rpc error: code = NotFound desc = could not find container \"889043e225f3f968efc7dfcf48a4fcb3822a1e4d8cf106f1af9b36a1cda57399\": container with ID starting with 889043e225f3f968efc7dfcf48a4fcb3822a1e4d8cf106f1af9b36a1cda57399 not found: ID does not exist" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.114982 4790 scope.go:117] "RemoveContainer" containerID="c71cf899ea7e4a2eff4caa8290abac4a40d0072f283f44d449955d488101621e" Nov 24 13:37:19 crc kubenswrapper[4790]: E1124 13:37:19.115386 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c71cf899ea7e4a2eff4caa8290abac4a40d0072f283f44d449955d488101621e\": container with ID starting with c71cf899ea7e4a2eff4caa8290abac4a40d0072f283f44d449955d488101621e not found: ID does not exist" containerID="c71cf899ea7e4a2eff4caa8290abac4a40d0072f283f44d449955d488101621e" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.115440 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c71cf899ea7e4a2eff4caa8290abac4a40d0072f283f44d449955d488101621e"} err="failed to get container status \"c71cf899ea7e4a2eff4caa8290abac4a40d0072f283f44d449955d488101621e\": rpc error: code = NotFound desc = could not find container \"c71cf899ea7e4a2eff4caa8290abac4a40d0072f283f44d449955d488101621e\": container with ID starting with c71cf899ea7e4a2eff4caa8290abac4a40d0072f283f44d449955d488101621e not found: ID does not exist" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.115469 4790 scope.go:117] "RemoveContainer" containerID="889043e225f3f968efc7dfcf48a4fcb3822a1e4d8cf106f1af9b36a1cda57399" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.115787 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"889043e225f3f968efc7dfcf48a4fcb3822a1e4d8cf106f1af9b36a1cda57399"} err="failed to get container status \"889043e225f3f968efc7dfcf48a4fcb3822a1e4d8cf106f1af9b36a1cda57399\": rpc error: code = NotFound desc = could not find container \"889043e225f3f968efc7dfcf48a4fcb3822a1e4d8cf106f1af9b36a1cda57399\": container with ID starting with 889043e225f3f968efc7dfcf48a4fcb3822a1e4d8cf106f1af9b36a1cda57399 not found: ID does not exist" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.115805 4790 scope.go:117] "RemoveContainer" containerID="c71cf899ea7e4a2eff4caa8290abac4a40d0072f283f44d449955d488101621e" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.116006 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c71cf899ea7e4a2eff4caa8290abac4a40d0072f283f44d449955d488101621e"} err="failed to get container status \"c71cf899ea7e4a2eff4caa8290abac4a40d0072f283f44d449955d488101621e\": rpc error: code = NotFound desc = could not find container \"c71cf899ea7e4a2eff4caa8290abac4a40d0072f283f44d449955d488101621e\": container with ID starting with c71cf899ea7e4a2eff4caa8290abac4a40d0072f283f44d449955d488101621e not found: ID does not exist" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.122544 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.123082 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="39d594c5-aae1-4cd7-a7de-9de435ad83b8" containerName="nova-cell1-conductor-conductor" containerID="cri-o://eb8a4e1ecddd7fb54e90dd1a525962762b3153943ebd954bb7544913710d2ec2" gracePeriod=30 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.126042 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.126199 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="7fdd5969-f22b-41e6-a0d9-f3eeb16689fd" containerName="nova-cell0-conductor-conductor" containerID="cri-o://d8eab96ea490fbfa774abdd6a272fffab4b7cacb733249bf3f282fa0dfcf8521" gracePeriod=30 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.127043 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="39508987-8655-46ed-861b-e1448652ddf6" containerName="rabbitmq" containerID="cri-o://921d14419dc93745c5bfa78ec576180ba71427f41b71bc188365cd076418023e" gracePeriod=604800 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.140584 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-s2mrt"] Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.147443 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9c5a116-6438-47b5-8462-207eb656b553-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "a9c5a116-6438-47b5-8462-207eb656b553" (UID: "a9c5a116-6438-47b5-8462-207eb656b553"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.158024 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-s2mrt"] Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.161098 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23e9fd30-8462-4b44-b9ae-11e3a367c814-combined-ca-bundle\") pod \"23e9fd30-8462-4b44-b9ae-11e3a367c814\" (UID: \"23e9fd30-8462-4b44-b9ae-11e3a367c814\") " Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.161164 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"23e9fd30-8462-4b44-b9ae-11e3a367c814\" (UID: \"23e9fd30-8462-4b44-b9ae-11e3a367c814\") " Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.161191 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/23e9fd30-8462-4b44-b9ae-11e3a367c814-metrics-certs-tls-certs\") pod \"23e9fd30-8462-4b44-b9ae-11e3a367c814\" (UID: \"23e9fd30-8462-4b44-b9ae-11e3a367c814\") " Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.161253 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/23e9fd30-8462-4b44-b9ae-11e3a367c814-ovsdbserver-sb-tls-certs\") pod \"23e9fd30-8462-4b44-b9ae-11e3a367c814\" (UID: \"23e9fd30-8462-4b44-b9ae-11e3a367c814\") " Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.161356 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/23e9fd30-8462-4b44-b9ae-11e3a367c814-scripts\") pod \"23e9fd30-8462-4b44-b9ae-11e3a367c814\" (UID: \"23e9fd30-8462-4b44-b9ae-11e3a367c814\") " Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.161459 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23e9fd30-8462-4b44-b9ae-11e3a367c814-config\") pod \"23e9fd30-8462-4b44-b9ae-11e3a367c814\" (UID: \"23e9fd30-8462-4b44-b9ae-11e3a367c814\") " Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.161510 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/23e9fd30-8462-4b44-b9ae-11e3a367c814-ovsdb-rundir\") pod \"23e9fd30-8462-4b44-b9ae-11e3a367c814\" (UID: \"23e9fd30-8462-4b44-b9ae-11e3a367c814\") " Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.161546 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnx7x\" (UniqueName: \"kubernetes.io/projected/23e9fd30-8462-4b44-b9ae-11e3a367c814-kube-api-access-bnx7x\") pod \"23e9fd30-8462-4b44-b9ae-11e3a367c814\" (UID: \"23e9fd30-8462-4b44-b9ae-11e3a367c814\") " Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.162155 4790 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8fa811d-92cf-46a2-a9fd-539130369eb1-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.162148 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23e9fd30-8462-4b44-b9ae-11e3a367c814-config" (OuterVolumeSpecName: "config") pod "23e9fd30-8462-4b44-b9ae-11e3a367c814" (UID: "23e9fd30-8462-4b44-b9ae-11e3a367c814"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.162172 4790 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9c5a116-6438-47b5-8462-207eb656b553-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.162982 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23e9fd30-8462-4b44-b9ae-11e3a367c814-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "23e9fd30-8462-4b44-b9ae-11e3a367c814" (UID: "23e9fd30-8462-4b44-b9ae-11e3a367c814"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.166546 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.166732 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="de6f5df1-eb57-493b-9d89-5e710c6f4304" containerName="nova-scheduler-scheduler" containerID="cri-o://6b8183ab06a7a4bd08c674c5788831aa9879a328f22b998c507e78d8d453b26d" gracePeriod=30 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.167368 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23e9fd30-8462-4b44-b9ae-11e3a367c814-scripts" (OuterVolumeSpecName: "scripts") pod "23e9fd30-8462-4b44-b9ae-11e3a367c814" (UID: "23e9fd30-8462-4b44-b9ae-11e3a367c814"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.179674 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23e9fd30-8462-4b44-b9ae-11e3a367c814-kube-api-access-bnx7x" (OuterVolumeSpecName: "kube-api-access-bnx7x") pod "23e9fd30-8462-4b44-b9ae-11e3a367c814" (UID: "23e9fd30-8462-4b44-b9ae-11e3a367c814"). InnerVolumeSpecName "kube-api-access-bnx7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.186506 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "23e9fd30-8462-4b44-b9ae-11e3a367c814" (UID: "23e9fd30-8462-4b44-b9ae-11e3a367c814"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.264286 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23e9fd30-8462-4b44-b9ae-11e3a367c814-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.264581 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/23e9fd30-8462-4b44-b9ae-11e3a367c814-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.264595 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnx7x\" (UniqueName: \"kubernetes.io/projected/23e9fd30-8462-4b44-b9ae-11e3a367c814-kube-api-access-bnx7x\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.264617 4790 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.264626 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/23e9fd30-8462-4b44-b9ae-11e3a367c814-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.278172 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23e9fd30-8462-4b44-b9ae-11e3a367c814-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "23e9fd30-8462-4b44-b9ae-11e3a367c814" (UID: "23e9fd30-8462-4b44-b9ae-11e3a367c814"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.281384 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23e9fd30-8462-4b44-b9ae-11e3a367c814-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "23e9fd30-8462-4b44-b9ae-11e3a367c814" (UID: "23e9fd30-8462-4b44-b9ae-11e3a367c814"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.284270 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23e9fd30-8462-4b44-b9ae-11e3a367c814-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "23e9fd30-8462-4b44-b9ae-11e3a367c814" (UID: "23e9fd30-8462-4b44-b9ae-11e3a367c814"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.291927 4790 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.333429 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="9c866267-37e8-48d7-83ca-53e3cbf8b762" containerName="galera" containerID="cri-o://bdeaeaffd01a11cc6124ff6c3f7d882d3b5bdbc38615b718c3eb5ed5e5f217f3" gracePeriod=30 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.371493 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/23e9fd30-8462-4b44-b9ae-11e3a367c814-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.371520 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23e9fd30-8462-4b44-b9ae-11e3a367c814-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.371534 4790 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.371546 4790 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/23e9fd30-8462-4b44-b9ae-11e3a367c814-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.378740 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-48rv7" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.393172 4790 generic.go:334] "Generic (PLEG): container finished" podID="a2ee8af7-26fb-4652-be37-594db62f1146" containerID="2b38eafa7ed5a3895c1dd40a15a595d2c6fed37ac22fb3628de8e8a277d110ed" exitCode=143 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.393298 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-58b5bcbcb-b6kbh" event={"ID":"a2ee8af7-26fb-4652-be37-594db62f1146","Type":"ContainerDied","Data":"2b38eafa7ed5a3895c1dd40a15a595d2c6fed37ac22fb3628de8e8a277d110ed"} Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.396602 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rkxr6" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.396853 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rkxr6" event={"ID":"c8fa811d-92cf-46a2-a9fd-539130369eb1","Type":"ContainerDied","Data":"48362faf8e69bd45114ae3862302df825a91d31f4faf109b679162750f6f967d"} Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.397213 4790 scope.go:117] "RemoveContainer" containerID="3e36b6d6b9a369ac071cdab1b1d854f1f347521f301b072cb644c63a4b575d0c" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.422102 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_23e9fd30-8462-4b44-b9ae-11e3a367c814/ovsdbserver-sb/0.log" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.422347 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"23e9fd30-8462-4b44-b9ae-11e3a367c814","Type":"ContainerDied","Data":"ff0a313f41ee312fd3dbafb956ceb2ec653b95ab2f5bd477fb801596d8544d55"} Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.422535 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.432577 4790 generic.go:334] "Generic (PLEG): container finished" podID="4d4e45de-9768-40e9-9c75-6abba8b38559" containerID="cef1cb69d1e5ecd83dedc3637e51913646334d098991071163c9a6574695cd3e" exitCode=0 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.432625 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-78cd478f4f-bqm94" event={"ID":"4d4e45de-9768-40e9-9c75-6abba8b38559","Type":"ContainerDied","Data":"cef1cb69d1e5ecd83dedc3637e51913646334d098991071163c9a6574695cd3e"} Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.440195 4790 generic.go:334] "Generic (PLEG): container finished" podID="c428221a-e989-4844-894c-de96782508a3" containerID="31d1bc29353608425771effac466332709c34ebd3c734237989fcb2339d732f9" exitCode=0 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.440255 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-48rv7" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.440277 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-48rv7" event={"ID":"c428221a-e989-4844-894c-de96782508a3","Type":"ContainerDied","Data":"31d1bc29353608425771effac466332709c34ebd3c734237989fcb2339d732f9"} Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.440309 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-48rv7" event={"ID":"c428221a-e989-4844-894c-de96782508a3","Type":"ContainerDied","Data":"a9e629ddc77758865107cad9072d8ed2001b03905c40bf6f5631cb19a96b4cd2"} Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.468338 4790 scope.go:117] "RemoveContainer" containerID="72c394197db4eb126b656feee3931bef19393201c83eb96112678edb0aa5ebd7" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.473274 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c428221a-e989-4844-894c-de96782508a3-ovsdbserver-sb\") pod \"c428221a-e989-4844-894c-de96782508a3\" (UID: \"c428221a-e989-4844-894c-de96782508a3\") " Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.473320 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c428221a-e989-4844-894c-de96782508a3-dns-swift-storage-0\") pod \"c428221a-e989-4844-894c-de96782508a3\" (UID: \"c428221a-e989-4844-894c-de96782508a3\") " Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.473351 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c428221a-e989-4844-894c-de96782508a3-ovsdbserver-nb\") pod \"c428221a-e989-4844-894c-de96782508a3\" (UID: \"c428221a-e989-4844-894c-de96782508a3\") " Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.473370 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c428221a-e989-4844-894c-de96782508a3-dns-svc\") pod \"c428221a-e989-4844-894c-de96782508a3\" (UID: \"c428221a-e989-4844-894c-de96782508a3\") " Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.473437 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c428221a-e989-4844-894c-de96782508a3-config\") pod \"c428221a-e989-4844-894c-de96782508a3\" (UID: \"c428221a-e989-4844-894c-de96782508a3\") " Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.473505 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfbd6\" (UniqueName: \"kubernetes.io/projected/c428221a-e989-4844-894c-de96782508a3-kube-api-access-wfbd6\") pod \"c428221a-e989-4844-894c-de96782508a3\" (UID: \"c428221a-e989-4844-894c-de96782508a3\") " Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.509171 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c428221a-e989-4844-894c-de96782508a3-kube-api-access-wfbd6" (OuterVolumeSpecName: "kube-api-access-wfbd6") pod "c428221a-e989-4844-894c-de96782508a3" (UID: "c428221a-e989-4844-894c-de96782508a3"). InnerVolumeSpecName "kube-api-access-wfbd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.510790 4790 generic.go:334] "Generic (PLEG): container finished" podID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerID="7d0ce12b3e15290c2d364a5d9b9eb84a5315993172cd4dded95d480621cad447" exitCode=0 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.510819 4790 generic.go:334] "Generic (PLEG): container finished" podID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerID="c2b3e00524e13f39df13207ebcb11976c8f2a411cb2cf3f7922cf052a42ee301" exitCode=0 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.510829 4790 generic.go:334] "Generic (PLEG): container finished" podID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerID="578285e8c4f233014516fdb85665a982ed0526d96886404924f292cd08a273f2" exitCode=0 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.510836 4790 generic.go:334] "Generic (PLEG): container finished" podID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerID="b471731f94d67b5b7e6123cfb1a2a5c179f4f9ec0d47b37c19690321409b4711" exitCode=0 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.510845 4790 generic.go:334] "Generic (PLEG): container finished" podID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerID="d23adafa420c3790e972ee180913c8f5a9453ac2fffd584b5675a60e8c5a418a" exitCode=0 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.510853 4790 generic.go:334] "Generic (PLEG): container finished" podID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerID="ca07858ec58622d34fa209b9cd36987ee9d21a975afb78c55e5eebc7646c3b1a" exitCode=0 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.510861 4790 generic.go:334] "Generic (PLEG): container finished" podID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerID="09f43e9aef5856b296a5ef62f8bc35fdd45d245130733c53d7e35377992a954b" exitCode=0 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.510869 4790 generic.go:334] "Generic (PLEG): container finished" podID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerID="c08ee946a46469851d49ebb63d97976087646f7f74bc00d1aba900aeb1b780eb" exitCode=0 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.510977 4790 generic.go:334] "Generic (PLEG): container finished" podID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerID="891362e93b7800953eed0ac6306309542589bcd92b30aa5ef910fdb82394fd3d" exitCode=0 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.510989 4790 generic.go:334] "Generic (PLEG): container finished" podID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerID="3122135b69c368fd946107ad82977792d44cf6ae8bfae58942ad6057773fa295" exitCode=0 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.510997 4790 generic.go:334] "Generic (PLEG): container finished" podID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerID="a9637198b1cf74e71f1558bec0b1a2c24a29c81f04cc33568637f3f6a30f9759" exitCode=0 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.511004 4790 generic.go:334] "Generic (PLEG): container finished" podID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerID="b3904f2398af315163d3ca972a5d587c4d9aa8034f444bc2b60c9b44aa285d9d" exitCode=0 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.511012 4790 generic.go:334] "Generic (PLEG): container finished" podID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerID="54b741f22ad4d511172f550e71786b130b052e565c75124d9ad50b4066de4deb" exitCode=0 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.511021 4790 generic.go:334] "Generic (PLEG): container finished" podID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerID="23963383dd944af8fb81fedb5a24119440fcae6b836f04def348d26a2568c09f" exitCode=0 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.511080 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63ce5126-cdab-4cdc-968a-acc26c6caf87","Type":"ContainerDied","Data":"7d0ce12b3e15290c2d364a5d9b9eb84a5315993172cd4dded95d480621cad447"} Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.511110 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63ce5126-cdab-4cdc-968a-acc26c6caf87","Type":"ContainerDied","Data":"c2b3e00524e13f39df13207ebcb11976c8f2a411cb2cf3f7922cf052a42ee301"} Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.511471 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63ce5126-cdab-4cdc-968a-acc26c6caf87","Type":"ContainerDied","Data":"578285e8c4f233014516fdb85665a982ed0526d96886404924f292cd08a273f2"} Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.511488 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63ce5126-cdab-4cdc-968a-acc26c6caf87","Type":"ContainerDied","Data":"b471731f94d67b5b7e6123cfb1a2a5c179f4f9ec0d47b37c19690321409b4711"} Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.511501 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.511519 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63ce5126-cdab-4cdc-968a-acc26c6caf87","Type":"ContainerDied","Data":"d23adafa420c3790e972ee180913c8f5a9453ac2fffd584b5675a60e8c5a418a"} Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.511531 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63ce5126-cdab-4cdc-968a-acc26c6caf87","Type":"ContainerDied","Data":"ca07858ec58622d34fa209b9cd36987ee9d21a975afb78c55e5eebc7646c3b1a"} Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.511542 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63ce5126-cdab-4cdc-968a-acc26c6caf87","Type":"ContainerDied","Data":"09f43e9aef5856b296a5ef62f8bc35fdd45d245130733c53d7e35377992a954b"} Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.511582 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63ce5126-cdab-4cdc-968a-acc26c6caf87","Type":"ContainerDied","Data":"c08ee946a46469851d49ebb63d97976087646f7f74bc00d1aba900aeb1b780eb"} Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.511595 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63ce5126-cdab-4cdc-968a-acc26c6caf87","Type":"ContainerDied","Data":"891362e93b7800953eed0ac6306309542589bcd92b30aa5ef910fdb82394fd3d"} Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.511606 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63ce5126-cdab-4cdc-968a-acc26c6caf87","Type":"ContainerDied","Data":"3122135b69c368fd946107ad82977792d44cf6ae8bfae58942ad6057773fa295"} Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.511618 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63ce5126-cdab-4cdc-968a-acc26c6caf87","Type":"ContainerDied","Data":"a9637198b1cf74e71f1558bec0b1a2c24a29c81f04cc33568637f3f6a30f9759"} Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.511630 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63ce5126-cdab-4cdc-968a-acc26c6caf87","Type":"ContainerDied","Data":"b3904f2398af315163d3ca972a5d587c4d9aa8034f444bc2b60c9b44aa285d9d"} Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.511640 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63ce5126-cdab-4cdc-968a-acc26c6caf87","Type":"ContainerDied","Data":"54b741f22ad4d511172f550e71786b130b052e565c75124d9ad50b4066de4deb"} Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.511650 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63ce5126-cdab-4cdc-968a-acc26c6caf87","Type":"ContainerDied","Data":"23963383dd944af8fb81fedb5a24119440fcae6b836f04def348d26a2568c09f"} Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.527332 4790 generic.go:334] "Generic (PLEG): container finished" podID="8f8fe1f4-7b63-455c-97d1-d3c78826c3c4" containerID="3e9b86f86ff0501669479b2ac2723096fea52e20d1dedb6c13ee82635ab6f1aa" exitCode=143 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.527434 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4","Type":"ContainerDied","Data":"3e9b86f86ff0501669479b2ac2723096fea52e20d1dedb6c13ee82635ab6f1aa"} Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.536380 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.543979 4790 generic.go:334] "Generic (PLEG): container finished" podID="f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c" containerID="0981452a5c31b7ac7be6aca3f35f652875423629b90d147c20e6f8b123ad006d" exitCode=137 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.547225 4790 generic.go:334] "Generic (PLEG): container finished" podID="1b507c50-0776-4d50-ad41-10f26f25fd5f" containerID="96fb30b3c69ce48bcf01e52ed956c73a35d60273c9db96fac4575e7dc8dd1004" exitCode=143 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.547335 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7957b6846-f8n7f" event={"ID":"1b507c50-0776-4d50-ad41-10f26f25fd5f","Type":"ContainerDied","Data":"96fb30b3c69ce48bcf01e52ed956c73a35d60273c9db96fac4575e7dc8dd1004"} Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.550174 4790 generic.go:334] "Generic (PLEG): container finished" podID="8311dfa1-6e87-493d-92e0-4eb69c792afe" containerID="5a96182e01bd62e46435ce7bc8d7492b1b623ec41db4e9fa84ec0792133db77d" exitCode=143 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.550283 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7d69688466-r2bq9" event={"ID":"8311dfa1-6e87-493d-92e0-4eb69c792afe","Type":"ContainerDied","Data":"5a96182e01bd62e46435ce7bc8d7492b1b623ec41db4e9fa84ec0792133db77d"} Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.553094 4790 generic.go:334] "Generic (PLEG): container finished" podID="c786f8b4-6946-45c1-8cde-00e56bf899ed" containerID="7e1e61d9e8e5a5c4687b8cc095ff2bfb83f045d184dc2b3cab025c89bd6960a1" exitCode=143 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.553214 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c786f8b4-6946-45c1-8cde-00e56bf899ed","Type":"ContainerDied","Data":"7e1e61d9e8e5a5c4687b8cc095ff2bfb83f045d184dc2b3cab025c89bd6960a1"} Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.555582 4790 generic.go:334] "Generic (PLEG): container finished" podID="81757ef9-7330-44bb-a35e-8e6f4c137c0e" containerID="6b1d4a6d406564bd94038ede12823e6303e9c099f716a3ed7065653038dff83e" exitCode=143 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.555652 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"81757ef9-7330-44bb-a35e-8e6f4c137c0e","Type":"ContainerDied","Data":"6b1d4a6d406564bd94038ede12823e6303e9c099f716a3ed7065653038dff83e"} Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.557279 4790 generic.go:334] "Generic (PLEG): container finished" podID="3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c" containerID="9191be68775994f407e5cef4ebbca1e4aed9cfffa519ea7b28b2742579b483ad" exitCode=143 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.557339 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-845f9d57f-2s48z" event={"ID":"3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c","Type":"ContainerDied","Data":"9191be68775994f407e5cef4ebbca1e4aed9cfffa519ea7b28b2742579b483ad"} Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.560320 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-z8gjd_a9c5a116-6438-47b5-8462-207eb656b553/openstack-network-exporter/0.log" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.560395 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-z8gjd" event={"ID":"a9c5a116-6438-47b5-8462-207eb656b553","Type":"ContainerDied","Data":"4822f6e99ac9559176292b80d7e95de938d66bc14872d70aaf7e01d84ccedf49"} Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.560756 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-z8gjd" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.566639 4790 generic.go:334] "Generic (PLEG): container finished" podID="4d6d5d88-adda-4679-80de-0df0e0847bcb" containerID="e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3" exitCode=0 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.566722 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-28cwj" event={"ID":"4d6d5d88-adda-4679-80de-0df0e0847bcb","Type":"ContainerDied","Data":"e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3"} Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.572868 4790 generic.go:334] "Generic (PLEG): container finished" podID="e6d696a7-d618-4416-b499-aae08088b079" containerID="7d8bdf173f30451f767a3b85fb31741d1552f46cecc70197217878b14326622c" exitCode=143 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.573005 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e6d696a7-d618-4416-b499-aae08088b079","Type":"ContainerDied","Data":"7d8bdf173f30451f767a3b85fb31741d1552f46cecc70197217878b14326622c"} Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.577608 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfk9j\" (UniqueName: \"kubernetes.io/projected/20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3-kube-api-access-qfk9j\") pod \"novacell1964b-account-delete-xhklw\" (UID: \"20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3\") " pod="openstack/novacell1964b-account-delete-xhklw" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.577653 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3-operator-scripts\") pod \"novacell1964b-account-delete-xhklw\" (UID: \"20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3\") " pod="openstack/novacell1964b-account-delete-xhklw" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.577873 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfbd6\" (UniqueName: \"kubernetes.io/projected/c428221a-e989-4844-894c-de96782508a3-kube-api-access-wfbd6\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:19 crc kubenswrapper[4790]: E1124 13:37:19.578128 4790 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 24 13:37:19 crc kubenswrapper[4790]: E1124 13:37:19.578163 4790 configmap.go:193] Couldn't get configMap openstack/openstack-cell1-scripts: configmap "openstack-cell1-scripts" not found Nov 24 13:37:19 crc kubenswrapper[4790]: E1124 13:37:19.578202 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/39508987-8655-46ed-861b-e1448652ddf6-config-data podName:39508987-8655-46ed-861b-e1448652ddf6 nodeName:}" failed. No retries permitted until 2025-11-24 13:37:21.578186207 +0000 UTC m=+1489.958079869 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/39508987-8655-46ed-861b-e1448652ddf6-config-data") pod "rabbitmq-server-0" (UID: "39508987-8655-46ed-861b-e1448652ddf6") : configmap "rabbitmq-config-data" not found Nov 24 13:37:19 crc kubenswrapper[4790]: E1124 13:37:19.578241 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3-operator-scripts podName:20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3 nodeName:}" failed. No retries permitted until 2025-11-24 13:37:21.578223698 +0000 UTC m=+1489.958117360 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3-operator-scripts") pod "novacell1964b-account-delete-xhklw" (UID: "20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3") : configmap "openstack-cell1-scripts" not found Nov 24 13:37:19 crc kubenswrapper[4790]: E1124 13:37:19.604895 4790 projected.go:194] Error preparing data for projected volume kube-api-access-qfk9j for pod openstack/novacell1964b-account-delete-xhklw: failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Nov 24 13:37:19 crc kubenswrapper[4790]: E1124 13:37:19.605217 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3-kube-api-access-qfk9j podName:20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3 nodeName:}" failed. No retries permitted until 2025-11-24 13:37:21.605167198 +0000 UTC m=+1489.985060860 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-qfk9j" (UniqueName: "kubernetes.io/projected/20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3-kube-api-access-qfk9j") pod "novacell1964b-account-delete-xhklw" (UID: "20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3") : failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.613941 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_08baac2a-300a-498f-8d0c-d1f261e6c8fc/ovsdbserver-nb/0.log" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.613987 4790 generic.go:334] "Generic (PLEG): container finished" podID="08baac2a-300a-498f-8d0c-d1f261e6c8fc" containerID="c70103120e0d00167ed2d2d0209229ea06b056321fe8833b38fe79ae3cb43775" exitCode=143 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.614298 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"08baac2a-300a-498f-8d0c-d1f261e6c8fc","Type":"ContainerDied","Data":"c70103120e0d00167ed2d2d0209229ea06b056321fe8833b38fe79ae3cb43775"} Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.615434 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-rkxr6"] Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.620399 4790 generic.go:334] "Generic (PLEG): container finished" podID="d0fa96c6-35fb-4e66-a7dc-31f6fa589132" containerID="fb2d4795a70584cebf1c4bedf0675e9e61a8fae6696ec9e5217fb4161288c248" exitCode=0 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.620643 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d0fa96c6-35fb-4e66-a7dc-31f6fa589132","Type":"ContainerDied","Data":"fb2d4795a70584cebf1c4bedf0675e9e61a8fae6696ec9e5217fb4161288c248"} Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.628772 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-rkxr6"] Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.631258 4790 generic.go:334] "Generic (PLEG): container finished" podID="8cbd6e74-a758-4bbf-81eb-acffdfa56955" containerID="1d279a19d65b0962ffcf048273f881a8a609e8c6c381d7ee412f985feb7393fb" exitCode=143 Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.631338 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell1964b-account-delete-xhklw" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.632060 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8cbd6e74-a758-4bbf-81eb-acffdfa56955","Type":"ContainerDied","Data":"1d279a19d65b0962ffcf048273f881a8a609e8c6c381d7ee412f985feb7393fb"} Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.659530 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement0fc6-account-delete-rl7sl"] Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.664863 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinderc394-account-delete-gnvr4"] Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.671308 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbicana018-account-delete-6z4kv"] Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.677284 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c428221a-e989-4844-894c-de96782508a3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c428221a-e989-4844-894c-de96782508a3" (UID: "c428221a-e989-4844-894c-de96782508a3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.680091 4790 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c428221a-e989-4844-894c-de96782508a3-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.682941 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c428221a-e989-4844-894c-de96782508a3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c428221a-e989-4844-894c-de96782508a3" (UID: "c428221a-e989-4844-894c-de96782508a3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.694496 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.707897 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-z8gjd"] Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.710524 4790 scope.go:117] "RemoveContainer" containerID="995eda8d4390f50082b69ba23ac398c85f53e9011ae0253de271653fcc9abd18" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.718927 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c428221a-e989-4844-894c-de96782508a3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c428221a-e989-4844-894c-de96782508a3" (UID: "c428221a-e989-4844-894c-de96782508a3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.721852 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-z8gjd"] Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.724869 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_08baac2a-300a-498f-8d0c-d1f261e6c8fc/ovsdbserver-nb/0.log" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.724967 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.728245 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell1964b-account-delete-xhklw" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.728479 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c428221a-e989-4844-894c-de96782508a3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c428221a-e989-4844-894c-de96782508a3" (UID: "c428221a-e989-4844-894c-de96782508a3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.739373 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c428221a-e989-4844-894c-de96782508a3-config" (OuterVolumeSpecName: "config") pod "c428221a-e989-4844-894c-de96782508a3" (UID: "c428221a-e989-4844-894c-de96782508a3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.785962 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08baac2a-300a-498f-8d0c-d1f261e6c8fc-combined-ca-bundle\") pod \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\" (UID: \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\") " Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.786010 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c-combined-ca-bundle\") pod \"f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c\" (UID: \"f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c\") " Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.786057 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4sxg\" (UniqueName: \"kubernetes.io/projected/08baac2a-300a-498f-8d0c-d1f261e6c8fc-kube-api-access-q4sxg\") pod \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\" (UID: \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\") " Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.786083 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/08baac2a-300a-498f-8d0c-d1f261e6c8fc-ovsdbserver-nb-tls-certs\") pod \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\" (UID: \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\") " Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.786129 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/08baac2a-300a-498f-8d0c-d1f261e6c8fc-scripts\") pod \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\" (UID: \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\") " Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.786151 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c-openstack-config-secret\") pod \"f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c\" (UID: \"f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c\") " Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.786211 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08baac2a-300a-498f-8d0c-d1f261e6c8fc-config\") pod \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\" (UID: \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\") " Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.786267 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c-openstack-config\") pod \"f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c\" (UID: \"f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c\") " Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.786286 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/08baac2a-300a-498f-8d0c-d1f261e6c8fc-metrics-certs-tls-certs\") pod \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\" (UID: \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\") " Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.786328 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\" (UID: \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\") " Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.786344 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5t4x\" (UniqueName: \"kubernetes.io/projected/f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c-kube-api-access-q5t4x\") pod \"f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c\" (UID: \"f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c\") " Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.786378 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/08baac2a-300a-498f-8d0c-d1f261e6c8fc-ovsdb-rundir\") pod \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\" (UID: \"08baac2a-300a-498f-8d0c-d1f261e6c8fc\") " Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.786770 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c428221a-e989-4844-894c-de96782508a3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.786791 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c428221a-e989-4844-894c-de96782508a3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.786801 4790 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c428221a-e989-4844-894c-de96782508a3-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.786812 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c428221a-e989-4844-894c-de96782508a3-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.794055 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08baac2a-300a-498f-8d0c-d1f261e6c8fc-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "08baac2a-300a-498f-8d0c-d1f261e6c8fc" (UID: "08baac2a-300a-498f-8d0c-d1f261e6c8fc"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.812279 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08baac2a-300a-498f-8d0c-d1f261e6c8fc-kube-api-access-q4sxg" (OuterVolumeSpecName: "kube-api-access-q4sxg") pod "08baac2a-300a-498f-8d0c-d1f261e6c8fc" (UID: "08baac2a-300a-498f-8d0c-d1f261e6c8fc"). InnerVolumeSpecName "kube-api-access-q4sxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.813639 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08baac2a-300a-498f-8d0c-d1f261e6c8fc-config" (OuterVolumeSpecName: "config") pod "08baac2a-300a-498f-8d0c-d1f261e6c8fc" (UID: "08baac2a-300a-498f-8d0c-d1f261e6c8fc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.813805 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c-kube-api-access-q5t4x" (OuterVolumeSpecName: "kube-api-access-q5t4x") pod "f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c" (UID: "f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c"). InnerVolumeSpecName "kube-api-access-q5t4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.815037 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08baac2a-300a-498f-8d0c-d1f261e6c8fc-scripts" (OuterVolumeSpecName: "scripts") pod "08baac2a-300a-498f-8d0c-d1f261e6c8fc" (UID: "08baac2a-300a-498f-8d0c-d1f261e6c8fc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.818333 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "08baac2a-300a-498f-8d0c-d1f261e6c8fc" (UID: "08baac2a-300a-498f-8d0c-d1f261e6c8fc"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.853090 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c" (UID: "f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.876531 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08baac2a-300a-498f-8d0c-d1f261e6c8fc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "08baac2a-300a-498f-8d0c-d1f261e6c8fc" (UID: "08baac2a-300a-498f-8d0c-d1f261e6c8fc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.882139 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c" (UID: "f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.890686 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/08baac2a-300a-498f-8d0c-d1f261e6c8fc-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.891018 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08baac2a-300a-498f-8d0c-d1f261e6c8fc-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.891167 4790 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.891292 4790 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.891374 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5t4x\" (UniqueName: \"kubernetes.io/projected/f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c-kube-api-access-q5t4x\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.891492 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/08baac2a-300a-498f-8d0c-d1f261e6c8fc-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.891588 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08baac2a-300a-498f-8d0c-d1f261e6c8fc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.891667 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.891851 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4sxg\" (UniqueName: \"kubernetes.io/projected/08baac2a-300a-498f-8d0c-d1f261e6c8fc-kube-api-access-q4sxg\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.895263 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-48rv7"] Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.937409 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-48rv7"] Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.937634 4790 scope.go:117] "RemoveContainer" containerID="31d1bc29353608425771effac466332709c34ebd3c734237989fcb2339d732f9" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.964113 4790 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.981487 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c" (UID: "f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.986358 4790 scope.go:117] "RemoveContainer" containerID="fc25723ed92be3d0368cbe18e81e746be59930af5dbf78c90256e70ecf2cbca4" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.995272 4790 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:19 crc kubenswrapper[4790]: I1124 13:37:19.995299 4790 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.028615 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.038531 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08baac2a-300a-498f-8d0c-d1f261e6c8fc-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "08baac2a-300a-498f-8d0c-d1f261e6c8fc" (UID: "08baac2a-300a-498f-8d0c-d1f261e6c8fc"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.049597 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="9c866267-37e8-48d7-83ca-53e3cbf8b762" containerName="galera" probeResult="failure" output="" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.082234 4790 scope.go:117] "RemoveContainer" containerID="31d1bc29353608425771effac466332709c34ebd3c734237989fcb2339d732f9" Nov 24 13:37:20 crc kubenswrapper[4790]: E1124 13:37:20.086413 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31d1bc29353608425771effac466332709c34ebd3c734237989fcb2339d732f9\": container with ID starting with 31d1bc29353608425771effac466332709c34ebd3c734237989fcb2339d732f9 not found: ID does not exist" containerID="31d1bc29353608425771effac466332709c34ebd3c734237989fcb2339d732f9" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.086447 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31d1bc29353608425771effac466332709c34ebd3c734237989fcb2339d732f9"} err="failed to get container status \"31d1bc29353608425771effac466332709c34ebd3c734237989fcb2339d732f9\": rpc error: code = NotFound desc = could not find container \"31d1bc29353608425771effac466332709c34ebd3c734237989fcb2339d732f9\": container with ID starting with 31d1bc29353608425771effac466332709c34ebd3c734237989fcb2339d732f9 not found: ID does not exist" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.086468 4790 scope.go:117] "RemoveContainer" containerID="fc25723ed92be3d0368cbe18e81e746be59930af5dbf78c90256e70ecf2cbca4" Nov 24 13:37:20 crc kubenswrapper[4790]: E1124 13:37:20.090126 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc25723ed92be3d0368cbe18e81e746be59930af5dbf78c90256e70ecf2cbca4\": container with ID starting with fc25723ed92be3d0368cbe18e81e746be59930af5dbf78c90256e70ecf2cbca4 not found: ID does not exist" containerID="fc25723ed92be3d0368cbe18e81e746be59930af5dbf78c90256e70ecf2cbca4" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.090153 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc25723ed92be3d0368cbe18e81e746be59930af5dbf78c90256e70ecf2cbca4"} err="failed to get container status \"fc25723ed92be3d0368cbe18e81e746be59930af5dbf78c90256e70ecf2cbca4\": rpc error: code = NotFound desc = could not find container \"fc25723ed92be3d0368cbe18e81e746be59930af5dbf78c90256e70ecf2cbca4\": container with ID starting with fc25723ed92be3d0368cbe18e81e746be59930af5dbf78c90256e70ecf2cbca4 not found: ID does not exist" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.090170 4790 scope.go:117] "RemoveContainer" containerID="c6902baf408c3891cff01a43b41f9f037b6b29b99555835a2a635032bcee8c85" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.097100 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/698a7113-712a-4940-aef9-aabd791b021d-nova-novncproxy-tls-certs\") pod \"698a7113-712a-4940-aef9-aabd791b021d\" (UID: \"698a7113-712a-4940-aef9-aabd791b021d\") " Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.097400 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/698a7113-712a-4940-aef9-aabd791b021d-vencrypt-tls-certs\") pod \"698a7113-712a-4940-aef9-aabd791b021d\" (UID: \"698a7113-712a-4940-aef9-aabd791b021d\") " Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.097576 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csgm4\" (UniqueName: \"kubernetes.io/projected/698a7113-712a-4940-aef9-aabd791b021d-kube-api-access-csgm4\") pod \"698a7113-712a-4940-aef9-aabd791b021d\" (UID: \"698a7113-712a-4940-aef9-aabd791b021d\") " Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.097762 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/698a7113-712a-4940-aef9-aabd791b021d-config-data\") pod \"698a7113-712a-4940-aef9-aabd791b021d\" (UID: \"698a7113-712a-4940-aef9-aabd791b021d\") " Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.097920 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/698a7113-712a-4940-aef9-aabd791b021d-combined-ca-bundle\") pod \"698a7113-712a-4940-aef9-aabd791b021d\" (UID: \"698a7113-712a-4940-aef9-aabd791b021d\") " Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.098623 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/08baac2a-300a-498f-8d0c-d1f261e6c8fc-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.101982 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08baac2a-300a-498f-8d0c-d1f261e6c8fc-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "08baac2a-300a-498f-8d0c-d1f261e6c8fc" (UID: "08baac2a-300a-498f-8d0c-d1f261e6c8fc"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.109218 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/698a7113-712a-4940-aef9-aabd791b021d-kube-api-access-csgm4" (OuterVolumeSpecName: "kube-api-access-csgm4") pod "698a7113-712a-4940-aef9-aabd791b021d" (UID: "698a7113-712a-4940-aef9-aabd791b021d"). InnerVolumeSpecName "kube-api-access-csgm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.189430 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutronf6b9-account-delete-tdltj"] Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.203829 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance388c-account-delete-8hfh2"] Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.206032 4790 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/08baac2a-300a-498f-8d0c-d1f261e6c8fc-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.206057 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csgm4\" (UniqueName: \"kubernetes.io/projected/698a7113-712a-4940-aef9-aabd791b021d-kube-api-access-csgm4\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.235317 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/698a7113-712a-4940-aef9-aabd791b021d-config-data" (OuterVolumeSpecName: "config-data") pod "698a7113-712a-4940-aef9-aabd791b021d" (UID: "698a7113-712a-4940-aef9-aabd791b021d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.267738 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-bf58ddff5-swc8b"] Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.267980 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-bf58ddff5-swc8b" podUID="9362fee7-a172-4948-8721-b4c83bf04a30" containerName="proxy-httpd" containerID="cri-o://3898a01261d357b913fd416fce34429e84d9cd9629907d6bc5716e0919fddbbd" gracePeriod=30 Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.268496 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-bf58ddff5-swc8b" podUID="9362fee7-a172-4948-8721-b4c83bf04a30" containerName="proxy-server" containerID="cri-o://2893ed9292031995f05a0b830dc78e77b0e8abbf5b78c8c24b7cdaa2e9f045de" gracePeriod=30 Nov 24 13:37:20 crc kubenswrapper[4790]: W1124 13:37:20.283648 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc8592b6_2527_449c_94a2_bf40f8ea4b81.slice/crio-dfe560b995610e1d584b18e97ab60b3041339b5edfbeb4766afda70e5abe61e1 WatchSource:0}: Error finding container dfe560b995610e1d584b18e97ab60b3041339b5edfbeb4766afda70e5abe61e1: Status 404 returned error can't find the container with id dfe560b995610e1d584b18e97ab60b3041339b5edfbeb4766afda70e5abe61e1 Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.298858 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell00c43-account-delete-dlpjf"] Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.308240 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/698a7113-712a-4940-aef9-aabd791b021d-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.338659 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23e9fd30-8462-4b44-b9ae-11e3a367c814" path="/var/lib/kubelet/pods/23e9fd30-8462-4b44-b9ae-11e3a367c814/volumes" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.339491 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5deb4907-763e-4f05-8901-4b577c8802cf" path="/var/lib/kubelet/pods/5deb4907-763e-4f05-8901-4b577c8802cf/volumes" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.340172 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6803c7f5-eace-4b73-8ed7-39609fc9334a" path="/var/lib/kubelet/pods/6803c7f5-eace-4b73-8ed7-39609fc9334a/volumes" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.343019 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76801ca0-a42b-4a25-84b8-43a689e929ff" path="/var/lib/kubelet/pods/76801ca0-a42b-4a25-84b8-43a689e929ff/volumes" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.343220 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/698a7113-712a-4940-aef9-aabd791b021d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "698a7113-712a-4940-aef9-aabd791b021d" (UID: "698a7113-712a-4940-aef9-aabd791b021d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.346900 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8451cd35-fbf0-429e-a323-c6534b64e2d6" path="/var/lib/kubelet/pods/8451cd35-fbf0-429e-a323-c6534b64e2d6/volumes" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.357032 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3fa97ac-7189-4828-8b95-664e852dea8e" path="/var/lib/kubelet/pods/a3fa97ac-7189-4828-8b95-664e852dea8e/volumes" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.362395 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9c5a116-6438-47b5-8462-207eb656b553" path="/var/lib/kubelet/pods/a9c5a116-6438-47b5-8462-207eb656b553/volumes" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.366466 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c428221a-e989-4844-894c-de96782508a3" path="/var/lib/kubelet/pods/c428221a-e989-4844-894c-de96782508a3/volumes" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.367330 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8fa811d-92cf-46a2-a9fd-539130369eb1" path="/var/lib/kubelet/pods/c8fa811d-92cf-46a2-a9fd-539130369eb1/volumes" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.367938 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5d01a68-e29a-44b3-93e6-f35b91673300" path="/var/lib/kubelet/pods/d5d01a68-e29a-44b3-93e6-f35b91673300/volumes" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.369542 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c" path="/var/lib/kubelet/pods/f8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c/volumes" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.385036 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-bf58ddff5-swc8b" podUID="9362fee7-a172-4948-8721-b4c83bf04a30" containerName="proxy-server" probeResult="failure" output="Get \"https://10.217.0.166:8080/healthcheck\": read tcp 10.217.0.2:42370->10.217.0.166:8080: read: connection reset by peer" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.385493 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-bf58ddff5-swc8b" podUID="9362fee7-a172-4948-8721-b4c83bf04a30" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.166:8080/healthcheck\": read tcp 10.217.0.2:42382->10.217.0.166:8080: read: connection reset by peer" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.410366 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/698a7113-712a-4940-aef9-aabd791b021d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.425930 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/698a7113-712a-4940-aef9-aabd791b021d-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "698a7113-712a-4940-aef9-aabd791b021d" (UID: "698a7113-712a-4940-aef9-aabd791b021d"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.458061 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/698a7113-712a-4940-aef9-aabd791b021d-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "698a7113-712a-4940-aef9-aabd791b021d" (UID: "698a7113-712a-4940-aef9-aabd791b021d"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.511921 4790 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/698a7113-712a-4940-aef9-aabd791b021d-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.511972 4790 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/698a7113-712a-4940-aef9-aabd791b021d-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:20 crc kubenswrapper[4790]: E1124 13:37:20.512051 4790 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 24 13:37:20 crc kubenswrapper[4790]: E1124 13:37:20.512108 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/85e04b92-ae17-4eb6-a736-395c5d44f563-config-data podName:85e04b92-ae17-4eb6-a736-395c5d44f563 nodeName:}" failed. No retries permitted until 2025-11-24 13:37:24.512090592 +0000 UTC m=+1492.891984254 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/85e04b92-ae17-4eb6-a736-395c5d44f563-config-data") pod "rabbitmq-cell1-server-0" (UID: "85e04b92-ae17-4eb6-a736-395c5d44f563") : configmap "rabbitmq-cell1-config-data" not found Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.616135 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapid63e-account-delete-k96dh"] Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.645939 4790 generic.go:334] "Generic (PLEG): container finished" podID="9362fee7-a172-4948-8721-b4c83bf04a30" containerID="3898a01261d357b913fd416fce34429e84d9cd9629907d6bc5716e0919fddbbd" exitCode=0 Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.646012 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-bf58ddff5-swc8b" event={"ID":"9362fee7-a172-4948-8721-b4c83bf04a30","Type":"ContainerDied","Data":"3898a01261d357b913fd416fce34429e84d9cd9629907d6bc5716e0919fddbbd"} Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.647506 4790 scope.go:117] "RemoveContainer" containerID="0981452a5c31b7ac7be6aca3f35f652875423629b90d147c20e6f8b123ad006d" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.647629 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.663125 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicana018-account-delete-6z4kv" event={"ID":"7f52457e-46da-4edd-8d76-dc4a28f02f0a","Type":"ContainerStarted","Data":"e47c7ea7b8258f1eb51e4adcc261a988e1bd70fb519246a5f17a5946416fdeb2"} Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.688661 4790 generic.go:334] "Generic (PLEG): container finished" podID="9c866267-37e8-48d7-83ca-53e3cbf8b762" containerID="bdeaeaffd01a11cc6124ff6c3f7d882d3b5bdbc38615b718c3eb5ed5e5f217f3" exitCode=0 Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.688719 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"9c866267-37e8-48d7-83ca-53e3cbf8b762","Type":"ContainerDied","Data":"bdeaeaffd01a11cc6124ff6c3f7d882d3b5bdbc38615b718c3eb5ed5e5f217f3"} Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.694483 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutronf6b9-account-delete-tdltj" event={"ID":"cc8592b6-2527-449c-94a2-bf40f8ea4b81","Type":"ContainerStarted","Data":"dfe560b995610e1d584b18e97ab60b3041339b5edfbeb4766afda70e5abe61e1"} Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.697588 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapid63e-account-delete-k96dh" event={"ID":"fef75227-ffd5-4c97-87c7-e05fe899eb25","Type":"ContainerStarted","Data":"a78c05c8e90f63d4b7ef543c7ffad290d2b6b1129e5497a3a0655c1020a34bb9"} Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.703668 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell00c43-account-delete-dlpjf" event={"ID":"dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008","Type":"ContainerStarted","Data":"afe769ca639c4456baf09b664f809cfb495a91771e552b6a23e357866b90aa8a"} Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.716258 4790 generic.go:334] "Generic (PLEG): container finished" podID="8095e4aa-7290-441d-ba67-79e48c23b8a6" containerID="8ed1474926abc1218abe6afb94d8fe954c475e5e1cfcc6a22fae6792a274eb6d" exitCode=0 Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.716338 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinderc394-account-delete-gnvr4" event={"ID":"8095e4aa-7290-441d-ba67-79e48c23b8a6","Type":"ContainerDied","Data":"8ed1474926abc1218abe6afb94d8fe954c475e5e1cfcc6a22fae6792a274eb6d"} Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.716367 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinderc394-account-delete-gnvr4" event={"ID":"8095e4aa-7290-441d-ba67-79e48c23b8a6","Type":"ContainerStarted","Data":"df1bb02e10b32a65fcedd5ae514e011f827429278eb8abd8e03c502f6be3bf01"} Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.771929 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_08baac2a-300a-498f-8d0c-d1f261e6c8fc/ovsdbserver-nb/0.log" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.772209 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"08baac2a-300a-498f-8d0c-d1f261e6c8fc","Type":"ContainerDied","Data":"3b21aedb6f04264926dc7ff26b8325a37ce70a1f9a469e53fcd12df7aac97be8"} Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.772243 4790 scope.go:117] "RemoveContainer" containerID="500eaeafad5bdd0b7c31ca6862cf936a5fd98d2a297af4a1a569abfdeafe28a7" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.772330 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.794043 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement0fc6-account-delete-rl7sl" event={"ID":"51552477-67be-4c73-ae48-aa131945caea","Type":"ContainerStarted","Data":"f4a2b5aa29d9b3a52a85e4c10a9cd130150455763a2d826c0311af280f0232e2"} Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.799398 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance388c-account-delete-8hfh2" event={"ID":"f76f637a-192d-4827-92b3-68e0fc348629","Type":"ContainerStarted","Data":"b7be72ad15d6f4d3f2bed56552507ebe9f11cece3ef4e837f619a38c9837bdd0"} Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.800586 4790 generic.go:334] "Generic (PLEG): container finished" podID="698a7113-712a-4940-aef9-aabd791b021d" containerID="a3bbb077cbcc6509c4a0bd5685f5bed2599daa1c8f9b2790688ae3b5133c5b0f" exitCode=0 Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.800632 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"698a7113-712a-4940-aef9-aabd791b021d","Type":"ContainerDied","Data":"a3bbb077cbcc6509c4a0bd5685f5bed2599daa1c8f9b2790688ae3b5133c5b0f"} Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.800650 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"698a7113-712a-4940-aef9-aabd791b021d","Type":"ContainerDied","Data":"3c27269dde176c00ab82bdbd7fd94332ddf4c94c67a1bae0d5a1cc1b7f27c4ca"} Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.800711 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.812524 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.817816 4790 scope.go:117] "RemoveContainer" containerID="c70103120e0d00167ed2d2d0209229ea06b056321fe8833b38fe79ae3cb43775" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.822306 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bh9qg" podUID="bb72fd7a-89c4-4a8a-a7b0-c758322a08b7" containerName="registry-server" containerID="cri-o://56961087bbae3f0602d7c24580f19c5fa8eba1503fda903df84c3561d57e0b05" gracePeriod=2 Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.822715 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell1964b-account-delete-xhklw" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.826440 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.857004 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.896440 4790 scope.go:117] "RemoveContainer" containerID="a3bbb077cbcc6509c4a0bd5685f5bed2599daa1c8f9b2790688ae3b5133c5b0f" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.897062 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell1964b-account-delete-xhklw"] Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.905986 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell1964b-account-delete-xhklw"] Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.930538 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9c866267-37e8-48d7-83ca-53e3cbf8b762-config-data-default\") pod \"9c866267-37e8-48d7-83ca-53e3cbf8b762\" (UID: \"9c866267-37e8-48d7-83ca-53e3cbf8b762\") " Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.931626 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c866267-37e8-48d7-83ca-53e3cbf8b762-operator-scripts\") pod \"9c866267-37e8-48d7-83ca-53e3cbf8b762\" (UID: \"9c866267-37e8-48d7-83ca-53e3cbf8b762\") " Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.931731 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kb2fg\" (UniqueName: \"kubernetes.io/projected/9c866267-37e8-48d7-83ca-53e3cbf8b762-kube-api-access-kb2fg\") pod \"9c866267-37e8-48d7-83ca-53e3cbf8b762\" (UID: \"9c866267-37e8-48d7-83ca-53e3cbf8b762\") " Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.931823 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9c866267-37e8-48d7-83ca-53e3cbf8b762-config-data-generated\") pod \"9c866267-37e8-48d7-83ca-53e3cbf8b762\" (UID: \"9c866267-37e8-48d7-83ca-53e3cbf8b762\") " Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.931980 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9c866267-37e8-48d7-83ca-53e3cbf8b762-kolla-config\") pod \"9c866267-37e8-48d7-83ca-53e3cbf8b762\" (UID: \"9c866267-37e8-48d7-83ca-53e3cbf8b762\") " Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.932089 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"9c866267-37e8-48d7-83ca-53e3cbf8b762\" (UID: \"9c866267-37e8-48d7-83ca-53e3cbf8b762\") " Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.932322 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c866267-37e8-48d7-83ca-53e3cbf8b762-galera-tls-certs\") pod \"9c866267-37e8-48d7-83ca-53e3cbf8b762\" (UID: \"9c866267-37e8-48d7-83ca-53e3cbf8b762\") " Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.932510 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c866267-37e8-48d7-83ca-53e3cbf8b762-combined-ca-bundle\") pod \"9c866267-37e8-48d7-83ca-53e3cbf8b762\" (UID: \"9c866267-37e8-48d7-83ca-53e3cbf8b762\") " Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.931458 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c866267-37e8-48d7-83ca-53e3cbf8b762-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "9c866267-37e8-48d7-83ca-53e3cbf8b762" (UID: "9c866267-37e8-48d7-83ca-53e3cbf8b762"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.938992 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.939022 4790 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9c866267-37e8-48d7-83ca-53e3cbf8b762-config-data-default\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.939031 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfk9j\" (UniqueName: \"kubernetes.io/projected/20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3-kube-api-access-qfk9j\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.940386 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.949302 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.952410 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c866267-37e8-48d7-83ca-53e3cbf8b762-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "9c866267-37e8-48d7-83ca-53e3cbf8b762" (UID: "9c866267-37e8-48d7-83ca-53e3cbf8b762"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.952563 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c866267-37e8-48d7-83ca-53e3cbf8b762-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9c866267-37e8-48d7-83ca-53e3cbf8b762" (UID: "9c866267-37e8-48d7-83ca-53e3cbf8b762"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.952581 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c866267-37e8-48d7-83ca-53e3cbf8b762-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "9c866267-37e8-48d7-83ca-53e3cbf8b762" (UID: "9c866267-37e8-48d7-83ca-53e3cbf8b762"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:20 crc kubenswrapper[4790]: I1124 13:37:20.988860 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c866267-37e8-48d7-83ca-53e3cbf8b762-kube-api-access-kb2fg" (OuterVolumeSpecName: "kube-api-access-kb2fg") pod "9c866267-37e8-48d7-83ca-53e3cbf8b762" (UID: "9c866267-37e8-48d7-83ca-53e3cbf8b762"). InnerVolumeSpecName "kube-api-access-kb2fg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.013893 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "mysql-db") pod "9c866267-37e8-48d7-83ca-53e3cbf8b762" (UID: "9c866267-37e8-48d7-83ca-53e3cbf8b762"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.020825 4790 scope.go:117] "RemoveContainer" containerID="a3bbb077cbcc6509c4a0bd5685f5bed2599daa1c8f9b2790688ae3b5133c5b0f" Nov 24 13:37:21 crc kubenswrapper[4790]: E1124 13:37:21.022128 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3bbb077cbcc6509c4a0bd5685f5bed2599daa1c8f9b2790688ae3b5133c5b0f\": container with ID starting with a3bbb077cbcc6509c4a0bd5685f5bed2599daa1c8f9b2790688ae3b5133c5b0f not found: ID does not exist" containerID="a3bbb077cbcc6509c4a0bd5685f5bed2599daa1c8f9b2790688ae3b5133c5b0f" Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.022160 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3bbb077cbcc6509c4a0bd5685f5bed2599daa1c8f9b2790688ae3b5133c5b0f"} err="failed to get container status \"a3bbb077cbcc6509c4a0bd5685f5bed2599daa1c8f9b2790688ae3b5133c5b0f\": rpc error: code = NotFound desc = could not find container \"a3bbb077cbcc6509c4a0bd5685f5bed2599daa1c8f9b2790688ae3b5133c5b0f\": container with ID starting with a3bbb077cbcc6509c4a0bd5685f5bed2599daa1c8f9b2790688ae3b5133c5b0f not found: ID does not exist" Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.043700 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c866267-37e8-48d7-83ca-53e3cbf8b762-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.043732 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kb2fg\" (UniqueName: \"kubernetes.io/projected/9c866267-37e8-48d7-83ca-53e3cbf8b762-kube-api-access-kb2fg\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.043745 4790 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9c866267-37e8-48d7-83ca-53e3cbf8b762-config-data-generated\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.043757 4790 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9c866267-37e8-48d7-83ca-53e3cbf8b762-kolla-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.043788 4790 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.142804 4790 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.147707 4790 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.229703 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c866267-37e8-48d7-83ca-53e3cbf8b762-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9c866267-37e8-48d7-83ca-53e3cbf8b762" (UID: "9c866267-37e8-48d7-83ca-53e3cbf8b762"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.239575 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c866267-37e8-48d7-83ca-53e3cbf8b762-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "9c866267-37e8-48d7-83ca-53e3cbf8b762" (UID: "9c866267-37e8-48d7-83ca-53e3cbf8b762"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.250129 4790 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c866267-37e8-48d7-83ca-53e3cbf8b762-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.250157 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c866267-37e8-48d7-83ca-53e3cbf8b762-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.309220 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/placement-7957b6846-f8n7f" podUID="1b507c50-0776-4d50-ad41-10f26f25fd5f" containerName="placement-log" probeResult="failure" output="Get \"https://10.217.0.148:8778/\": read tcp 10.217.0.2:33578->10.217.0.148:8778: read: connection reset by peer" Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.309497 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/placement-7957b6846-f8n7f" podUID="1b507c50-0776-4d50-ad41-10f26f25fd5f" containerName="placement-api" probeResult="failure" output="Get \"https://10.217.0.148:8778/\": read tcp 10.217.0.2:33576->10.217.0.148:8778: read: connection reset by peer" Nov 24 13:37:21 crc kubenswrapper[4790]: E1124 13:37:21.660048 4790 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 24 13:37:21 crc kubenswrapper[4790]: E1124 13:37:21.660131 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/39508987-8655-46ed-861b-e1448652ddf6-config-data podName:39508987-8655-46ed-861b-e1448652ddf6 nodeName:}" failed. No retries permitted until 2025-11-24 13:37:25.660114236 +0000 UTC m=+1494.040007898 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/39508987-8655-46ed-861b-e1448652ddf6-config-data") pod "rabbitmq-server-0" (UID: "39508987-8655-46ed-861b-e1448652ddf6") : configmap "rabbitmq-config-data" not found Nov 24 13:37:21 crc kubenswrapper[4790]: E1124 13:37:21.828059 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b8183ab06a7a4bd08c674c5788831aa9879a328f22b998c507e78d8d453b26d is running failed: container process not found" containerID="6b8183ab06a7a4bd08c674c5788831aa9879a328f22b998c507e78d8d453b26d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 13:37:21 crc kubenswrapper[4790]: E1124 13:37:21.829063 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b8183ab06a7a4bd08c674c5788831aa9879a328f22b998c507e78d8d453b26d is running failed: container process not found" containerID="6b8183ab06a7a4bd08c674c5788831aa9879a328f22b998c507e78d8d453b26d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 13:37:21 crc kubenswrapper[4790]: E1124 13:37:21.829554 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b8183ab06a7a4bd08c674c5788831aa9879a328f22b998c507e78d8d453b26d is running failed: container process not found" containerID="6b8183ab06a7a4bd08c674c5788831aa9879a328f22b998c507e78d8d453b26d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 13:37:21 crc kubenswrapper[4790]: E1124 13:37:21.829602 4790 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b8183ab06a7a4bd08c674c5788831aa9879a328f22b998c507e78d8d453b26d is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="de6f5df1-eb57-493b-9d89-5e710c6f4304" containerName="nova-scheduler-scheduler" Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.860843 4790 generic.go:334] "Generic (PLEG): container finished" podID="7f52457e-46da-4edd-8d76-dc4a28f02f0a" containerID="49eb35e3cab0fee6082b7f8ae966428557fc6407e903621826b70befb0756a67" exitCode=0 Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.861177 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicana018-account-delete-6z4kv" event={"ID":"7f52457e-46da-4edd-8d76-dc4a28f02f0a","Type":"ContainerDied","Data":"49eb35e3cab0fee6082b7f8ae966428557fc6407e903621826b70befb0756a67"} Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.882088 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.884749 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.885050 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3fc4314a-6a31-49d1-8652-f4de6da97916" containerName="ceilometer-central-agent" containerID="cri-o://63bc861e59ed27e8b2dec969395af49861e48d187c85af757c5022b25349826a" gracePeriod=30 Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.885174 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3fc4314a-6a31-49d1-8652-f4de6da97916" containerName="proxy-httpd" containerID="cri-o://3bdd8dbb30e64b7458fddef45148be02e9280fe559bf6018d442a5bf9eb8dada" gracePeriod=30 Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.885215 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3fc4314a-6a31-49d1-8652-f4de6da97916" containerName="sg-core" containerID="cri-o://b5e492786f34b12c4fccd95cb12360cbb4386005a0414f30c8e1be783d5252e0" gracePeriod=30 Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.885245 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3fc4314a-6a31-49d1-8652-f4de6da97916" containerName="ceilometer-notification-agent" containerID="cri-o://593dfb8aab769d099082292784d20c9dd7fc610d3ac9c0659d6fd1635968dfa7" gracePeriod=30 Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.886303 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bh9qg" Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.895953 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-bf58ddff5-swc8b" Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.896344 4790 generic.go:334] "Generic (PLEG): container finished" podID="bb72fd7a-89c4-4a8a-a7b0-c758322a08b7" containerID="56961087bbae3f0602d7c24580f19c5fa8eba1503fda903df84c3561d57e0b05" exitCode=0 Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.896392 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bh9qg" event={"ID":"bb72fd7a-89c4-4a8a-a7b0-c758322a08b7","Type":"ContainerDied","Data":"56961087bbae3f0602d7c24580f19c5fa8eba1503fda903df84c3561d57e0b05"} Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.896424 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bh9qg" event={"ID":"bb72fd7a-89c4-4a8a-a7b0-c758322a08b7","Type":"ContainerDied","Data":"64380177ab7771b2d7b3c1ad694c3450d1fb45313e2b57caf0b4716126ab5d21"} Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.896440 4790 scope.go:117] "RemoveContainer" containerID="56961087bbae3f0602d7c24580f19c5fa8eba1503fda903df84c3561d57e0b05" Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.909956 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.935259 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"9c866267-37e8-48d7-83ca-53e3cbf8b762","Type":"ContainerDied","Data":"9d54c6d89a4aa563bf6e3e34eb7285204e3ed1ccfd61c3285c7cc567678f7c48"} Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.935531 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.943321 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7957b6846-f8n7f" Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.955126 4790 generic.go:334] "Generic (PLEG): container finished" podID="f76f637a-192d-4827-92b3-68e0fc348629" containerID="93d8fe1bd57c2a04bf2a36122f12d68f6d367d8fe7a99c2f41e508c7615f68a4" exitCode=0 Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.955213 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance388c-account-delete-8hfh2" event={"ID":"f76f637a-192d-4827-92b3-68e0fc348629","Type":"ContainerDied","Data":"93d8fe1bd57c2a04bf2a36122f12d68f6d367d8fe7a99c2f41e508c7615f68a4"} Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.963285 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9362fee7-a172-4948-8721-b4c83bf04a30-public-tls-certs\") pod \"9362fee7-a172-4948-8721-b4c83bf04a30\" (UID: \"9362fee7-a172-4948-8721-b4c83bf04a30\") " Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.963320 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9362fee7-a172-4948-8721-b4c83bf04a30-run-httpd\") pod \"9362fee7-a172-4948-8721-b4c83bf04a30\" (UID: \"9362fee7-a172-4948-8721-b4c83bf04a30\") " Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.963341 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b507c50-0776-4d50-ad41-10f26f25fd5f-logs\") pod \"1b507c50-0776-4d50-ad41-10f26f25fd5f\" (UID: \"1b507c50-0776-4d50-ad41-10f26f25fd5f\") " Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.963390 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b507c50-0776-4d50-ad41-10f26f25fd5f-public-tls-certs\") pod \"1b507c50-0776-4d50-ad41-10f26f25fd5f\" (UID: \"1b507c50-0776-4d50-ad41-10f26f25fd5f\") " Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.963423 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de6f5df1-eb57-493b-9d89-5e710c6f4304-config-data\") pod \"de6f5df1-eb57-493b-9d89-5e710c6f4304\" (UID: \"de6f5df1-eb57-493b-9d89-5e710c6f4304\") " Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.963449 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b507c50-0776-4d50-ad41-10f26f25fd5f-config-data\") pod \"1b507c50-0776-4d50-ad41-10f26f25fd5f\" (UID: \"1b507c50-0776-4d50-ad41-10f26f25fd5f\") " Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.963471 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de6f5df1-eb57-493b-9d89-5e710c6f4304-combined-ca-bundle\") pod \"de6f5df1-eb57-493b-9d89-5e710c6f4304\" (UID: \"de6f5df1-eb57-493b-9d89-5e710c6f4304\") " Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.963541 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b507c50-0776-4d50-ad41-10f26f25fd5f-internal-tls-certs\") pod \"1b507c50-0776-4d50-ad41-10f26f25fd5f\" (UID: \"1b507c50-0776-4d50-ad41-10f26f25fd5f\") " Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.963588 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9362fee7-a172-4948-8721-b4c83bf04a30-internal-tls-certs\") pod \"9362fee7-a172-4948-8721-b4c83bf04a30\" (UID: \"9362fee7-a172-4948-8721-b4c83bf04a30\") " Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.963616 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rqhq\" (UniqueName: \"kubernetes.io/projected/9362fee7-a172-4948-8721-b4c83bf04a30-kube-api-access-7rqhq\") pod \"9362fee7-a172-4948-8721-b4c83bf04a30\" (UID: \"9362fee7-a172-4948-8721-b4c83bf04a30\") " Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.963631 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9362fee7-a172-4948-8721-b4c83bf04a30-log-httpd\") pod \"9362fee7-a172-4948-8721-b4c83bf04a30\" (UID: \"9362fee7-a172-4948-8721-b4c83bf04a30\") " Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.963656 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb72fd7a-89c4-4a8a-a7b0-c758322a08b7-utilities\") pod \"bb72fd7a-89c4-4a8a-a7b0-c758322a08b7\" (UID: \"bb72fd7a-89c4-4a8a-a7b0-c758322a08b7\") " Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.963672 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9362fee7-a172-4948-8721-b4c83bf04a30-combined-ca-bundle\") pod \"9362fee7-a172-4948-8721-b4c83bf04a30\" (UID: \"9362fee7-a172-4948-8721-b4c83bf04a30\") " Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.963705 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lr2bq\" (UniqueName: \"kubernetes.io/projected/1b507c50-0776-4d50-ad41-10f26f25fd5f-kube-api-access-lr2bq\") pod \"1b507c50-0776-4d50-ad41-10f26f25fd5f\" (UID: \"1b507c50-0776-4d50-ad41-10f26f25fd5f\") " Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.963721 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b507c50-0776-4d50-ad41-10f26f25fd5f-scripts\") pod \"1b507c50-0776-4d50-ad41-10f26f25fd5f\" (UID: \"1b507c50-0776-4d50-ad41-10f26f25fd5f\") " Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.963766 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b507c50-0776-4d50-ad41-10f26f25fd5f-combined-ca-bundle\") pod \"1b507c50-0776-4d50-ad41-10f26f25fd5f\" (UID: \"1b507c50-0776-4d50-ad41-10f26f25fd5f\") " Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.963786 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9362fee7-a172-4948-8721-b4c83bf04a30-config-data\") pod \"9362fee7-a172-4948-8721-b4c83bf04a30\" (UID: \"9362fee7-a172-4948-8721-b4c83bf04a30\") " Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.963812 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6p57k\" (UniqueName: \"kubernetes.io/projected/de6f5df1-eb57-493b-9d89-5e710c6f4304-kube-api-access-6p57k\") pod \"de6f5df1-eb57-493b-9d89-5e710c6f4304\" (UID: \"de6f5df1-eb57-493b-9d89-5e710c6f4304\") " Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.963841 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9362fee7-a172-4948-8721-b4c83bf04a30-etc-swift\") pod \"9362fee7-a172-4948-8721-b4c83bf04a30\" (UID: \"9362fee7-a172-4948-8721-b4c83bf04a30\") " Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.963857 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb72fd7a-89c4-4a8a-a7b0-c758322a08b7-catalog-content\") pod \"bb72fd7a-89c4-4a8a-a7b0-c758322a08b7\" (UID: \"bb72fd7a-89c4-4a8a-a7b0-c758322a08b7\") " Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.963891 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrwvv\" (UniqueName: \"kubernetes.io/projected/bb72fd7a-89c4-4a8a-a7b0-c758322a08b7-kube-api-access-xrwvv\") pod \"bb72fd7a-89c4-4a8a-a7b0-c758322a08b7\" (UID: \"bb72fd7a-89c4-4a8a-a7b0-c758322a08b7\") " Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.968323 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9362fee7-a172-4948-8721-b4c83bf04a30-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9362fee7-a172-4948-8721-b4c83bf04a30" (UID: "9362fee7-a172-4948-8721-b4c83bf04a30"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.970127 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb72fd7a-89c4-4a8a-a7b0-c758322a08b7-utilities" (OuterVolumeSpecName: "utilities") pod "bb72fd7a-89c4-4a8a-a7b0-c758322a08b7" (UID: "bb72fd7a-89c4-4a8a-a7b0-c758322a08b7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.970553 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b507c50-0776-4d50-ad41-10f26f25fd5f-logs" (OuterVolumeSpecName: "logs") pod "1b507c50-0776-4d50-ad41-10f26f25fd5f" (UID: "1b507c50-0776-4d50-ad41-10f26f25fd5f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.972514 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b507c50-0776-4d50-ad41-10f26f25fd5f-scripts" (OuterVolumeSpecName: "scripts") pod "1b507c50-0776-4d50-ad41-10f26f25fd5f" (UID: "1b507c50-0776-4d50-ad41-10f26f25fd5f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.976912 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9362fee7-a172-4948-8721-b4c83bf04a30-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9362fee7-a172-4948-8721-b4c83bf04a30" (UID: "9362fee7-a172-4948-8721-b4c83bf04a30"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.980472 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9362fee7-a172-4948-8721-b4c83bf04a30-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "9362fee7-a172-4948-8721-b4c83bf04a30" (UID: "9362fee7-a172-4948-8721-b4c83bf04a30"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.987793 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb72fd7a-89c4-4a8a-a7b0-c758322a08b7-kube-api-access-xrwvv" (OuterVolumeSpecName: "kube-api-access-xrwvv") pod "bb72fd7a-89c4-4a8a-a7b0-c758322a08b7" (UID: "bb72fd7a-89c4-4a8a-a7b0-c758322a08b7"). InnerVolumeSpecName "kube-api-access-xrwvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.987921 4790 scope.go:117] "RemoveContainer" containerID="212a63776bee7e4aa0a7c001871fb866bdc261b22ca9ead26adc668341a14cd7" Nov 24 13:37:21 crc kubenswrapper[4790]: I1124 13:37:21.990960 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9362fee7-a172-4948-8721-b4c83bf04a30-kube-api-access-7rqhq" (OuterVolumeSpecName: "kube-api-access-7rqhq") pod "9362fee7-a172-4948-8721-b4c83bf04a30" (UID: "9362fee7-a172-4948-8721-b4c83bf04a30"). InnerVolumeSpecName "kube-api-access-7rqhq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.000474 4790 generic.go:334] "Generic (PLEG): container finished" podID="dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008" containerID="8f04a853d966a711ba4815311768aa3e3f970e90d66e7b8eada63e04a25b99f2" exitCode=0 Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.000500 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b507c50-0776-4d50-ad41-10f26f25fd5f-kube-api-access-lr2bq" (OuterVolumeSpecName: "kube-api-access-lr2bq") pod "1b507c50-0776-4d50-ad41-10f26f25fd5f" (UID: "1b507c50-0776-4d50-ad41-10f26f25fd5f"). InnerVolumeSpecName "kube-api-access-lr2bq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.000538 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell00c43-account-delete-dlpjf" event={"ID":"dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008","Type":"ContainerDied","Data":"8f04a853d966a711ba4815311768aa3e3f970e90d66e7b8eada63e04a25b99f2"} Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.008481 4790 generic.go:334] "Generic (PLEG): container finished" podID="9362fee7-a172-4948-8721-b4c83bf04a30" containerID="2893ed9292031995f05a0b830dc78e77b0e8abbf5b78c8c24b7cdaa2e9f045de" exitCode=0 Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.008540 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-bf58ddff5-swc8b" event={"ID":"9362fee7-a172-4948-8721-b4c83bf04a30","Type":"ContainerDied","Data":"2893ed9292031995f05a0b830dc78e77b0e8abbf5b78c8c24b7cdaa2e9f045de"} Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.008566 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-bf58ddff5-swc8b" event={"ID":"9362fee7-a172-4948-8721-b4c83bf04a30","Type":"ContainerDied","Data":"a19ffc076b57ad785b185695066fd43a17b7546b1762de4a4f70d8894cb8ee26"} Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.008623 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-bf58ddff5-swc8b" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.012227 4790 generic.go:334] "Generic (PLEG): container finished" podID="e6d696a7-d618-4416-b499-aae08088b079" containerID="002e858954a60800481c9819989042b1ad82aae0f1762f13550da43a5f262b74" exitCode=0 Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.012322 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e6d696a7-d618-4416-b499-aae08088b079","Type":"ContainerDied","Data":"002e858954a60800481c9819989042b1ad82aae0f1762f13550da43a5f262b74"} Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.024793 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb72fd7a-89c4-4a8a-a7b0-c758322a08b7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bb72fd7a-89c4-4a8a-a7b0-c758322a08b7" (UID: "bb72fd7a-89c4-4a8a-a7b0-c758322a08b7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.025794 4790 generic.go:334] "Generic (PLEG): container finished" podID="cc8592b6-2527-449c-94a2-bf40f8ea4b81" containerID="ce87db4947942f196094e19acebc86c6075c26a08fa34743ddae754d898e5938" exitCode=0 Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.025902 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutronf6b9-account-delete-tdltj" event={"ID":"cc8592b6-2527-449c-94a2-bf40f8ea4b81","Type":"ContainerDied","Data":"ce87db4947942f196094e19acebc86c6075c26a08fa34743ddae754d898e5938"} Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.039235 4790 generic.go:334] "Generic (PLEG): container finished" podID="51552477-67be-4c73-ae48-aa131945caea" containerID="9add9dc189b49006e7c669c2830732c7550855a82a0cb3552df890b1e86ae064" exitCode=0 Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.039340 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement0fc6-account-delete-rl7sl" event={"ID":"51552477-67be-4c73-ae48-aa131945caea","Type":"ContainerDied","Data":"9add9dc189b49006e7c669c2830732c7550855a82a0cb3552df890b1e86ae064"} Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.041346 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de6f5df1-eb57-493b-9d89-5e710c6f4304-kube-api-access-6p57k" (OuterVolumeSpecName: "kube-api-access-6p57k") pod "de6f5df1-eb57-493b-9d89-5e710c6f4304" (UID: "de6f5df1-eb57-493b-9d89-5e710c6f4304"). InnerVolumeSpecName "kube-api-access-6p57k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.068208 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rqhq\" (UniqueName: \"kubernetes.io/projected/9362fee7-a172-4948-8721-b4c83bf04a30-kube-api-access-7rqhq\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.068245 4790 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9362fee7-a172-4948-8721-b4c83bf04a30-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.068258 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb72fd7a-89c4-4a8a-a7b0-c758322a08b7-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.068269 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lr2bq\" (UniqueName: \"kubernetes.io/projected/1b507c50-0776-4d50-ad41-10f26f25fd5f-kube-api-access-lr2bq\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.068280 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b507c50-0776-4d50-ad41-10f26f25fd5f-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.068292 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6p57k\" (UniqueName: \"kubernetes.io/projected/de6f5df1-eb57-493b-9d89-5e710c6f4304-kube-api-access-6p57k\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.068303 4790 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9362fee7-a172-4948-8721-b4c83bf04a30-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.068316 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb72fd7a-89c4-4a8a-a7b0-c758322a08b7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.068327 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrwvv\" (UniqueName: \"kubernetes.io/projected/bb72fd7a-89c4-4a8a-a7b0-c758322a08b7-kube-api-access-xrwvv\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.068338 4790 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9362fee7-a172-4948-8721-b4c83bf04a30-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.068371 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b507c50-0776-4d50-ad41-10f26f25fd5f-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.080497 4790 generic.go:334] "Generic (PLEG): container finished" podID="1b507c50-0776-4d50-ad41-10f26f25fd5f" containerID="2f816c8e5f36c0c446e29e140b1f74aeb5df0880612029daa8ef51dfaee75bee" exitCode=0 Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.080579 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7957b6846-f8n7f" event={"ID":"1b507c50-0776-4d50-ad41-10f26f25fd5f","Type":"ContainerDied","Data":"2f816c8e5f36c0c446e29e140b1f74aeb5df0880612029daa8ef51dfaee75bee"} Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.080612 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7957b6846-f8n7f" event={"ID":"1b507c50-0776-4d50-ad41-10f26f25fd5f","Type":"ContainerDied","Data":"d67870725bbb86c25d9c408bb9a7732e56318396fd04d3647257357971eef254"} Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.080677 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7957b6846-f8n7f" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.107115 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="81757ef9-7330-44bb-a35e-8e6f4c137c0e" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": read tcp 10.217.0.2:40516->10.217.0.201:8775: read: connection reset by peer" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.107402 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="81757ef9-7330-44bb-a35e-8e6f4c137c0e" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": read tcp 10.217.0.2:40506->10.217.0.201:8775: read: connection reset by peer" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.117276 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.117445 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="56d67de4-df9a-44fa-92b2-cbb168c4ae87" containerName="memcached" containerID="cri-o://6501f4d3c19251ce46a87c27719040d3bcdb4ba9e7bcdb6dbb40bc282a09c44d" gracePeriod=30 Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.134911 4790 generic.go:334] "Generic (PLEG): container finished" podID="de6f5df1-eb57-493b-9d89-5e710c6f4304" containerID="6b8183ab06a7a4bd08c674c5788831aa9879a328f22b998c507e78d8d453b26d" exitCode=0 Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.135008 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"de6f5df1-eb57-493b-9d89-5e710c6f4304","Type":"ContainerDied","Data":"6b8183ab06a7a4bd08c674c5788831aa9879a328f22b998c507e78d8d453b26d"} Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.135034 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"de6f5df1-eb57-493b-9d89-5e710c6f4304","Type":"ContainerDied","Data":"42f4b21d70f156369f559c7e6ef7a64596f9aba627a9c1ab988a7e9aee79943c"} Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.135086 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.147302 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.156736 4790 generic.go:334] "Generic (PLEG): container finished" podID="fef75227-ffd5-4c97-87c7-e05fe899eb25" containerID="cfc22827f7595f647048bc319094a1a8720f3f344164f8db5947014cfaf081fc" exitCode=0 Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.156899 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="23704906-0f33-4a4e-8bbb-a796a9a1733f" containerName="kube-state-metrics" containerID="cri-o://7eb4e19d335efc1382a2f1934fe4914e9bce012cc0d8c8dd38049967fe85a33d" gracePeriod=30 Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.157009 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapid63e-account-delete-k96dh" event={"ID":"fef75227-ffd5-4c97-87c7-e05fe899eb25","Type":"ContainerDied","Data":"cfc22827f7595f647048bc319094a1a8720f3f344164f8db5947014cfaf081fc"} Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.168786 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.186009 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-m27jn"] Nov 24 13:37:22 crc kubenswrapper[4790]: E1124 13:37:22.195866 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="eb8a4e1ecddd7fb54e90dd1a525962762b3153943ebd954bb7544913710d2ec2" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.204341 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-m27jn"] Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.207261 4790 scope.go:117] "RemoveContainer" containerID="a0bab3581d08313eb5fb228fef8c8dc47ab3fb05a7b2805a913fce1b4d13047e" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.215063 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de6f5df1-eb57-493b-9d89-5e710c6f4304-config-data" (OuterVolumeSpecName: "config-data") pod "de6f5df1-eb57-493b-9d89-5e710c6f4304" (UID: "de6f5df1-eb57-493b-9d89-5e710c6f4304"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.217828 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de6f5df1-eb57-493b-9d89-5e710c6f4304-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "de6f5df1-eb57-493b-9d89-5e710c6f4304" (UID: "de6f5df1-eb57-493b-9d89-5e710c6f4304"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.222938 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-zqb2l"] Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.255051 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-zqb2l"] Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.271587 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de6f5df1-eb57-493b-9d89-5e710c6f4304-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.271612 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de6f5df1-eb57-493b-9d89-5e710c6f4304-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:22 crc kubenswrapper[4790]: E1124 13:37:22.287225 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="eb8a4e1ecddd7fb54e90dd1a525962762b3153943ebd954bb7544913710d2ec2" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 24 13:37:22 crc kubenswrapper[4790]: E1124 13:37:22.293113 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="eb8a4e1ecddd7fb54e90dd1a525962762b3153943ebd954bb7544913710d2ec2" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 24 13:37:22 crc kubenswrapper[4790]: E1124 13:37:22.293180 4790 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="39d594c5-aae1-4cd7-a7de-9de435ad83b8" containerName="nova-cell1-conductor-conductor" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.305409 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-7ddd868cf8-lc6qt"] Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.305654 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-7ddd868cf8-lc6qt" podUID="336619dc-13bf-4168-a58b-b5c724759890" containerName="keystone-api" containerID="cri-o://07ac1c04efe6c3d8ada271427bad943fb80399185792ecba62e72b13d34fc58d" gracePeriod=30 Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.404067 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08baac2a-300a-498f-8d0c-d1f261e6c8fc" path="/var/lib/kubelet/pods/08baac2a-300a-498f-8d0c-d1f261e6c8fc/volumes" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.405158 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3" path="/var/lib/kubelet/pods/20fb35f9-dff5-4c3f-84bf-668f6f4ba7b3/volumes" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.405472 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="482e2d7c-282b-490c-b476-955ddc498707" path="/var/lib/kubelet/pods/482e2d7c-282b-490c-b476-955ddc498707/volumes" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.406043 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="698a7113-712a-4940-aef9-aabd791b021d" path="/var/lib/kubelet/pods/698a7113-712a-4940-aef9-aabd791b021d/volumes" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.407467 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f04f724-1dc9-4e6a-80ab-bd139438d1b8" path="/var/lib/kubelet/pods/7f04f724-1dc9-4e6a-80ab-bd139438d1b8/volumes" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.408521 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c866267-37e8-48d7-83ca-53e3cbf8b762" path="/var/lib/kubelet/pods/9c866267-37e8-48d7-83ca-53e3cbf8b762/volumes" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.455468 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7d69688466-r2bq9" podUID="8311dfa1-6e87-493d-92e0-4eb69c792afe" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:38688->10.217.0.160:9311: read: connection reset by peer" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.456142 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7d69688466-r2bq9" podUID="8311dfa1-6e87-493d-92e0-4eb69c792afe" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:38692->10.217.0.160:9311: read: connection reset by peer" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.472135 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9362fee7-a172-4948-8721-b4c83bf04a30-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "9362fee7-a172-4948-8721-b4c83bf04a30" (UID: "9362fee7-a172-4948-8721-b4c83bf04a30"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.472245 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9362fee7-a172-4948-8721-b4c83bf04a30-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "9362fee7-a172-4948-8721-b4c83bf04a30" (UID: "9362fee7-a172-4948-8721-b4c83bf04a30"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.490629 4790 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9362fee7-a172-4948-8721-b4c83bf04a30-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.490665 4790 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9362fee7-a172-4948-8721-b4c83bf04a30-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.595018 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b507c50-0776-4d50-ad41-10f26f25fd5f-config-data" (OuterVolumeSpecName: "config-data") pod "1b507c50-0776-4d50-ad41-10f26f25fd5f" (UID: "1b507c50-0776-4d50-ad41-10f26f25fd5f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.603235 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b507c50-0776-4d50-ad41-10f26f25fd5f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "1b507c50-0776-4d50-ad41-10f26f25fd5f" (UID: "1b507c50-0776-4d50-ad41-10f26f25fd5f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.603803 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9362fee7-a172-4948-8721-b4c83bf04a30-config-data" (OuterVolumeSpecName: "config-data") pod "9362fee7-a172-4948-8721-b4c83bf04a30" (UID: "9362fee7-a172-4948-8721-b4c83bf04a30"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.604331 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b507c50-0776-4d50-ad41-10f26f25fd5f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1b507c50-0776-4d50-ad41-10f26f25fd5f" (UID: "1b507c50-0776-4d50-ad41-10f26f25fd5f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.605529 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9362fee7-a172-4948-8721-b4c83bf04a30-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9362fee7-a172-4948-8721-b4c83bf04a30" (UID: "9362fee7-a172-4948-8721-b4c83bf04a30"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.625785 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.625833 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-86tcg"] Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.625852 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-86tcg"] Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.625866 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutronf6b9-account-delete-tdltj"] Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.625876 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-f6b9-account-create-zfn8m"] Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.625904 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-f6b9-account-create-zfn8m"] Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.625914 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-xsphw"] Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.679618 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-xsphw"] Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.687750 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b507c50-0776-4d50-ad41-10f26f25fd5f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "1b507c50-0776-4d50-ad41-10f26f25fd5f" (UID: "1b507c50-0776-4d50-ad41-10f26f25fd5f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.694989 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b507c50-0776-4d50-ad41-10f26f25fd5f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.695026 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9362fee7-a172-4948-8721-b4c83bf04a30-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.695035 4790 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b507c50-0776-4d50-ad41-10f26f25fd5f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.695043 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b507c50-0776-4d50-ad41-10f26f25fd5f-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.695052 4790 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b507c50-0776-4d50-ad41-10f26f25fd5f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.695066 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9362fee7-a172-4948-8721-b4c83bf04a30-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.696207 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.726314 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-6aea-account-create-hfz2v"] Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.736942 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-6aea-account-create-hfz2v"] Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.749660 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-4qj2r"] Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.758802 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-4qj2r"] Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.807255 4790 scope.go:117] "RemoveContainer" containerID="56961087bbae3f0602d7c24580f19c5fa8eba1503fda903df84c3561d57e0b05" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.811503 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdxq7\" (UniqueName: \"kubernetes.io/projected/e6d696a7-d618-4416-b499-aae08088b079-kube-api-access-mdxq7\") pod \"e6d696a7-d618-4416-b499-aae08088b079\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.811623 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-internal-tls-certs\") pod \"e6d696a7-d618-4416-b499-aae08088b079\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.811673 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6d696a7-d618-4416-b499-aae08088b079-logs\") pod \"e6d696a7-d618-4416-b499-aae08088b079\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.811712 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-config-data-custom\") pod \"e6d696a7-d618-4416-b499-aae08088b079\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.811759 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e6d696a7-d618-4416-b499-aae08088b079-etc-machine-id\") pod \"e6d696a7-d618-4416-b499-aae08088b079\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.811775 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-combined-ca-bundle\") pod \"e6d696a7-d618-4416-b499-aae08088b079\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.811791 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-scripts\") pod \"e6d696a7-d618-4416-b499-aae08088b079\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.811810 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-public-tls-certs\") pod \"e6d696a7-d618-4416-b499-aae08088b079\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.811914 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-config-data\") pod \"e6d696a7-d618-4416-b499-aae08088b079\" (UID: \"e6d696a7-d618-4416-b499-aae08088b079\") " Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.819063 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-d63e-account-create-vhfd4"] Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.823568 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e6d696a7-d618-4416-b499-aae08088b079-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e6d696a7-d618-4416-b499-aae08088b079" (UID: "e6d696a7-d618-4416-b499-aae08088b079"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.825339 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6d696a7-d618-4416-b499-aae08088b079-logs" (OuterVolumeSpecName: "logs") pod "e6d696a7-d618-4416-b499-aae08088b079" (UID: "e6d696a7-d618-4416-b499-aae08088b079"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.828977 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-d63e-account-create-vhfd4"] Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.830554 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e6d696a7-d618-4416-b499-aae08088b079" (UID: "e6d696a7-d618-4416-b499-aae08088b079"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:22 crc kubenswrapper[4790]: E1124 13:37:22.831302 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56961087bbae3f0602d7c24580f19c5fa8eba1503fda903df84c3561d57e0b05\": container with ID starting with 56961087bbae3f0602d7c24580f19c5fa8eba1503fda903df84c3561d57e0b05 not found: ID does not exist" containerID="56961087bbae3f0602d7c24580f19c5fa8eba1503fda903df84c3561d57e0b05" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.834349 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56961087bbae3f0602d7c24580f19c5fa8eba1503fda903df84c3561d57e0b05"} err="failed to get container status \"56961087bbae3f0602d7c24580f19c5fa8eba1503fda903df84c3561d57e0b05\": rpc error: code = NotFound desc = could not find container \"56961087bbae3f0602d7c24580f19c5fa8eba1503fda903df84c3561d57e0b05\": container with ID starting with 56961087bbae3f0602d7c24580f19c5fa8eba1503fda903df84c3561d57e0b05 not found: ID does not exist" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.834558 4790 scope.go:117] "RemoveContainer" containerID="212a63776bee7e4aa0a7c001871fb866bdc261b22ca9ead26adc668341a14cd7" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.833961 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapid63e-account-delete-k96dh"] Nov 24 13:37:22 crc kubenswrapper[4790]: E1124 13:37:22.838080 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"212a63776bee7e4aa0a7c001871fb866bdc261b22ca9ead26adc668341a14cd7\": container with ID starting with 212a63776bee7e4aa0a7c001871fb866bdc261b22ca9ead26adc668341a14cd7 not found: ID does not exist" containerID="212a63776bee7e4aa0a7c001871fb866bdc261b22ca9ead26adc668341a14cd7" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.838137 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"212a63776bee7e4aa0a7c001871fb866bdc261b22ca9ead26adc668341a14cd7"} err="failed to get container status \"212a63776bee7e4aa0a7c001871fb866bdc261b22ca9ead26adc668341a14cd7\": rpc error: code = NotFound desc = could not find container \"212a63776bee7e4aa0a7c001871fb866bdc261b22ca9ead26adc668341a14cd7\": container with ID starting with 212a63776bee7e4aa0a7c001871fb866bdc261b22ca9ead26adc668341a14cd7 not found: ID does not exist" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.838170 4790 scope.go:117] "RemoveContainer" containerID="a0bab3581d08313eb5fb228fef8c8dc47ab3fb05a7b2805a913fce1b4d13047e" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.841709 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-94scw"] Nov 24 13:37:22 crc kubenswrapper[4790]: E1124 13:37:22.843025 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0bab3581d08313eb5fb228fef8c8dc47ab3fb05a7b2805a913fce1b4d13047e\": container with ID starting with a0bab3581d08313eb5fb228fef8c8dc47ab3fb05a7b2805a913fce1b4d13047e not found: ID does not exist" containerID="a0bab3581d08313eb5fb228fef8c8dc47ab3fb05a7b2805a913fce1b4d13047e" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.843077 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0bab3581d08313eb5fb228fef8c8dc47ab3fb05a7b2805a913fce1b4d13047e"} err="failed to get container status \"a0bab3581d08313eb5fb228fef8c8dc47ab3fb05a7b2805a913fce1b4d13047e\": rpc error: code = NotFound desc = could not find container \"a0bab3581d08313eb5fb228fef8c8dc47ab3fb05a7b2805a913fce1b4d13047e\": container with ID starting with a0bab3581d08313eb5fb228fef8c8dc47ab3fb05a7b2805a913fce1b4d13047e not found: ID does not exist" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.843103 4790 scope.go:117] "RemoveContainer" containerID="bdeaeaffd01a11cc6124ff6c3f7d882d3b5bdbc38615b718c3eb5ed5e5f217f3" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.852368 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-94scw"] Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.859952 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-0c43-account-create-l9mrr"] Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.869954 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-0c43-account-create-l9mrr"] Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.877177 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-scripts" (OuterVolumeSpecName: "scripts") pod "e6d696a7-d618-4416-b499-aae08088b079" (UID: "e6d696a7-d618-4416-b499-aae08088b079"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.884445 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6d696a7-d618-4416-b499-aae08088b079-kube-api-access-mdxq7" (OuterVolumeSpecName: "kube-api-access-mdxq7") pod "e6d696a7-d618-4416-b499-aae08088b079" (UID: "e6d696a7-d618-4416-b499-aae08088b079"). InnerVolumeSpecName "kube-api-access-mdxq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.903525 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell00c43-account-delete-dlpjf"] Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.913530 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdxq7\" (UniqueName: \"kubernetes.io/projected/e6d696a7-d618-4416-b499-aae08088b079-kube-api-access-mdxq7\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.913564 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6d696a7-d618-4416-b499-aae08088b079-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.913576 4790 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.913588 4790 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e6d696a7-d618-4416-b499-aae08088b079-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.913599 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.923059 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.930803 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.937703 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-bf58ddff5-swc8b"] Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.943037 4790 scope.go:117] "RemoveContainer" containerID="ac2a207b8fd870760b58312e8f38e10ea17fb9654e5b83ea853f3eeae396f453" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.962992 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-bf58ddff5-swc8b"] Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.967846 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinderc394-account-delete-gnvr4" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.969826 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e6d696a7-d618-4416-b499-aae08088b079" (UID: "e6d696a7-d618-4416-b499-aae08088b079"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.974890 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e6d696a7-d618-4416-b499-aae08088b079" (UID: "e6d696a7-d618-4416-b499-aae08088b079"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.984763 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e6d696a7-d618-4416-b499-aae08088b079" (UID: "e6d696a7-d618-4416-b499-aae08088b079"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:22 crc kubenswrapper[4790]: I1124 13:37:22.995893 4790 scope.go:117] "RemoveContainer" containerID="2893ed9292031995f05a0b830dc78e77b0e8abbf5b78c8c24b7cdaa2e9f045de" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.000288 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-config-data" (OuterVolumeSpecName: "config-data") pod "e6d696a7-d618-4416-b499-aae08088b079" (UID: "e6d696a7-d618-4416-b499-aae08088b079"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.027373 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.027627 4790 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.027723 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.027779 4790 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6d696a7-d618-4416-b499-aae08088b079-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.129812 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8095e4aa-7290-441d-ba67-79e48c23b8a6-operator-scripts\") pod \"8095e4aa-7290-441d-ba67-79e48c23b8a6\" (UID: \"8095e4aa-7290-441d-ba67-79e48c23b8a6\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.130283 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pw9pq\" (UniqueName: \"kubernetes.io/projected/8095e4aa-7290-441d-ba67-79e48c23b8a6-kube-api-access-pw9pq\") pod \"8095e4aa-7290-441d-ba67-79e48c23b8a6\" (UID: \"8095e4aa-7290-441d-ba67-79e48c23b8a6\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.135784 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8095e4aa-7290-441d-ba67-79e48c23b8a6-kube-api-access-pw9pq" (OuterVolumeSpecName: "kube-api-access-pw9pq") pod "8095e4aa-7290-441d-ba67-79e48c23b8a6" (UID: "8095e4aa-7290-441d-ba67-79e48c23b8a6"). InnerVolumeSpecName "kube-api-access-pw9pq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.150275 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8095e4aa-7290-441d-ba67-79e48c23b8a6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8095e4aa-7290-441d-ba67-79e48c23b8a6" (UID: "8095e4aa-7290-441d-ba67-79e48c23b8a6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.197617 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="5a434802-b18e-4d14-8415-4879eaa44191" containerName="galera" containerID="cri-o://3da69fd56ac70e4cd6a7e0c488e7105251f570721cf9c47f92c46924f483efc6" gracePeriod=30 Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.224737 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinderc394-account-delete-gnvr4" event={"ID":"8095e4aa-7290-441d-ba67-79e48c23b8a6","Type":"ContainerDied","Data":"df1bb02e10b32a65fcedd5ae514e011f827429278eb8abd8e03c502f6be3bf01"} Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.224811 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df1bb02e10b32a65fcedd5ae514e011f827429278eb8abd8e03c502f6be3bf01" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.224749 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinderc394-account-delete-gnvr4" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.232401 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pw9pq\" (UniqueName: \"kubernetes.io/projected/8095e4aa-7290-441d-ba67-79e48c23b8a6-kube-api-access-pw9pq\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.232440 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8095e4aa-7290-441d-ba67-79e48c23b8a6-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.242211 4790 generic.go:334] "Generic (PLEG): container finished" podID="8f8fe1f4-7b63-455c-97d1-d3c78826c3c4" containerID="77d68f10a4550d5451c7aeeccb710c9bb4cd97f4057890444850ad09cff25f6b" exitCode=0 Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.242305 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4","Type":"ContainerDied","Data":"77d68f10a4550d5451c7aeeccb710c9bb4cd97f4057890444850ad09cff25f6b"} Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.247172 4790 generic.go:334] "Generic (PLEG): container finished" podID="81757ef9-7330-44bb-a35e-8e6f4c137c0e" containerID="6485471927d5a223e611cbd164262141a1a0f94b13c44e9d6ea257f4db80ed9a" exitCode=0 Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.247232 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"81757ef9-7330-44bb-a35e-8e6f4c137c0e","Type":"ContainerDied","Data":"6485471927d5a223e611cbd164262141a1a0f94b13c44e9d6ea257f4db80ed9a"} Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.248597 4790 generic.go:334] "Generic (PLEG): container finished" podID="3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c" containerID="32dece0aef9f5c5f700993f3c33e9dc9b09465378d0d7abd138bd90e7d5c6319" exitCode=0 Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.248638 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-845f9d57f-2s48z" event={"ID":"3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c","Type":"ContainerDied","Data":"32dece0aef9f5c5f700993f3c33e9dc9b09465378d0d7abd138bd90e7d5c6319"} Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.250358 4790 generic.go:334] "Generic (PLEG): container finished" podID="8cbd6e74-a758-4bbf-81eb-acffdfa56955" containerID="8b87dbc9b7ac42a8486a73feacf6d51b270b4dd6af8a81539a4ea922eb8a0e1b" exitCode=0 Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.250410 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8cbd6e74-a758-4bbf-81eb-acffdfa56955","Type":"ContainerDied","Data":"8b87dbc9b7ac42a8486a73feacf6d51b270b4dd6af8a81539a4ea922eb8a0e1b"} Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.267079 4790 generic.go:334] "Generic (PLEG): container finished" podID="a2ee8af7-26fb-4652-be37-594db62f1146" containerID="ef998419eb802553e1ad108205b17dff1eada33d519b61e4493aea07e0efefc0" exitCode=0 Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.267120 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-58b5bcbcb-b6kbh" event={"ID":"a2ee8af7-26fb-4652-be37-594db62f1146","Type":"ContainerDied","Data":"ef998419eb802553e1ad108205b17dff1eada33d519b61e4493aea07e0efefc0"} Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.274757 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e6d696a7-d618-4416-b499-aae08088b079","Type":"ContainerDied","Data":"2ce5dd9abe30c2b33028a60ed6174141ddfe8cc25750d7d125c466efb0a51375"} Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.274844 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.301807 4790 generic.go:334] "Generic (PLEG): container finished" podID="7fdd5969-f22b-41e6-a0d9-f3eeb16689fd" containerID="d8eab96ea490fbfa774abdd6a272fffab4b7cacb733249bf3f282fa0dfcf8521" exitCode=0 Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.303976 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"7fdd5969-f22b-41e6-a0d9-f3eeb16689fd","Type":"ContainerDied","Data":"d8eab96ea490fbfa774abdd6a272fffab4b7cacb733249bf3f282fa0dfcf8521"} Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.331717 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bh9qg" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.362243 4790 generic.go:334] "Generic (PLEG): container finished" podID="39d594c5-aae1-4cd7-a7de-9de435ad83b8" containerID="eb8a4e1ecddd7fb54e90dd1a525962762b3153943ebd954bb7544913710d2ec2" exitCode=0 Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.362307 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"39d594c5-aae1-4cd7-a7de-9de435ad83b8","Type":"ContainerDied","Data":"eb8a4e1ecddd7fb54e90dd1a525962762b3153943ebd954bb7544913710d2ec2"} Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.447989 4790 scope.go:117] "RemoveContainer" containerID="3898a01261d357b913fd416fce34429e84d9cd9629907d6bc5716e0919fddbbd" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.471027 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.473461 4790 generic.go:334] "Generic (PLEG): container finished" podID="3fc4314a-6a31-49d1-8652-f4de6da97916" containerID="3bdd8dbb30e64b7458fddef45148be02e9280fe559bf6018d442a5bf9eb8dada" exitCode=0 Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.473493 4790 generic.go:334] "Generic (PLEG): container finished" podID="3fc4314a-6a31-49d1-8652-f4de6da97916" containerID="b5e492786f34b12c4fccd95cb12360cbb4386005a0414f30c8e1be783d5252e0" exitCode=2 Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.473500 4790 generic.go:334] "Generic (PLEG): container finished" podID="3fc4314a-6a31-49d1-8652-f4de6da97916" containerID="593dfb8aab769d099082292784d20c9dd7fc610d3ac9c0659d6fd1635968dfa7" exitCode=0 Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.473506 4790 generic.go:334] "Generic (PLEG): container finished" podID="3fc4314a-6a31-49d1-8652-f4de6da97916" containerID="63bc861e59ed27e8b2dec969395af49861e48d187c85af757c5022b25349826a" exitCode=0 Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.473539 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3fc4314a-6a31-49d1-8652-f4de6da97916","Type":"ContainerDied","Data":"3bdd8dbb30e64b7458fddef45148be02e9280fe559bf6018d442a5bf9eb8dada"} Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.473569 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3fc4314a-6a31-49d1-8652-f4de6da97916","Type":"ContainerDied","Data":"b5e492786f34b12c4fccd95cb12360cbb4386005a0414f30c8e1be783d5252e0"} Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.473583 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3fc4314a-6a31-49d1-8652-f4de6da97916","Type":"ContainerDied","Data":"593dfb8aab769d099082292784d20c9dd7fc610d3ac9c0659d6fd1635968dfa7"} Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.473593 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3fc4314a-6a31-49d1-8652-f4de6da97916","Type":"ContainerDied","Data":"63bc861e59ed27e8b2dec969395af49861e48d187c85af757c5022b25349826a"} Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.480494 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.504451 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.505924 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-7957b6846-f8n7f"] Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.521146 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.526032 4790 scope.go:117] "RemoveContainer" containerID="2893ed9292031995f05a0b830dc78e77b0e8abbf5b78c8c24b7cdaa2e9f045de" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.533270 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-7957b6846-f8n7f"] Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.540163 4790 generic.go:334] "Generic (PLEG): container finished" podID="8311dfa1-6e87-493d-92e0-4eb69c792afe" containerID="d07411834060975e29533a55785dd1993dfe65d5e3ce92d8eb62b08cba9dc0d1" exitCode=0 Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.540252 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7d69688466-r2bq9" event={"ID":"8311dfa1-6e87-493d-92e0-4eb69c792afe","Type":"ContainerDied","Data":"d07411834060975e29533a55785dd1993dfe65d5e3ce92d8eb62b08cba9dc0d1"} Nov 24 13:37:23 crc kubenswrapper[4790]: E1124 13:37:23.540350 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2893ed9292031995f05a0b830dc78e77b0e8abbf5b78c8c24b7cdaa2e9f045de\": container with ID starting with 2893ed9292031995f05a0b830dc78e77b0e8abbf5b78c8c24b7cdaa2e9f045de not found: ID does not exist" containerID="2893ed9292031995f05a0b830dc78e77b0e8abbf5b78c8c24b7cdaa2e9f045de" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.540370 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2893ed9292031995f05a0b830dc78e77b0e8abbf5b78c8c24b7cdaa2e9f045de"} err="failed to get container status \"2893ed9292031995f05a0b830dc78e77b0e8abbf5b78c8c24b7cdaa2e9f045de\": rpc error: code = NotFound desc = could not find container \"2893ed9292031995f05a0b830dc78e77b0e8abbf5b78c8c24b7cdaa2e9f045de\": container with ID starting with 2893ed9292031995f05a0b830dc78e77b0e8abbf5b78c8c24b7cdaa2e9f045de not found: ID does not exist" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.540388 4790 scope.go:117] "RemoveContainer" containerID="3898a01261d357b913fd416fce34429e84d9cd9629907d6bc5716e0919fddbbd" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.541709 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23704906-0f33-4a4e-8bbb-a796a9a1733f-combined-ca-bundle\") pod \"23704906-0f33-4a4e-8bbb-a796a9a1733f\" (UID: \"23704906-0f33-4a4e-8bbb-a796a9a1733f\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.541741 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6bbx\" (UniqueName: \"kubernetes.io/projected/23704906-0f33-4a4e-8bbb-a796a9a1733f-kube-api-access-t6bbx\") pod \"23704906-0f33-4a4e-8bbb-a796a9a1733f\" (UID: \"23704906-0f33-4a4e-8bbb-a796a9a1733f\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.541806 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/23704906-0f33-4a4e-8bbb-a796a9a1733f-kube-state-metrics-tls-config\") pod \"23704906-0f33-4a4e-8bbb-a796a9a1733f\" (UID: \"23704906-0f33-4a4e-8bbb-a796a9a1733f\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.541875 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/23704906-0f33-4a4e-8bbb-a796a9a1733f-kube-state-metrics-tls-certs\") pod \"23704906-0f33-4a4e-8bbb-a796a9a1733f\" (UID: \"23704906-0f33-4a4e-8bbb-a796a9a1733f\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.546420 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bh9qg"] Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.550762 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23704906-0f33-4a4e-8bbb-a796a9a1733f-kube-api-access-t6bbx" (OuterVolumeSpecName: "kube-api-access-t6bbx") pod "23704906-0f33-4a4e-8bbb-a796a9a1733f" (UID: "23704906-0f33-4a4e-8bbb-a796a9a1733f"). InnerVolumeSpecName "kube-api-access-t6bbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:23 crc kubenswrapper[4790]: E1124 13:37:23.551426 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3898a01261d357b913fd416fce34429e84d9cd9629907d6bc5716e0919fddbbd\": container with ID starting with 3898a01261d357b913fd416fce34429e84d9cd9629907d6bc5716e0919fddbbd not found: ID does not exist" containerID="3898a01261d357b913fd416fce34429e84d9cd9629907d6bc5716e0919fddbbd" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.551472 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3898a01261d357b913fd416fce34429e84d9cd9629907d6bc5716e0919fddbbd"} err="failed to get container status \"3898a01261d357b913fd416fce34429e84d9cd9629907d6bc5716e0919fddbbd\": rpc error: code = NotFound desc = could not find container \"3898a01261d357b913fd416fce34429e84d9cd9629907d6bc5716e0919fddbbd\": container with ID starting with 3898a01261d357b913fd416fce34429e84d9cd9629907d6bc5716e0919fddbbd not found: ID does not exist" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.551505 4790 scope.go:117] "RemoveContainer" containerID="2f816c8e5f36c0c446e29e140b1f74aeb5df0880612029daa8ef51dfaee75bee" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.556185 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bh9qg"] Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.560170 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7d69688466-r2bq9" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.560292 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.560206 4790 generic.go:334] "Generic (PLEG): container finished" podID="23704906-0f33-4a4e-8bbb-a796a9a1733f" containerID="7eb4e19d335efc1382a2f1934fe4914e9bce012cc0d8c8dd38049967fe85a33d" exitCode=2 Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.560231 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"23704906-0f33-4a4e-8bbb-a796a9a1733f","Type":"ContainerDied","Data":"7eb4e19d335efc1382a2f1934fe4914e9bce012cc0d8c8dd38049967fe85a33d"} Nov 24 13:37:23 crc kubenswrapper[4790]: E1124 13:37:23.577268 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3 is running failed: container process not found" containerID="e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 13:37:23 crc kubenswrapper[4790]: E1124 13:37:23.578316 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3 is running failed: container process not found" containerID="e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 13:37:23 crc kubenswrapper[4790]: E1124 13:37:23.578720 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3 is running failed: container process not found" containerID="e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 13:37:23 crc kubenswrapper[4790]: E1124 13:37:23.578795 4790 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-28cwj" podUID="4d6d5d88-adda-4679-80de-0df0e0847bcb" containerName="ovsdb-server" Nov 24 13:37:23 crc kubenswrapper[4790]: E1124 13:37:23.580074 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="46182cf796e680922c3d2c23de93f19051dee361d1797948d997345bb2652287" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.580194 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 13:37:23 crc kubenswrapper[4790]: E1124 13:37:23.586532 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="46182cf796e680922c3d2c23de93f19051dee361d1797948d997345bb2652287" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.589124 4790 scope.go:117] "RemoveContainer" containerID="96fb30b3c69ce48bcf01e52ed956c73a35d60273c9db96fac4575e7dc8dd1004" Nov 24 13:37:23 crc kubenswrapper[4790]: E1124 13:37:23.592737 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="46182cf796e680922c3d2c23de93f19051dee361d1797948d997345bb2652287" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 13:37:23 crc kubenswrapper[4790]: E1124 13:37:23.592798 4790 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-28cwj" podUID="4d6d5d88-adda-4679-80de-0df0e0847bcb" containerName="ovs-vswitchd" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.598141 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.605126 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23704906-0f33-4a4e-8bbb-a796a9a1733f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "23704906-0f33-4a4e-8bbb-a796a9a1733f" (UID: "23704906-0f33-4a4e-8bbb-a796a9a1733f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.616725 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23704906-0f33-4a4e-8bbb-a796a9a1733f-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "23704906-0f33-4a4e-8bbb-a796a9a1733f" (UID: "23704906-0f33-4a4e-8bbb-a796a9a1733f"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.645444 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-httpd-run\") pod \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\" (UID: \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.645492 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cbd6e74-a758-4bbf-81eb-acffdfa56955-config-data\") pod \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\" (UID: \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.645539 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xsvp6\" (UniqueName: \"kubernetes.io/projected/8cbd6e74-a758-4bbf-81eb-acffdfa56955-kube-api-access-xsvp6\") pod \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\" (UID: \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.645558 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cbd6e74-a758-4bbf-81eb-acffdfa56955-logs\") pod \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\" (UID: \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.645573 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cbd6e74-a758-4bbf-81eb-acffdfa56955-combined-ca-bundle\") pod \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\" (UID: \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.645639 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cbd6e74-a758-4bbf-81eb-acffdfa56955-public-tls-certs\") pod \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\" (UID: \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.645663 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8311dfa1-6e87-493d-92e0-4eb69c792afe-combined-ca-bundle\") pod \"8311dfa1-6e87-493d-92e0-4eb69c792afe\" (UID: \"8311dfa1-6e87-493d-92e0-4eb69c792afe\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.645686 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8311dfa1-6e87-493d-92e0-4eb69c792afe-public-tls-certs\") pod \"8311dfa1-6e87-493d-92e0-4eb69c792afe\" (UID: \"8311dfa1-6e87-493d-92e0-4eb69c792afe\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.645713 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\" (UID: \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.645734 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h829l\" (UniqueName: \"kubernetes.io/projected/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-kube-api-access-h829l\") pod \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\" (UID: \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.645748 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\" (UID: \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.645770 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-combined-ca-bundle\") pod \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\" (UID: \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.645789 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81757ef9-7330-44bb-a35e-8e6f4c137c0e-config-data\") pod \"81757ef9-7330-44bb-a35e-8e6f4c137c0e\" (UID: \"81757ef9-7330-44bb-a35e-8e6f4c137c0e\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.645812 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dchb\" (UniqueName: \"kubernetes.io/projected/81757ef9-7330-44bb-a35e-8e6f4c137c0e-kube-api-access-4dchb\") pod \"81757ef9-7330-44bb-a35e-8e6f4c137c0e\" (UID: \"81757ef9-7330-44bb-a35e-8e6f4c137c0e\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.645839 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81757ef9-7330-44bb-a35e-8e6f4c137c0e-combined-ca-bundle\") pod \"81757ef9-7330-44bb-a35e-8e6f4c137c0e\" (UID: \"81757ef9-7330-44bb-a35e-8e6f4c137c0e\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.645860 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-scripts\") pod \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\" (UID: \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.645894 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8311dfa1-6e87-493d-92e0-4eb69c792afe-logs\") pod \"8311dfa1-6e87-493d-92e0-4eb69c792afe\" (UID: \"8311dfa1-6e87-493d-92e0-4eb69c792afe\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.645910 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8311dfa1-6e87-493d-92e0-4eb69c792afe-config-data-custom\") pod \"8311dfa1-6e87-493d-92e0-4eb69c792afe\" (UID: \"8311dfa1-6e87-493d-92e0-4eb69c792afe\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.645929 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-internal-tls-certs\") pod \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\" (UID: \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.645964 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81757ef9-7330-44bb-a35e-8e6f4c137c0e-logs\") pod \"81757ef9-7330-44bb-a35e-8e6f4c137c0e\" (UID: \"81757ef9-7330-44bb-a35e-8e6f4c137c0e\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.645991 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8311dfa1-6e87-493d-92e0-4eb69c792afe-config-data\") pod \"8311dfa1-6e87-493d-92e0-4eb69c792afe\" (UID: \"8311dfa1-6e87-493d-92e0-4eb69c792afe\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.646026 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8cbd6e74-a758-4bbf-81eb-acffdfa56955-httpd-run\") pod \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\" (UID: \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.646047 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8311dfa1-6e87-493d-92e0-4eb69c792afe-internal-tls-certs\") pod \"8311dfa1-6e87-493d-92e0-4eb69c792afe\" (UID: \"8311dfa1-6e87-493d-92e0-4eb69c792afe\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.646137 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-logs\") pod \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\" (UID: \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.646214 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-config-data\") pod \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\" (UID: \"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.646252 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nr64t\" (UniqueName: \"kubernetes.io/projected/8311dfa1-6e87-493d-92e0-4eb69c792afe-kube-api-access-nr64t\") pod \"8311dfa1-6e87-493d-92e0-4eb69c792afe\" (UID: \"8311dfa1-6e87-493d-92e0-4eb69c792afe\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.646279 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/81757ef9-7330-44bb-a35e-8e6f4c137c0e-nova-metadata-tls-certs\") pod \"81757ef9-7330-44bb-a35e-8e6f4c137c0e\" (UID: \"81757ef9-7330-44bb-a35e-8e6f4c137c0e\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.646300 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cbd6e74-a758-4bbf-81eb-acffdfa56955-scripts\") pod \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\" (UID: \"8cbd6e74-a758-4bbf-81eb-acffdfa56955\") " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.646960 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23704906-0f33-4a4e-8bbb-a796a9a1733f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.646980 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6bbx\" (UniqueName: \"kubernetes.io/projected/23704906-0f33-4a4e-8bbb-a796a9a1733f-kube-api-access-t6bbx\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.646991 4790 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/23704906-0f33-4a4e-8bbb-a796a9a1733f-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.650798 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8f8fe1f4-7b63-455c-97d1-d3c78826c3c4" (UID: "8f8fe1f4-7b63-455c-97d1-d3c78826c3c4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.650833 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8cbd6e74-a758-4bbf-81eb-acffdfa56955-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8cbd6e74-a758-4bbf-81eb-acffdfa56955" (UID: "8cbd6e74-a758-4bbf-81eb-acffdfa56955"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.651321 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8cbd6e74-a758-4bbf-81eb-acffdfa56955-logs" (OuterVolumeSpecName: "logs") pod "8cbd6e74-a758-4bbf-81eb-acffdfa56955" (UID: "8cbd6e74-a758-4bbf-81eb-acffdfa56955"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.654292 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8311dfa1-6e87-493d-92e0-4eb69c792afe-logs" (OuterVolumeSpecName: "logs") pod "8311dfa1-6e87-493d-92e0-4eb69c792afe" (UID: "8311dfa1-6e87-493d-92e0-4eb69c792afe"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.653787 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81757ef9-7330-44bb-a35e-8e6f4c137c0e-kube-api-access-4dchb" (OuterVolumeSpecName: "kube-api-access-4dchb") pod "81757ef9-7330-44bb-a35e-8e6f4c137c0e" (UID: "81757ef9-7330-44bb-a35e-8e6f4c137c0e"). InnerVolumeSpecName "kube-api-access-4dchb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.655177 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cbd6e74-a758-4bbf-81eb-acffdfa56955-scripts" (OuterVolumeSpecName: "scripts") pod "8cbd6e74-a758-4bbf-81eb-acffdfa56955" (UID: "8cbd6e74-a758-4bbf-81eb-acffdfa56955"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.655169 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23704906-0f33-4a4e-8bbb-a796a9a1733f-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "23704906-0f33-4a4e-8bbb-a796a9a1733f" (UID: "23704906-0f33-4a4e-8bbb-a796a9a1733f"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.657109 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-logs" (OuterVolumeSpecName: "logs") pod "8f8fe1f4-7b63-455c-97d1-d3c78826c3c4" (UID: "8f8fe1f4-7b63-455c-97d1-d3c78826c3c4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.660895 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81757ef9-7330-44bb-a35e-8e6f4c137c0e-logs" (OuterVolumeSpecName: "logs") pod "81757ef9-7330-44bb-a35e-8e6f4c137c0e" (UID: "81757ef9-7330-44bb-a35e-8e6f4c137c0e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.666379 4790 scope.go:117] "RemoveContainer" containerID="2f816c8e5f36c0c446e29e140b1f74aeb5df0880612029daa8ef51dfaee75bee" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.669626 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cbd6e74-a758-4bbf-81eb-acffdfa56955-kube-api-access-xsvp6" (OuterVolumeSpecName: "kube-api-access-xsvp6") pod "8cbd6e74-a758-4bbf-81eb-acffdfa56955" (UID: "8cbd6e74-a758-4bbf-81eb-acffdfa56955"). InnerVolumeSpecName "kube-api-access-xsvp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:23 crc kubenswrapper[4790]: E1124 13:37:23.670143 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f816c8e5f36c0c446e29e140b1f74aeb5df0880612029daa8ef51dfaee75bee\": container with ID starting with 2f816c8e5f36c0c446e29e140b1f74aeb5df0880612029daa8ef51dfaee75bee not found: ID does not exist" containerID="2f816c8e5f36c0c446e29e140b1f74aeb5df0880612029daa8ef51dfaee75bee" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.670179 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f816c8e5f36c0c446e29e140b1f74aeb5df0880612029daa8ef51dfaee75bee"} err="failed to get container status \"2f816c8e5f36c0c446e29e140b1f74aeb5df0880612029daa8ef51dfaee75bee\": rpc error: code = NotFound desc = could not find container \"2f816c8e5f36c0c446e29e140b1f74aeb5df0880612029daa8ef51dfaee75bee\": container with ID starting with 2f816c8e5f36c0c446e29e140b1f74aeb5df0880612029daa8ef51dfaee75bee not found: ID does not exist" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.670203 4790 scope.go:117] "RemoveContainer" containerID="96fb30b3c69ce48bcf01e52ed956c73a35d60273c9db96fac4575e7dc8dd1004" Nov 24 13:37:23 crc kubenswrapper[4790]: E1124 13:37:23.674048 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96fb30b3c69ce48bcf01e52ed956c73a35d60273c9db96fac4575e7dc8dd1004\": container with ID starting with 96fb30b3c69ce48bcf01e52ed956c73a35d60273c9db96fac4575e7dc8dd1004 not found: ID does not exist" containerID="96fb30b3c69ce48bcf01e52ed956c73a35d60273c9db96fac4575e7dc8dd1004" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.674091 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96fb30b3c69ce48bcf01e52ed956c73a35d60273c9db96fac4575e7dc8dd1004"} err="failed to get container status \"96fb30b3c69ce48bcf01e52ed956c73a35d60273c9db96fac4575e7dc8dd1004\": rpc error: code = NotFound desc = could not find container \"96fb30b3c69ce48bcf01e52ed956c73a35d60273c9db96fac4575e7dc8dd1004\": container with ID starting with 96fb30b3c69ce48bcf01e52ed956c73a35d60273c9db96fac4575e7dc8dd1004 not found: ID does not exist" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.674115 4790 scope.go:117] "RemoveContainer" containerID="6b8183ab06a7a4bd08c674c5788831aa9879a328f22b998c507e78d8d453b26d" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.684322 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "8f8fe1f4-7b63-455c-97d1-d3c78826c3c4" (UID: "8f8fe1f4-7b63-455c-97d1-d3c78826c3c4"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.697839 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-kube-api-access-h829l" (OuterVolumeSpecName: "kube-api-access-h829l") pod "8f8fe1f4-7b63-455c-97d1-d3c78826c3c4" (UID: "8f8fe1f4-7b63-455c-97d1-d3c78826c3c4"). InnerVolumeSpecName "kube-api-access-h829l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.698285 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8311dfa1-6e87-493d-92e0-4eb69c792afe-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8311dfa1-6e87-493d-92e0-4eb69c792afe" (UID: "8311dfa1-6e87-493d-92e0-4eb69c792afe"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.698765 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8311dfa1-6e87-493d-92e0-4eb69c792afe-kube-api-access-nr64t" (OuterVolumeSpecName: "kube-api-access-nr64t") pod "8311dfa1-6e87-493d-92e0-4eb69c792afe" (UID: "8311dfa1-6e87-493d-92e0-4eb69c792afe"). InnerVolumeSpecName "kube-api-access-nr64t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.703477 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-scripts" (OuterVolumeSpecName: "scripts") pod "8f8fe1f4-7b63-455c-97d1-d3c78826c3c4" (UID: "8f8fe1f4-7b63-455c-97d1-d3c78826c3c4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:23 crc kubenswrapper[4790]: E1124 13:37:23.715801 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d8eab96ea490fbfa774abdd6a272fffab4b7cacb733249bf3f282fa0dfcf8521 is running failed: container process not found" containerID="d8eab96ea490fbfa774abdd6a272fffab4b7cacb733249bf3f282fa0dfcf8521" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 24 13:37:23 crc kubenswrapper[4790]: E1124 13:37:23.717104 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d8eab96ea490fbfa774abdd6a272fffab4b7cacb733249bf3f282fa0dfcf8521 is running failed: container process not found" containerID="d8eab96ea490fbfa774abdd6a272fffab4b7cacb733249bf3f282fa0dfcf8521" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 24 13:37:23 crc kubenswrapper[4790]: E1124 13:37:23.719122 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d8eab96ea490fbfa774abdd6a272fffab4b7cacb733249bf3f282fa0dfcf8521 is running failed: container process not found" containerID="d8eab96ea490fbfa774abdd6a272fffab4b7cacb733249bf3f282fa0dfcf8521" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 24 13:37:23 crc kubenswrapper[4790]: E1124 13:37:23.719242 4790 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d8eab96ea490fbfa774abdd6a272fffab4b7cacb733249bf3f282fa0dfcf8521 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="7fdd5969-f22b-41e6-a0d9-f3eeb16689fd" containerName="nova-cell0-conductor-conductor" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.720084 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "8cbd6e74-a758-4bbf-81eb-acffdfa56955" (UID: "8cbd6e74-a758-4bbf-81eb-acffdfa56955"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.730771 4790 scope.go:117] "RemoveContainer" containerID="6b8183ab06a7a4bd08c674c5788831aa9879a328f22b998c507e78d8d453b26d" Nov 24 13:37:23 crc kubenswrapper[4790]: E1124 13:37:23.734810 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b8183ab06a7a4bd08c674c5788831aa9879a328f22b998c507e78d8d453b26d\": container with ID starting with 6b8183ab06a7a4bd08c674c5788831aa9879a328f22b998c507e78d8d453b26d not found: ID does not exist" containerID="6b8183ab06a7a4bd08c674c5788831aa9879a328f22b998c507e78d8d453b26d" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.735012 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b8183ab06a7a4bd08c674c5788831aa9879a328f22b998c507e78d8d453b26d"} err="failed to get container status \"6b8183ab06a7a4bd08c674c5788831aa9879a328f22b998c507e78d8d453b26d\": rpc error: code = NotFound desc = could not find container \"6b8183ab06a7a4bd08c674c5788831aa9879a328f22b998c507e78d8d453b26d\": container with ID starting with 6b8183ab06a7a4bd08c674c5788831aa9879a328f22b998c507e78d8d453b26d not found: ID does not exist" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.735134 4790 scope.go:117] "RemoveContainer" containerID="002e858954a60800481c9819989042b1ad82aae0f1762f13550da43a5f262b74" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.746652 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cbd6e74-a758-4bbf-81eb-acffdfa56955-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8cbd6e74-a758-4bbf-81eb-acffdfa56955" (UID: "8cbd6e74-a758-4bbf-81eb-acffdfa56955"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.748298 4790 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/23704906-0f33-4a4e-8bbb-a796a9a1733f-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.748338 4790 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.748349 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h829l\" (UniqueName: \"kubernetes.io/projected/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-kube-api-access-h829l\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.748365 4790 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.748377 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dchb\" (UniqueName: \"kubernetes.io/projected/81757ef9-7330-44bb-a35e-8e6f4c137c0e-kube-api-access-4dchb\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.759303 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.759317 4790 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8311dfa1-6e87-493d-92e0-4eb69c792afe-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.759331 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8311dfa1-6e87-493d-92e0-4eb69c792afe-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.759340 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81757ef9-7330-44bb-a35e-8e6f4c137c0e-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.759349 4790 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8cbd6e74-a758-4bbf-81eb-acffdfa56955-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.759359 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.759367 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nr64t\" (UniqueName: \"kubernetes.io/projected/8311dfa1-6e87-493d-92e0-4eb69c792afe-kube-api-access-nr64t\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.759376 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cbd6e74-a758-4bbf-81eb-acffdfa56955-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.759383 4790 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.759395 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xsvp6\" (UniqueName: \"kubernetes.io/projected/8cbd6e74-a758-4bbf-81eb-acffdfa56955-kube-api-access-xsvp6\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.759403 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cbd6e74-a758-4bbf-81eb-acffdfa56955-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.759411 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cbd6e74-a758-4bbf-81eb-acffdfa56955-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.823728 4790 scope.go:117] "RemoveContainer" containerID="7d8bdf173f30451f767a3b85fb31741d1552f46cecc70197217878b14326622c" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.860137 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8f8fe1f4-7b63-455c-97d1-d3c78826c3c4" (UID: "8f8fe1f4-7b63-455c-97d1-d3c78826c3c4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.861331 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.862328 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81757ef9-7330-44bb-a35e-8e6f4c137c0e-config-data" (OuterVolumeSpecName: "config-data") pod "81757ef9-7330-44bb-a35e-8e6f4c137c0e" (UID: "81757ef9-7330-44bb-a35e-8e6f4c137c0e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.884163 4790 scope.go:117] "RemoveContainer" containerID="7eb4e19d335efc1382a2f1934fe4914e9bce012cc0d8c8dd38049967fe85a33d" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.938972 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.939422 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-58b5bcbcb-b6kbh" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.940649 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.960864 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.961583 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "8f8fe1f4-7b63-455c-97d1-d3c78826c3c4" (UID: "8f8fe1f4-7b63-455c-97d1-d3c78826c3c4"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.963143 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81757ef9-7330-44bb-a35e-8e6f4c137c0e-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:23 crc kubenswrapper[4790]: I1124 13:37:23.963328 4790 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.066340 4790 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.072672 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a2ee8af7-26fb-4652-be37-594db62f1146-config-data-custom\") pod \"a2ee8af7-26fb-4652-be37-594db62f1146\" (UID: \"a2ee8af7-26fb-4652-be37-594db62f1146\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.072794 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39d594c5-aae1-4cd7-a7de-9de435ad83b8-config-data\") pod \"39d594c5-aae1-4cd7-a7de-9de435ad83b8\" (UID: \"39d594c5-aae1-4cd7-a7de-9de435ad83b8\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.072832 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2ee8af7-26fb-4652-be37-594db62f1146-config-data\") pod \"a2ee8af7-26fb-4652-be37-594db62f1146\" (UID: \"a2ee8af7-26fb-4652-be37-594db62f1146\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.072854 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39d594c5-aae1-4cd7-a7de-9de435ad83b8-combined-ca-bundle\") pod \"39d594c5-aae1-4cd7-a7de-9de435ad83b8\" (UID: \"39d594c5-aae1-4cd7-a7de-9de435ad83b8\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.072902 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqgn6\" (UniqueName: \"kubernetes.io/projected/a2ee8af7-26fb-4652-be37-594db62f1146-kube-api-access-hqgn6\") pod \"a2ee8af7-26fb-4652-be37-594db62f1146\" (UID: \"a2ee8af7-26fb-4652-be37-594db62f1146\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.073004 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p524t\" (UniqueName: \"kubernetes.io/projected/39d594c5-aae1-4cd7-a7de-9de435ad83b8-kube-api-access-p524t\") pod \"39d594c5-aae1-4cd7-a7de-9de435ad83b8\" (UID: \"39d594c5-aae1-4cd7-a7de-9de435ad83b8\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.073067 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2ee8af7-26fb-4652-be37-594db62f1146-combined-ca-bundle\") pod \"a2ee8af7-26fb-4652-be37-594db62f1146\" (UID: \"a2ee8af7-26fb-4652-be37-594db62f1146\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.073254 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2ee8af7-26fb-4652-be37-594db62f1146-logs\") pod \"a2ee8af7-26fb-4652-be37-594db62f1146\" (UID: \"a2ee8af7-26fb-4652-be37-594db62f1146\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.074147 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8311dfa1-6e87-493d-92e0-4eb69c792afe-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "8311dfa1-6e87-493d-92e0-4eb69c792afe" (UID: "8311dfa1-6e87-493d-92e0-4eb69c792afe"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.075129 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8311dfa1-6e87-493d-92e0-4eb69c792afe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8311dfa1-6e87-493d-92e0-4eb69c792afe" (UID: "8311dfa1-6e87-493d-92e0-4eb69c792afe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.078289 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39d594c5-aae1-4cd7-a7de-9de435ad83b8-kube-api-access-p524t" (OuterVolumeSpecName: "kube-api-access-p524t") pod "39d594c5-aae1-4cd7-a7de-9de435ad83b8" (UID: "39d594c5-aae1-4cd7-a7de-9de435ad83b8"). InnerVolumeSpecName "kube-api-access-p524t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.079017 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2ee8af7-26fb-4652-be37-594db62f1146-logs" (OuterVolumeSpecName: "logs") pod "a2ee8af7-26fb-4652-be37-594db62f1146" (UID: "a2ee8af7-26fb-4652-be37-594db62f1146"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.086172 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8311dfa1-6e87-493d-92e0-4eb69c792afe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.086202 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p524t\" (UniqueName: \"kubernetes.io/projected/39d594c5-aae1-4cd7-a7de-9de435ad83b8-kube-api-access-p524t\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.086215 4790 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.086224 4790 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8311dfa1-6e87-493d-92e0-4eb69c792afe-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.086233 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2ee8af7-26fb-4652-be37-594db62f1146-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.091356 4790 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.097334 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2ee8af7-26fb-4652-be37-594db62f1146-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a2ee8af7-26fb-4652-be37-594db62f1146" (UID: "a2ee8af7-26fb-4652-be37-594db62f1146"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.105261 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2ee8af7-26fb-4652-be37-594db62f1146-kube-api-access-hqgn6" (OuterVolumeSpecName: "kube-api-access-hqgn6") pod "a2ee8af7-26fb-4652-be37-594db62f1146" (UID: "a2ee8af7-26fb-4652-be37-594db62f1146"). InnerVolumeSpecName "kube-api-access-hqgn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.123843 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81757ef9-7330-44bb-a35e-8e6f4c137c0e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "81757ef9-7330-44bb-a35e-8e6f4c137c0e" (UID: "81757ef9-7330-44bb-a35e-8e6f4c137c0e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.124945 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-config-data" (OuterVolumeSpecName: "config-data") pod "8f8fe1f4-7b63-455c-97d1-d3c78826c3c4" (UID: "8f8fe1f4-7b63-455c-97d1-d3c78826c3c4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.132721 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8311dfa1-6e87-493d-92e0-4eb69c792afe-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8311dfa1-6e87-493d-92e0-4eb69c792afe" (UID: "8311dfa1-6e87-493d-92e0-4eb69c792afe"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.152428 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8311dfa1-6e87-493d-92e0-4eb69c792afe-config-data" (OuterVolumeSpecName: "config-data") pod "8311dfa1-6e87-493d-92e0-4eb69c792afe" (UID: "8311dfa1-6e87-493d-92e0-4eb69c792afe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.153152 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81757ef9-7330-44bb-a35e-8e6f4c137c0e-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "81757ef9-7330-44bb-a35e-8e6f4c137c0e" (UID: "81757ef9-7330-44bb-a35e-8e6f4c137c0e"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.157305 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-cd5cbd7b9-48rv7" podUID="c428221a-e989-4844-894c-de96782508a3" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.196:5353: i/o timeout" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.159851 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2ee8af7-26fb-4652-be37-594db62f1146-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a2ee8af7-26fb-4652-be37-594db62f1146" (UID: "a2ee8af7-26fb-4652-be37-594db62f1146"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.167581 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cbd6e74-a758-4bbf-81eb-acffdfa56955-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8cbd6e74-a758-4bbf-81eb-acffdfa56955" (UID: "8cbd6e74-a758-4bbf-81eb-acffdfa56955"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.187349 4790 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a2ee8af7-26fb-4652-be37-594db62f1146-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.187382 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqgn6\" (UniqueName: \"kubernetes.io/projected/a2ee8af7-26fb-4652-be37-594db62f1146-kube-api-access-hqgn6\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.187394 4790 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cbd6e74-a758-4bbf-81eb-acffdfa56955-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.187403 4790 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8311dfa1-6e87-493d-92e0-4eb69c792afe-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.187411 4790 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.187420 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2ee8af7-26fb-4652-be37-594db62f1146-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.187428 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81757ef9-7330-44bb-a35e-8e6f4c137c0e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.187436 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8311dfa1-6e87-493d-92e0-4eb69c792afe-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.187444 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.187452 4790 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/81757ef9-7330-44bb-a35e-8e6f4c137c0e-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.199667 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39d594c5-aae1-4cd7-a7de-9de435ad83b8-config-data" (OuterVolumeSpecName: "config-data") pod "39d594c5-aae1-4cd7-a7de-9de435ad83b8" (UID: "39d594c5-aae1-4cd7-a7de-9de435ad83b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.201221 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cbd6e74-a758-4bbf-81eb-acffdfa56955-config-data" (OuterVolumeSpecName: "config-data") pod "8cbd6e74-a758-4bbf-81eb-acffdfa56955" (UID: "8cbd6e74-a758-4bbf-81eb-acffdfa56955"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.206532 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2ee8af7-26fb-4652-be37-594db62f1146-config-data" (OuterVolumeSpecName: "config-data") pod "a2ee8af7-26fb-4652-be37-594db62f1146" (UID: "a2ee8af7-26fb-4652-be37-594db62f1146"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.207039 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39d594c5-aae1-4cd7-a7de-9de435ad83b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "39d594c5-aae1-4cd7-a7de-9de435ad83b8" (UID: "39d594c5-aae1-4cd7-a7de-9de435ad83b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.209175 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-78cd478f4f-bqm94" podUID="4d4e45de-9768-40e9-9c75-6abba8b38559" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.151:9696/\": dial tcp 10.217.0.151:9696: connect: connection refused" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.289842 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cbd6e74-a758-4bbf-81eb-acffdfa56955-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.290055 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39d594c5-aae1-4cd7-a7de-9de435ad83b8-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.290113 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2ee8af7-26fb-4652-be37-594db62f1146-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.290166 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39d594c5-aae1-4cd7-a7de-9de435ad83b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.356113 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b507c50-0776-4d50-ad41-10f26f25fd5f" path="/var/lib/kubelet/pods/1b507c50-0776-4d50-ad41-10f26f25fd5f/volumes" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.356950 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23704906-0f33-4a4e-8bbb-a796a9a1733f" path="/var/lib/kubelet/pods/23704906-0f33-4a4e-8bbb-a796a9a1733f/volumes" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.357501 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="294bfffa-4d3e-43ff-b86c-26cc00b81b19" path="/var/lib/kubelet/pods/294bfffa-4d3e-43ff-b86c-26cc00b81b19/volumes" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.360682 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36466fc4-4db9-473c-9372-dce30dbac728" path="/var/lib/kubelet/pods/36466fc4-4db9-473c-9372-dce30dbac728/volumes" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.361202 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85671e80-3009-49dc-a308-e126b08847a2" path="/var/lib/kubelet/pods/85671e80-3009-49dc-a308-e126b08847a2/volumes" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.361726 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a30bb00-53b7-4128-954f-d5183fec7675" path="/var/lib/kubelet/pods/8a30bb00-53b7-4128-954f-d5183fec7675/volumes" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.362797 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9362fee7-a172-4948-8721-b4c83bf04a30" path="/var/lib/kubelet/pods/9362fee7-a172-4948-8721-b4c83bf04a30/volumes" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.363396 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97cce111-7b68-4ff8-bafe-e76b7f0da75a" path="/var/lib/kubelet/pods/97cce111-7b68-4ff8-bafe-e76b7f0da75a/volumes" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.363848 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cef4bc7-1efa-4f25-a73e-9cff140e0e09" path="/var/lib/kubelet/pods/9cef4bc7-1efa-4f25-a73e-9cff140e0e09/volumes" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.364366 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad24288a-7c41-4e73-b1f3-c403b8646ff7" path="/var/lib/kubelet/pods/ad24288a-7c41-4e73-b1f3-c403b8646ff7/volumes" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.365302 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb72fd7a-89c4-4a8a-a7b0-c758322a08b7" path="/var/lib/kubelet/pods/bb72fd7a-89c4-4a8a-a7b0-c758322a08b7/volumes" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.366049 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db12b3e5-072f-4b17-8ffe-f2861edc47e8" path="/var/lib/kubelet/pods/db12b3e5-072f-4b17-8ffe-f2861edc47e8/volumes" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.366497 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de6f5df1-eb57-493b-9d89-5e710c6f4304" path="/var/lib/kubelet/pods/de6f5df1-eb57-493b-9d89-5e710c6f4304/volumes" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.378965 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6d696a7-d618-4416-b499-aae08088b079" path="/var/lib/kubelet/pods/e6d696a7-d618-4416-b499-aae08088b079/volumes" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.399987 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.403280 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-845f9d57f-2s48z" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.499293 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.530122 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbicana018-account-delete-6z4kv" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.535136 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance388c-account-delete-8hfh2" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.545246 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutronf6b9-account-delete-tdltj" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.553517 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell00c43-account-delete-dlpjf" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.574954 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapid63e-account-delete-k96dh" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.588943 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapid63e-account-delete-k96dh" event={"ID":"fef75227-ffd5-4c97-87c7-e05fe899eb25","Type":"ContainerDied","Data":"a78c05c8e90f63d4b7ef543c7ffad290d2b6b1129e5497a3a0655c1020a34bb9"} Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.588980 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a78c05c8e90f63d4b7ef543c7ffad290d2b6b1129e5497a3a0655c1020a34bb9" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.589050 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapid63e-account-delete-k96dh" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.593610 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell00c43-account-delete-dlpjf" event={"ID":"dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008","Type":"ContainerDied","Data":"afe769ca639c4456baf09b664f809cfb495a91771e552b6a23e357866b90aa8a"} Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.593653 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afe769ca639c4456baf09b664f809cfb495a91771e552b6a23e357866b90aa8a" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.593654 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell00c43-account-delete-dlpjf" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.595951 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement0fc6-account-delete-rl7sl" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.598251 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fc4314a-6a31-49d1-8652-f4de6da97916-ceilometer-tls-certs\") pod \"3fc4314a-6a31-49d1-8652-f4de6da97916\" (UID: \"3fc4314a-6a31-49d1-8652-f4de6da97916\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.598294 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c-config-data\") pod \"3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c\" (UID: \"3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.598318 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcxrv\" (UniqueName: \"kubernetes.io/projected/51552477-67be-4c73-ae48-aa131945caea-kube-api-access-hcxrv\") pod \"51552477-67be-4c73-ae48-aa131945caea\" (UID: \"51552477-67be-4c73-ae48-aa131945caea\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.598336 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fc4314a-6a31-49d1-8652-f4de6da97916-combined-ca-bundle\") pod \"3fc4314a-6a31-49d1-8652-f4de6da97916\" (UID: \"3fc4314a-6a31-49d1-8652-f4de6da97916\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.598352 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3fc4314a-6a31-49d1-8652-f4de6da97916-log-httpd\") pod \"3fc4314a-6a31-49d1-8652-f4de6da97916\" (UID: \"3fc4314a-6a31-49d1-8652-f4de6da97916\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.598369 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51552477-67be-4c73-ae48-aa131945caea-operator-scripts\") pod \"51552477-67be-4c73-ae48-aa131945caea\" (UID: \"51552477-67be-4c73-ae48-aa131945caea\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.598390 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3fc4314a-6a31-49d1-8652-f4de6da97916-scripts\") pod \"3fc4314a-6a31-49d1-8652-f4de6da97916\" (UID: \"3fc4314a-6a31-49d1-8652-f4de6da97916\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.598413 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkrrs\" (UniqueName: \"kubernetes.io/projected/3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c-kube-api-access-fkrrs\") pod \"3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c\" (UID: \"3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.598439 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c-combined-ca-bundle\") pod \"3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c\" (UID: \"3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.598455 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3fc4314a-6a31-49d1-8652-f4de6da97916-sg-core-conf-yaml\") pod \"3fc4314a-6a31-49d1-8652-f4de6da97916\" (UID: \"3fc4314a-6a31-49d1-8652-f4de6da97916\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.598473 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f76f637a-192d-4827-92b3-68e0fc348629-operator-scripts\") pod \"f76f637a-192d-4827-92b3-68e0fc348629\" (UID: \"f76f637a-192d-4827-92b3-68e0fc348629\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.598497 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szr6v\" (UniqueName: \"kubernetes.io/projected/fef75227-ffd5-4c97-87c7-e05fe899eb25-kube-api-access-szr6v\") pod \"fef75227-ffd5-4c97-87c7-e05fe899eb25\" (UID: \"fef75227-ffd5-4c97-87c7-e05fe899eb25\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.598523 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbb22\" (UniqueName: \"kubernetes.io/projected/dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008-kube-api-access-mbb22\") pod \"dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008\" (UID: \"dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.598545 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fef75227-ffd5-4c97-87c7-e05fe899eb25-operator-scripts\") pod \"fef75227-ffd5-4c97-87c7-e05fe899eb25\" (UID: \"fef75227-ffd5-4c97-87c7-e05fe899eb25\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.598604 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c-logs\") pod \"3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c\" (UID: \"3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.598624 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fc4314a-6a31-49d1-8652-f4de6da97916-config-data\") pod \"3fc4314a-6a31-49d1-8652-f4de6da97916\" (UID: \"3fc4314a-6a31-49d1-8652-f4de6da97916\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.598643 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3fc4314a-6a31-49d1-8652-f4de6da97916-run-httpd\") pod \"3fc4314a-6a31-49d1-8652-f4de6da97916\" (UID: \"3fc4314a-6a31-49d1-8652-f4de6da97916\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.598664 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f52457e-46da-4edd-8d76-dc4a28f02f0a-operator-scripts\") pod \"7f52457e-46da-4edd-8d76-dc4a28f02f0a\" (UID: \"7f52457e-46da-4edd-8d76-dc4a28f02f0a\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.598681 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qv5lh\" (UniqueName: \"kubernetes.io/projected/f76f637a-192d-4827-92b3-68e0fc348629-kube-api-access-qv5lh\") pod \"f76f637a-192d-4827-92b3-68e0fc348629\" (UID: \"f76f637a-192d-4827-92b3-68e0fc348629\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.598699 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008-operator-scripts\") pod \"dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008\" (UID: \"dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.598718 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9d8z\" (UniqueName: \"kubernetes.io/projected/3fc4314a-6a31-49d1-8652-f4de6da97916-kube-api-access-m9d8z\") pod \"3fc4314a-6a31-49d1-8652-f4de6da97916\" (UID: \"3fc4314a-6a31-49d1-8652-f4de6da97916\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.598734 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-852vf\" (UniqueName: \"kubernetes.io/projected/7fdd5969-f22b-41e6-a0d9-f3eeb16689fd-kube-api-access-852vf\") pod \"7fdd5969-f22b-41e6-a0d9-f3eeb16689fd\" (UID: \"7fdd5969-f22b-41e6-a0d9-f3eeb16689fd\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.598750 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mw5sg\" (UniqueName: \"kubernetes.io/projected/7f52457e-46da-4edd-8d76-dc4a28f02f0a-kube-api-access-mw5sg\") pod \"7f52457e-46da-4edd-8d76-dc4a28f02f0a\" (UID: \"7f52457e-46da-4edd-8d76-dc4a28f02f0a\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.598852 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8psfq\" (UniqueName: \"kubernetes.io/projected/cc8592b6-2527-449c-94a2-bf40f8ea4b81-kube-api-access-8psfq\") pod \"cc8592b6-2527-449c-94a2-bf40f8ea4b81\" (UID: \"cc8592b6-2527-449c-94a2-bf40f8ea4b81\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.598903 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc8592b6-2527-449c-94a2-bf40f8ea4b81-operator-scripts\") pod \"cc8592b6-2527-449c-94a2-bf40f8ea4b81\" (UID: \"cc8592b6-2527-449c-94a2-bf40f8ea4b81\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.598922 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fdd5969-f22b-41e6-a0d9-f3eeb16689fd-combined-ca-bundle\") pod \"7fdd5969-f22b-41e6-a0d9-f3eeb16689fd\" (UID: \"7fdd5969-f22b-41e6-a0d9-f3eeb16689fd\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.598950 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c-config-data-custom\") pod \"3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c\" (UID: \"3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.598964 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fdd5969-f22b-41e6-a0d9-f3eeb16689fd-config-data\") pod \"7fdd5969-f22b-41e6-a0d9-f3eeb16689fd\" (UID: \"7fdd5969-f22b-41e6-a0d9-f3eeb16689fd\") " Nov 24 13:37:24 crc kubenswrapper[4790]: E1124 13:37:24.599253 4790 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 24 13:37:24 crc kubenswrapper[4790]: E1124 13:37:24.599321 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/85e04b92-ae17-4eb6-a736-395c5d44f563-config-data podName:85e04b92-ae17-4eb6-a736-395c5d44f563 nodeName:}" failed. No retries permitted until 2025-11-24 13:37:32.599291618 +0000 UTC m=+1500.979185280 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/85e04b92-ae17-4eb6-a736-395c5d44f563-config-data") pod "rabbitmq-cell1-server-0" (UID: "85e04b92-ae17-4eb6-a736-395c5d44f563") : configmap "rabbitmq-cell1-config-data" not found Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.600050 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51552477-67be-4c73-ae48-aa131945caea-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "51552477-67be-4c73-ae48-aa131945caea" (UID: "51552477-67be-4c73-ae48-aa131945caea"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.600612 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fc4314a-6a31-49d1-8652-f4de6da97916-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3fc4314a-6a31-49d1-8652-f4de6da97916" (UID: "3fc4314a-6a31-49d1-8652-f4de6da97916"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.607896 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c-kube-api-access-fkrrs" (OuterVolumeSpecName: "kube-api-access-fkrrs") pod "3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c" (UID: "3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c"). InnerVolumeSpecName "kube-api-access-fkrrs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.609488 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fef75227-ffd5-4c97-87c7-e05fe899eb25-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fef75227-ffd5-4c97-87c7-e05fe899eb25" (UID: "fef75227-ffd5-4c97-87c7-e05fe899eb25"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.610913 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f76f637a-192d-4827-92b3-68e0fc348629-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f76f637a-192d-4827-92b3-68e0fc348629" (UID: "f76f637a-192d-4827-92b3-68e0fc348629"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.611070 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"81757ef9-7330-44bb-a35e-8e6f4c137c0e","Type":"ContainerDied","Data":"dd52f7be4dc7b7ab84f9f1447bde4fa92633674da2d16d7841751f839f7e2b13"} Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.611118 4790 scope.go:117] "RemoveContainer" containerID="6485471927d5a223e611cbd164262141a1a0f94b13c44e9d6ea257f4db80ed9a" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.611247 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.613358 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fc4314a-6a31-49d1-8652-f4de6da97916-scripts" (OuterVolumeSpecName: "scripts") pod "3fc4314a-6a31-49d1-8652-f4de6da97916" (UID: "3fc4314a-6a31-49d1-8652-f4de6da97916"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.616955 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f76f637a-192d-4827-92b3-68e0fc348629-kube-api-access-qv5lh" (OuterVolumeSpecName: "kube-api-access-qv5lh") pod "f76f637a-192d-4827-92b3-68e0fc348629" (UID: "f76f637a-192d-4827-92b3-68e0fc348629"). InnerVolumeSpecName "kube-api-access-qv5lh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.617127 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fef75227-ffd5-4c97-87c7-e05fe899eb25-kube-api-access-szr6v" (OuterVolumeSpecName: "kube-api-access-szr6v") pod "fef75227-ffd5-4c97-87c7-e05fe899eb25" (UID: "fef75227-ffd5-4c97-87c7-e05fe899eb25"). InnerVolumeSpecName "kube-api-access-szr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.617348 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c-logs" (OuterVolumeSpecName: "logs") pod "3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c" (UID: "3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.625397 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fc4314a-6a31-49d1-8652-f4de6da97916-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3fc4314a-6a31-49d1-8652-f4de6da97916" (UID: "3fc4314a-6a31-49d1-8652-f4de6da97916"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.626032 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f52457e-46da-4edd-8d76-dc4a28f02f0a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7f52457e-46da-4edd-8d76-dc4a28f02f0a" (UID: "7f52457e-46da-4edd-8d76-dc4a28f02f0a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.634608 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51552477-67be-4c73-ae48-aa131945caea-kube-api-access-hcxrv" (OuterVolumeSpecName: "kube-api-access-hcxrv") pod "51552477-67be-4c73-ae48-aa131945caea" (UID: "51552477-67be-4c73-ae48-aa131945caea"). InnerVolumeSpecName "kube-api-access-hcxrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.635114 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008-kube-api-access-mbb22" (OuterVolumeSpecName: "kube-api-access-mbb22") pod "dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008" (UID: "dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008"). InnerVolumeSpecName "kube-api-access-mbb22". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.639383 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fdd5969-f22b-41e6-a0d9-f3eeb16689fd-kube-api-access-852vf" (OuterVolumeSpecName: "kube-api-access-852vf") pod "7fdd5969-f22b-41e6-a0d9-f3eeb16689fd" (UID: "7fdd5969-f22b-41e6-a0d9-f3eeb16689fd"). InnerVolumeSpecName "kube-api-access-852vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.645059 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc8592b6-2527-449c-94a2-bf40f8ea4b81-kube-api-access-8psfq" (OuterVolumeSpecName: "kube-api-access-8psfq") pod "cc8592b6-2527-449c-94a2-bf40f8ea4b81" (UID: "cc8592b6-2527-449c-94a2-bf40f8ea4b81"). InnerVolumeSpecName "kube-api-access-8psfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.645356 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8cbd6e74-a758-4bbf-81eb-acffdfa56955","Type":"ContainerDied","Data":"9fff344ac0471c407ef0a075373fb989e75368750bedada8926fbfb9ee7adfe0"} Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.645490 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.645922 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c" (UID: "3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.646343 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc8592b6-2527-449c-94a2-bf40f8ea4b81-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cc8592b6-2527-449c-94a2-bf40f8ea4b81" (UID: "cc8592b6-2527-449c-94a2-bf40f8ea4b81"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.647394 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fc4314a-6a31-49d1-8652-f4de6da97916-kube-api-access-m9d8z" (OuterVolumeSpecName: "kube-api-access-m9d8z") pod "3fc4314a-6a31-49d1-8652-f4de6da97916" (UID: "3fc4314a-6a31-49d1-8652-f4de6da97916"). InnerVolumeSpecName "kube-api-access-m9d8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.647786 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008" (UID: "dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.649566 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f52457e-46da-4edd-8d76-dc4a28f02f0a-kube-api-access-mw5sg" (OuterVolumeSpecName: "kube-api-access-mw5sg") pod "7f52457e-46da-4edd-8d76-dc4a28f02f0a" (UID: "7f52457e-46da-4edd-8d76-dc4a28f02f0a"). InnerVolumeSpecName "kube-api-access-mw5sg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.658966 4790 scope.go:117] "RemoveContainer" containerID="6b1d4a6d406564bd94038ede12823e6303e9c099f716a3ed7065653038dff83e" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.660140 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement0fc6-account-delete-rl7sl" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.660154 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement0fc6-account-delete-rl7sl" event={"ID":"51552477-67be-4c73-ae48-aa131945caea","Type":"ContainerDied","Data":"f4a2b5aa29d9b3a52a85e4c10a9cd130150455763a2d826c0311af280f0232e2"} Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.660181 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4a2b5aa29d9b3a52a85e4c10a9cd130150455763a2d826c0311af280f0232e2" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.665067 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"39d594c5-aae1-4cd7-a7de-9de435ad83b8","Type":"ContainerDied","Data":"846877fb35e003c0d6457b8ae7c84f49884f7e6761533a1bfde34d0a491f0b51"} Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.665980 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.669770 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c" (UID: "3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.671898 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7d69688466-r2bq9" event={"ID":"8311dfa1-6e87-493d-92e0-4eb69c792afe","Type":"ContainerDied","Data":"fd82fc411383d3237786b63929394f537a76680df3850407c10f53055832007b"} Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.672003 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7d69688466-r2bq9" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.686667 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8f8fe1f4-7b63-455c-97d1-d3c78826c3c4","Type":"ContainerDied","Data":"189abe70c405aaf3f7415caeafe81469f1527b66c9fcf4bdd2d54cc3e040a222"} Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.686765 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.689591 4790 scope.go:117] "RemoveContainer" containerID="8b87dbc9b7ac42a8486a73feacf6d51b270b4dd6af8a81539a4ea922eb8a0e1b" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.692176 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutronf6b9-account-delete-tdltj" event={"ID":"cc8592b6-2527-449c-94a2-bf40f8ea4b81","Type":"ContainerDied","Data":"dfe560b995610e1d584b18e97ab60b3041339b5edfbeb4766afda70e5abe61e1"} Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.692206 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dfe560b995610e1d584b18e97ab60b3041339b5edfbeb4766afda70e5abe61e1" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.692256 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutronf6b9-account-delete-tdltj" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.702440 4790 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.702465 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcxrv\" (UniqueName: \"kubernetes.io/projected/51552477-67be-4c73-ae48-aa131945caea-kube-api-access-hcxrv\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.702474 4790 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3fc4314a-6a31-49d1-8652-f4de6da97916-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.702485 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51552477-67be-4c73-ae48-aa131945caea-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.702495 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3fc4314a-6a31-49d1-8652-f4de6da97916-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.702503 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkrrs\" (UniqueName: \"kubernetes.io/projected/3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c-kube-api-access-fkrrs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.702511 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.702521 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f76f637a-192d-4827-92b3-68e0fc348629-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.702622 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szr6v\" (UniqueName: \"kubernetes.io/projected/fef75227-ffd5-4c97-87c7-e05fe899eb25-kube-api-access-szr6v\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.702632 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbb22\" (UniqueName: \"kubernetes.io/projected/dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008-kube-api-access-mbb22\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.702641 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fef75227-ffd5-4c97-87c7-e05fe899eb25-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.702651 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.702659 4790 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3fc4314a-6a31-49d1-8652-f4de6da97916-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.702668 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f52457e-46da-4edd-8d76-dc4a28f02f0a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.702676 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qv5lh\" (UniqueName: \"kubernetes.io/projected/f76f637a-192d-4827-92b3-68e0fc348629-kube-api-access-qv5lh\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.702692 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.702710 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9d8z\" (UniqueName: \"kubernetes.io/projected/3fc4314a-6a31-49d1-8652-f4de6da97916-kube-api-access-m9d8z\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.702720 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-852vf\" (UniqueName: \"kubernetes.io/projected/7fdd5969-f22b-41e6-a0d9-f3eeb16689fd-kube-api-access-852vf\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.702732 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mw5sg\" (UniqueName: \"kubernetes.io/projected/7f52457e-46da-4edd-8d76-dc4a28f02f0a-kube-api-access-mw5sg\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.702744 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8psfq\" (UniqueName: \"kubernetes.io/projected/cc8592b6-2527-449c-94a2-bf40f8ea4b81-kube-api-access-8psfq\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.702756 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc8592b6-2527-449c-94a2-bf40f8ea4b81-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.709053 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fc4314a-6a31-49d1-8652-f4de6da97916-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3fc4314a-6a31-49d1-8652-f4de6da97916" (UID: "3fc4314a-6a31-49d1-8652-f4de6da97916"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.714058 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance388c-account-delete-8hfh2" event={"ID":"f76f637a-192d-4827-92b3-68e0fc348629","Type":"ContainerDied","Data":"b7be72ad15d6f4d3f2bed56552507ebe9f11cece3ef4e837f619a38c9837bdd0"} Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.714093 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7be72ad15d6f4d3f2bed56552507ebe9f11cece3ef4e837f619a38c9837bdd0" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.714147 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance388c-account-delete-8hfh2" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.721553 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbicana018-account-delete-6z4kv" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.721844 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicana018-account-delete-6z4kv" event={"ID":"7f52457e-46da-4edd-8d76-dc4a28f02f0a","Type":"ContainerDied","Data":"e47c7ea7b8258f1eb51e4adcc261a988e1bd70fb519246a5f17a5946416fdeb2"} Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.721899 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e47c7ea7b8258f1eb51e4adcc261a988e1bd70fb519246a5f17a5946416fdeb2" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.735404 4790 generic.go:334] "Generic (PLEG): container finished" podID="56d67de4-df9a-44fa-92b2-cbb168c4ae87" containerID="6501f4d3c19251ce46a87c27719040d3bcdb4ba9e7bcdb6dbb40bc282a09c44d" exitCode=0 Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.735483 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"56d67de4-df9a-44fa-92b2-cbb168c4ae87","Type":"ContainerDied","Data":"6501f4d3c19251ce46a87c27719040d3bcdb4ba9e7bcdb6dbb40bc282a09c44d"} Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.742298 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fdd5969-f22b-41e6-a0d9-f3eeb16689fd-config-data" (OuterVolumeSpecName: "config-data") pod "7fdd5969-f22b-41e6-a0d9-f3eeb16689fd" (UID: "7fdd5969-f22b-41e6-a0d9-f3eeb16689fd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.747844 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fdd5969-f22b-41e6-a0d9-f3eeb16689fd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7fdd5969-f22b-41e6-a0d9-f3eeb16689fd" (UID: "7fdd5969-f22b-41e6-a0d9-f3eeb16689fd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.748586 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.749127 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-58b5bcbcb-b6kbh" event={"ID":"a2ee8af7-26fb-4652-be37-594db62f1146","Type":"ContainerDied","Data":"194a06eb24751737741f835e766a9edc998ff73eeed814aa6bd2b16d57534f4e"} Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.749489 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-58b5bcbcb-b6kbh" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.759784 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.765782 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.766315 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3fc4314a-6a31-49d1-8652-f4de6da97916","Type":"ContainerDied","Data":"e8a1a03fbff93f503eaa9aa5d709b529d2dbba23cfdbd560e8362b99eb1a00ea"} Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.766403 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.770578 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"7fdd5969-f22b-41e6-a0d9-f3eeb16689fd","Type":"ContainerDied","Data":"e05aa9b0eb724d4a31ae22ff76a39eeca75170714d76458a140c2bb64d160a41"} Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.770648 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.774069 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.788100 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7d69688466-r2bq9"] Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.788173 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-7d69688466-r2bq9"] Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.788437 4790 scope.go:117] "RemoveContainer" containerID="1d279a19d65b0962ffcf048273f881a8a609e8c6c381d7ee412f985feb7393fb" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.795409 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.800003 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.807932 4790 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3fc4314a-6a31-49d1-8652-f4de6da97916-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.807969 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fdd5969-f22b-41e6-a0d9-f3eeb16689fd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.807981 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fdd5969-f22b-41e6-a0d9-f3eeb16689fd-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.809537 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.814730 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.819260 4790 generic.go:334] "Generic (PLEG): container finished" podID="c786f8b4-6946-45c1-8cde-00e56bf899ed" containerID="ddb81a6c66eee6e319c0646f64dd78e67d54faaf3b3d4b87630354653dbf2c4c" exitCode=0 Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.819351 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c786f8b4-6946-45c1-8cde-00e56bf899ed","Type":"ContainerDied","Data":"ddb81a6c66eee6e319c0646f64dd78e67d54faaf3b3d4b87630354653dbf2c4c"} Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.819609 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fc4314a-6a31-49d1-8652-f4de6da97916-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3fc4314a-6a31-49d1-8652-f4de6da97916" (UID: "3fc4314a-6a31-49d1-8652-f4de6da97916"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.833195 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutronf6b9-account-delete-tdltj"] Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.833898 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.834252 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-845f9d57f-2s48z" event={"ID":"3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c","Type":"ContainerDied","Data":"4484a8a800584754e0bc823ba32652af95c8926cd5358a8c3f54ea000f9ae805"} Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.834317 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-845f9d57f-2s48z" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.841294 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutronf6b9-account-delete-tdltj"] Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.856215 4790 scope.go:117] "RemoveContainer" containerID="eb8a4e1ecddd7fb54e90dd1a525962762b3153943ebd954bb7544913710d2ec2" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.856365 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-58b5bcbcb-b6kbh"] Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.861339 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fc4314a-6a31-49d1-8652-f4de6da97916-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "3fc4314a-6a31-49d1-8652-f4de6da97916" (UID: "3fc4314a-6a31-49d1-8652-f4de6da97916"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.861399 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-58b5bcbcb-b6kbh"] Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.866160 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fc4314a-6a31-49d1-8652-f4de6da97916-config-data" (OuterVolumeSpecName: "config-data") pod "3fc4314a-6a31-49d1-8652-f4de6da97916" (UID: "3fc4314a-6a31-49d1-8652-f4de6da97916"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.866984 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.873511 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.875037 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c-config-data" (OuterVolumeSpecName: "config-data") pod "3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c" (UID: "3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.881370 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.890742 4790 scope.go:117] "RemoveContainer" containerID="d07411834060975e29533a55785dd1993dfe65d5e3ce92d8eb62b08cba9dc0d1" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.908853 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/56d67de4-df9a-44fa-92b2-cbb168c4ae87-memcached-tls-certs\") pod \"56d67de4-df9a-44fa-92b2-cbb168c4ae87\" (UID: \"56d67de4-df9a-44fa-92b2-cbb168c4ae87\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.908976 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c786f8b4-6946-45c1-8cde-00e56bf899ed-logs\") pod \"c786f8b4-6946-45c1-8cde-00e56bf899ed\" (UID: \"c786f8b4-6946-45c1-8cde-00e56bf899ed\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.909005 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56d67de4-df9a-44fa-92b2-cbb168c4ae87-combined-ca-bundle\") pod \"56d67de4-df9a-44fa-92b2-cbb168c4ae87\" (UID: \"56d67de4-df9a-44fa-92b2-cbb168c4ae87\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.909030 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c786f8b4-6946-45c1-8cde-00e56bf899ed-combined-ca-bundle\") pod \"c786f8b4-6946-45c1-8cde-00e56bf899ed\" (UID: \"c786f8b4-6946-45c1-8cde-00e56bf899ed\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.909053 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pt66d\" (UniqueName: \"kubernetes.io/projected/56d67de4-df9a-44fa-92b2-cbb168c4ae87-kube-api-access-pt66d\") pod \"56d67de4-df9a-44fa-92b2-cbb168c4ae87\" (UID: \"56d67de4-df9a-44fa-92b2-cbb168c4ae87\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.909100 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c786f8b4-6946-45c1-8cde-00e56bf899ed-config-data\") pod \"c786f8b4-6946-45c1-8cde-00e56bf899ed\" (UID: \"c786f8b4-6946-45c1-8cde-00e56bf899ed\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.909143 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56d67de4-df9a-44fa-92b2-cbb168c4ae87-config-data\") pod \"56d67de4-df9a-44fa-92b2-cbb168c4ae87\" (UID: \"56d67de4-df9a-44fa-92b2-cbb168c4ae87\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.909168 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/56d67de4-df9a-44fa-92b2-cbb168c4ae87-kolla-config\") pod \"56d67de4-df9a-44fa-92b2-cbb168c4ae87\" (UID: \"56d67de4-df9a-44fa-92b2-cbb168c4ae87\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.909189 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c786f8b4-6946-45c1-8cde-00e56bf899ed-public-tls-certs\") pod \"c786f8b4-6946-45c1-8cde-00e56bf899ed\" (UID: \"c786f8b4-6946-45c1-8cde-00e56bf899ed\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.909234 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c786f8b4-6946-45c1-8cde-00e56bf899ed-internal-tls-certs\") pod \"c786f8b4-6946-45c1-8cde-00e56bf899ed\" (UID: \"c786f8b4-6946-45c1-8cde-00e56bf899ed\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.909258 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6gxb\" (UniqueName: \"kubernetes.io/projected/c786f8b4-6946-45c1-8cde-00e56bf899ed-kube-api-access-m6gxb\") pod \"c786f8b4-6946-45c1-8cde-00e56bf899ed\" (UID: \"c786f8b4-6946-45c1-8cde-00e56bf899ed\") " Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.909555 4790 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fc4314a-6a31-49d1-8652-f4de6da97916-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.909574 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.909569 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c786f8b4-6946-45c1-8cde-00e56bf899ed-logs" (OuterVolumeSpecName: "logs") pod "c786f8b4-6946-45c1-8cde-00e56bf899ed" (UID: "c786f8b4-6946-45c1-8cde-00e56bf899ed"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.909584 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fc4314a-6a31-49d1-8652-f4de6da97916-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.909628 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fc4314a-6a31-49d1-8652-f4de6da97916-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.910139 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56d67de4-df9a-44fa-92b2-cbb168c4ae87-config-data" (OuterVolumeSpecName: "config-data") pod "56d67de4-df9a-44fa-92b2-cbb168c4ae87" (UID: "56d67de4-df9a-44fa-92b2-cbb168c4ae87"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.910984 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56d67de4-df9a-44fa-92b2-cbb168c4ae87-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "56d67de4-df9a-44fa-92b2-cbb168c4ae87" (UID: "56d67de4-df9a-44fa-92b2-cbb168c4ae87"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.912651 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c786f8b4-6946-45c1-8cde-00e56bf899ed-kube-api-access-m6gxb" (OuterVolumeSpecName: "kube-api-access-m6gxb") pod "c786f8b4-6946-45c1-8cde-00e56bf899ed" (UID: "c786f8b4-6946-45c1-8cde-00e56bf899ed"). InnerVolumeSpecName "kube-api-access-m6gxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.913932 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56d67de4-df9a-44fa-92b2-cbb168c4ae87-kube-api-access-pt66d" (OuterVolumeSpecName: "kube-api-access-pt66d") pod "56d67de4-df9a-44fa-92b2-cbb168c4ae87" (UID: "56d67de4-df9a-44fa-92b2-cbb168c4ae87"). InnerVolumeSpecName "kube-api-access-pt66d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.923256 4790 scope.go:117] "RemoveContainer" containerID="5a96182e01bd62e46435ce7bc8d7492b1b623ec41db4e9fa84ec0792133db77d" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.959352 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56d67de4-df9a-44fa-92b2-cbb168c4ae87-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "56d67de4-df9a-44fa-92b2-cbb168c4ae87" (UID: "56d67de4-df9a-44fa-92b2-cbb168c4ae87"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.959462 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c786f8b4-6946-45c1-8cde-00e56bf899ed-config-data" (OuterVolumeSpecName: "config-data") pod "c786f8b4-6946-45c1-8cde-00e56bf899ed" (UID: "c786f8b4-6946-45c1-8cde-00e56bf899ed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.968008 4790 scope.go:117] "RemoveContainer" containerID="77d68f10a4550d5451c7aeeccb710c9bb4cd97f4057890444850ad09cff25f6b" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.974969 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapid63e-account-delete-k96dh"] Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.978745 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c786f8b4-6946-45c1-8cde-00e56bf899ed-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c786f8b4-6946-45c1-8cde-00e56bf899ed" (UID: "c786f8b4-6946-45c1-8cde-00e56bf899ed"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.983543 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novaapid63e-account-delete-k96dh"] Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.992080 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56d67de4-df9a-44fa-92b2-cbb168c4ae87-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "56d67de4-df9a-44fa-92b2-cbb168c4ae87" (UID: "56d67de4-df9a-44fa-92b2-cbb168c4ae87"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.992084 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c786f8b4-6946-45c1-8cde-00e56bf899ed-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c786f8b4-6946-45c1-8cde-00e56bf899ed" (UID: "c786f8b4-6946-45c1-8cde-00e56bf899ed"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.992673 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c786f8b4-6946-45c1-8cde-00e56bf899ed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c786f8b4-6946-45c1-8cde-00e56bf899ed" (UID: "c786f8b4-6946-45c1-8cde-00e56bf899ed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4790]: I1124 13:37:24.995061 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell00c43-account-delete-dlpjf"] Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.001087 4790 scope.go:117] "RemoveContainer" containerID="3e9b86f86ff0501669479b2ac2723096fea52e20d1dedb6c13ee82635ab6f1aa" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.004307 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell00c43-account-delete-dlpjf"] Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.022663 4790 scope.go:117] "RemoveContainer" containerID="ef998419eb802553e1ad108205b17dff1eada33d519b61e4493aea07e0efefc0" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.023032 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c786f8b4-6946-45c1-8cde-00e56bf899ed-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.023072 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56d67de4-df9a-44fa-92b2-cbb168c4ae87-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.023105 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c786f8b4-6946-45c1-8cde-00e56bf899ed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.023116 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pt66d\" (UniqueName: \"kubernetes.io/projected/56d67de4-df9a-44fa-92b2-cbb168c4ae87-kube-api-access-pt66d\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.023127 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c786f8b4-6946-45c1-8cde-00e56bf899ed-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.023137 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56d67de4-df9a-44fa-92b2-cbb168c4ae87-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.023147 4790 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/56d67de4-df9a-44fa-92b2-cbb168c4ae87-kolla-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.023183 4790 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c786f8b4-6946-45c1-8cde-00e56bf899ed-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.023193 4790 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c786f8b4-6946-45c1-8cde-00e56bf899ed-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.023202 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6gxb\" (UniqueName: \"kubernetes.io/projected/c786f8b4-6946-45c1-8cde-00e56bf899ed-kube-api-access-m6gxb\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.023213 4790 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/56d67de4-df9a-44fa-92b2-cbb168c4ae87-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.045110 4790 scope.go:117] "RemoveContainer" containerID="2b38eafa7ed5a3895c1dd40a15a595d2c6fed37ac22fb3628de8e8a277d110ed" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.066834 4790 scope.go:117] "RemoveContainer" containerID="3bdd8dbb30e64b7458fddef45148be02e9280fe559bf6018d442a5bf9eb8dada" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.085227 4790 scope.go:117] "RemoveContainer" containerID="b5e492786f34b12c4fccd95cb12360cbb4386005a0414f30c8e1be783d5252e0" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.105554 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.107660 4790 scope.go:117] "RemoveContainer" containerID="593dfb8aab769d099082292784d20c9dd7fc610d3ac9c0659d6fd1635968dfa7" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.110205 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.133503 4790 scope.go:117] "RemoveContainer" containerID="63bc861e59ed27e8b2dec969395af49861e48d187c85af757c5022b25349826a" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.250848 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-845f9d57f-2s48z"] Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.253736 4790 scope.go:117] "RemoveContainer" containerID="d8eab96ea490fbfa774abdd6a272fffab4b7cacb733249bf3f282fa0dfcf8521" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.265502 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-845f9d57f-2s48z"] Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.298939 4790 scope.go:117] "RemoveContainer" containerID="32dece0aef9f5c5f700993f3c33e9dc9b09465378d0d7abd138bd90e7d5c6319" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.462342 4790 scope.go:117] "RemoveContainer" containerID="9191be68775994f407e5cef4ebbca1e4aed9cfffa519ea7b28b2742579b483ad" Nov 24 13:37:25 crc kubenswrapper[4790]: E1124 13:37:25.748534 4790 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 24 13:37:25 crc kubenswrapper[4790]: E1124 13:37:25.748596 4790 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/39508987-8655-46ed-861b-e1448652ddf6-config-data podName:39508987-8655-46ed-861b-e1448652ddf6 nodeName:}" failed. No retries permitted until 2025-11-24 13:37:33.748581816 +0000 UTC m=+1502.128475478 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/39508987-8655-46ed-861b-e1448652ddf6-config-data") pod "rabbitmq-server-0" (UID: "39508987-8655-46ed-861b-e1448652ddf6") : configmap "rabbitmq-config-data" not found Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.850633 4790 generic.go:334] "Generic (PLEG): container finished" podID="85e04b92-ae17-4eb6-a736-395c5d44f563" containerID="8c7b1f8ad061938b73c9c94a4ef4bfe3303a40fe30294f1a92f9a97da7d69a3e" exitCode=0 Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.850693 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"85e04b92-ae17-4eb6-a736-395c5d44f563","Type":"ContainerDied","Data":"8c7b1f8ad061938b73c9c94a4ef4bfe3303a40fe30294f1a92f9a97da7d69a3e"} Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.850719 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"85e04b92-ae17-4eb6-a736-395c5d44f563","Type":"ContainerDied","Data":"bc9f2037bb2695552eaa4fead244895c9466a76a2ae724a8942e06cc71d82ae3"} Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.850730 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc9f2037bb2695552eaa4fead244895c9466a76a2ae724a8942e06cc71d82ae3" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.852584 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.870762 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c786f8b4-6946-45c1-8cde-00e56bf899ed","Type":"ContainerDied","Data":"967fad38c66d7248a13455d3a6db225ef5fd1da2a231cbb2ae3462876c679f42"} Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.870827 4790 scope.go:117] "RemoveContainer" containerID="ddb81a6c66eee6e319c0646f64dd78e67d54faaf3b3d4b87630354653dbf2c4c" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.870976 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.879959 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.887238 4790 generic.go:334] "Generic (PLEG): container finished" podID="39508987-8655-46ed-861b-e1448652ddf6" containerID="921d14419dc93745c5bfa78ec576180ba71427f41b71bc188365cd076418023e" exitCode=0 Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.887326 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"39508987-8655-46ed-861b-e1448652ddf6","Type":"ContainerDied","Data":"921d14419dc93745c5bfa78ec576180ba71427f41b71bc188365cd076418023e"} Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.887348 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"39508987-8655-46ed-861b-e1448652ddf6","Type":"ContainerDied","Data":"51ec4a5ea0e39374fa24505f9849fedadea6ce01615a1254dbd15f98bbefcab0"} Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.887415 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.895589 4790 generic.go:334] "Generic (PLEG): container finished" podID="5a434802-b18e-4d14-8415-4879eaa44191" containerID="3da69fd56ac70e4cd6a7e0c488e7105251f570721cf9c47f92c46924f483efc6" exitCode=0 Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.895712 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5a434802-b18e-4d14-8415-4879eaa44191","Type":"ContainerDied","Data":"3da69fd56ac70e4cd6a7e0c488e7105251f570721cf9c47f92c46924f483efc6"} Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.899408 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"56d67de4-df9a-44fa-92b2-cbb168c4ae87","Type":"ContainerDied","Data":"5b0c3aba8fa6e8f3f97dd7758dadf64532bd3c89036783eb8c58ac1dff1546a6"} Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.899437 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.905149 4790 scope.go:117] "RemoveContainer" containerID="7e1e61d9e8e5a5c4687b8cc095ff2bfb83f045d184dc2b3cab025c89bd6960a1" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.910665 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.943189 4790 scope.go:117] "RemoveContainer" containerID="921d14419dc93745c5bfa78ec576180ba71427f41b71bc188365cd076418023e" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.951339 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/39508987-8655-46ed-861b-e1448652ddf6-plugins-conf\") pod \"39508987-8655-46ed-861b-e1448652ddf6\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.951372 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/85e04b92-ae17-4eb6-a736-395c5d44f563-erlang-cookie-secret\") pod \"85e04b92-ae17-4eb6-a736-395c5d44f563\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.951406 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/39508987-8655-46ed-861b-e1448652ddf6-server-conf\") pod \"39508987-8655-46ed-861b-e1448652ddf6\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.951431 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/85e04b92-ae17-4eb6-a736-395c5d44f563-config-data\") pod \"85e04b92-ae17-4eb6-a736-395c5d44f563\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.951445 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/85e04b92-ae17-4eb6-a736-395c5d44f563-plugins-conf\") pod \"85e04b92-ae17-4eb6-a736-395c5d44f563\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.951477 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/39508987-8655-46ed-861b-e1448652ddf6-rabbitmq-tls\") pod \"39508987-8655-46ed-861b-e1448652ddf6\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.951493 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbsrv\" (UniqueName: \"kubernetes.io/projected/39508987-8655-46ed-861b-e1448652ddf6-kube-api-access-vbsrv\") pod \"39508987-8655-46ed-861b-e1448652ddf6\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.951510 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/39508987-8655-46ed-861b-e1448652ddf6-pod-info\") pod \"39508987-8655-46ed-861b-e1448652ddf6\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.951523 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/85e04b92-ae17-4eb6-a736-395c5d44f563-pod-info\") pod \"85e04b92-ae17-4eb6-a736-395c5d44f563\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.951556 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a434802-b18e-4d14-8415-4879eaa44191-combined-ca-bundle\") pod \"5a434802-b18e-4d14-8415-4879eaa44191\" (UID: \"5a434802-b18e-4d14-8415-4879eaa44191\") " Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.951576 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"39508987-8655-46ed-861b-e1448652ddf6\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.951597 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/39508987-8655-46ed-861b-e1448652ddf6-erlang-cookie-secret\") pod \"39508987-8655-46ed-861b-e1448652ddf6\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.951642 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/85e04b92-ae17-4eb6-a736-395c5d44f563-rabbitmq-confd\") pod \"85e04b92-ae17-4eb6-a736-395c5d44f563\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.951663 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5a434802-b18e-4d14-8415-4879eaa44191-config-data-default\") pod \"5a434802-b18e-4d14-8415-4879eaa44191\" (UID: \"5a434802-b18e-4d14-8415-4879eaa44191\") " Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.951681 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/85e04b92-ae17-4eb6-a736-395c5d44f563-rabbitmq-tls\") pod \"85e04b92-ae17-4eb6-a736-395c5d44f563\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.951720 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/39508987-8655-46ed-861b-e1448652ddf6-rabbitmq-confd\") pod \"39508987-8655-46ed-861b-e1448652ddf6\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.951758 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/85e04b92-ae17-4eb6-a736-395c5d44f563-rabbitmq-plugins\") pod \"85e04b92-ae17-4eb6-a736-395c5d44f563\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.951797 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/85e04b92-ae17-4eb6-a736-395c5d44f563-server-conf\") pod \"85e04b92-ae17-4eb6-a736-395c5d44f563\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.951813 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/39508987-8655-46ed-861b-e1448652ddf6-rabbitmq-plugins\") pod \"39508987-8655-46ed-861b-e1448652ddf6\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.951828 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"85e04b92-ae17-4eb6-a736-395c5d44f563\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.951863 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/39508987-8655-46ed-861b-e1448652ddf6-config-data\") pod \"39508987-8655-46ed-861b-e1448652ddf6\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.951889 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5a434802-b18e-4d14-8415-4879eaa44191-config-data-generated\") pod \"5a434802-b18e-4d14-8415-4879eaa44191\" (UID: \"5a434802-b18e-4d14-8415-4879eaa44191\") " Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.951960 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"5a434802-b18e-4d14-8415-4879eaa44191\" (UID: \"5a434802-b18e-4d14-8415-4879eaa44191\") " Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.951992 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/39508987-8655-46ed-861b-e1448652ddf6-rabbitmq-erlang-cookie\") pod \"39508987-8655-46ed-861b-e1448652ddf6\" (UID: \"39508987-8655-46ed-861b-e1448652ddf6\") " Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.951988 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.953430 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/85e04b92-ae17-4eb6-a736-395c5d44f563-rabbitmq-erlang-cookie\") pod \"85e04b92-ae17-4eb6-a736-395c5d44f563\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.953495 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llwqf\" (UniqueName: \"kubernetes.io/projected/5a434802-b18e-4d14-8415-4879eaa44191-kube-api-access-llwqf\") pod \"5a434802-b18e-4d14-8415-4879eaa44191\" (UID: \"5a434802-b18e-4d14-8415-4879eaa44191\") " Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.953830 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zxlr\" (UniqueName: \"kubernetes.io/projected/85e04b92-ae17-4eb6-a736-395c5d44f563-kube-api-access-7zxlr\") pod \"85e04b92-ae17-4eb6-a736-395c5d44f563\" (UID: \"85e04b92-ae17-4eb6-a736-395c5d44f563\") " Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.953851 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5a434802-b18e-4d14-8415-4879eaa44191-kolla-config\") pod \"5a434802-b18e-4d14-8415-4879eaa44191\" (UID: \"5a434802-b18e-4d14-8415-4879eaa44191\") " Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.953899 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a434802-b18e-4d14-8415-4879eaa44191-operator-scripts\") pod \"5a434802-b18e-4d14-8415-4879eaa44191\" (UID: \"5a434802-b18e-4d14-8415-4879eaa44191\") " Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.953918 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a434802-b18e-4d14-8415-4879eaa44191-galera-tls-certs\") pod \"5a434802-b18e-4d14-8415-4879eaa44191\" (UID: \"5a434802-b18e-4d14-8415-4879eaa44191\") " Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.960835 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a434802-b18e-4d14-8415-4879eaa44191-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "5a434802-b18e-4d14-8415-4879eaa44191" (UID: "5a434802-b18e-4d14-8415-4879eaa44191"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.961480 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85e04b92-ae17-4eb6-a736-395c5d44f563-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "85e04b92-ae17-4eb6-a736-395c5d44f563" (UID: "85e04b92-ae17-4eb6-a736-395c5d44f563"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.961839 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39508987-8655-46ed-861b-e1448652ddf6-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "39508987-8655-46ed-861b-e1448652ddf6" (UID: "39508987-8655-46ed-861b-e1448652ddf6"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.962267 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85e04b92-ae17-4eb6-a736-395c5d44f563-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "85e04b92-ae17-4eb6-a736-395c5d44f563" (UID: "85e04b92-ae17-4eb6-a736-395c5d44f563"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.962827 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39508987-8655-46ed-861b-e1448652ddf6-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "39508987-8655-46ed-861b-e1448652ddf6" (UID: "39508987-8655-46ed-861b-e1448652ddf6"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.971386 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.971435 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.973329 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39508987-8655-46ed-861b-e1448652ddf6-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "39508987-8655-46ed-861b-e1448652ddf6" (UID: "39508987-8655-46ed-861b-e1448652ddf6"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.974368 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.974483 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "39508987-8655-46ed-861b-e1448652ddf6" (UID: "39508987-8655-46ed-861b-e1448652ddf6"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.974930 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a434802-b18e-4d14-8415-4879eaa44191-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "5a434802-b18e-4d14-8415-4879eaa44191" (UID: "5a434802-b18e-4d14-8415-4879eaa44191"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.975027 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85e04b92-ae17-4eb6-a736-395c5d44f563-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "85e04b92-ae17-4eb6-a736-395c5d44f563" (UID: "85e04b92-ae17-4eb6-a736-395c5d44f563"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.984005 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "85e04b92-ae17-4eb6-a736-395c5d44f563" (UID: "85e04b92-ae17-4eb6-a736-395c5d44f563"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.985980 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39508987-8655-46ed-861b-e1448652ddf6-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "39508987-8655-46ed-861b-e1448652ddf6" (UID: "39508987-8655-46ed-861b-e1448652ddf6"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.987285 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85e04b92-ae17-4eb6-a736-395c5d44f563-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "85e04b92-ae17-4eb6-a736-395c5d44f563" (UID: "85e04b92-ae17-4eb6-a736-395c5d44f563"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.988596 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85e04b92-ae17-4eb6-a736-395c5d44f563-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "85e04b92-ae17-4eb6-a736-395c5d44f563" (UID: "85e04b92-ae17-4eb6-a736-395c5d44f563"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.988627 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a434802-b18e-4d14-8415-4879eaa44191-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5a434802-b18e-4d14-8415-4879eaa44191" (UID: "5a434802-b18e-4d14-8415-4879eaa44191"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.989440 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a434802-b18e-4d14-8415-4879eaa44191-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "5a434802-b18e-4d14-8415-4879eaa44191" (UID: "5a434802-b18e-4d14-8415-4879eaa44191"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:25 crc kubenswrapper[4790]: I1124 13:37:25.999109 4790 scope.go:117] "RemoveContainer" containerID="9ed4735bba7ba4c006885cc1b29d1af4db7e93db39ebe812c28f73a898f8fa68" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.000038 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39508987-8655-46ed-861b-e1448652ddf6-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "39508987-8655-46ed-861b-e1448652ddf6" (UID: "39508987-8655-46ed-861b-e1448652ddf6"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.010352 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39508987-8655-46ed-861b-e1448652ddf6-kube-api-access-vbsrv" (OuterVolumeSpecName: "kube-api-access-vbsrv") pod "39508987-8655-46ed-861b-e1448652ddf6" (UID: "39508987-8655-46ed-861b-e1448652ddf6"). InnerVolumeSpecName "kube-api-access-vbsrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.011380 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/85e04b92-ae17-4eb6-a736-395c5d44f563-pod-info" (OuterVolumeSpecName: "pod-info") pod "85e04b92-ae17-4eb6-a736-395c5d44f563" (UID: "85e04b92-ae17-4eb6-a736-395c5d44f563"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.011556 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/39508987-8655-46ed-861b-e1448652ddf6-pod-info" (OuterVolumeSpecName: "pod-info") pod "39508987-8655-46ed-861b-e1448652ddf6" (UID: "39508987-8655-46ed-861b-e1448652ddf6"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.011583 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85e04b92-ae17-4eb6-a736-395c5d44f563-kube-api-access-7zxlr" (OuterVolumeSpecName: "kube-api-access-7zxlr") pod "85e04b92-ae17-4eb6-a736-395c5d44f563" (UID: "85e04b92-ae17-4eb6-a736-395c5d44f563"). InnerVolumeSpecName "kube-api-access-7zxlr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.016171 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a434802-b18e-4d14-8415-4879eaa44191-kube-api-access-llwqf" (OuterVolumeSpecName: "kube-api-access-llwqf") pod "5a434802-b18e-4d14-8415-4879eaa44191" (UID: "5a434802-b18e-4d14-8415-4879eaa44191"). InnerVolumeSpecName "kube-api-access-llwqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.027005 4790 scope.go:117] "RemoveContainer" containerID="921d14419dc93745c5bfa78ec576180ba71427f41b71bc188365cd076418023e" Nov 24 13:37:26 crc kubenswrapper[4790]: E1124 13:37:26.027492 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"921d14419dc93745c5bfa78ec576180ba71427f41b71bc188365cd076418023e\": container with ID starting with 921d14419dc93745c5bfa78ec576180ba71427f41b71bc188365cd076418023e not found: ID does not exist" containerID="921d14419dc93745c5bfa78ec576180ba71427f41b71bc188365cd076418023e" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.027525 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"921d14419dc93745c5bfa78ec576180ba71427f41b71bc188365cd076418023e"} err="failed to get container status \"921d14419dc93745c5bfa78ec576180ba71427f41b71bc188365cd076418023e\": rpc error: code = NotFound desc = could not find container \"921d14419dc93745c5bfa78ec576180ba71427f41b71bc188365cd076418023e\": container with ID starting with 921d14419dc93745c5bfa78ec576180ba71427f41b71bc188365cd076418023e not found: ID does not exist" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.027547 4790 scope.go:117] "RemoveContainer" containerID="9ed4735bba7ba4c006885cc1b29d1af4db7e93db39ebe812c28f73a898f8fa68" Nov 24 13:37:26 crc kubenswrapper[4790]: E1124 13:37:26.027821 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ed4735bba7ba4c006885cc1b29d1af4db7e93db39ebe812c28f73a898f8fa68\": container with ID starting with 9ed4735bba7ba4c006885cc1b29d1af4db7e93db39ebe812c28f73a898f8fa68 not found: ID does not exist" containerID="9ed4735bba7ba4c006885cc1b29d1af4db7e93db39ebe812c28f73a898f8fa68" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.027847 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ed4735bba7ba4c006885cc1b29d1af4db7e93db39ebe812c28f73a898f8fa68"} err="failed to get container status \"9ed4735bba7ba4c006885cc1b29d1af4db7e93db39ebe812c28f73a898f8fa68\": rpc error: code = NotFound desc = could not find container \"9ed4735bba7ba4c006885cc1b29d1af4db7e93db39ebe812c28f73a898f8fa68\": container with ID starting with 9ed4735bba7ba4c006885cc1b29d1af4db7e93db39ebe812c28f73a898f8fa68 not found: ID does not exist" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.027869 4790 scope.go:117] "RemoveContainer" containerID="6501f4d3c19251ce46a87c27719040d3bcdb4ba9e7bcdb6dbb40bc282a09c44d" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.036315 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "mysql-db") pod "5a434802-b18e-4d14-8415-4879eaa44191" (UID: "5a434802-b18e-4d14-8415-4879eaa44191"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.054595 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85e04b92-ae17-4eb6-a736-395c5d44f563-server-conf" (OuterVolumeSpecName: "server-conf") pod "85e04b92-ae17-4eb6-a736-395c5d44f563" (UID: "85e04b92-ae17-4eb6-a736-395c5d44f563"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.055393 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a434802-b18e-4d14-8415-4879eaa44191-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "5a434802-b18e-4d14-8415-4879eaa44191" (UID: "5a434802-b18e-4d14-8415-4879eaa44191"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.055745 4790 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.055772 4790 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5a434802-b18e-4d14-8415-4879eaa44191-config-data-generated\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.055792 4790 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.055805 4790 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/39508987-8655-46ed-861b-e1448652ddf6-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.055817 4790 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/85e04b92-ae17-4eb6-a736-395c5d44f563-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.055829 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llwqf\" (UniqueName: \"kubernetes.io/projected/5a434802-b18e-4d14-8415-4879eaa44191-kube-api-access-llwqf\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.055840 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zxlr\" (UniqueName: \"kubernetes.io/projected/85e04b92-ae17-4eb6-a736-395c5d44f563-kube-api-access-7zxlr\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.055851 4790 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5a434802-b18e-4d14-8415-4879eaa44191-kolla-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.055862 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a434802-b18e-4d14-8415-4879eaa44191-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.056135 4790 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a434802-b18e-4d14-8415-4879eaa44191-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.056153 4790 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/39508987-8655-46ed-861b-e1448652ddf6-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.056165 4790 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/85e04b92-ae17-4eb6-a736-395c5d44f563-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.056178 4790 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/85e04b92-ae17-4eb6-a736-395c5d44f563-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.056191 4790 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/39508987-8655-46ed-861b-e1448652ddf6-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.056203 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbsrv\" (UniqueName: \"kubernetes.io/projected/39508987-8655-46ed-861b-e1448652ddf6-kube-api-access-vbsrv\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.056213 4790 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/39508987-8655-46ed-861b-e1448652ddf6-pod-info\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.056224 4790 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/85e04b92-ae17-4eb6-a736-395c5d44f563-pod-info\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.056243 4790 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.056256 4790 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/39508987-8655-46ed-861b-e1448652ddf6-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.056267 4790 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5a434802-b18e-4d14-8415-4879eaa44191-config-data-default\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.056278 4790 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/85e04b92-ae17-4eb6-a736-395c5d44f563-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.056289 4790 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/85e04b92-ae17-4eb6-a736-395c5d44f563-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.056300 4790 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/85e04b92-ae17-4eb6-a736-395c5d44f563-server-conf\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.056310 4790 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/39508987-8655-46ed-861b-e1448652ddf6-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.062693 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a434802-b18e-4d14-8415-4879eaa44191-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5a434802-b18e-4d14-8415-4879eaa44191" (UID: "5a434802-b18e-4d14-8415-4879eaa44191"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.068809 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39508987-8655-46ed-861b-e1448652ddf6-config-data" (OuterVolumeSpecName: "config-data") pod "39508987-8655-46ed-861b-e1448652ddf6" (UID: "39508987-8655-46ed-861b-e1448652ddf6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.079049 4790 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.080055 4790 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.099060 4790 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.140746 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85e04b92-ae17-4eb6-a736-395c5d44f563-config-data" (OuterVolumeSpecName: "config-data") pod "85e04b92-ae17-4eb6-a736-395c5d44f563" (UID: "85e04b92-ae17-4eb6-a736-395c5d44f563"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.143218 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39508987-8655-46ed-861b-e1448652ddf6-server-conf" (OuterVolumeSpecName: "server-conf") pod "39508987-8655-46ed-861b-e1448652ddf6" (UID: "39508987-8655-46ed-861b-e1448652ddf6"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.143455 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39508987-8655-46ed-861b-e1448652ddf6-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "39508987-8655-46ed-861b-e1448652ddf6" (UID: "39508987-8655-46ed-861b-e1448652ddf6"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.158467 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/39508987-8655-46ed-861b-e1448652ddf6-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.158500 4790 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.158509 4790 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/39508987-8655-46ed-861b-e1448652ddf6-server-conf\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.158518 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/85e04b92-ae17-4eb6-a736-395c5d44f563-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.158527 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a434802-b18e-4d14-8415-4879eaa44191-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.158535 4790 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.158543 4790 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/39508987-8655-46ed-861b-e1448652ddf6-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.158551 4790 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.164292 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85e04b92-ae17-4eb6-a736-395c5d44f563-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "85e04b92-ae17-4eb6-a736-395c5d44f563" (UID: "85e04b92-ae17-4eb6-a736-395c5d44f563"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.260089 4790 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/85e04b92-ae17-4eb6-a736-395c5d44f563-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.276101 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.279125 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7ddd868cf8-lc6qt" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.281653 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.324895 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c" path="/var/lib/kubelet/pods/3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c/volumes" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.325628 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39508987-8655-46ed-861b-e1448652ddf6" path="/var/lib/kubelet/pods/39508987-8655-46ed-861b-e1448652ddf6/volumes" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.326186 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39d594c5-aae1-4cd7-a7de-9de435ad83b8" path="/var/lib/kubelet/pods/39d594c5-aae1-4cd7-a7de-9de435ad83b8/volumes" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.327560 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fc4314a-6a31-49d1-8652-f4de6da97916" path="/var/lib/kubelet/pods/3fc4314a-6a31-49d1-8652-f4de6da97916/volumes" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.328474 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56d67de4-df9a-44fa-92b2-cbb168c4ae87" path="/var/lib/kubelet/pods/56d67de4-df9a-44fa-92b2-cbb168c4ae87/volumes" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.329636 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fdd5969-f22b-41e6-a0d9-f3eeb16689fd" path="/var/lib/kubelet/pods/7fdd5969-f22b-41e6-a0d9-f3eeb16689fd/volumes" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.330143 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81757ef9-7330-44bb-a35e-8e6f4c137c0e" path="/var/lib/kubelet/pods/81757ef9-7330-44bb-a35e-8e6f4c137c0e/volumes" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.330745 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8311dfa1-6e87-493d-92e0-4eb69c792afe" path="/var/lib/kubelet/pods/8311dfa1-6e87-493d-92e0-4eb69c792afe/volumes" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.335510 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cbd6e74-a758-4bbf-81eb-acffdfa56955" path="/var/lib/kubelet/pods/8cbd6e74-a758-4bbf-81eb-acffdfa56955/volumes" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.337170 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f8fe1f4-7b63-455c-97d1-d3c78826c3c4" path="/var/lib/kubelet/pods/8f8fe1f4-7b63-455c-97d1-d3c78826c3c4/volumes" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.338199 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2ee8af7-26fb-4652-be37-594db62f1146" path="/var/lib/kubelet/pods/a2ee8af7-26fb-4652-be37-594db62f1146/volumes" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.339456 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c786f8b4-6946-45c1-8cde-00e56bf899ed" path="/var/lib/kubelet/pods/c786f8b4-6946-45c1-8cde-00e56bf899ed/volumes" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.340412 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc8592b6-2527-449c-94a2-bf40f8ea4b81" path="/var/lib/kubelet/pods/cc8592b6-2527-449c-94a2-bf40f8ea4b81/volumes" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.342280 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008" path="/var/lib/kubelet/pods/dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008/volumes" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.342832 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fef75227-ffd5-4c97-87c7-e05fe899eb25" path="/var/lib/kubelet/pods/fef75227-ffd5-4c97-87c7-e05fe899eb25/volumes" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.462421 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-public-tls-certs\") pod \"336619dc-13bf-4168-a58b-b5c724759890\" (UID: \"336619dc-13bf-4168-a58b-b5c724759890\") " Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.462490 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-fernet-keys\") pod \"336619dc-13bf-4168-a58b-b5c724759890\" (UID: \"336619dc-13bf-4168-a58b-b5c724759890\") " Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.462526 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-config-data\") pod \"336619dc-13bf-4168-a58b-b5c724759890\" (UID: \"336619dc-13bf-4168-a58b-b5c724759890\") " Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.462546 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-scripts\") pod \"336619dc-13bf-4168-a58b-b5c724759890\" (UID: \"336619dc-13bf-4168-a58b-b5c724759890\") " Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.462568 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-credential-keys\") pod \"336619dc-13bf-4168-a58b-b5c724759890\" (UID: \"336619dc-13bf-4168-a58b-b5c724759890\") " Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.462623 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-internal-tls-certs\") pod \"336619dc-13bf-4168-a58b-b5c724759890\" (UID: \"336619dc-13bf-4168-a58b-b5c724759890\") " Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.462667 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-combined-ca-bundle\") pod \"336619dc-13bf-4168-a58b-b5c724759890\" (UID: \"336619dc-13bf-4168-a58b-b5c724759890\") " Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.462725 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dv4jm\" (UniqueName: \"kubernetes.io/projected/336619dc-13bf-4168-a58b-b5c724759890-kube-api-access-dv4jm\") pod \"336619dc-13bf-4168-a58b-b5c724759890\" (UID: \"336619dc-13bf-4168-a58b-b5c724759890\") " Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.466347 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "336619dc-13bf-4168-a58b-b5c724759890" (UID: "336619dc-13bf-4168-a58b-b5c724759890"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.469036 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "336619dc-13bf-4168-a58b-b5c724759890" (UID: "336619dc-13bf-4168-a58b-b5c724759890"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.471440 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/336619dc-13bf-4168-a58b-b5c724759890-kube-api-access-dv4jm" (OuterVolumeSpecName: "kube-api-access-dv4jm") pod "336619dc-13bf-4168-a58b-b5c724759890" (UID: "336619dc-13bf-4168-a58b-b5c724759890"). InnerVolumeSpecName "kube-api-access-dv4jm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.480483 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-scripts" (OuterVolumeSpecName: "scripts") pod "336619dc-13bf-4168-a58b-b5c724759890" (UID: "336619dc-13bf-4168-a58b-b5c724759890"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.484042 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-config-data" (OuterVolumeSpecName: "config-data") pod "336619dc-13bf-4168-a58b-b5c724759890" (UID: "336619dc-13bf-4168-a58b-b5c724759890"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.488502 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "336619dc-13bf-4168-a58b-b5c724759890" (UID: "336619dc-13bf-4168-a58b-b5c724759890"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.501742 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "336619dc-13bf-4168-a58b-b5c724759890" (UID: "336619dc-13bf-4168-a58b-b5c724759890"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.513367 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "336619dc-13bf-4168-a58b-b5c724759890" (UID: "336619dc-13bf-4168-a58b-b5c724759890"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.564848 4790 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.564916 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.564930 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dv4jm\" (UniqueName: \"kubernetes.io/projected/336619dc-13bf-4168-a58b-b5c724759890-kube-api-access-dv4jm\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.564947 4790 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.564959 4790 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.564971 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.564982 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.564994 4790 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/336619dc-13bf-4168-a58b-b5c724759890-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.724921 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-67mw7"] Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.731501 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-67mw7"] Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.741808 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement0fc6-account-delete-rl7sl"] Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.749450 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement0fc6-account-delete-rl7sl"] Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.758405 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-0fc6-account-create-7zc7c"] Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.768771 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-0fc6-account-create-7zc7c"] Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.883168 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-mlh9g"] Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.894948 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-mlh9g"] Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.927080 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-a018-account-create-q2z5q"] Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.932714 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbicana018-account-delete-6z4kv"] Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.939092 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbicana018-account-delete-6z4kv"] Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.948540 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-a018-account-create-q2z5q"] Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.999428 4790 generic.go:334] "Generic (PLEG): container finished" podID="336619dc-13bf-4168-a58b-b5c724759890" containerID="07ac1c04efe6c3d8ada271427bad943fb80399185792ecba62e72b13d34fc58d" exitCode=0 Nov 24 13:37:26 crc kubenswrapper[4790]: I1124 13:37:26.999621 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7ddd868cf8-lc6qt" Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.004080 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7ddd868cf8-lc6qt" event={"ID":"336619dc-13bf-4168-a58b-b5c724759890","Type":"ContainerDied","Data":"07ac1c04efe6c3d8ada271427bad943fb80399185792ecba62e72b13d34fc58d"} Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.004134 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-8vcj4"] Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.004155 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7ddd868cf8-lc6qt" event={"ID":"336619dc-13bf-4168-a58b-b5c724759890","Type":"ContainerDied","Data":"1a431a97315bb5dee273ee36c96be5d5d31b8ac7dddb0b81940b402108384302"} Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.004176 4790 scope.go:117] "RemoveContainer" containerID="07ac1c04efe6c3d8ada271427bad943fb80399185792ecba62e72b13d34fc58d" Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.006196 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-8vcj4"] Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.053172 4790 generic.go:334] "Generic (PLEG): container finished" podID="d0fa96c6-35fb-4e66-a7dc-31f6fa589132" containerID="92ffb21f402a267f3a3f49d256f427879130f4670ece784d82a260628c06ed0d" exitCode=0 Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.053258 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d0fa96c6-35fb-4e66-a7dc-31f6fa589132","Type":"ContainerDied","Data":"92ffb21f402a267f3a3f49d256f427879130f4670ece784d82a260628c06ed0d"} Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.063733 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-c394-account-create-6kqhg"] Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.067061 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5a434802-b18e-4d14-8415-4879eaa44191","Type":"ContainerDied","Data":"0f939fb4108c2d62d2ca6653bb9bf2cfe97165eb15af2979d0b39e7b7325bd4f"} Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.067175 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.068848 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinderc394-account-delete-gnvr4"] Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.072783 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.076771 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-c394-account-create-6kqhg"] Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.081911 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinderc394-account-delete-gnvr4"] Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.082496 4790 scope.go:117] "RemoveContainer" containerID="07ac1c04efe6c3d8ada271427bad943fb80399185792ecba62e72b13d34fc58d" Nov 24 13:37:27 crc kubenswrapper[4790]: E1124 13:37:27.082812 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07ac1c04efe6c3d8ada271427bad943fb80399185792ecba62e72b13d34fc58d\": container with ID starting with 07ac1c04efe6c3d8ada271427bad943fb80399185792ecba62e72b13d34fc58d not found: ID does not exist" containerID="07ac1c04efe6c3d8ada271427bad943fb80399185792ecba62e72b13d34fc58d" Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.082847 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07ac1c04efe6c3d8ada271427bad943fb80399185792ecba62e72b13d34fc58d"} err="failed to get container status \"07ac1c04efe6c3d8ada271427bad943fb80399185792ecba62e72b13d34fc58d\": rpc error: code = NotFound desc = could not find container \"07ac1c04efe6c3d8ada271427bad943fb80399185792ecba62e72b13d34fc58d\": container with ID starting with 07ac1c04efe6c3d8ada271427bad943fb80399185792ecba62e72b13d34fc58d not found: ID does not exist" Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.082874 4790 scope.go:117] "RemoveContainer" containerID="3da69fd56ac70e4cd6a7e0c488e7105251f570721cf9c47f92c46924f483efc6" Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.090627 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-7ddd868cf8-lc6qt"] Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.098011 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-7ddd868cf8-lc6qt"] Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.119905 4790 scope.go:117] "RemoveContainer" containerID="31d8225618eff635579b4cb575f6c144d5c9c01d087c36178d30076b744ae0fd" Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.146458 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.153180 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.165629 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.178957 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.180829 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-ff7zf"] Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.191753 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-ff7zf"] Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.200466 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-388c-account-create-p6bcq"] Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.225406 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance388c-account-delete-8hfh2"] Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.228051 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-388c-account-create-p6bcq"] Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.232491 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance388c-account-delete-8hfh2"] Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.391173 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.487601 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-config-data-custom\") pod \"d0fa96c6-35fb-4e66-a7dc-31f6fa589132\" (UID: \"d0fa96c6-35fb-4e66-a7dc-31f6fa589132\") " Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.487667 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-scripts\") pod \"d0fa96c6-35fb-4e66-a7dc-31f6fa589132\" (UID: \"d0fa96c6-35fb-4e66-a7dc-31f6fa589132\") " Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.487685 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-combined-ca-bundle\") pod \"d0fa96c6-35fb-4e66-a7dc-31f6fa589132\" (UID: \"d0fa96c6-35fb-4e66-a7dc-31f6fa589132\") " Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.487866 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdzgl\" (UniqueName: \"kubernetes.io/projected/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-kube-api-access-mdzgl\") pod \"d0fa96c6-35fb-4e66-a7dc-31f6fa589132\" (UID: \"d0fa96c6-35fb-4e66-a7dc-31f6fa589132\") " Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.487907 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-etc-machine-id\") pod \"d0fa96c6-35fb-4e66-a7dc-31f6fa589132\" (UID: \"d0fa96c6-35fb-4e66-a7dc-31f6fa589132\") " Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.488089 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d0fa96c6-35fb-4e66-a7dc-31f6fa589132" (UID: "d0fa96c6-35fb-4e66-a7dc-31f6fa589132"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.488434 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-config-data\") pod \"d0fa96c6-35fb-4e66-a7dc-31f6fa589132\" (UID: \"d0fa96c6-35fb-4e66-a7dc-31f6fa589132\") " Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.488648 4790 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.505309 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-kube-api-access-mdzgl" (OuterVolumeSpecName: "kube-api-access-mdzgl") pod "d0fa96c6-35fb-4e66-a7dc-31f6fa589132" (UID: "d0fa96c6-35fb-4e66-a7dc-31f6fa589132"). InnerVolumeSpecName "kube-api-access-mdzgl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.505309 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d0fa96c6-35fb-4e66-a7dc-31f6fa589132" (UID: "d0fa96c6-35fb-4e66-a7dc-31f6fa589132"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.505415 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-scripts" (OuterVolumeSpecName: "scripts") pod "d0fa96c6-35fb-4e66-a7dc-31f6fa589132" (UID: "d0fa96c6-35fb-4e66-a7dc-31f6fa589132"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.534992 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d0fa96c6-35fb-4e66-a7dc-31f6fa589132" (UID: "d0fa96c6-35fb-4e66-a7dc-31f6fa589132"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.569479 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-config-data" (OuterVolumeSpecName: "config-data") pod "d0fa96c6-35fb-4e66-a7dc-31f6fa589132" (UID: "d0fa96c6-35fb-4e66-a7dc-31f6fa589132"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.590360 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.590385 4790 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.590395 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.590403 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:27 crc kubenswrapper[4790]: I1124 13:37:27.590413 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdzgl\" (UniqueName: \"kubernetes.io/projected/d0fa96c6-35fb-4e66-a7dc-31f6fa589132-kube-api-access-mdzgl\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:28 crc kubenswrapper[4790]: I1124 13:37:28.088731 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d0fa96c6-35fb-4e66-a7dc-31f6fa589132","Type":"ContainerDied","Data":"5519e85bf1eaf040d0b597f6c16478311b609bfa922465b207a9b70aea556814"} Nov 24 13:37:28 crc kubenswrapper[4790]: I1124 13:37:28.088783 4790 scope.go:117] "RemoveContainer" containerID="fb2d4795a70584cebf1c4bedf0675e9e61a8fae6696ec9e5217fb4161288c248" Nov 24 13:37:28 crc kubenswrapper[4790]: I1124 13:37:28.088899 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 13:37:28 crc kubenswrapper[4790]: I1124 13:37:28.125521 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 13:37:28 crc kubenswrapper[4790]: I1124 13:37:28.132322 4790 scope.go:117] "RemoveContainer" containerID="92ffb21f402a267f3a3f49d256f427879130f4670ece784d82a260628c06ed0d" Nov 24 13:37:28 crc kubenswrapper[4790]: I1124 13:37:28.136064 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 13:37:28 crc kubenswrapper[4790]: I1124 13:37:28.344390 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="336619dc-13bf-4168-a58b-b5c724759890" path="/var/lib/kubelet/pods/336619dc-13bf-4168-a58b-b5c724759890/volumes" Nov 24 13:37:28 crc kubenswrapper[4790]: I1124 13:37:28.345347 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c2d52e6-25f1-4ef0-a50e-6a8564c5302a" path="/var/lib/kubelet/pods/4c2d52e6-25f1-4ef0-a50e-6a8564c5302a/volumes" Nov 24 13:37:28 crc kubenswrapper[4790]: I1124 13:37:28.345975 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51552477-67be-4c73-ae48-aa131945caea" path="/var/lib/kubelet/pods/51552477-67be-4c73-ae48-aa131945caea/volumes" Nov 24 13:37:28 crc kubenswrapper[4790]: I1124 13:37:28.347352 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a434802-b18e-4d14-8415-4879eaa44191" path="/var/lib/kubelet/pods/5a434802-b18e-4d14-8415-4879eaa44191/volumes" Nov 24 13:37:28 crc kubenswrapper[4790]: I1124 13:37:28.348378 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f52457e-46da-4edd-8d76-dc4a28f02f0a" path="/var/lib/kubelet/pods/7f52457e-46da-4edd-8d76-dc4a28f02f0a/volumes" Nov 24 13:37:28 crc kubenswrapper[4790]: I1124 13:37:28.349241 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8095e4aa-7290-441d-ba67-79e48c23b8a6" path="/var/lib/kubelet/pods/8095e4aa-7290-441d-ba67-79e48c23b8a6/volumes" Nov 24 13:37:28 crc kubenswrapper[4790]: I1124 13:37:28.350846 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85e04b92-ae17-4eb6-a736-395c5d44f563" path="/var/lib/kubelet/pods/85e04b92-ae17-4eb6-a736-395c5d44f563/volumes" Nov 24 13:37:28 crc kubenswrapper[4790]: I1124 13:37:28.352171 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88b90d01-f375-4158-9dd6-8042fdfbbd74" path="/var/lib/kubelet/pods/88b90d01-f375-4158-9dd6-8042fdfbbd74/volumes" Nov 24 13:37:28 crc kubenswrapper[4790]: I1124 13:37:28.352820 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="937f8268-ef33-440e-a297-d7c1b1cbefc0" path="/var/lib/kubelet/pods/937f8268-ef33-440e-a297-d7c1b1cbefc0/volumes" Nov 24 13:37:28 crc kubenswrapper[4790]: I1124 13:37:28.354684 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="967da6fe-b0e0-4559-8db1-eea829746d13" path="/var/lib/kubelet/pods/967da6fe-b0e0-4559-8db1-eea829746d13/volumes" Nov 24 13:37:28 crc kubenswrapper[4790]: I1124 13:37:28.355466 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5bee7f9-da20-4e5d-9816-62bb3e9d6bce" path="/var/lib/kubelet/pods/a5bee7f9-da20-4e5d-9816-62bb3e9d6bce/volumes" Nov 24 13:37:28 crc kubenswrapper[4790]: I1124 13:37:28.356161 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0fa96c6-35fb-4e66-a7dc-31f6fa589132" path="/var/lib/kubelet/pods/d0fa96c6-35fb-4e66-a7dc-31f6fa589132/volumes" Nov 24 13:37:28 crc kubenswrapper[4790]: I1124 13:37:28.357470 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee85107b-89f0-48c3-aecf-05295449f025" path="/var/lib/kubelet/pods/ee85107b-89f0-48c3-aecf-05295449f025/volumes" Nov 24 13:37:28 crc kubenswrapper[4790]: I1124 13:37:28.358317 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f14beb2d-e165-43a0-a038-4cb2a3a86491" path="/var/lib/kubelet/pods/f14beb2d-e165-43a0-a038-4cb2a3a86491/volumes" Nov 24 13:37:28 crc kubenswrapper[4790]: I1124 13:37:28.359088 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f76f637a-192d-4827-92b3-68e0fc348629" path="/var/lib/kubelet/pods/f76f637a-192d-4827-92b3-68e0fc348629/volumes" Nov 24 13:37:28 crc kubenswrapper[4790]: I1124 13:37:28.359694 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f95eea4d-a9fd-49e3-98d3-3a871e945550" path="/var/lib/kubelet/pods/f95eea4d-a9fd-49e3-98d3-3a871e945550/volumes" Nov 24 13:37:28 crc kubenswrapper[4790]: E1124 13:37:28.574270 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3 is running failed: container process not found" containerID="e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 13:37:28 crc kubenswrapper[4790]: E1124 13:37:28.574811 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3 is running failed: container process not found" containerID="e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 13:37:28 crc kubenswrapper[4790]: E1124 13:37:28.575641 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3 is running failed: container process not found" containerID="e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 13:37:28 crc kubenswrapper[4790]: E1124 13:37:28.575703 4790 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-28cwj" podUID="4d6d5d88-adda-4679-80de-0df0e0847bcb" containerName="ovsdb-server" Nov 24 13:37:28 crc kubenswrapper[4790]: E1124 13:37:28.575769 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="46182cf796e680922c3d2c23de93f19051dee361d1797948d997345bb2652287" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 13:37:28 crc kubenswrapper[4790]: E1124 13:37:28.577108 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="46182cf796e680922c3d2c23de93f19051dee361d1797948d997345bb2652287" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 13:37:28 crc kubenswrapper[4790]: E1124 13:37:28.579921 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="46182cf796e680922c3d2c23de93f19051dee361d1797948d997345bb2652287" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 13:37:28 crc kubenswrapper[4790]: E1124 13:37:28.579990 4790 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-28cwj" podUID="4d6d5d88-adda-4679-80de-0df0e0847bcb" containerName="ovs-vswitchd" Nov 24 13:37:29 crc kubenswrapper[4790]: I1124 13:37:29.669015 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/memcached-0" podUID="56d67de4-df9a-44fa-92b2-cbb168c4ae87" containerName="memcached" probeResult="failure" output="dial tcp 10.217.0.104:11211: i/o timeout" Nov 24 13:37:33 crc kubenswrapper[4790]: E1124 13:37:33.574285 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3 is running failed: container process not found" containerID="e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 13:37:33 crc kubenswrapper[4790]: E1124 13:37:33.576032 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3 is running failed: container process not found" containerID="e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 13:37:33 crc kubenswrapper[4790]: E1124 13:37:33.576317 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3 is running failed: container process not found" containerID="e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 13:37:33 crc kubenswrapper[4790]: E1124 13:37:33.576428 4790 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-28cwj" podUID="4d6d5d88-adda-4679-80de-0df0e0847bcb" containerName="ovsdb-server" Nov 24 13:37:33 crc kubenswrapper[4790]: E1124 13:37:33.578824 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="46182cf796e680922c3d2c23de93f19051dee361d1797948d997345bb2652287" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 13:37:33 crc kubenswrapper[4790]: E1124 13:37:33.580424 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="46182cf796e680922c3d2c23de93f19051dee361d1797948d997345bb2652287" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 13:37:33 crc kubenswrapper[4790]: E1124 13:37:33.581925 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="46182cf796e680922c3d2c23de93f19051dee361d1797948d997345bb2652287" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 13:37:33 crc kubenswrapper[4790]: E1124 13:37:33.582041 4790 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-28cwj" podUID="4d6d5d88-adda-4679-80de-0df0e0847bcb" containerName="ovs-vswitchd" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.461502 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dbvgd"] Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462031 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008" containerName="mariadb-account-delete" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462054 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008" containerName="mariadb-account-delete" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462071 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cbd6e74-a758-4bbf-81eb-acffdfa56955" containerName="glance-httpd" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462078 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cbd6e74-a758-4bbf-81eb-acffdfa56955" containerName="glance-httpd" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462095 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39508987-8655-46ed-861b-e1448652ddf6" containerName="setup-container" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462103 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="39508987-8655-46ed-861b-e1448652ddf6" containerName="setup-container" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462115 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9c5a116-6438-47b5-8462-207eb656b553" containerName="openstack-network-exporter" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462122 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9c5a116-6438-47b5-8462-207eb656b553" containerName="openstack-network-exporter" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462130 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c428221a-e989-4844-894c-de96782508a3" containerName="init" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462137 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="c428221a-e989-4844-894c-de96782508a3" containerName="init" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462152 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6d696a7-d618-4416-b499-aae08088b079" containerName="cinder-api" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462159 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6d696a7-d618-4416-b499-aae08088b079" containerName="cinder-api" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462172 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23704906-0f33-4a4e-8bbb-a796a9a1733f" containerName="kube-state-metrics" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462179 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="23704906-0f33-4a4e-8bbb-a796a9a1733f" containerName="kube-state-metrics" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462190 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08baac2a-300a-498f-8d0c-d1f261e6c8fc" containerName="openstack-network-exporter" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462197 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="08baac2a-300a-498f-8d0c-d1f261e6c8fc" containerName="openstack-network-exporter" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462209 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f8fe1f4-7b63-455c-97d1-d3c78826c3c4" containerName="glance-log" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462216 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f8fe1f4-7b63-455c-97d1-d3c78826c3c4" containerName="glance-log" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462229 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9362fee7-a172-4948-8721-b4c83bf04a30" containerName="proxy-httpd" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462236 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="9362fee7-a172-4948-8721-b4c83bf04a30" containerName="proxy-httpd" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462249 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9362fee7-a172-4948-8721-b4c83bf04a30" containerName="proxy-server" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462256 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="9362fee7-a172-4948-8721-b4c83bf04a30" containerName="proxy-server" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462269 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81757ef9-7330-44bb-a35e-8e6f4c137c0e" containerName="nova-metadata-log" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462276 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="81757ef9-7330-44bb-a35e-8e6f4c137c0e" containerName="nova-metadata-log" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462288 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f76f637a-192d-4827-92b3-68e0fc348629" containerName="mariadb-account-delete" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462296 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="f76f637a-192d-4827-92b3-68e0fc348629" containerName="mariadb-account-delete" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462305 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23e9fd30-8462-4b44-b9ae-11e3a367c814" containerName="openstack-network-exporter" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462312 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="23e9fd30-8462-4b44-b9ae-11e3a367c814" containerName="openstack-network-exporter" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462324 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de6f5df1-eb57-493b-9d89-5e710c6f4304" containerName="nova-scheduler-scheduler" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462331 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="de6f5df1-eb57-493b-9d89-5e710c6f4304" containerName="nova-scheduler-scheduler" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462340 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8095e4aa-7290-441d-ba67-79e48c23b8a6" containerName="mariadb-account-delete" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462347 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="8095e4aa-7290-441d-ba67-79e48c23b8a6" containerName="mariadb-account-delete" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462358 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2ee8af7-26fb-4652-be37-594db62f1146" containerName="barbican-keystone-listener-log" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462365 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2ee8af7-26fb-4652-be37-594db62f1146" containerName="barbican-keystone-listener-log" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462379 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23e9fd30-8462-4b44-b9ae-11e3a367c814" containerName="ovsdbserver-sb" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462386 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="23e9fd30-8462-4b44-b9ae-11e3a367c814" containerName="ovsdbserver-sb" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462408 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2ee8af7-26fb-4652-be37-594db62f1146" containerName="barbican-keystone-listener" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462415 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2ee8af7-26fb-4652-be37-594db62f1146" containerName="barbican-keystone-listener" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462429 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8311dfa1-6e87-493d-92e0-4eb69c792afe" containerName="barbican-api" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462436 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="8311dfa1-6e87-493d-92e0-4eb69c792afe" containerName="barbican-api" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462447 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc8592b6-2527-449c-94a2-bf40f8ea4b81" containerName="mariadb-account-delete" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462455 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc8592b6-2527-449c-94a2-bf40f8ea4b81" containerName="mariadb-account-delete" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462470 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c" containerName="barbican-worker-log" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462477 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c" containerName="barbican-worker-log" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462489 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39508987-8655-46ed-861b-e1448652ddf6" containerName="rabbitmq" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462496 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="39508987-8655-46ed-861b-e1448652ddf6" containerName="rabbitmq" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462510 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8fa811d-92cf-46a2-a9fd-539130369eb1" containerName="ovn-controller" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462517 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8fa811d-92cf-46a2-a9fd-539130369eb1" containerName="ovn-controller" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462529 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c866267-37e8-48d7-83ca-53e3cbf8b762" containerName="galera" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462547 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c866267-37e8-48d7-83ca-53e3cbf8b762" containerName="galera" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462555 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a434802-b18e-4d14-8415-4879eaa44191" containerName="mysql-bootstrap" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462562 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a434802-b18e-4d14-8415-4879eaa44191" containerName="mysql-bootstrap" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462572 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08baac2a-300a-498f-8d0c-d1f261e6c8fc" containerName="ovsdbserver-nb" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462578 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="08baac2a-300a-498f-8d0c-d1f261e6c8fc" containerName="ovsdbserver-nb" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462590 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0fa96c6-35fb-4e66-a7dc-31f6fa589132" containerName="cinder-scheduler" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462598 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0fa96c6-35fb-4e66-a7dc-31f6fa589132" containerName="cinder-scheduler" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462607 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fc4314a-6a31-49d1-8652-f4de6da97916" containerName="ceilometer-notification-agent" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462614 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fc4314a-6a31-49d1-8652-f4de6da97916" containerName="ceilometer-notification-agent" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462626 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81757ef9-7330-44bb-a35e-8e6f4c137c0e" containerName="nova-metadata-metadata" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462633 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="81757ef9-7330-44bb-a35e-8e6f4c137c0e" containerName="nova-metadata-metadata" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462643 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="336619dc-13bf-4168-a58b-b5c724759890" containerName="keystone-api" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462651 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="336619dc-13bf-4168-a58b-b5c724759890" containerName="keystone-api" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462659 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c786f8b4-6946-45c1-8cde-00e56bf899ed" containerName="nova-api-log" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462666 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="c786f8b4-6946-45c1-8cde-00e56bf899ed" containerName="nova-api-log" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462680 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a434802-b18e-4d14-8415-4879eaa44191" containerName="galera" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462687 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a434802-b18e-4d14-8415-4879eaa44191" containerName="galera" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462698 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb72fd7a-89c4-4a8a-a7b0-c758322a08b7" containerName="extract-content" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462705 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb72fd7a-89c4-4a8a-a7b0-c758322a08b7" containerName="extract-content" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462715 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb72fd7a-89c4-4a8a-a7b0-c758322a08b7" containerName="registry-server" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462722 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb72fd7a-89c4-4a8a-a7b0-c758322a08b7" containerName="registry-server" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462737 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8311dfa1-6e87-493d-92e0-4eb69c792afe" containerName="barbican-api-log" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462744 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="8311dfa1-6e87-493d-92e0-4eb69c792afe" containerName="barbican-api-log" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462753 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39d594c5-aae1-4cd7-a7de-9de435ad83b8" containerName="nova-cell1-conductor-conductor" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462760 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="39d594c5-aae1-4cd7-a7de-9de435ad83b8" containerName="nova-cell1-conductor-conductor" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462771 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb72fd7a-89c4-4a8a-a7b0-c758322a08b7" containerName="extract-utilities" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462778 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb72fd7a-89c4-4a8a-a7b0-c758322a08b7" containerName="extract-utilities" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462789 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85e04b92-ae17-4eb6-a736-395c5d44f563" containerName="rabbitmq" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462796 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="85e04b92-ae17-4eb6-a736-395c5d44f563" containerName="rabbitmq" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462804 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b507c50-0776-4d50-ad41-10f26f25fd5f" containerName="placement-api" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462811 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b507c50-0776-4d50-ad41-10f26f25fd5f" containerName="placement-api" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462821 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c786f8b4-6946-45c1-8cde-00e56bf899ed" containerName="nova-api-api" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462828 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="c786f8b4-6946-45c1-8cde-00e56bf899ed" containerName="nova-api-api" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462836 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="698a7113-712a-4940-aef9-aabd791b021d" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462843 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="698a7113-712a-4940-aef9-aabd791b021d" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462853 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51552477-67be-4c73-ae48-aa131945caea" containerName="mariadb-account-delete" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462859 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="51552477-67be-4c73-ae48-aa131945caea" containerName="mariadb-account-delete" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462873 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c428221a-e989-4844-894c-de96782508a3" containerName="dnsmasq-dns" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462908 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="c428221a-e989-4844-894c-de96782508a3" containerName="dnsmasq-dns" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462919 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c866267-37e8-48d7-83ca-53e3cbf8b762" containerName="mysql-bootstrap" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462927 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c866267-37e8-48d7-83ca-53e3cbf8b762" containerName="mysql-bootstrap" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462935 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f52457e-46da-4edd-8d76-dc4a28f02f0a" containerName="mariadb-account-delete" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462943 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f52457e-46da-4edd-8d76-dc4a28f02f0a" containerName="mariadb-account-delete" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462953 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f8fe1f4-7b63-455c-97d1-d3c78826c3c4" containerName="glance-httpd" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462961 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f8fe1f4-7b63-455c-97d1-d3c78826c3c4" containerName="glance-httpd" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462979 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85e04b92-ae17-4eb6-a736-395c5d44f563" containerName="setup-container" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.462986 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="85e04b92-ae17-4eb6-a736-395c5d44f563" containerName="setup-container" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.462999 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6d696a7-d618-4416-b499-aae08088b079" containerName="cinder-api-log" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463006 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6d696a7-d618-4416-b499-aae08088b079" containerName="cinder-api-log" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.463020 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fef75227-ffd5-4c97-87c7-e05fe899eb25" containerName="mariadb-account-delete" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463027 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="fef75227-ffd5-4c97-87c7-e05fe899eb25" containerName="mariadb-account-delete" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.463038 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fc4314a-6a31-49d1-8652-f4de6da97916" containerName="sg-core" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463044 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fc4314a-6a31-49d1-8652-f4de6da97916" containerName="sg-core" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.463057 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fc4314a-6a31-49d1-8652-f4de6da97916" containerName="ceilometer-central-agent" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463064 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fc4314a-6a31-49d1-8652-f4de6da97916" containerName="ceilometer-central-agent" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.463076 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fc4314a-6a31-49d1-8652-f4de6da97916" containerName="proxy-httpd" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463083 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fc4314a-6a31-49d1-8652-f4de6da97916" containerName="proxy-httpd" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.463093 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0fa96c6-35fb-4e66-a7dc-31f6fa589132" containerName="probe" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463100 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0fa96c6-35fb-4e66-a7dc-31f6fa589132" containerName="probe" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.463112 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fdd5969-f22b-41e6-a0d9-f3eeb16689fd" containerName="nova-cell0-conductor-conductor" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463118 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fdd5969-f22b-41e6-a0d9-f3eeb16689fd" containerName="nova-cell0-conductor-conductor" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.463128 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56d67de4-df9a-44fa-92b2-cbb168c4ae87" containerName="memcached" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463133 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="56d67de4-df9a-44fa-92b2-cbb168c4ae87" containerName="memcached" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.463143 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cbd6e74-a758-4bbf-81eb-acffdfa56955" containerName="glance-log" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463161 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cbd6e74-a758-4bbf-81eb-acffdfa56955" containerName="glance-log" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.463171 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c" containerName="barbican-worker" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463178 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c" containerName="barbican-worker" Nov 24 13:37:34 crc kubenswrapper[4790]: E1124 13:37:34.463186 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b507c50-0776-4d50-ad41-10f26f25fd5f" containerName="placement-log" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463191 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b507c50-0776-4d50-ad41-10f26f25fd5f" containerName="placement-log" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463317 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f52457e-46da-4edd-8d76-dc4a28f02f0a" containerName="mariadb-account-delete" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463330 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="8095e4aa-7290-441d-ba67-79e48c23b8a6" containerName="mariadb-account-delete" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463341 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0fa96c6-35fb-4e66-a7dc-31f6fa589132" containerName="probe" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463347 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb72fd7a-89c4-4a8a-a7b0-c758322a08b7" containerName="registry-server" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463358 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="9362fee7-a172-4948-8721-b4c83bf04a30" containerName="proxy-server" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463365 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="de6f5df1-eb57-493b-9d89-5e710c6f4304" containerName="nova-scheduler-scheduler" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463376 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="fef75227-ffd5-4c97-87c7-e05fe899eb25" containerName="mariadb-account-delete" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463387 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="39508987-8655-46ed-861b-e1448652ddf6" containerName="rabbitmq" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463396 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f8fe1f4-7b63-455c-97d1-d3c78826c3c4" containerName="glance-log" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463405 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fc4314a-6a31-49d1-8652-f4de6da97916" containerName="sg-core" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463412 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="8311dfa1-6e87-493d-92e0-4eb69c792afe" containerName="barbican-api" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463420 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="dae4ae7d-d780-4ae0-8ab7-8ca5dd6a8008" containerName="mariadb-account-delete" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463431 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8fa811d-92cf-46a2-a9fd-539130369eb1" containerName="ovn-controller" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463437 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a434802-b18e-4d14-8415-4879eaa44191" containerName="galera" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463445 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fc4314a-6a31-49d1-8652-f4de6da97916" containerName="proxy-httpd" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463452 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="698a7113-712a-4940-aef9-aabd791b021d" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463462 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="08baac2a-300a-498f-8d0c-d1f261e6c8fc" containerName="openstack-network-exporter" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463472 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6d696a7-d618-4416-b499-aae08088b079" containerName="cinder-api" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463478 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="51552477-67be-4c73-ae48-aa131945caea" containerName="mariadb-account-delete" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463487 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b507c50-0776-4d50-ad41-10f26f25fd5f" containerName="placement-api" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463494 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc8592b6-2527-449c-94a2-bf40f8ea4b81" containerName="mariadb-account-delete" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463502 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="c428221a-e989-4844-894c-de96782508a3" containerName="dnsmasq-dns" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463508 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cbd6e74-a758-4bbf-81eb-acffdfa56955" containerName="glance-log" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463514 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="23e9fd30-8462-4b44-b9ae-11e3a367c814" containerName="openstack-network-exporter" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463521 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="81757ef9-7330-44bb-a35e-8e6f4c137c0e" containerName="nova-metadata-metadata" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463531 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2ee8af7-26fb-4652-be37-594db62f1146" containerName="barbican-keystone-listener-log" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463538 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b507c50-0776-4d50-ad41-10f26f25fd5f" containerName="placement-log" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463546 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0fa96c6-35fb-4e66-a7dc-31f6fa589132" containerName="cinder-scheduler" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463557 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="8311dfa1-6e87-493d-92e0-4eb69c792afe" containerName="barbican-api-log" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463568 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="08baac2a-300a-498f-8d0c-d1f261e6c8fc" containerName="ovsdbserver-nb" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463581 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="c786f8b4-6946-45c1-8cde-00e56bf899ed" containerName="nova-api-log" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463593 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="56d67de4-df9a-44fa-92b2-cbb168c4ae87" containerName="memcached" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463602 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c" containerName="barbican-worker" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463611 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="39d594c5-aae1-4cd7-a7de-9de435ad83b8" containerName="nova-cell1-conductor-conductor" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463622 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="23704906-0f33-4a4e-8bbb-a796a9a1733f" containerName="kube-state-metrics" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463630 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="81757ef9-7330-44bb-a35e-8e6f4c137c0e" containerName="nova-metadata-log" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463641 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9c5a116-6438-47b5-8462-207eb656b553" containerName="openstack-network-exporter" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463654 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="85e04b92-ae17-4eb6-a736-395c5d44f563" containerName="rabbitmq" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463667 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="c786f8b4-6946-45c1-8cde-00e56bf899ed" containerName="nova-api-api" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463674 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="f76f637a-192d-4827-92b3-68e0fc348629" containerName="mariadb-account-delete" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463685 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f8fe1f4-7b63-455c-97d1-d3c78826c3c4" containerName="glance-httpd" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463694 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fc4314a-6a31-49d1-8652-f4de6da97916" containerName="ceilometer-central-agent" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463705 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c866267-37e8-48d7-83ca-53e3cbf8b762" containerName="galera" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463715 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fdd5969-f22b-41e6-a0d9-f3eeb16689fd" containerName="nova-cell0-conductor-conductor" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463724 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="23e9fd30-8462-4b44-b9ae-11e3a367c814" containerName="ovsdbserver-sb" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463733 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="3229a3f1-aa2f-44d4-b9a2-f8b4595d6e0c" containerName="barbican-worker-log" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463743 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fc4314a-6a31-49d1-8652-f4de6da97916" containerName="ceilometer-notification-agent" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463751 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6d696a7-d618-4416-b499-aae08088b079" containerName="cinder-api-log" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463761 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="336619dc-13bf-4168-a58b-b5c724759890" containerName="keystone-api" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463768 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2ee8af7-26fb-4652-be37-594db62f1146" containerName="barbican-keystone-listener" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463778 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cbd6e74-a758-4bbf-81eb-acffdfa56955" containerName="glance-httpd" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.463788 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="9362fee7-a172-4948-8721-b4c83bf04a30" containerName="proxy-httpd" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.464808 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dbvgd" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.480109 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dbvgd"] Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.584750 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3add0b78-c01f-43d6-b9a4-85ae9a461710-utilities\") pod \"certified-operators-dbvgd\" (UID: \"3add0b78-c01f-43d6-b9a4-85ae9a461710\") " pod="openshift-marketplace/certified-operators-dbvgd" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.584813 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvfdz\" (UniqueName: \"kubernetes.io/projected/3add0b78-c01f-43d6-b9a4-85ae9a461710-kube-api-access-qvfdz\") pod \"certified-operators-dbvgd\" (UID: \"3add0b78-c01f-43d6-b9a4-85ae9a461710\") " pod="openshift-marketplace/certified-operators-dbvgd" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.584898 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3add0b78-c01f-43d6-b9a4-85ae9a461710-catalog-content\") pod \"certified-operators-dbvgd\" (UID: \"3add0b78-c01f-43d6-b9a4-85ae9a461710\") " pod="openshift-marketplace/certified-operators-dbvgd" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.686425 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3add0b78-c01f-43d6-b9a4-85ae9a461710-utilities\") pod \"certified-operators-dbvgd\" (UID: \"3add0b78-c01f-43d6-b9a4-85ae9a461710\") " pod="openshift-marketplace/certified-operators-dbvgd" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.686690 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvfdz\" (UniqueName: \"kubernetes.io/projected/3add0b78-c01f-43d6-b9a4-85ae9a461710-kube-api-access-qvfdz\") pod \"certified-operators-dbvgd\" (UID: \"3add0b78-c01f-43d6-b9a4-85ae9a461710\") " pod="openshift-marketplace/certified-operators-dbvgd" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.686846 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3add0b78-c01f-43d6-b9a4-85ae9a461710-catalog-content\") pod \"certified-operators-dbvgd\" (UID: \"3add0b78-c01f-43d6-b9a4-85ae9a461710\") " pod="openshift-marketplace/certified-operators-dbvgd" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.687429 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3add0b78-c01f-43d6-b9a4-85ae9a461710-utilities\") pod \"certified-operators-dbvgd\" (UID: \"3add0b78-c01f-43d6-b9a4-85ae9a461710\") " pod="openshift-marketplace/certified-operators-dbvgd" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.687625 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3add0b78-c01f-43d6-b9a4-85ae9a461710-catalog-content\") pod \"certified-operators-dbvgd\" (UID: \"3add0b78-c01f-43d6-b9a4-85ae9a461710\") " pod="openshift-marketplace/certified-operators-dbvgd" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.729212 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvfdz\" (UniqueName: \"kubernetes.io/projected/3add0b78-c01f-43d6-b9a4-85ae9a461710-kube-api-access-qvfdz\") pod \"certified-operators-dbvgd\" (UID: \"3add0b78-c01f-43d6-b9a4-85ae9a461710\") " pod="openshift-marketplace/certified-operators-dbvgd" Nov 24 13:37:34 crc kubenswrapper[4790]: I1124 13:37:34.783623 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dbvgd" Nov 24 13:37:35 crc kubenswrapper[4790]: I1124 13:37:35.343360 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dbvgd"] Nov 24 13:37:36 crc kubenswrapper[4790]: I1124 13:37:36.155924 4790 generic.go:334] "Generic (PLEG): container finished" podID="3add0b78-c01f-43d6-b9a4-85ae9a461710" containerID="b97fef37ff4b08e344426b0d24b01d10f65c44609a58e019873fccb890284cae" exitCode=0 Nov 24 13:37:36 crc kubenswrapper[4790]: I1124 13:37:36.156035 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbvgd" event={"ID":"3add0b78-c01f-43d6-b9a4-85ae9a461710","Type":"ContainerDied","Data":"b97fef37ff4b08e344426b0d24b01d10f65c44609a58e019873fccb890284cae"} Nov 24 13:37:36 crc kubenswrapper[4790]: I1124 13:37:36.156241 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbvgd" event={"ID":"3add0b78-c01f-43d6-b9a4-85ae9a461710","Type":"ContainerStarted","Data":"ddaa9122ab0292a4bf3bf1414ef53bd78dfefc8ca572f0607efdae1e76ea8e7b"} Nov 24 13:37:36 crc kubenswrapper[4790]: I1124 13:37:36.870125 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-78cd478f4f-bqm94" Nov 24 13:37:37 crc kubenswrapper[4790]: I1124 13:37:37.016930 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-internal-tls-certs\") pod \"4d4e45de-9768-40e9-9c75-6abba8b38559\" (UID: \"4d4e45de-9768-40e9-9c75-6abba8b38559\") " Nov 24 13:37:37 crc kubenswrapper[4790]: I1124 13:37:37.017101 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-config\") pod \"4d4e45de-9768-40e9-9c75-6abba8b38559\" (UID: \"4d4e45de-9768-40e9-9c75-6abba8b38559\") " Nov 24 13:37:37 crc kubenswrapper[4790]: I1124 13:37:37.017200 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-combined-ca-bundle\") pod \"4d4e45de-9768-40e9-9c75-6abba8b38559\" (UID: \"4d4e45de-9768-40e9-9c75-6abba8b38559\") " Nov 24 13:37:37 crc kubenswrapper[4790]: I1124 13:37:37.017253 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-ovndb-tls-certs\") pod \"4d4e45de-9768-40e9-9c75-6abba8b38559\" (UID: \"4d4e45de-9768-40e9-9c75-6abba8b38559\") " Nov 24 13:37:37 crc kubenswrapper[4790]: I1124 13:37:37.017278 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzgcq\" (UniqueName: \"kubernetes.io/projected/4d4e45de-9768-40e9-9c75-6abba8b38559-kube-api-access-zzgcq\") pod \"4d4e45de-9768-40e9-9c75-6abba8b38559\" (UID: \"4d4e45de-9768-40e9-9c75-6abba8b38559\") " Nov 24 13:37:37 crc kubenswrapper[4790]: I1124 13:37:37.017342 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-public-tls-certs\") pod \"4d4e45de-9768-40e9-9c75-6abba8b38559\" (UID: \"4d4e45de-9768-40e9-9c75-6abba8b38559\") " Nov 24 13:37:37 crc kubenswrapper[4790]: I1124 13:37:37.017376 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-httpd-config\") pod \"4d4e45de-9768-40e9-9c75-6abba8b38559\" (UID: \"4d4e45de-9768-40e9-9c75-6abba8b38559\") " Nov 24 13:37:37 crc kubenswrapper[4790]: I1124 13:37:37.023099 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "4d4e45de-9768-40e9-9c75-6abba8b38559" (UID: "4d4e45de-9768-40e9-9c75-6abba8b38559"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:37 crc kubenswrapper[4790]: I1124 13:37:37.027758 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d4e45de-9768-40e9-9c75-6abba8b38559-kube-api-access-zzgcq" (OuterVolumeSpecName: "kube-api-access-zzgcq") pod "4d4e45de-9768-40e9-9c75-6abba8b38559" (UID: "4d4e45de-9768-40e9-9c75-6abba8b38559"). InnerVolumeSpecName "kube-api-access-zzgcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:37 crc kubenswrapper[4790]: I1124 13:37:37.069194 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "4d4e45de-9768-40e9-9c75-6abba8b38559" (UID: "4d4e45de-9768-40e9-9c75-6abba8b38559"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:37 crc kubenswrapper[4790]: I1124 13:37:37.069666 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d4e45de-9768-40e9-9c75-6abba8b38559" (UID: "4d4e45de-9768-40e9-9c75-6abba8b38559"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:37 crc kubenswrapper[4790]: I1124 13:37:37.070769 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-config" (OuterVolumeSpecName: "config") pod "4d4e45de-9768-40e9-9c75-6abba8b38559" (UID: "4d4e45de-9768-40e9-9c75-6abba8b38559"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:37 crc kubenswrapper[4790]: I1124 13:37:37.075727 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "4d4e45de-9768-40e9-9c75-6abba8b38559" (UID: "4d4e45de-9768-40e9-9c75-6abba8b38559"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:37 crc kubenswrapper[4790]: I1124 13:37:37.097329 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "4d4e45de-9768-40e9-9c75-6abba8b38559" (UID: "4d4e45de-9768-40e9-9c75-6abba8b38559"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:37 crc kubenswrapper[4790]: I1124 13:37:37.118749 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:37 crc kubenswrapper[4790]: I1124 13:37:37.118782 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzgcq\" (UniqueName: \"kubernetes.io/projected/4d4e45de-9768-40e9-9c75-6abba8b38559-kube-api-access-zzgcq\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:37 crc kubenswrapper[4790]: I1124 13:37:37.118797 4790 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:37 crc kubenswrapper[4790]: I1124 13:37:37.118807 4790 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:37 crc kubenswrapper[4790]: I1124 13:37:37.118818 4790 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:37 crc kubenswrapper[4790]: I1124 13:37:37.118828 4790 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:37 crc kubenswrapper[4790]: I1124 13:37:37.118838 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/4d4e45de-9768-40e9-9c75-6abba8b38559-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:37 crc kubenswrapper[4790]: I1124 13:37:37.164970 4790 generic.go:334] "Generic (PLEG): container finished" podID="4d4e45de-9768-40e9-9c75-6abba8b38559" containerID="deee7a82aed113ecc740dcca448cfcb53642b959cd494942c6732db6590cb066" exitCode=0 Nov 24 13:37:37 crc kubenswrapper[4790]: I1124 13:37:37.165013 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-78cd478f4f-bqm94" event={"ID":"4d4e45de-9768-40e9-9c75-6abba8b38559","Type":"ContainerDied","Data":"deee7a82aed113ecc740dcca448cfcb53642b959cd494942c6732db6590cb066"} Nov 24 13:37:37 crc kubenswrapper[4790]: I1124 13:37:37.165040 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-78cd478f4f-bqm94" event={"ID":"4d4e45de-9768-40e9-9c75-6abba8b38559","Type":"ContainerDied","Data":"fcfab0bd76919bb2d0cb9a4d6fe3601824fa84b37b36fb0578f01279ecef5776"} Nov 24 13:37:37 crc kubenswrapper[4790]: I1124 13:37:37.165063 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-78cd478f4f-bqm94" Nov 24 13:37:37 crc kubenswrapper[4790]: I1124 13:37:37.165071 4790 scope.go:117] "RemoveContainer" containerID="cef1cb69d1e5ecd83dedc3637e51913646334d098991071163c9a6574695cd3e" Nov 24 13:37:37 crc kubenswrapper[4790]: I1124 13:37:37.199192 4790 scope.go:117] "RemoveContainer" containerID="deee7a82aed113ecc740dcca448cfcb53642b959cd494942c6732db6590cb066" Nov 24 13:37:37 crc kubenswrapper[4790]: I1124 13:37:37.204668 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-78cd478f4f-bqm94"] Nov 24 13:37:37 crc kubenswrapper[4790]: I1124 13:37:37.210431 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-78cd478f4f-bqm94"] Nov 24 13:37:37 crc kubenswrapper[4790]: I1124 13:37:37.234198 4790 scope.go:117] "RemoveContainer" containerID="cef1cb69d1e5ecd83dedc3637e51913646334d098991071163c9a6574695cd3e" Nov 24 13:37:37 crc kubenswrapper[4790]: E1124 13:37:37.235165 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cef1cb69d1e5ecd83dedc3637e51913646334d098991071163c9a6574695cd3e\": container with ID starting with cef1cb69d1e5ecd83dedc3637e51913646334d098991071163c9a6574695cd3e not found: ID does not exist" containerID="cef1cb69d1e5ecd83dedc3637e51913646334d098991071163c9a6574695cd3e" Nov 24 13:37:37 crc kubenswrapper[4790]: I1124 13:37:37.235259 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cef1cb69d1e5ecd83dedc3637e51913646334d098991071163c9a6574695cd3e"} err="failed to get container status \"cef1cb69d1e5ecd83dedc3637e51913646334d098991071163c9a6574695cd3e\": rpc error: code = NotFound desc = could not find container \"cef1cb69d1e5ecd83dedc3637e51913646334d098991071163c9a6574695cd3e\": container with ID starting with cef1cb69d1e5ecd83dedc3637e51913646334d098991071163c9a6574695cd3e not found: ID does not exist" Nov 24 13:37:37 crc kubenswrapper[4790]: I1124 13:37:37.235286 4790 scope.go:117] "RemoveContainer" containerID="deee7a82aed113ecc740dcca448cfcb53642b959cd494942c6732db6590cb066" Nov 24 13:37:37 crc kubenswrapper[4790]: E1124 13:37:37.235554 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"deee7a82aed113ecc740dcca448cfcb53642b959cd494942c6732db6590cb066\": container with ID starting with deee7a82aed113ecc740dcca448cfcb53642b959cd494942c6732db6590cb066 not found: ID does not exist" containerID="deee7a82aed113ecc740dcca448cfcb53642b959cd494942c6732db6590cb066" Nov 24 13:37:37 crc kubenswrapper[4790]: I1124 13:37:37.235596 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"deee7a82aed113ecc740dcca448cfcb53642b959cd494942c6732db6590cb066"} err="failed to get container status \"deee7a82aed113ecc740dcca448cfcb53642b959cd494942c6732db6590cb066\": rpc error: code = NotFound desc = could not find container \"deee7a82aed113ecc740dcca448cfcb53642b959cd494942c6732db6590cb066\": container with ID starting with deee7a82aed113ecc740dcca448cfcb53642b959cd494942c6732db6590cb066 not found: ID does not exist" Nov 24 13:37:38 crc kubenswrapper[4790]: I1124 13:37:38.177212 4790 generic.go:334] "Generic (PLEG): container finished" podID="3add0b78-c01f-43d6-b9a4-85ae9a461710" containerID="f3de1fac7cf5b2507bb1c3d3eeb3d1a3ddf1321ae52a67b9e7e55e6c2212f7c0" exitCode=0 Nov 24 13:37:38 crc kubenswrapper[4790]: I1124 13:37:38.177258 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbvgd" event={"ID":"3add0b78-c01f-43d6-b9a4-85ae9a461710","Type":"ContainerDied","Data":"f3de1fac7cf5b2507bb1c3d3eeb3d1a3ddf1321ae52a67b9e7e55e6c2212f7c0"} Nov 24 13:37:38 crc kubenswrapper[4790]: I1124 13:37:38.324471 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d4e45de-9768-40e9-9c75-6abba8b38559" path="/var/lib/kubelet/pods/4d4e45de-9768-40e9-9c75-6abba8b38559/volumes" Nov 24 13:37:38 crc kubenswrapper[4790]: E1124 13:37:38.574146 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3 is running failed: container process not found" containerID="e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 13:37:38 crc kubenswrapper[4790]: E1124 13:37:38.574811 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3 is running failed: container process not found" containerID="e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 13:37:38 crc kubenswrapper[4790]: E1124 13:37:38.575450 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3 is running failed: container process not found" containerID="e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 13:37:38 crc kubenswrapper[4790]: E1124 13:37:38.575494 4790 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-28cwj" podUID="4d6d5d88-adda-4679-80de-0df0e0847bcb" containerName="ovsdb-server" Nov 24 13:37:38 crc kubenswrapper[4790]: E1124 13:37:38.575665 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="46182cf796e680922c3d2c23de93f19051dee361d1797948d997345bb2652287" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 13:37:38 crc kubenswrapper[4790]: E1124 13:37:38.577795 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="46182cf796e680922c3d2c23de93f19051dee361d1797948d997345bb2652287" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 13:37:38 crc kubenswrapper[4790]: E1124 13:37:38.579256 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="46182cf796e680922c3d2c23de93f19051dee361d1797948d997345bb2652287" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 13:37:38 crc kubenswrapper[4790]: E1124 13:37:38.579352 4790 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-28cwj" podUID="4d6d5d88-adda-4679-80de-0df0e0847bcb" containerName="ovs-vswitchd" Nov 24 13:37:39 crc kubenswrapper[4790]: I1124 13:37:39.186984 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbvgd" event={"ID":"3add0b78-c01f-43d6-b9a4-85ae9a461710","Type":"ContainerStarted","Data":"ac48177c4c2a0074d2f547c976b08222a6f22c678deaac4beb8834f738cfdcfb"} Nov 24 13:37:39 crc kubenswrapper[4790]: I1124 13:37:39.212011 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dbvgd" podStartSLOduration=2.6973336310000002 podStartE2EDuration="5.211981972s" podCreationTimestamp="2025-11-24 13:37:34 +0000 UTC" firstStartedPulling="2025-11-24 13:37:36.157594148 +0000 UTC m=+1504.537487810" lastFinishedPulling="2025-11-24 13:37:38.672242489 +0000 UTC m=+1507.052136151" observedRunningTime="2025-11-24 13:37:39.204509008 +0000 UTC m=+1507.584402670" watchObservedRunningTime="2025-11-24 13:37:39.211981972 +0000 UTC m=+1507.591875654" Nov 24 13:37:43 crc kubenswrapper[4790]: E1124 13:37:43.574622 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3 is running failed: container process not found" containerID="e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 13:37:43 crc kubenswrapper[4790]: E1124 13:37:43.575547 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3 is running failed: container process not found" containerID="e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 13:37:43 crc kubenswrapper[4790]: E1124 13:37:43.575619 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="46182cf796e680922c3d2c23de93f19051dee361d1797948d997345bb2652287" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 13:37:43 crc kubenswrapper[4790]: E1124 13:37:43.576080 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3 is running failed: container process not found" containerID="e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 13:37:43 crc kubenswrapper[4790]: E1124 13:37:43.576112 4790 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-28cwj" podUID="4d6d5d88-adda-4679-80de-0df0e0847bcb" containerName="ovsdb-server" Nov 24 13:37:43 crc kubenswrapper[4790]: E1124 13:37:43.576730 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="46182cf796e680922c3d2c23de93f19051dee361d1797948d997345bb2652287" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 13:37:43 crc kubenswrapper[4790]: E1124 13:37:43.577873 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="46182cf796e680922c3d2c23de93f19051dee361d1797948d997345bb2652287" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 13:37:43 crc kubenswrapper[4790]: E1124 13:37:43.577922 4790 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-28cwj" podUID="4d6d5d88-adda-4679-80de-0df0e0847bcb" containerName="ovs-vswitchd" Nov 24 13:37:44 crc kubenswrapper[4790]: I1124 13:37:44.784753 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dbvgd" Nov 24 13:37:44 crc kubenswrapper[4790]: I1124 13:37:44.784795 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dbvgd" Nov 24 13:37:44 crc kubenswrapper[4790]: I1124 13:37:44.825978 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dbvgd" Nov 24 13:37:45 crc kubenswrapper[4790]: I1124 13:37:45.308064 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dbvgd" Nov 24 13:37:45 crc kubenswrapper[4790]: I1124 13:37:45.347763 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dbvgd"] Nov 24 13:37:47 crc kubenswrapper[4790]: I1124 13:37:47.277767 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dbvgd" podUID="3add0b78-c01f-43d6-b9a4-85ae9a461710" containerName="registry-server" containerID="cri-o://ac48177c4c2a0074d2f547c976b08222a6f22c678deaac4beb8834f738cfdcfb" gracePeriod=2 Nov 24 13:37:47 crc kubenswrapper[4790]: I1124 13:37:47.716993 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dbvgd" Nov 24 13:37:47 crc kubenswrapper[4790]: I1124 13:37:47.855374 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3add0b78-c01f-43d6-b9a4-85ae9a461710-utilities\") pod \"3add0b78-c01f-43d6-b9a4-85ae9a461710\" (UID: \"3add0b78-c01f-43d6-b9a4-85ae9a461710\") " Nov 24 13:37:47 crc kubenswrapper[4790]: I1124 13:37:47.855491 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvfdz\" (UniqueName: \"kubernetes.io/projected/3add0b78-c01f-43d6-b9a4-85ae9a461710-kube-api-access-qvfdz\") pod \"3add0b78-c01f-43d6-b9a4-85ae9a461710\" (UID: \"3add0b78-c01f-43d6-b9a4-85ae9a461710\") " Nov 24 13:37:47 crc kubenswrapper[4790]: I1124 13:37:47.855608 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3add0b78-c01f-43d6-b9a4-85ae9a461710-catalog-content\") pod \"3add0b78-c01f-43d6-b9a4-85ae9a461710\" (UID: \"3add0b78-c01f-43d6-b9a4-85ae9a461710\") " Nov 24 13:37:47 crc kubenswrapper[4790]: I1124 13:37:47.856479 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3add0b78-c01f-43d6-b9a4-85ae9a461710-utilities" (OuterVolumeSpecName: "utilities") pod "3add0b78-c01f-43d6-b9a4-85ae9a461710" (UID: "3add0b78-c01f-43d6-b9a4-85ae9a461710"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:47 crc kubenswrapper[4790]: I1124 13:37:47.864086 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3add0b78-c01f-43d6-b9a4-85ae9a461710-kube-api-access-qvfdz" (OuterVolumeSpecName: "kube-api-access-qvfdz") pod "3add0b78-c01f-43d6-b9a4-85ae9a461710" (UID: "3add0b78-c01f-43d6-b9a4-85ae9a461710"). InnerVolumeSpecName "kube-api-access-qvfdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:47 crc kubenswrapper[4790]: I1124 13:37:47.867293 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-28cwj_4d6d5d88-adda-4679-80de-0df0e0847bcb/ovs-vswitchd/0.log" Nov 24 13:37:47 crc kubenswrapper[4790]: I1124 13:37:47.868181 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-28cwj" Nov 24 13:37:47 crc kubenswrapper[4790]: I1124 13:37:47.914932 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3add0b78-c01f-43d6-b9a4-85ae9a461710-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3add0b78-c01f-43d6-b9a4-85ae9a461710" (UID: "3add0b78-c01f-43d6-b9a4-85ae9a461710"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:47 crc kubenswrapper[4790]: I1124 13:37:47.956827 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/4d6d5d88-adda-4679-80de-0df0e0847bcb-etc-ovs\") pod \"4d6d5d88-adda-4679-80de-0df0e0847bcb\" (UID: \"4d6d5d88-adda-4679-80de-0df0e0847bcb\") " Nov 24 13:37:47 crc kubenswrapper[4790]: I1124 13:37:47.956983 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4d6d5d88-adda-4679-80de-0df0e0847bcb-scripts\") pod \"4d6d5d88-adda-4679-80de-0df0e0847bcb\" (UID: \"4d6d5d88-adda-4679-80de-0df0e0847bcb\") " Nov 24 13:37:47 crc kubenswrapper[4790]: I1124 13:37:47.957050 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/4d6d5d88-adda-4679-80de-0df0e0847bcb-var-log\") pod \"4d6d5d88-adda-4679-80de-0df0e0847bcb\" (UID: \"4d6d5d88-adda-4679-80de-0df0e0847bcb\") " Nov 24 13:37:47 crc kubenswrapper[4790]: I1124 13:37:47.957074 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4d6d5d88-adda-4679-80de-0df0e0847bcb-var-run\") pod \"4d6d5d88-adda-4679-80de-0df0e0847bcb\" (UID: \"4d6d5d88-adda-4679-80de-0df0e0847bcb\") " Nov 24 13:37:47 crc kubenswrapper[4790]: I1124 13:37:47.957051 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d6d5d88-adda-4679-80de-0df0e0847bcb-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "4d6d5d88-adda-4679-80de-0df0e0847bcb" (UID: "4d6d5d88-adda-4679-80de-0df0e0847bcb"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:37:47 crc kubenswrapper[4790]: I1124 13:37:47.957095 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/4d6d5d88-adda-4679-80de-0df0e0847bcb-var-lib\") pod \"4d6d5d88-adda-4679-80de-0df0e0847bcb\" (UID: \"4d6d5d88-adda-4679-80de-0df0e0847bcb\") " Nov 24 13:37:47 crc kubenswrapper[4790]: I1124 13:37:47.957114 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d6d5d88-adda-4679-80de-0df0e0847bcb-var-log" (OuterVolumeSpecName: "var-log") pod "4d6d5d88-adda-4679-80de-0df0e0847bcb" (UID: "4d6d5d88-adda-4679-80de-0df0e0847bcb"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:37:47 crc kubenswrapper[4790]: I1124 13:37:47.957133 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d6d5d88-adda-4679-80de-0df0e0847bcb-var-run" (OuterVolumeSpecName: "var-run") pod "4d6d5d88-adda-4679-80de-0df0e0847bcb" (UID: "4d6d5d88-adda-4679-80de-0df0e0847bcb"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:37:47 crc kubenswrapper[4790]: I1124 13:37:47.957150 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2g5hd\" (UniqueName: \"kubernetes.io/projected/4d6d5d88-adda-4679-80de-0df0e0847bcb-kube-api-access-2g5hd\") pod \"4d6d5d88-adda-4679-80de-0df0e0847bcb\" (UID: \"4d6d5d88-adda-4679-80de-0df0e0847bcb\") " Nov 24 13:37:47 crc kubenswrapper[4790]: I1124 13:37:47.957157 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d6d5d88-adda-4679-80de-0df0e0847bcb-var-lib" (OuterVolumeSpecName: "var-lib") pod "4d6d5d88-adda-4679-80de-0df0e0847bcb" (UID: "4d6d5d88-adda-4679-80de-0df0e0847bcb"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:37:47 crc kubenswrapper[4790]: I1124 13:37:47.957421 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3add0b78-c01f-43d6-b9a4-85ae9a461710-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:47 crc kubenswrapper[4790]: I1124 13:37:47.957452 4790 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/4d6d5d88-adda-4679-80de-0df0e0847bcb-var-log\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:47 crc kubenswrapper[4790]: I1124 13:37:47.957460 4790 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4d6d5d88-adda-4679-80de-0df0e0847bcb-var-run\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:47 crc kubenswrapper[4790]: I1124 13:37:47.957468 4790 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/4d6d5d88-adda-4679-80de-0df0e0847bcb-var-lib\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:47 crc kubenswrapper[4790]: I1124 13:37:47.957477 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvfdz\" (UniqueName: \"kubernetes.io/projected/3add0b78-c01f-43d6-b9a4-85ae9a461710-kube-api-access-qvfdz\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:47 crc kubenswrapper[4790]: I1124 13:37:47.957487 4790 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/4d6d5d88-adda-4679-80de-0df0e0847bcb-etc-ovs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:47 crc kubenswrapper[4790]: I1124 13:37:47.957498 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3add0b78-c01f-43d6-b9a4-85ae9a461710-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:47 crc kubenswrapper[4790]: I1124 13:37:47.958835 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d6d5d88-adda-4679-80de-0df0e0847bcb-scripts" (OuterVolumeSpecName: "scripts") pod "4d6d5d88-adda-4679-80de-0df0e0847bcb" (UID: "4d6d5d88-adda-4679-80de-0df0e0847bcb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:47 crc kubenswrapper[4790]: I1124 13:37:47.960166 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d6d5d88-adda-4679-80de-0df0e0847bcb-kube-api-access-2g5hd" (OuterVolumeSpecName: "kube-api-access-2g5hd") pod "4d6d5d88-adda-4679-80de-0df0e0847bcb" (UID: "4d6d5d88-adda-4679-80de-0df0e0847bcb"). InnerVolumeSpecName "kube-api-access-2g5hd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.058807 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2g5hd\" (UniqueName: \"kubernetes.io/projected/4d6d5d88-adda-4679-80de-0df0e0847bcb-kube-api-access-2g5hd\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.058855 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4d6d5d88-adda-4679-80de-0df0e0847bcb-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.288928 4790 generic.go:334] "Generic (PLEG): container finished" podID="3add0b78-c01f-43d6-b9a4-85ae9a461710" containerID="ac48177c4c2a0074d2f547c976b08222a6f22c678deaac4beb8834f738cfdcfb" exitCode=0 Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.288996 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dbvgd" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.289032 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbvgd" event={"ID":"3add0b78-c01f-43d6-b9a4-85ae9a461710","Type":"ContainerDied","Data":"ac48177c4c2a0074d2f547c976b08222a6f22c678deaac4beb8834f738cfdcfb"} Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.289088 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbvgd" event={"ID":"3add0b78-c01f-43d6-b9a4-85ae9a461710","Type":"ContainerDied","Data":"ddaa9122ab0292a4bf3bf1414ef53bd78dfefc8ca572f0607efdae1e76ea8e7b"} Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.289119 4790 scope.go:117] "RemoveContainer" containerID="ac48177c4c2a0074d2f547c976b08222a6f22c678deaac4beb8834f738cfdcfb" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.291104 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-28cwj_4d6d5d88-adda-4679-80de-0df0e0847bcb/ovs-vswitchd/0.log" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.292144 4790 generic.go:334] "Generic (PLEG): container finished" podID="4d6d5d88-adda-4679-80de-0df0e0847bcb" containerID="46182cf796e680922c3d2c23de93f19051dee361d1797948d997345bb2652287" exitCode=137 Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.292179 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-28cwj" event={"ID":"4d6d5d88-adda-4679-80de-0df0e0847bcb","Type":"ContainerDied","Data":"46182cf796e680922c3d2c23de93f19051dee361d1797948d997345bb2652287"} Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.292200 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-28cwj" event={"ID":"4d6d5d88-adda-4679-80de-0df0e0847bcb","Type":"ContainerDied","Data":"dbc82dc390d1fdd1005a9ed6190bb3b0ffc70585af9231cca01bec44931c045e"} Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.292216 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-28cwj" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.314918 4790 scope.go:117] "RemoveContainer" containerID="f3de1fac7cf5b2507bb1c3d3eeb3d1a3ddf1321ae52a67b9e7e55e6c2212f7c0" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.340645 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dbvgd"] Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.346336 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dbvgd"] Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.353805 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-28cwj"] Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.361001 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-28cwj"] Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.366101 4790 scope.go:117] "RemoveContainer" containerID="b97fef37ff4b08e344426b0d24b01d10f65c44609a58e019873fccb890284cae" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.454679 4790 scope.go:117] "RemoveContainer" containerID="ac48177c4c2a0074d2f547c976b08222a6f22c678deaac4beb8834f738cfdcfb" Nov 24 13:37:48 crc kubenswrapper[4790]: E1124 13:37:48.455725 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac48177c4c2a0074d2f547c976b08222a6f22c678deaac4beb8834f738cfdcfb\": container with ID starting with ac48177c4c2a0074d2f547c976b08222a6f22c678deaac4beb8834f738cfdcfb not found: ID does not exist" containerID="ac48177c4c2a0074d2f547c976b08222a6f22c678deaac4beb8834f738cfdcfb" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.455779 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac48177c4c2a0074d2f547c976b08222a6f22c678deaac4beb8834f738cfdcfb"} err="failed to get container status \"ac48177c4c2a0074d2f547c976b08222a6f22c678deaac4beb8834f738cfdcfb\": rpc error: code = NotFound desc = could not find container \"ac48177c4c2a0074d2f547c976b08222a6f22c678deaac4beb8834f738cfdcfb\": container with ID starting with ac48177c4c2a0074d2f547c976b08222a6f22c678deaac4beb8834f738cfdcfb not found: ID does not exist" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.455812 4790 scope.go:117] "RemoveContainer" containerID="f3de1fac7cf5b2507bb1c3d3eeb3d1a3ddf1321ae52a67b9e7e55e6c2212f7c0" Nov 24 13:37:48 crc kubenswrapper[4790]: E1124 13:37:48.456339 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3de1fac7cf5b2507bb1c3d3eeb3d1a3ddf1321ae52a67b9e7e55e6c2212f7c0\": container with ID starting with f3de1fac7cf5b2507bb1c3d3eeb3d1a3ddf1321ae52a67b9e7e55e6c2212f7c0 not found: ID does not exist" containerID="f3de1fac7cf5b2507bb1c3d3eeb3d1a3ddf1321ae52a67b9e7e55e6c2212f7c0" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.456361 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3de1fac7cf5b2507bb1c3d3eeb3d1a3ddf1321ae52a67b9e7e55e6c2212f7c0"} err="failed to get container status \"f3de1fac7cf5b2507bb1c3d3eeb3d1a3ddf1321ae52a67b9e7e55e6c2212f7c0\": rpc error: code = NotFound desc = could not find container \"f3de1fac7cf5b2507bb1c3d3eeb3d1a3ddf1321ae52a67b9e7e55e6c2212f7c0\": container with ID starting with f3de1fac7cf5b2507bb1c3d3eeb3d1a3ddf1321ae52a67b9e7e55e6c2212f7c0 not found: ID does not exist" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.456374 4790 scope.go:117] "RemoveContainer" containerID="b97fef37ff4b08e344426b0d24b01d10f65c44609a58e019873fccb890284cae" Nov 24 13:37:48 crc kubenswrapper[4790]: E1124 13:37:48.456616 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b97fef37ff4b08e344426b0d24b01d10f65c44609a58e019873fccb890284cae\": container with ID starting with b97fef37ff4b08e344426b0d24b01d10f65c44609a58e019873fccb890284cae not found: ID does not exist" containerID="b97fef37ff4b08e344426b0d24b01d10f65c44609a58e019873fccb890284cae" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.456637 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b97fef37ff4b08e344426b0d24b01d10f65c44609a58e019873fccb890284cae"} err="failed to get container status \"b97fef37ff4b08e344426b0d24b01d10f65c44609a58e019873fccb890284cae\": rpc error: code = NotFound desc = could not find container \"b97fef37ff4b08e344426b0d24b01d10f65c44609a58e019873fccb890284cae\": container with ID starting with b97fef37ff4b08e344426b0d24b01d10f65c44609a58e019873fccb890284cae not found: ID does not exist" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.456649 4790 scope.go:117] "RemoveContainer" containerID="46182cf796e680922c3d2c23de93f19051dee361d1797948d997345bb2652287" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.479447 4790 scope.go:117] "RemoveContainer" containerID="e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.504273 4790 scope.go:117] "RemoveContainer" containerID="d8259b93cb0ee7e3a787272e19de83ebfd12257522109ed7dfb63ae1d006ab09" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.551506 4790 scope.go:117] "RemoveContainer" containerID="46182cf796e680922c3d2c23de93f19051dee361d1797948d997345bb2652287" Nov 24 13:37:48 crc kubenswrapper[4790]: E1124 13:37:48.552010 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46182cf796e680922c3d2c23de93f19051dee361d1797948d997345bb2652287\": container with ID starting with 46182cf796e680922c3d2c23de93f19051dee361d1797948d997345bb2652287 not found: ID does not exist" containerID="46182cf796e680922c3d2c23de93f19051dee361d1797948d997345bb2652287" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.552088 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46182cf796e680922c3d2c23de93f19051dee361d1797948d997345bb2652287"} err="failed to get container status \"46182cf796e680922c3d2c23de93f19051dee361d1797948d997345bb2652287\": rpc error: code = NotFound desc = could not find container \"46182cf796e680922c3d2c23de93f19051dee361d1797948d997345bb2652287\": container with ID starting with 46182cf796e680922c3d2c23de93f19051dee361d1797948d997345bb2652287 not found: ID does not exist" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.552137 4790 scope.go:117] "RemoveContainer" containerID="e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3" Nov 24 13:37:48 crc kubenswrapper[4790]: E1124 13:37:48.552630 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3\": container with ID starting with e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3 not found: ID does not exist" containerID="e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.552652 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3"} err="failed to get container status \"e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3\": rpc error: code = NotFound desc = could not find container \"e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3\": container with ID starting with e934d8d9dab2b84e4d6cbe80aaf8f5a23f2e21d0a45e04917adafb37dd4f0fb3 not found: ID does not exist" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.552670 4790 scope.go:117] "RemoveContainer" containerID="d8259b93cb0ee7e3a787272e19de83ebfd12257522109ed7dfb63ae1d006ab09" Nov 24 13:37:48 crc kubenswrapper[4790]: E1124 13:37:48.553267 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8259b93cb0ee7e3a787272e19de83ebfd12257522109ed7dfb63ae1d006ab09\": container with ID starting with d8259b93cb0ee7e3a787272e19de83ebfd12257522109ed7dfb63ae1d006ab09 not found: ID does not exist" containerID="d8259b93cb0ee7e3a787272e19de83ebfd12257522109ed7dfb63ae1d006ab09" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.553303 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8259b93cb0ee7e3a787272e19de83ebfd12257522109ed7dfb63ae1d006ab09"} err="failed to get container status \"d8259b93cb0ee7e3a787272e19de83ebfd12257522109ed7dfb63ae1d006ab09\": rpc error: code = NotFound desc = could not find container \"d8259b93cb0ee7e3a787272e19de83ebfd12257522109ed7dfb63ae1d006ab09\": container with ID starting with d8259b93cb0ee7e3a787272e19de83ebfd12257522109ed7dfb63ae1d006ab09 not found: ID does not exist" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.711503 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.869540 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/63ce5126-cdab-4cdc-968a-acc26c6caf87-cache\") pod \"63ce5126-cdab-4cdc-968a-acc26c6caf87\" (UID: \"63ce5126-cdab-4cdc-968a-acc26c6caf87\") " Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.869712 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/63ce5126-cdab-4cdc-968a-acc26c6caf87-etc-swift\") pod \"63ce5126-cdab-4cdc-968a-acc26c6caf87\" (UID: \"63ce5126-cdab-4cdc-968a-acc26c6caf87\") " Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.869759 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5jk4\" (UniqueName: \"kubernetes.io/projected/63ce5126-cdab-4cdc-968a-acc26c6caf87-kube-api-access-n5jk4\") pod \"63ce5126-cdab-4cdc-968a-acc26c6caf87\" (UID: \"63ce5126-cdab-4cdc-968a-acc26c6caf87\") " Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.869791 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"63ce5126-cdab-4cdc-968a-acc26c6caf87\" (UID: \"63ce5126-cdab-4cdc-968a-acc26c6caf87\") " Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.869823 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/63ce5126-cdab-4cdc-968a-acc26c6caf87-lock\") pod \"63ce5126-cdab-4cdc-968a-acc26c6caf87\" (UID: \"63ce5126-cdab-4cdc-968a-acc26c6caf87\") " Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.870244 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63ce5126-cdab-4cdc-968a-acc26c6caf87-cache" (OuterVolumeSpecName: "cache") pod "63ce5126-cdab-4cdc-968a-acc26c6caf87" (UID: "63ce5126-cdab-4cdc-968a-acc26c6caf87"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.870287 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63ce5126-cdab-4cdc-968a-acc26c6caf87-lock" (OuterVolumeSpecName: "lock") pod "63ce5126-cdab-4cdc-968a-acc26c6caf87" (UID: "63ce5126-cdab-4cdc-968a-acc26c6caf87"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.874330 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63ce5126-cdab-4cdc-968a-acc26c6caf87-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "63ce5126-cdab-4cdc-968a-acc26c6caf87" (UID: "63ce5126-cdab-4cdc-968a-acc26c6caf87"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.876031 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "swift") pod "63ce5126-cdab-4cdc-968a-acc26c6caf87" (UID: "63ce5126-cdab-4cdc-968a-acc26c6caf87"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.877183 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63ce5126-cdab-4cdc-968a-acc26c6caf87-kube-api-access-n5jk4" (OuterVolumeSpecName: "kube-api-access-n5jk4") pod "63ce5126-cdab-4cdc-968a-acc26c6caf87" (UID: "63ce5126-cdab-4cdc-968a-acc26c6caf87"). InnerVolumeSpecName "kube-api-access-n5jk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.971801 4790 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/63ce5126-cdab-4cdc-968a-acc26c6caf87-cache\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.971846 4790 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/63ce5126-cdab-4cdc-968a-acc26c6caf87-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.971858 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5jk4\" (UniqueName: \"kubernetes.io/projected/63ce5126-cdab-4cdc-968a-acc26c6caf87-kube-api-access-n5jk4\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.971903 4790 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.971916 4790 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/63ce5126-cdab-4cdc-968a-acc26c6caf87-lock\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:48 crc kubenswrapper[4790]: I1124 13:37:48.985373 4790 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.073026 4790 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.311147 4790 generic.go:334] "Generic (PLEG): container finished" podID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerID="b7e87dedf7cf3279e0cd55e6ed5885652aac1f5a97cb5db5d608e23cb07f7dc4" exitCode=137 Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.311219 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63ce5126-cdab-4cdc-968a-acc26c6caf87","Type":"ContainerDied","Data":"b7e87dedf7cf3279e0cd55e6ed5885652aac1f5a97cb5db5d608e23cb07f7dc4"} Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.311255 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63ce5126-cdab-4cdc-968a-acc26c6caf87","Type":"ContainerDied","Data":"141512db35629d7b3d10ad972903d58fad16a3d6829a08d08bd2eddc6dffaf3d"} Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.311289 4790 scope.go:117] "RemoveContainer" containerID="b7e87dedf7cf3279e0cd55e6ed5885652aac1f5a97cb5db5d608e23cb07f7dc4" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.311617 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.336000 4790 scope.go:117] "RemoveContainer" containerID="7d0ce12b3e15290c2d364a5d9b9eb84a5315993172cd4dded95d480621cad447" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.353772 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.358058 4790 scope.go:117] "RemoveContainer" containerID="c2b3e00524e13f39df13207ebcb11976c8f2a411cb2cf3f7922cf052a42ee301" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.361014 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.373996 4790 scope.go:117] "RemoveContainer" containerID="578285e8c4f233014516fdb85665a982ed0526d96886404924f292cd08a273f2" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.391493 4790 scope.go:117] "RemoveContainer" containerID="b471731f94d67b5b7e6123cfb1a2a5c179f4f9ec0d47b37c19690321409b4711" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.410297 4790 scope.go:117] "RemoveContainer" containerID="d23adafa420c3790e972ee180913c8f5a9453ac2fffd584b5675a60e8c5a418a" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.449824 4790 scope.go:117] "RemoveContainer" containerID="ca07858ec58622d34fa209b9cd36987ee9d21a975afb78c55e5eebc7646c3b1a" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.471553 4790 scope.go:117] "RemoveContainer" containerID="09f43e9aef5856b296a5ef62f8bc35fdd45d245130733c53d7e35377992a954b" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.486903 4790 scope.go:117] "RemoveContainer" containerID="c08ee946a46469851d49ebb63d97976087646f7f74bc00d1aba900aeb1b780eb" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.502248 4790 scope.go:117] "RemoveContainer" containerID="891362e93b7800953eed0ac6306309542589bcd92b30aa5ef910fdb82394fd3d" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.522864 4790 scope.go:117] "RemoveContainer" containerID="3122135b69c368fd946107ad82977792d44cf6ae8bfae58942ad6057773fa295" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.540917 4790 scope.go:117] "RemoveContainer" containerID="a9637198b1cf74e71f1558bec0b1a2c24a29c81f04cc33568637f3f6a30f9759" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.560973 4790 scope.go:117] "RemoveContainer" containerID="b3904f2398af315163d3ca972a5d587c4d9aa8034f444bc2b60c9b44aa285d9d" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.584094 4790 scope.go:117] "RemoveContainer" containerID="54b741f22ad4d511172f550e71786b130b052e565c75124d9ad50b4066de4deb" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.603184 4790 scope.go:117] "RemoveContainer" containerID="23963383dd944af8fb81fedb5a24119440fcae6b836f04def348d26a2568c09f" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.622537 4790 scope.go:117] "RemoveContainer" containerID="b7e87dedf7cf3279e0cd55e6ed5885652aac1f5a97cb5db5d608e23cb07f7dc4" Nov 24 13:37:49 crc kubenswrapper[4790]: E1124 13:37:49.623056 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7e87dedf7cf3279e0cd55e6ed5885652aac1f5a97cb5db5d608e23cb07f7dc4\": container with ID starting with b7e87dedf7cf3279e0cd55e6ed5885652aac1f5a97cb5db5d608e23cb07f7dc4 not found: ID does not exist" containerID="b7e87dedf7cf3279e0cd55e6ed5885652aac1f5a97cb5db5d608e23cb07f7dc4" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.623132 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7e87dedf7cf3279e0cd55e6ed5885652aac1f5a97cb5db5d608e23cb07f7dc4"} err="failed to get container status \"b7e87dedf7cf3279e0cd55e6ed5885652aac1f5a97cb5db5d608e23cb07f7dc4\": rpc error: code = NotFound desc = could not find container \"b7e87dedf7cf3279e0cd55e6ed5885652aac1f5a97cb5db5d608e23cb07f7dc4\": container with ID starting with b7e87dedf7cf3279e0cd55e6ed5885652aac1f5a97cb5db5d608e23cb07f7dc4 not found: ID does not exist" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.623159 4790 scope.go:117] "RemoveContainer" containerID="7d0ce12b3e15290c2d364a5d9b9eb84a5315993172cd4dded95d480621cad447" Nov 24 13:37:49 crc kubenswrapper[4790]: E1124 13:37:49.623454 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d0ce12b3e15290c2d364a5d9b9eb84a5315993172cd4dded95d480621cad447\": container with ID starting with 7d0ce12b3e15290c2d364a5d9b9eb84a5315993172cd4dded95d480621cad447 not found: ID does not exist" containerID="7d0ce12b3e15290c2d364a5d9b9eb84a5315993172cd4dded95d480621cad447" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.623482 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d0ce12b3e15290c2d364a5d9b9eb84a5315993172cd4dded95d480621cad447"} err="failed to get container status \"7d0ce12b3e15290c2d364a5d9b9eb84a5315993172cd4dded95d480621cad447\": rpc error: code = NotFound desc = could not find container \"7d0ce12b3e15290c2d364a5d9b9eb84a5315993172cd4dded95d480621cad447\": container with ID starting with 7d0ce12b3e15290c2d364a5d9b9eb84a5315993172cd4dded95d480621cad447 not found: ID does not exist" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.623497 4790 scope.go:117] "RemoveContainer" containerID="c2b3e00524e13f39df13207ebcb11976c8f2a411cb2cf3f7922cf052a42ee301" Nov 24 13:37:49 crc kubenswrapper[4790]: E1124 13:37:49.623903 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2b3e00524e13f39df13207ebcb11976c8f2a411cb2cf3f7922cf052a42ee301\": container with ID starting with c2b3e00524e13f39df13207ebcb11976c8f2a411cb2cf3f7922cf052a42ee301 not found: ID does not exist" containerID="c2b3e00524e13f39df13207ebcb11976c8f2a411cb2cf3f7922cf052a42ee301" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.623966 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2b3e00524e13f39df13207ebcb11976c8f2a411cb2cf3f7922cf052a42ee301"} err="failed to get container status \"c2b3e00524e13f39df13207ebcb11976c8f2a411cb2cf3f7922cf052a42ee301\": rpc error: code = NotFound desc = could not find container \"c2b3e00524e13f39df13207ebcb11976c8f2a411cb2cf3f7922cf052a42ee301\": container with ID starting with c2b3e00524e13f39df13207ebcb11976c8f2a411cb2cf3f7922cf052a42ee301 not found: ID does not exist" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.624012 4790 scope.go:117] "RemoveContainer" containerID="578285e8c4f233014516fdb85665a982ed0526d96886404924f292cd08a273f2" Nov 24 13:37:49 crc kubenswrapper[4790]: E1124 13:37:49.624340 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"578285e8c4f233014516fdb85665a982ed0526d96886404924f292cd08a273f2\": container with ID starting with 578285e8c4f233014516fdb85665a982ed0526d96886404924f292cd08a273f2 not found: ID does not exist" containerID="578285e8c4f233014516fdb85665a982ed0526d96886404924f292cd08a273f2" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.624364 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"578285e8c4f233014516fdb85665a982ed0526d96886404924f292cd08a273f2"} err="failed to get container status \"578285e8c4f233014516fdb85665a982ed0526d96886404924f292cd08a273f2\": rpc error: code = NotFound desc = could not find container \"578285e8c4f233014516fdb85665a982ed0526d96886404924f292cd08a273f2\": container with ID starting with 578285e8c4f233014516fdb85665a982ed0526d96886404924f292cd08a273f2 not found: ID does not exist" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.624380 4790 scope.go:117] "RemoveContainer" containerID="b471731f94d67b5b7e6123cfb1a2a5c179f4f9ec0d47b37c19690321409b4711" Nov 24 13:37:49 crc kubenswrapper[4790]: E1124 13:37:49.624675 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b471731f94d67b5b7e6123cfb1a2a5c179f4f9ec0d47b37c19690321409b4711\": container with ID starting with b471731f94d67b5b7e6123cfb1a2a5c179f4f9ec0d47b37c19690321409b4711 not found: ID does not exist" containerID="b471731f94d67b5b7e6123cfb1a2a5c179f4f9ec0d47b37c19690321409b4711" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.624715 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b471731f94d67b5b7e6123cfb1a2a5c179f4f9ec0d47b37c19690321409b4711"} err="failed to get container status \"b471731f94d67b5b7e6123cfb1a2a5c179f4f9ec0d47b37c19690321409b4711\": rpc error: code = NotFound desc = could not find container \"b471731f94d67b5b7e6123cfb1a2a5c179f4f9ec0d47b37c19690321409b4711\": container with ID starting with b471731f94d67b5b7e6123cfb1a2a5c179f4f9ec0d47b37c19690321409b4711 not found: ID does not exist" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.624742 4790 scope.go:117] "RemoveContainer" containerID="d23adafa420c3790e972ee180913c8f5a9453ac2fffd584b5675a60e8c5a418a" Nov 24 13:37:49 crc kubenswrapper[4790]: E1124 13:37:49.625049 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d23adafa420c3790e972ee180913c8f5a9453ac2fffd584b5675a60e8c5a418a\": container with ID starting with d23adafa420c3790e972ee180913c8f5a9453ac2fffd584b5675a60e8c5a418a not found: ID does not exist" containerID="d23adafa420c3790e972ee180913c8f5a9453ac2fffd584b5675a60e8c5a418a" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.625084 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d23adafa420c3790e972ee180913c8f5a9453ac2fffd584b5675a60e8c5a418a"} err="failed to get container status \"d23adafa420c3790e972ee180913c8f5a9453ac2fffd584b5675a60e8c5a418a\": rpc error: code = NotFound desc = could not find container \"d23adafa420c3790e972ee180913c8f5a9453ac2fffd584b5675a60e8c5a418a\": container with ID starting with d23adafa420c3790e972ee180913c8f5a9453ac2fffd584b5675a60e8c5a418a not found: ID does not exist" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.625106 4790 scope.go:117] "RemoveContainer" containerID="ca07858ec58622d34fa209b9cd36987ee9d21a975afb78c55e5eebc7646c3b1a" Nov 24 13:37:49 crc kubenswrapper[4790]: E1124 13:37:49.625418 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca07858ec58622d34fa209b9cd36987ee9d21a975afb78c55e5eebc7646c3b1a\": container with ID starting with ca07858ec58622d34fa209b9cd36987ee9d21a975afb78c55e5eebc7646c3b1a not found: ID does not exist" containerID="ca07858ec58622d34fa209b9cd36987ee9d21a975afb78c55e5eebc7646c3b1a" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.625440 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca07858ec58622d34fa209b9cd36987ee9d21a975afb78c55e5eebc7646c3b1a"} err="failed to get container status \"ca07858ec58622d34fa209b9cd36987ee9d21a975afb78c55e5eebc7646c3b1a\": rpc error: code = NotFound desc = could not find container \"ca07858ec58622d34fa209b9cd36987ee9d21a975afb78c55e5eebc7646c3b1a\": container with ID starting with ca07858ec58622d34fa209b9cd36987ee9d21a975afb78c55e5eebc7646c3b1a not found: ID does not exist" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.625454 4790 scope.go:117] "RemoveContainer" containerID="09f43e9aef5856b296a5ef62f8bc35fdd45d245130733c53d7e35377992a954b" Nov 24 13:37:49 crc kubenswrapper[4790]: E1124 13:37:49.625684 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09f43e9aef5856b296a5ef62f8bc35fdd45d245130733c53d7e35377992a954b\": container with ID starting with 09f43e9aef5856b296a5ef62f8bc35fdd45d245130733c53d7e35377992a954b not found: ID does not exist" containerID="09f43e9aef5856b296a5ef62f8bc35fdd45d245130733c53d7e35377992a954b" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.625710 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09f43e9aef5856b296a5ef62f8bc35fdd45d245130733c53d7e35377992a954b"} err="failed to get container status \"09f43e9aef5856b296a5ef62f8bc35fdd45d245130733c53d7e35377992a954b\": rpc error: code = NotFound desc = could not find container \"09f43e9aef5856b296a5ef62f8bc35fdd45d245130733c53d7e35377992a954b\": container with ID starting with 09f43e9aef5856b296a5ef62f8bc35fdd45d245130733c53d7e35377992a954b not found: ID does not exist" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.625727 4790 scope.go:117] "RemoveContainer" containerID="c08ee946a46469851d49ebb63d97976087646f7f74bc00d1aba900aeb1b780eb" Nov 24 13:37:49 crc kubenswrapper[4790]: E1124 13:37:49.625977 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c08ee946a46469851d49ebb63d97976087646f7f74bc00d1aba900aeb1b780eb\": container with ID starting with c08ee946a46469851d49ebb63d97976087646f7f74bc00d1aba900aeb1b780eb not found: ID does not exist" containerID="c08ee946a46469851d49ebb63d97976087646f7f74bc00d1aba900aeb1b780eb" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.626030 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c08ee946a46469851d49ebb63d97976087646f7f74bc00d1aba900aeb1b780eb"} err="failed to get container status \"c08ee946a46469851d49ebb63d97976087646f7f74bc00d1aba900aeb1b780eb\": rpc error: code = NotFound desc = could not find container \"c08ee946a46469851d49ebb63d97976087646f7f74bc00d1aba900aeb1b780eb\": container with ID starting with c08ee946a46469851d49ebb63d97976087646f7f74bc00d1aba900aeb1b780eb not found: ID does not exist" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.626054 4790 scope.go:117] "RemoveContainer" containerID="891362e93b7800953eed0ac6306309542589bcd92b30aa5ef910fdb82394fd3d" Nov 24 13:37:49 crc kubenswrapper[4790]: E1124 13:37:49.626319 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"891362e93b7800953eed0ac6306309542589bcd92b30aa5ef910fdb82394fd3d\": container with ID starting with 891362e93b7800953eed0ac6306309542589bcd92b30aa5ef910fdb82394fd3d not found: ID does not exist" containerID="891362e93b7800953eed0ac6306309542589bcd92b30aa5ef910fdb82394fd3d" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.626343 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"891362e93b7800953eed0ac6306309542589bcd92b30aa5ef910fdb82394fd3d"} err="failed to get container status \"891362e93b7800953eed0ac6306309542589bcd92b30aa5ef910fdb82394fd3d\": rpc error: code = NotFound desc = could not find container \"891362e93b7800953eed0ac6306309542589bcd92b30aa5ef910fdb82394fd3d\": container with ID starting with 891362e93b7800953eed0ac6306309542589bcd92b30aa5ef910fdb82394fd3d not found: ID does not exist" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.626358 4790 scope.go:117] "RemoveContainer" containerID="3122135b69c368fd946107ad82977792d44cf6ae8bfae58942ad6057773fa295" Nov 24 13:37:49 crc kubenswrapper[4790]: E1124 13:37:49.626580 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3122135b69c368fd946107ad82977792d44cf6ae8bfae58942ad6057773fa295\": container with ID starting with 3122135b69c368fd946107ad82977792d44cf6ae8bfae58942ad6057773fa295 not found: ID does not exist" containerID="3122135b69c368fd946107ad82977792d44cf6ae8bfae58942ad6057773fa295" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.626609 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3122135b69c368fd946107ad82977792d44cf6ae8bfae58942ad6057773fa295"} err="failed to get container status \"3122135b69c368fd946107ad82977792d44cf6ae8bfae58942ad6057773fa295\": rpc error: code = NotFound desc = could not find container \"3122135b69c368fd946107ad82977792d44cf6ae8bfae58942ad6057773fa295\": container with ID starting with 3122135b69c368fd946107ad82977792d44cf6ae8bfae58942ad6057773fa295 not found: ID does not exist" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.626660 4790 scope.go:117] "RemoveContainer" containerID="a9637198b1cf74e71f1558bec0b1a2c24a29c81f04cc33568637f3f6a30f9759" Nov 24 13:37:49 crc kubenswrapper[4790]: E1124 13:37:49.626916 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9637198b1cf74e71f1558bec0b1a2c24a29c81f04cc33568637f3f6a30f9759\": container with ID starting with a9637198b1cf74e71f1558bec0b1a2c24a29c81f04cc33568637f3f6a30f9759 not found: ID does not exist" containerID="a9637198b1cf74e71f1558bec0b1a2c24a29c81f04cc33568637f3f6a30f9759" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.626943 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9637198b1cf74e71f1558bec0b1a2c24a29c81f04cc33568637f3f6a30f9759"} err="failed to get container status \"a9637198b1cf74e71f1558bec0b1a2c24a29c81f04cc33568637f3f6a30f9759\": rpc error: code = NotFound desc = could not find container \"a9637198b1cf74e71f1558bec0b1a2c24a29c81f04cc33568637f3f6a30f9759\": container with ID starting with a9637198b1cf74e71f1558bec0b1a2c24a29c81f04cc33568637f3f6a30f9759 not found: ID does not exist" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.626960 4790 scope.go:117] "RemoveContainer" containerID="b3904f2398af315163d3ca972a5d587c4d9aa8034f444bc2b60c9b44aa285d9d" Nov 24 13:37:49 crc kubenswrapper[4790]: E1124 13:37:49.627193 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3904f2398af315163d3ca972a5d587c4d9aa8034f444bc2b60c9b44aa285d9d\": container with ID starting with b3904f2398af315163d3ca972a5d587c4d9aa8034f444bc2b60c9b44aa285d9d not found: ID does not exist" containerID="b3904f2398af315163d3ca972a5d587c4d9aa8034f444bc2b60c9b44aa285d9d" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.627216 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3904f2398af315163d3ca972a5d587c4d9aa8034f444bc2b60c9b44aa285d9d"} err="failed to get container status \"b3904f2398af315163d3ca972a5d587c4d9aa8034f444bc2b60c9b44aa285d9d\": rpc error: code = NotFound desc = could not find container \"b3904f2398af315163d3ca972a5d587c4d9aa8034f444bc2b60c9b44aa285d9d\": container with ID starting with b3904f2398af315163d3ca972a5d587c4d9aa8034f444bc2b60c9b44aa285d9d not found: ID does not exist" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.627232 4790 scope.go:117] "RemoveContainer" containerID="54b741f22ad4d511172f550e71786b130b052e565c75124d9ad50b4066de4deb" Nov 24 13:37:49 crc kubenswrapper[4790]: E1124 13:37:49.627516 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54b741f22ad4d511172f550e71786b130b052e565c75124d9ad50b4066de4deb\": container with ID starting with 54b741f22ad4d511172f550e71786b130b052e565c75124d9ad50b4066de4deb not found: ID does not exist" containerID="54b741f22ad4d511172f550e71786b130b052e565c75124d9ad50b4066de4deb" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.627539 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54b741f22ad4d511172f550e71786b130b052e565c75124d9ad50b4066de4deb"} err="failed to get container status \"54b741f22ad4d511172f550e71786b130b052e565c75124d9ad50b4066de4deb\": rpc error: code = NotFound desc = could not find container \"54b741f22ad4d511172f550e71786b130b052e565c75124d9ad50b4066de4deb\": container with ID starting with 54b741f22ad4d511172f550e71786b130b052e565c75124d9ad50b4066de4deb not found: ID does not exist" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.627555 4790 scope.go:117] "RemoveContainer" containerID="23963383dd944af8fb81fedb5a24119440fcae6b836f04def348d26a2568c09f" Nov 24 13:37:49 crc kubenswrapper[4790]: E1124 13:37:49.627783 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23963383dd944af8fb81fedb5a24119440fcae6b836f04def348d26a2568c09f\": container with ID starting with 23963383dd944af8fb81fedb5a24119440fcae6b836f04def348d26a2568c09f not found: ID does not exist" containerID="23963383dd944af8fb81fedb5a24119440fcae6b836f04def348d26a2568c09f" Nov 24 13:37:49 crc kubenswrapper[4790]: I1124 13:37:49.627809 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23963383dd944af8fb81fedb5a24119440fcae6b836f04def348d26a2568c09f"} err="failed to get container status \"23963383dd944af8fb81fedb5a24119440fcae6b836f04def348d26a2568c09f\": rpc error: code = NotFound desc = could not find container \"23963383dd944af8fb81fedb5a24119440fcae6b836f04def348d26a2568c09f\": container with ID starting with 23963383dd944af8fb81fedb5a24119440fcae6b836f04def348d26a2568c09f not found: ID does not exist" Nov 24 13:37:50 crc kubenswrapper[4790]: I1124 13:37:50.324282 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3add0b78-c01f-43d6-b9a4-85ae9a461710" path="/var/lib/kubelet/pods/3add0b78-c01f-43d6-b9a4-85ae9a461710/volumes" Nov 24 13:37:50 crc kubenswrapper[4790]: I1124 13:37:50.325307 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d6d5d88-adda-4679-80de-0df0e0847bcb" path="/var/lib/kubelet/pods/4d6d5d88-adda-4679-80de-0df0e0847bcb/volumes" Nov 24 13:37:50 crc kubenswrapper[4790]: I1124 13:37:50.325988 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" path="/var/lib/kubelet/pods/63ce5126-cdab-4cdc-968a-acc26c6caf87/volumes" Nov 24 13:37:50 crc kubenswrapper[4790]: I1124 13:37:50.685716 4790 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","podf8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort podf8be67b1-1f0a-4ddb-82f1-906fbbf6ac7c] : Timed out while waiting for systemd to remove kubepods-besteffort-podf8be67b1_1f0a_4ddb_82f1_906fbbf6ac7c.slice" Nov 24 13:38:43 crc kubenswrapper[4790]: I1124 13:38:43.528262 4790 scope.go:117] "RemoveContainer" containerID="d03a01bc9550ec4a70da4d90c581b8b702ead5abff5bbee34ae2f7e4898a6b05" Nov 24 13:38:43 crc kubenswrapper[4790]: I1124 13:38:43.557749 4790 scope.go:117] "RemoveContainer" containerID="66e7fe0a6042aa8f4790b3a22b5fd027de4aef544a4fdf42731e2c840ffd4cf2" Nov 24 13:38:43 crc kubenswrapper[4790]: I1124 13:38:43.586104 4790 scope.go:117] "RemoveContainer" containerID="569c178276bd7deaa7ff29136e16e02f63f79ee3c9e59aae1ab276151057e59e" Nov 24 13:38:43 crc kubenswrapper[4790]: I1124 13:38:43.622481 4790 scope.go:117] "RemoveContainer" containerID="169855f1ee71f39432956cfa61f341f14290544d596d6f99664ca020d983892a" Nov 24 13:38:43 crc kubenswrapper[4790]: I1124 13:38:43.647808 4790 scope.go:117] "RemoveContainer" containerID="e2a955b6a87db6036e41d12bbbbf6162cfd89d2b0bb6d235a325aafe02fc51ca" Nov 24 13:38:43 crc kubenswrapper[4790]: I1124 13:38:43.675636 4790 scope.go:117] "RemoveContainer" containerID="8c7b1f8ad061938b73c9c94a4ef4bfe3303a40fe30294f1a92f9a97da7d69a3e" Nov 24 13:38:43 crc kubenswrapper[4790]: I1124 13:38:43.696570 4790 scope.go:117] "RemoveContainer" containerID="7976704e51ab332073a36acf2fadd9c4a5760783d9fc94a68e424b56646d342e" Nov 24 13:38:43 crc kubenswrapper[4790]: I1124 13:38:43.717415 4790 scope.go:117] "RemoveContainer" containerID="489363e812cab6ae0ad579bf95670ebf0ce1d05dfc24e582a5b1891a50435049" Nov 24 13:38:43 crc kubenswrapper[4790]: I1124 13:38:43.748856 4790 scope.go:117] "RemoveContainer" containerID="ad9512a19986dcbf34683ed61d8fb2ead36a72f27f516136c67f069e610076cf" Nov 24 13:38:43 crc kubenswrapper[4790]: I1124 13:38:43.793375 4790 scope.go:117] "RemoveContainer" containerID="2d459199980f3814665ba6c630a8bf310ef6efb41a3f2368c9edbeb004c62897" Nov 24 13:38:43 crc kubenswrapper[4790]: I1124 13:38:43.823744 4790 scope.go:117] "RemoveContainer" containerID="eab70d8990c01fc8d4d58c7a888bef266d435891d2a2affc19e30cc46f143f8a" Nov 24 13:39:43 crc kubenswrapper[4790]: I1124 13:39:43.938501 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:39:43 crc kubenswrapper[4790]: I1124 13:39:43.938933 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:39:44 crc kubenswrapper[4790]: I1124 13:39:44.057097 4790 scope.go:117] "RemoveContainer" containerID="147300b703a9e3feb3b2b670b1b8b7a5e8e9817668e01ca6b34aaceb64126d0d" Nov 24 13:39:44 crc kubenswrapper[4790]: I1124 13:39:44.097579 4790 scope.go:117] "RemoveContainer" containerID="f4493288d1d8aa4448843cf480499bfdbc6588351887c3da3c497a1384b9f401" Nov 24 13:39:44 crc kubenswrapper[4790]: I1124 13:39:44.135770 4790 scope.go:117] "RemoveContainer" containerID="4e4860f11978bcccff47dec8491807d068e5a1c38c0c9e47770b0e3c265e76f7" Nov 24 13:39:44 crc kubenswrapper[4790]: I1124 13:39:44.158265 4790 scope.go:117] "RemoveContainer" containerID="1cdea1e01c6022d8752115c1dbd9b9cd87f839392ad6920de4f0d664706cf7e2" Nov 24 13:39:44 crc kubenswrapper[4790]: I1124 13:39:44.198329 4790 scope.go:117] "RemoveContainer" containerID="a92ec3cde75257483e0a8e30ef0c90a883c89a010f5b7ce77f5f75db1ebfab68" Nov 24 13:39:44 crc kubenswrapper[4790]: I1124 13:39:44.230825 4790 scope.go:117] "RemoveContainer" containerID="2c6c3b4bcd309e9bbfe6304002cfac512cde3ab86acfec0e937b2af1a62d7ef2" Nov 24 13:39:44 crc kubenswrapper[4790]: I1124 13:39:44.259099 4790 scope.go:117] "RemoveContainer" containerID="cc10010eead4062b594ab18421178a8aa7d04ff50aeac7bc2ef5d887df82e45c" Nov 24 13:39:44 crc kubenswrapper[4790]: I1124 13:39:44.289410 4790 scope.go:117] "RemoveContainer" containerID="8e0b47428124851c84a07cc1375f10d5a8146d6c54766f0c2f6d6b97e1386e0f" Nov 24 13:39:44 crc kubenswrapper[4790]: I1124 13:39:44.320515 4790 scope.go:117] "RemoveContainer" containerID="cf5790b66698273019b8f16b970dbdbffc943209fdca11bb7061f46cb7418370" Nov 24 13:39:44 crc kubenswrapper[4790]: I1124 13:39:44.357362 4790 scope.go:117] "RemoveContainer" containerID="5da42ceece6abe14f86b2b9b233db8ef83224e750e7ddae6f4086373f0ef1127" Nov 24 13:39:44 crc kubenswrapper[4790]: I1124 13:39:44.373585 4790 scope.go:117] "RemoveContainer" containerID="3ba62451b6b4646aa85e946c0688dc0e1bb36dca41e3cd99558373a169b528f8" Nov 24 13:39:44 crc kubenswrapper[4790]: I1124 13:39:44.394417 4790 scope.go:117] "RemoveContainer" containerID="d7420f1383875329ff22e1b2127be77d6b3b85841829420c95d5e97849a31ae4" Nov 24 13:39:44 crc kubenswrapper[4790]: I1124 13:39:44.410950 4790 scope.go:117] "RemoveContainer" containerID="215ea9dbc1c1da6e42c9fce4e68a1f3b57fb111ed332cc067f9ccf293122ec60" Nov 24 13:39:44 crc kubenswrapper[4790]: I1124 13:39:44.430414 4790 scope.go:117] "RemoveContainer" containerID="1080017b2c864b1e6ae073dbeb6f0f37bbb80e9d0e0f77e3c13cf7a6fbb41f70" Nov 24 13:39:44 crc kubenswrapper[4790]: I1124 13:39:44.459736 4790 scope.go:117] "RemoveContainer" containerID="66f5ea547ee89aef5100d0c34e763ebff615070dd32d43842f8ab4c00325722c" Nov 24 13:39:44 crc kubenswrapper[4790]: I1124 13:39:44.477981 4790 scope.go:117] "RemoveContainer" containerID="5b52e40c19049d488eb9bca51afc3064c16e94264443da3acadd6e63b4102e3b" Nov 24 13:40:13 crc kubenswrapper[4790]: I1124 13:40:13.938741 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:40:13 crc kubenswrapper[4790]: I1124 13:40:13.939307 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:40:43 crc kubenswrapper[4790]: I1124 13:40:43.939153 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:40:43 crc kubenswrapper[4790]: I1124 13:40:43.939681 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:40:43 crc kubenswrapper[4790]: I1124 13:40:43.939733 4790 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 13:40:43 crc kubenswrapper[4790]: I1124 13:40:43.940455 4790 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0e39e707a2686ae4ea3e31dae8bfe7595a875cadfa529bb73dd02722c8084b9e"} pod="openshift-machine-config-operator/machine-config-daemon-xz49t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 13:40:43 crc kubenswrapper[4790]: I1124 13:40:43.940508 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" containerID="cri-o://0e39e707a2686ae4ea3e31dae8bfe7595a875cadfa529bb73dd02722c8084b9e" gracePeriod=600 Nov 24 13:40:44 crc kubenswrapper[4790]: E1124 13:40:44.074000 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:40:44 crc kubenswrapper[4790]: I1124 13:40:44.674446 4790 scope.go:117] "RemoveContainer" containerID="d45798957a393567d315abf87c95bac024e5a20af18679a63ea606b0ce35e55e" Nov 24 13:40:44 crc kubenswrapper[4790]: I1124 13:40:44.708954 4790 scope.go:117] "RemoveContainer" containerID="d54ef9df4953d96ae51b1400969f9bf8a83e542d0262e0cb2e78aa602b92ae44" Nov 24 13:40:44 crc kubenswrapper[4790]: I1124 13:40:44.808711 4790 generic.go:334] "Generic (PLEG): container finished" podID="0d73b133-48f0-455f-8f6a-742e633f631a" containerID="0e39e707a2686ae4ea3e31dae8bfe7595a875cadfa529bb73dd02722c8084b9e" exitCode=0 Nov 24 13:40:44 crc kubenswrapper[4790]: I1124 13:40:44.808800 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerDied","Data":"0e39e707a2686ae4ea3e31dae8bfe7595a875cadfa529bb73dd02722c8084b9e"} Nov 24 13:40:44 crc kubenswrapper[4790]: I1124 13:40:44.809084 4790 scope.go:117] "RemoveContainer" containerID="8544370c5579779a1c61d01762c06ad3cb058bc1368085e6b5750e300ddbf449" Nov 24 13:40:44 crc kubenswrapper[4790]: I1124 13:40:44.809641 4790 scope.go:117] "RemoveContainer" containerID="0e39e707a2686ae4ea3e31dae8bfe7595a875cadfa529bb73dd02722c8084b9e" Nov 24 13:40:44 crc kubenswrapper[4790]: E1124 13:40:44.810086 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:40:57 crc kubenswrapper[4790]: I1124 13:40:57.315112 4790 scope.go:117] "RemoveContainer" containerID="0e39e707a2686ae4ea3e31dae8bfe7595a875cadfa529bb73dd02722c8084b9e" Nov 24 13:40:57 crc kubenswrapper[4790]: E1124 13:40:57.316398 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:41:12 crc kubenswrapper[4790]: I1124 13:41:12.320654 4790 scope.go:117] "RemoveContainer" containerID="0e39e707a2686ae4ea3e31dae8bfe7595a875cadfa529bb73dd02722c8084b9e" Nov 24 13:41:12 crc kubenswrapper[4790]: E1124 13:41:12.321283 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:41:27 crc kubenswrapper[4790]: I1124 13:41:27.314722 4790 scope.go:117] "RemoveContainer" containerID="0e39e707a2686ae4ea3e31dae8bfe7595a875cadfa529bb73dd02722c8084b9e" Nov 24 13:41:27 crc kubenswrapper[4790]: E1124 13:41:27.315485 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:41:38 crc kubenswrapper[4790]: I1124 13:41:38.314701 4790 scope.go:117] "RemoveContainer" containerID="0e39e707a2686ae4ea3e31dae8bfe7595a875cadfa529bb73dd02722c8084b9e" Nov 24 13:41:38 crc kubenswrapper[4790]: E1124 13:41:38.315673 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:41:44 crc kubenswrapper[4790]: I1124 13:41:44.803562 4790 scope.go:117] "RemoveContainer" containerID="82a2835a200a5782ef25e715c28b47b86306977b719a0b3e260fe10df86d647c" Nov 24 13:41:44 crc kubenswrapper[4790]: I1124 13:41:44.865670 4790 scope.go:117] "RemoveContainer" containerID="1ce47c3d1e1960a319da58bba01d387b5ca2f97570e10c6078a4e5e9f4e33040" Nov 24 13:41:44 crc kubenswrapper[4790]: I1124 13:41:44.884016 4790 scope.go:117] "RemoveContainer" containerID="f9e4f39302ec3fef535c4d941a4deb560f0999311d673edd1adcbdd80b3f09fa" Nov 24 13:41:44 crc kubenswrapper[4790]: I1124 13:41:44.906549 4790 scope.go:117] "RemoveContainer" containerID="99ebe4f359d29dd9dadccf947e629f98fdb917d1ee6164cc3c18632420cc9128" Nov 24 13:41:44 crc kubenswrapper[4790]: I1124 13:41:44.943058 4790 scope.go:117] "RemoveContainer" containerID="13ff2a6bf7913ecdacaf058d5f469f70a2b6c6e055b55607d3a365e30aaf58d9" Nov 24 13:41:44 crc kubenswrapper[4790]: I1124 13:41:44.983493 4790 scope.go:117] "RemoveContainer" containerID="9a7848c9a709ae9ddd3f799bb95363637b474062de0cad438d73ae38726dea36" Nov 24 13:41:44 crc kubenswrapper[4790]: I1124 13:41:44.998153 4790 scope.go:117] "RemoveContainer" containerID="73738b88616d24d9d38a7dde3fbe8536abf120812c9d699ed3f1dae25824b738" Nov 24 13:41:45 crc kubenswrapper[4790]: I1124 13:41:45.029921 4790 scope.go:117] "RemoveContainer" containerID="4c149f4ac8815779e6d7cb3c9896f38e1386454d59e01a601935ffd9f4da470b" Nov 24 13:41:45 crc kubenswrapper[4790]: I1124 13:41:45.043792 4790 scope.go:117] "RemoveContainer" containerID="559b39f1cb8cb24bfe5aee6eebf0bd4b5f7ea54f034e3fab965ea795a0c3e81d" Nov 24 13:41:49 crc kubenswrapper[4790]: I1124 13:41:49.314765 4790 scope.go:117] "RemoveContainer" containerID="0e39e707a2686ae4ea3e31dae8bfe7595a875cadfa529bb73dd02722c8084b9e" Nov 24 13:41:49 crc kubenswrapper[4790]: E1124 13:41:49.315504 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:42:01 crc kubenswrapper[4790]: I1124 13:42:01.315348 4790 scope.go:117] "RemoveContainer" containerID="0e39e707a2686ae4ea3e31dae8bfe7595a875cadfa529bb73dd02722c8084b9e" Nov 24 13:42:01 crc kubenswrapper[4790]: E1124 13:42:01.316074 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:42:13 crc kubenswrapper[4790]: I1124 13:42:13.315226 4790 scope.go:117] "RemoveContainer" containerID="0e39e707a2686ae4ea3e31dae8bfe7595a875cadfa529bb73dd02722c8084b9e" Nov 24 13:42:13 crc kubenswrapper[4790]: E1124 13:42:13.316003 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:42:28 crc kubenswrapper[4790]: I1124 13:42:28.314136 4790 scope.go:117] "RemoveContainer" containerID="0e39e707a2686ae4ea3e31dae8bfe7595a875cadfa529bb73dd02722c8084b9e" Nov 24 13:42:28 crc kubenswrapper[4790]: E1124 13:42:28.314808 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:42:43 crc kubenswrapper[4790]: I1124 13:42:43.314234 4790 scope.go:117] "RemoveContainer" containerID="0e39e707a2686ae4ea3e31dae8bfe7595a875cadfa529bb73dd02722c8084b9e" Nov 24 13:42:43 crc kubenswrapper[4790]: E1124 13:42:43.314862 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:42:45 crc kubenswrapper[4790]: I1124 13:42:45.183129 4790 scope.go:117] "RemoveContainer" containerID="2461d3602b2db4efea3d420eb4e138e451c6048fe390dc3d4b479c8e5a584628" Nov 24 13:42:58 crc kubenswrapper[4790]: I1124 13:42:58.314474 4790 scope.go:117] "RemoveContainer" containerID="0e39e707a2686ae4ea3e31dae8bfe7595a875cadfa529bb73dd02722c8084b9e" Nov 24 13:42:58 crc kubenswrapper[4790]: E1124 13:42:58.315284 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:43:09 crc kubenswrapper[4790]: I1124 13:43:09.314795 4790 scope.go:117] "RemoveContainer" containerID="0e39e707a2686ae4ea3e31dae8bfe7595a875cadfa529bb73dd02722c8084b9e" Nov 24 13:43:09 crc kubenswrapper[4790]: E1124 13:43:09.315603 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:43:23 crc kubenswrapper[4790]: I1124 13:43:23.314293 4790 scope.go:117] "RemoveContainer" containerID="0e39e707a2686ae4ea3e31dae8bfe7595a875cadfa529bb73dd02722c8084b9e" Nov 24 13:43:23 crc kubenswrapper[4790]: E1124 13:43:23.315979 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:43:38 crc kubenswrapper[4790]: I1124 13:43:38.314732 4790 scope.go:117] "RemoveContainer" containerID="0e39e707a2686ae4ea3e31dae8bfe7595a875cadfa529bb73dd02722c8084b9e" Nov 24 13:43:38 crc kubenswrapper[4790]: E1124 13:43:38.315720 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:43:45 crc kubenswrapper[4790]: I1124 13:43:45.257795 4790 scope.go:117] "RemoveContainer" containerID="8f04a853d966a711ba4815311768aa3e3f970e90d66e7b8eada63e04a25b99f2" Nov 24 13:43:45 crc kubenswrapper[4790]: I1124 13:43:45.281113 4790 scope.go:117] "RemoveContainer" containerID="ce87db4947942f196094e19acebc86c6075c26a08fa34743ddae754d898e5938" Nov 24 13:43:45 crc kubenswrapper[4790]: I1124 13:43:45.318791 4790 scope.go:117] "RemoveContainer" containerID="9add9dc189b49006e7c669c2830732c7550855a82a0cb3552df890b1e86ae064" Nov 24 13:43:45 crc kubenswrapper[4790]: I1124 13:43:45.338458 4790 scope.go:117] "RemoveContainer" containerID="49eb35e3cab0fee6082b7f8ae966428557fc6407e903621826b70befb0756a67" Nov 24 13:43:45 crc kubenswrapper[4790]: I1124 13:43:45.359777 4790 scope.go:117] "RemoveContainer" containerID="93d8fe1bd57c2a04bf2a36122f12d68f6d367d8fe7a99c2f41e508c7615f68a4" Nov 24 13:43:45 crc kubenswrapper[4790]: I1124 13:43:45.381611 4790 scope.go:117] "RemoveContainer" containerID="8ed1474926abc1218abe6afb94d8fe954c475e5e1cfcc6a22fae6792a274eb6d" Nov 24 13:43:45 crc kubenswrapper[4790]: I1124 13:43:45.403425 4790 scope.go:117] "RemoveContainer" containerID="cfc22827f7595f647048bc319094a1a8720f3f344164f8db5947014cfaf081fc" Nov 24 13:43:51 crc kubenswrapper[4790]: I1124 13:43:51.314734 4790 scope.go:117] "RemoveContainer" containerID="0e39e707a2686ae4ea3e31dae8bfe7595a875cadfa529bb73dd02722c8084b9e" Nov 24 13:43:51 crc kubenswrapper[4790]: E1124 13:43:51.315565 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:44:05 crc kubenswrapper[4790]: I1124 13:44:05.314997 4790 scope.go:117] "RemoveContainer" containerID="0e39e707a2686ae4ea3e31dae8bfe7595a875cadfa529bb73dd02722c8084b9e" Nov 24 13:44:05 crc kubenswrapper[4790]: E1124 13:44:05.315518 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:44:19 crc kubenswrapper[4790]: I1124 13:44:19.315168 4790 scope.go:117] "RemoveContainer" containerID="0e39e707a2686ae4ea3e31dae8bfe7595a875cadfa529bb73dd02722c8084b9e" Nov 24 13:44:19 crc kubenswrapper[4790]: E1124 13:44:19.315860 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:44:33 crc kubenswrapper[4790]: I1124 13:44:33.314500 4790 scope.go:117] "RemoveContainer" containerID="0e39e707a2686ae4ea3e31dae8bfe7595a875cadfa529bb73dd02722c8084b9e" Nov 24 13:44:33 crc kubenswrapper[4790]: E1124 13:44:33.315284 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:44:45 crc kubenswrapper[4790]: I1124 13:44:45.314489 4790 scope.go:117] "RemoveContainer" containerID="0e39e707a2686ae4ea3e31dae8bfe7595a875cadfa529bb73dd02722c8084b9e" Nov 24 13:44:45 crc kubenswrapper[4790]: E1124 13:44:45.315147 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:44:56 crc kubenswrapper[4790]: I1124 13:44:56.315616 4790 scope.go:117] "RemoveContainer" containerID="0e39e707a2686ae4ea3e31dae8bfe7595a875cadfa529bb73dd02722c8084b9e" Nov 24 13:44:56 crc kubenswrapper[4790]: E1124 13:44:56.316729 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.144513 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399865-tdf2x"] Nov 24 13:45:00 crc kubenswrapper[4790]: E1124 13:45:00.145247 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3add0b78-c01f-43d6-b9a4-85ae9a461710" containerName="extract-utilities" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.145265 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="3add0b78-c01f-43d6-b9a4-85ae9a461710" containerName="extract-utilities" Nov 24 13:45:00 crc kubenswrapper[4790]: E1124 13:45:00.145278 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="account-auditor" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.145298 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="account-auditor" Nov 24 13:45:00 crc kubenswrapper[4790]: E1124 13:45:00.145305 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="rsync" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.145312 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="rsync" Nov 24 13:45:00 crc kubenswrapper[4790]: E1124 13:45:00.145318 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d6d5d88-adda-4679-80de-0df0e0847bcb" containerName="ovsdb-server-init" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.145324 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d6d5d88-adda-4679-80de-0df0e0847bcb" containerName="ovsdb-server-init" Nov 24 13:45:00 crc kubenswrapper[4790]: E1124 13:45:00.145332 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="object-expirer" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.145339 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="object-expirer" Nov 24 13:45:00 crc kubenswrapper[4790]: E1124 13:45:00.145348 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="container-server" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.145358 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="container-server" Nov 24 13:45:00 crc kubenswrapper[4790]: E1124 13:45:00.145419 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="object-updater" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.145428 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="object-updater" Nov 24 13:45:00 crc kubenswrapper[4790]: E1124 13:45:00.145438 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="container-replicator" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.145446 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="container-replicator" Nov 24 13:45:00 crc kubenswrapper[4790]: E1124 13:45:00.145482 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="object-server" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.145490 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="object-server" Nov 24 13:45:00 crc kubenswrapper[4790]: E1124 13:45:00.145498 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="object-auditor" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.145504 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="object-auditor" Nov 24 13:45:00 crc kubenswrapper[4790]: E1124 13:45:00.145514 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d6d5d88-adda-4679-80de-0df0e0847bcb" containerName="ovsdb-server" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.145521 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d6d5d88-adda-4679-80de-0df0e0847bcb" containerName="ovsdb-server" Nov 24 13:45:00 crc kubenswrapper[4790]: E1124 13:45:00.145530 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="container-auditor" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.145556 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="container-auditor" Nov 24 13:45:00 crc kubenswrapper[4790]: E1124 13:45:00.145575 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="object-replicator" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.145581 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="object-replicator" Nov 24 13:45:00 crc kubenswrapper[4790]: E1124 13:45:00.145594 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="account-replicator" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.145603 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="account-replicator" Nov 24 13:45:00 crc kubenswrapper[4790]: E1124 13:45:00.145636 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d6d5d88-adda-4679-80de-0df0e0847bcb" containerName="ovs-vswitchd" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.145643 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d6d5d88-adda-4679-80de-0df0e0847bcb" containerName="ovs-vswitchd" Nov 24 13:45:00 crc kubenswrapper[4790]: E1124 13:45:00.145656 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="account-reaper" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.145664 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="account-reaper" Nov 24 13:45:00 crc kubenswrapper[4790]: E1124 13:45:00.145679 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d4e45de-9768-40e9-9c75-6abba8b38559" containerName="neutron-api" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.145686 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d4e45de-9768-40e9-9c75-6abba8b38559" containerName="neutron-api" Nov 24 13:45:00 crc kubenswrapper[4790]: E1124 13:45:00.145717 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3add0b78-c01f-43d6-b9a4-85ae9a461710" containerName="extract-content" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.145725 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="3add0b78-c01f-43d6-b9a4-85ae9a461710" containerName="extract-content" Nov 24 13:45:00 crc kubenswrapper[4790]: E1124 13:45:00.145736 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d4e45de-9768-40e9-9c75-6abba8b38559" containerName="neutron-httpd" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.145743 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d4e45de-9768-40e9-9c75-6abba8b38559" containerName="neutron-httpd" Nov 24 13:45:00 crc kubenswrapper[4790]: E1124 13:45:00.145755 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3add0b78-c01f-43d6-b9a4-85ae9a461710" containerName="registry-server" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.145762 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="3add0b78-c01f-43d6-b9a4-85ae9a461710" containerName="registry-server" Nov 24 13:45:00 crc kubenswrapper[4790]: E1124 13:45:00.145790 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="swift-recon-cron" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.145798 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="swift-recon-cron" Nov 24 13:45:00 crc kubenswrapper[4790]: E1124 13:45:00.145810 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="account-server" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.145816 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="account-server" Nov 24 13:45:00 crc kubenswrapper[4790]: E1124 13:45:00.145823 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="container-updater" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.145829 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="container-updater" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.146077 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="container-updater" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.146113 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="object-replicator" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.146122 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="rsync" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.146132 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d4e45de-9768-40e9-9c75-6abba8b38559" containerName="neutron-httpd" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.146141 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d6d5d88-adda-4679-80de-0df0e0847bcb" containerName="ovs-vswitchd" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.146148 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="account-server" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.146161 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="container-server" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.146170 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="object-updater" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.146224 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="object-auditor" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.146234 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d4e45de-9768-40e9-9c75-6abba8b38559" containerName="neutron-api" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.146241 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="container-auditor" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.146250 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="3add0b78-c01f-43d6-b9a4-85ae9a461710" containerName="registry-server" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.146283 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="account-replicator" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.146306 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="swift-recon-cron" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.146322 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="account-reaper" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.146332 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="container-replicator" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.146362 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="account-auditor" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.146372 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d6d5d88-adda-4679-80de-0df0e0847bcb" containerName="ovsdb-server" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.146380 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="object-expirer" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.146390 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="63ce5126-cdab-4cdc-968a-acc26c6caf87" containerName="object-server" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.153087 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tdf2x" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.155349 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.155406 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.155475 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399865-tdf2x"] Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.194705 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b-config-volume\") pod \"collect-profiles-29399865-tdf2x\" (UID: \"12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tdf2x" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.194756 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mxvp\" (UniqueName: \"kubernetes.io/projected/12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b-kube-api-access-8mxvp\") pod \"collect-profiles-29399865-tdf2x\" (UID: \"12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tdf2x" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.194931 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b-secret-volume\") pod \"collect-profiles-29399865-tdf2x\" (UID: \"12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tdf2x" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.296622 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b-secret-volume\") pod \"collect-profiles-29399865-tdf2x\" (UID: \"12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tdf2x" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.296697 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b-config-volume\") pod \"collect-profiles-29399865-tdf2x\" (UID: \"12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tdf2x" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.296722 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mxvp\" (UniqueName: \"kubernetes.io/projected/12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b-kube-api-access-8mxvp\") pod \"collect-profiles-29399865-tdf2x\" (UID: \"12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tdf2x" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.298067 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b-config-volume\") pod \"collect-profiles-29399865-tdf2x\" (UID: \"12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tdf2x" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.302523 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b-secret-volume\") pod \"collect-profiles-29399865-tdf2x\" (UID: \"12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tdf2x" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.326663 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mxvp\" (UniqueName: \"kubernetes.io/projected/12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b-kube-api-access-8mxvp\") pod \"collect-profiles-29399865-tdf2x\" (UID: \"12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tdf2x" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.482171 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tdf2x" Nov 24 13:45:00 crc kubenswrapper[4790]: I1124 13:45:00.903579 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399865-tdf2x"] Nov 24 13:45:01 crc kubenswrapper[4790]: I1124 13:45:01.078548 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tdf2x" event={"ID":"12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b","Type":"ContainerStarted","Data":"3f3ceba70968dd466f787b443c993444af51db55029a014cb89bd77d14331f7b"} Nov 24 13:45:01 crc kubenswrapper[4790]: I1124 13:45:01.078931 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tdf2x" event={"ID":"12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b","Type":"ContainerStarted","Data":"b9f3ce50eafeeb55c9fd965eed6aa91e0f272a395bc087f13db96fe10f7595ab"} Nov 24 13:45:01 crc kubenswrapper[4790]: I1124 13:45:01.095160 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tdf2x" podStartSLOduration=1.095141355 podStartE2EDuration="1.095141355s" podCreationTimestamp="2025-11-24 13:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:45:01.092039413 +0000 UTC m=+1949.471933095" watchObservedRunningTime="2025-11-24 13:45:01.095141355 +0000 UTC m=+1949.475035017" Nov 24 13:45:02 crc kubenswrapper[4790]: I1124 13:45:02.085981 4790 generic.go:334] "Generic (PLEG): container finished" podID="12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b" containerID="3f3ceba70968dd466f787b443c993444af51db55029a014cb89bd77d14331f7b" exitCode=0 Nov 24 13:45:02 crc kubenswrapper[4790]: I1124 13:45:02.086026 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tdf2x" event={"ID":"12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b","Type":"ContainerDied","Data":"3f3ceba70968dd466f787b443c993444af51db55029a014cb89bd77d14331f7b"} Nov 24 13:45:03 crc kubenswrapper[4790]: I1124 13:45:03.402379 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tdf2x" Nov 24 13:45:03 crc kubenswrapper[4790]: I1124 13:45:03.436381 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mxvp\" (UniqueName: \"kubernetes.io/projected/12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b-kube-api-access-8mxvp\") pod \"12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b\" (UID: \"12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b\") " Nov 24 13:45:03 crc kubenswrapper[4790]: I1124 13:45:03.436460 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b-config-volume\") pod \"12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b\" (UID: \"12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b\") " Nov 24 13:45:03 crc kubenswrapper[4790]: I1124 13:45:03.436591 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b-secret-volume\") pod \"12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b\" (UID: \"12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b\") " Nov 24 13:45:03 crc kubenswrapper[4790]: I1124 13:45:03.437257 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b-config-volume" (OuterVolumeSpecName: "config-volume") pod "12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b" (UID: "12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:45:03 crc kubenswrapper[4790]: I1124 13:45:03.441749 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b-kube-api-access-8mxvp" (OuterVolumeSpecName: "kube-api-access-8mxvp") pod "12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b" (UID: "12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b"). InnerVolumeSpecName "kube-api-access-8mxvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:45:03 crc kubenswrapper[4790]: I1124 13:45:03.442073 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b" (UID: "12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:45:03 crc kubenswrapper[4790]: I1124 13:45:03.538570 4790 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 13:45:03 crc kubenswrapper[4790]: I1124 13:45:03.538618 4790 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 13:45:03 crc kubenswrapper[4790]: I1124 13:45:03.538633 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mxvp\" (UniqueName: \"kubernetes.io/projected/12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b-kube-api-access-8mxvp\") on node \"crc\" DevicePath \"\"" Nov 24 13:45:04 crc kubenswrapper[4790]: I1124 13:45:04.102734 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tdf2x" event={"ID":"12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b","Type":"ContainerDied","Data":"b9f3ce50eafeeb55c9fd965eed6aa91e0f272a395bc087f13db96fe10f7595ab"} Nov 24 13:45:04 crc kubenswrapper[4790]: I1124 13:45:04.102796 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9f3ce50eafeeb55c9fd965eed6aa91e0f272a395bc087f13db96fe10f7595ab" Nov 24 13:45:04 crc kubenswrapper[4790]: I1124 13:45:04.102873 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tdf2x" Nov 24 13:45:09 crc kubenswrapper[4790]: I1124 13:45:09.315045 4790 scope.go:117] "RemoveContainer" containerID="0e39e707a2686ae4ea3e31dae8bfe7595a875cadfa529bb73dd02722c8084b9e" Nov 24 13:45:09 crc kubenswrapper[4790]: E1124 13:45:09.315611 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:45:20 crc kubenswrapper[4790]: I1124 13:45:20.315518 4790 scope.go:117] "RemoveContainer" containerID="0e39e707a2686ae4ea3e31dae8bfe7595a875cadfa529bb73dd02722c8084b9e" Nov 24 13:45:20 crc kubenswrapper[4790]: E1124 13:45:20.316412 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:45:31 crc kubenswrapper[4790]: I1124 13:45:31.314257 4790 scope.go:117] "RemoveContainer" containerID="0e39e707a2686ae4ea3e31dae8bfe7595a875cadfa529bb73dd02722c8084b9e" Nov 24 13:45:31 crc kubenswrapper[4790]: E1124 13:45:31.314854 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:45:45 crc kubenswrapper[4790]: I1124 13:45:45.314977 4790 scope.go:117] "RemoveContainer" containerID="0e39e707a2686ae4ea3e31dae8bfe7595a875cadfa529bb73dd02722c8084b9e" Nov 24 13:45:46 crc kubenswrapper[4790]: I1124 13:45:46.395032 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"d22c23244bf6119211e0ddeceeef48f1aa920868cf60d45e67811b7753e963eb"} Nov 24 13:46:15 crc kubenswrapper[4790]: I1124 13:46:15.428162 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vvpjq"] Nov 24 13:46:15 crc kubenswrapper[4790]: E1124 13:46:15.428966 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b" containerName="collect-profiles" Nov 24 13:46:15 crc kubenswrapper[4790]: I1124 13:46:15.428980 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b" containerName="collect-profiles" Nov 24 13:46:15 crc kubenswrapper[4790]: I1124 13:46:15.429111 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b" containerName="collect-profiles" Nov 24 13:46:15 crc kubenswrapper[4790]: I1124 13:46:15.430089 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vvpjq" Nov 24 13:46:15 crc kubenswrapper[4790]: I1124 13:46:15.442012 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vvpjq"] Nov 24 13:46:15 crc kubenswrapper[4790]: I1124 13:46:15.524439 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aafebbee-d404-4313-9898-d1baec8e31be-utilities\") pod \"redhat-operators-vvpjq\" (UID: \"aafebbee-d404-4313-9898-d1baec8e31be\") " pod="openshift-marketplace/redhat-operators-vvpjq" Nov 24 13:46:15 crc kubenswrapper[4790]: I1124 13:46:15.524484 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lp9ct\" (UniqueName: \"kubernetes.io/projected/aafebbee-d404-4313-9898-d1baec8e31be-kube-api-access-lp9ct\") pod \"redhat-operators-vvpjq\" (UID: \"aafebbee-d404-4313-9898-d1baec8e31be\") " pod="openshift-marketplace/redhat-operators-vvpjq" Nov 24 13:46:15 crc kubenswrapper[4790]: I1124 13:46:15.524652 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aafebbee-d404-4313-9898-d1baec8e31be-catalog-content\") pod \"redhat-operators-vvpjq\" (UID: \"aafebbee-d404-4313-9898-d1baec8e31be\") " pod="openshift-marketplace/redhat-operators-vvpjq" Nov 24 13:46:15 crc kubenswrapper[4790]: I1124 13:46:15.625580 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aafebbee-d404-4313-9898-d1baec8e31be-utilities\") pod \"redhat-operators-vvpjq\" (UID: \"aafebbee-d404-4313-9898-d1baec8e31be\") " pod="openshift-marketplace/redhat-operators-vvpjq" Nov 24 13:46:15 crc kubenswrapper[4790]: I1124 13:46:15.625629 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lp9ct\" (UniqueName: \"kubernetes.io/projected/aafebbee-d404-4313-9898-d1baec8e31be-kube-api-access-lp9ct\") pod \"redhat-operators-vvpjq\" (UID: \"aafebbee-d404-4313-9898-d1baec8e31be\") " pod="openshift-marketplace/redhat-operators-vvpjq" Nov 24 13:46:15 crc kubenswrapper[4790]: I1124 13:46:15.625672 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aafebbee-d404-4313-9898-d1baec8e31be-catalog-content\") pod \"redhat-operators-vvpjq\" (UID: \"aafebbee-d404-4313-9898-d1baec8e31be\") " pod="openshift-marketplace/redhat-operators-vvpjq" Nov 24 13:46:15 crc kubenswrapper[4790]: I1124 13:46:15.626246 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aafebbee-d404-4313-9898-d1baec8e31be-utilities\") pod \"redhat-operators-vvpjq\" (UID: \"aafebbee-d404-4313-9898-d1baec8e31be\") " pod="openshift-marketplace/redhat-operators-vvpjq" Nov 24 13:46:15 crc kubenswrapper[4790]: I1124 13:46:15.626278 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aafebbee-d404-4313-9898-d1baec8e31be-catalog-content\") pod \"redhat-operators-vvpjq\" (UID: \"aafebbee-d404-4313-9898-d1baec8e31be\") " pod="openshift-marketplace/redhat-operators-vvpjq" Nov 24 13:46:15 crc kubenswrapper[4790]: I1124 13:46:15.646116 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lp9ct\" (UniqueName: \"kubernetes.io/projected/aafebbee-d404-4313-9898-d1baec8e31be-kube-api-access-lp9ct\") pod \"redhat-operators-vvpjq\" (UID: \"aafebbee-d404-4313-9898-d1baec8e31be\") " pod="openshift-marketplace/redhat-operators-vvpjq" Nov 24 13:46:15 crc kubenswrapper[4790]: I1124 13:46:15.748094 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vvpjq" Nov 24 13:46:16 crc kubenswrapper[4790]: I1124 13:46:16.204221 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vvpjq"] Nov 24 13:46:16 crc kubenswrapper[4790]: E1124 13:46:16.546209 4790 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaafebbee_d404_4313_9898_d1baec8e31be.slice/crio-b77631e8b4a1e4d5176d3bcff651aea2e0c7c38c95f7be884e04648d6fa532c6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaafebbee_d404_4313_9898_d1baec8e31be.slice/crio-conmon-b77631e8b4a1e4d5176d3bcff651aea2e0c7c38c95f7be884e04648d6fa532c6.scope\": RecentStats: unable to find data in memory cache]" Nov 24 13:46:16 crc kubenswrapper[4790]: I1124 13:46:16.587973 4790 generic.go:334] "Generic (PLEG): container finished" podID="aafebbee-d404-4313-9898-d1baec8e31be" containerID="b77631e8b4a1e4d5176d3bcff651aea2e0c7c38c95f7be884e04648d6fa532c6" exitCode=0 Nov 24 13:46:16 crc kubenswrapper[4790]: I1124 13:46:16.588022 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vvpjq" event={"ID":"aafebbee-d404-4313-9898-d1baec8e31be","Type":"ContainerDied","Data":"b77631e8b4a1e4d5176d3bcff651aea2e0c7c38c95f7be884e04648d6fa532c6"} Nov 24 13:46:16 crc kubenswrapper[4790]: I1124 13:46:16.588058 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vvpjq" event={"ID":"aafebbee-d404-4313-9898-d1baec8e31be","Type":"ContainerStarted","Data":"9b26284e6a59ea690dcd17b8fc1075f56d7d6dc80507f0baa3ab59f1b3277f28"} Nov 24 13:46:16 crc kubenswrapper[4790]: I1124 13:46:16.590234 4790 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 13:46:18 crc kubenswrapper[4790]: I1124 13:46:18.606202 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vvpjq" event={"ID":"aafebbee-d404-4313-9898-d1baec8e31be","Type":"ContainerStarted","Data":"6a7bc6f92a5a7b82ec951f86bdd82158919f071b8af354e127d47565e5419a8b"} Nov 24 13:46:19 crc kubenswrapper[4790]: I1124 13:46:19.615468 4790 generic.go:334] "Generic (PLEG): container finished" podID="aafebbee-d404-4313-9898-d1baec8e31be" containerID="6a7bc6f92a5a7b82ec951f86bdd82158919f071b8af354e127d47565e5419a8b" exitCode=0 Nov 24 13:46:19 crc kubenswrapper[4790]: I1124 13:46:19.615582 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vvpjq" event={"ID":"aafebbee-d404-4313-9898-d1baec8e31be","Type":"ContainerDied","Data":"6a7bc6f92a5a7b82ec951f86bdd82158919f071b8af354e127d47565e5419a8b"} Nov 24 13:46:20 crc kubenswrapper[4790]: I1124 13:46:20.643347 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vvpjq" event={"ID":"aafebbee-d404-4313-9898-d1baec8e31be","Type":"ContainerStarted","Data":"7da17141a8af0fd7f716c504953498e62e118b8808bf77aa8282e5759fcf839d"} Nov 24 13:46:20 crc kubenswrapper[4790]: I1124 13:46:20.666553 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vvpjq" podStartSLOduration=2.095726778 podStartE2EDuration="5.666535353s" podCreationTimestamp="2025-11-24 13:46:15 +0000 UTC" firstStartedPulling="2025-11-24 13:46:16.590062408 +0000 UTC m=+2024.969956070" lastFinishedPulling="2025-11-24 13:46:20.160870953 +0000 UTC m=+2028.540764645" observedRunningTime="2025-11-24 13:46:20.660435498 +0000 UTC m=+2029.040329150" watchObservedRunningTime="2025-11-24 13:46:20.666535353 +0000 UTC m=+2029.046429015" Nov 24 13:46:25 crc kubenswrapper[4790]: I1124 13:46:25.749286 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vvpjq" Nov 24 13:46:25 crc kubenswrapper[4790]: I1124 13:46:25.749681 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vvpjq" Nov 24 13:46:25 crc kubenswrapper[4790]: I1124 13:46:25.801707 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vvpjq" Nov 24 13:46:26 crc kubenswrapper[4790]: I1124 13:46:26.720852 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vvpjq" Nov 24 13:46:26 crc kubenswrapper[4790]: I1124 13:46:26.760101 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vvpjq"] Nov 24 13:46:28 crc kubenswrapper[4790]: I1124 13:46:28.705253 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vvpjq" podUID="aafebbee-d404-4313-9898-d1baec8e31be" containerName="registry-server" containerID="cri-o://7da17141a8af0fd7f716c504953498e62e118b8808bf77aa8282e5759fcf839d" gracePeriod=2 Nov 24 13:46:30 crc kubenswrapper[4790]: I1124 13:46:30.237277 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vvpjq" Nov 24 13:46:30 crc kubenswrapper[4790]: I1124 13:46:30.383964 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aafebbee-d404-4313-9898-d1baec8e31be-utilities\") pod \"aafebbee-d404-4313-9898-d1baec8e31be\" (UID: \"aafebbee-d404-4313-9898-d1baec8e31be\") " Nov 24 13:46:30 crc kubenswrapper[4790]: I1124 13:46:30.384100 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lp9ct\" (UniqueName: \"kubernetes.io/projected/aafebbee-d404-4313-9898-d1baec8e31be-kube-api-access-lp9ct\") pod \"aafebbee-d404-4313-9898-d1baec8e31be\" (UID: \"aafebbee-d404-4313-9898-d1baec8e31be\") " Nov 24 13:46:30 crc kubenswrapper[4790]: I1124 13:46:30.384171 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aafebbee-d404-4313-9898-d1baec8e31be-catalog-content\") pod \"aafebbee-d404-4313-9898-d1baec8e31be\" (UID: \"aafebbee-d404-4313-9898-d1baec8e31be\") " Nov 24 13:46:30 crc kubenswrapper[4790]: I1124 13:46:30.386785 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aafebbee-d404-4313-9898-d1baec8e31be-utilities" (OuterVolumeSpecName: "utilities") pod "aafebbee-d404-4313-9898-d1baec8e31be" (UID: "aafebbee-d404-4313-9898-d1baec8e31be"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:46:30 crc kubenswrapper[4790]: I1124 13:46:30.390678 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aafebbee-d404-4313-9898-d1baec8e31be-kube-api-access-lp9ct" (OuterVolumeSpecName: "kube-api-access-lp9ct") pod "aafebbee-d404-4313-9898-d1baec8e31be" (UID: "aafebbee-d404-4313-9898-d1baec8e31be"). InnerVolumeSpecName "kube-api-access-lp9ct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:30 crc kubenswrapper[4790]: I1124 13:46:30.495171 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lp9ct\" (UniqueName: \"kubernetes.io/projected/aafebbee-d404-4313-9898-d1baec8e31be-kube-api-access-lp9ct\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:30 crc kubenswrapper[4790]: I1124 13:46:30.495222 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aafebbee-d404-4313-9898-d1baec8e31be-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:30 crc kubenswrapper[4790]: I1124 13:46:30.511035 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aafebbee-d404-4313-9898-d1baec8e31be-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aafebbee-d404-4313-9898-d1baec8e31be" (UID: "aafebbee-d404-4313-9898-d1baec8e31be"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:46:30 crc kubenswrapper[4790]: I1124 13:46:30.596242 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aafebbee-d404-4313-9898-d1baec8e31be-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:30 crc kubenswrapper[4790]: I1124 13:46:30.727359 4790 generic.go:334] "Generic (PLEG): container finished" podID="aafebbee-d404-4313-9898-d1baec8e31be" containerID="7da17141a8af0fd7f716c504953498e62e118b8808bf77aa8282e5759fcf839d" exitCode=0 Nov 24 13:46:30 crc kubenswrapper[4790]: I1124 13:46:30.727421 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vvpjq" event={"ID":"aafebbee-d404-4313-9898-d1baec8e31be","Type":"ContainerDied","Data":"7da17141a8af0fd7f716c504953498e62e118b8808bf77aa8282e5759fcf839d"} Nov 24 13:46:30 crc kubenswrapper[4790]: I1124 13:46:30.727462 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vvpjq" Nov 24 13:46:30 crc kubenswrapper[4790]: I1124 13:46:30.727494 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vvpjq" event={"ID":"aafebbee-d404-4313-9898-d1baec8e31be","Type":"ContainerDied","Data":"9b26284e6a59ea690dcd17b8fc1075f56d7d6dc80507f0baa3ab59f1b3277f28"} Nov 24 13:46:30 crc kubenswrapper[4790]: I1124 13:46:30.727528 4790 scope.go:117] "RemoveContainer" containerID="7da17141a8af0fd7f716c504953498e62e118b8808bf77aa8282e5759fcf839d" Nov 24 13:46:30 crc kubenswrapper[4790]: I1124 13:46:30.743278 4790 scope.go:117] "RemoveContainer" containerID="6a7bc6f92a5a7b82ec951f86bdd82158919f071b8af354e127d47565e5419a8b" Nov 24 13:46:30 crc kubenswrapper[4790]: I1124 13:46:30.765087 4790 scope.go:117] "RemoveContainer" containerID="b77631e8b4a1e4d5176d3bcff651aea2e0c7c38c95f7be884e04648d6fa532c6" Nov 24 13:46:30 crc kubenswrapper[4790]: I1124 13:46:30.812261 4790 scope.go:117] "RemoveContainer" containerID="7da17141a8af0fd7f716c504953498e62e118b8808bf77aa8282e5759fcf839d" Nov 24 13:46:30 crc kubenswrapper[4790]: E1124 13:46:30.822194 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7da17141a8af0fd7f716c504953498e62e118b8808bf77aa8282e5759fcf839d\": container with ID starting with 7da17141a8af0fd7f716c504953498e62e118b8808bf77aa8282e5759fcf839d not found: ID does not exist" containerID="7da17141a8af0fd7f716c504953498e62e118b8808bf77aa8282e5759fcf839d" Nov 24 13:46:30 crc kubenswrapper[4790]: I1124 13:46:30.822251 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7da17141a8af0fd7f716c504953498e62e118b8808bf77aa8282e5759fcf839d"} err="failed to get container status \"7da17141a8af0fd7f716c504953498e62e118b8808bf77aa8282e5759fcf839d\": rpc error: code = NotFound desc = could not find container \"7da17141a8af0fd7f716c504953498e62e118b8808bf77aa8282e5759fcf839d\": container with ID starting with 7da17141a8af0fd7f716c504953498e62e118b8808bf77aa8282e5759fcf839d not found: ID does not exist" Nov 24 13:46:30 crc kubenswrapper[4790]: I1124 13:46:30.822289 4790 scope.go:117] "RemoveContainer" containerID="6a7bc6f92a5a7b82ec951f86bdd82158919f071b8af354e127d47565e5419a8b" Nov 24 13:46:30 crc kubenswrapper[4790]: I1124 13:46:30.823048 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vvpjq"] Nov 24 13:46:30 crc kubenswrapper[4790]: E1124 13:46:30.823466 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a7bc6f92a5a7b82ec951f86bdd82158919f071b8af354e127d47565e5419a8b\": container with ID starting with 6a7bc6f92a5a7b82ec951f86bdd82158919f071b8af354e127d47565e5419a8b not found: ID does not exist" containerID="6a7bc6f92a5a7b82ec951f86bdd82158919f071b8af354e127d47565e5419a8b" Nov 24 13:46:30 crc kubenswrapper[4790]: I1124 13:46:30.823507 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a7bc6f92a5a7b82ec951f86bdd82158919f071b8af354e127d47565e5419a8b"} err="failed to get container status \"6a7bc6f92a5a7b82ec951f86bdd82158919f071b8af354e127d47565e5419a8b\": rpc error: code = NotFound desc = could not find container \"6a7bc6f92a5a7b82ec951f86bdd82158919f071b8af354e127d47565e5419a8b\": container with ID starting with 6a7bc6f92a5a7b82ec951f86bdd82158919f071b8af354e127d47565e5419a8b not found: ID does not exist" Nov 24 13:46:30 crc kubenswrapper[4790]: I1124 13:46:30.823529 4790 scope.go:117] "RemoveContainer" containerID="b77631e8b4a1e4d5176d3bcff651aea2e0c7c38c95f7be884e04648d6fa532c6" Nov 24 13:46:30 crc kubenswrapper[4790]: E1124 13:46:30.823839 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b77631e8b4a1e4d5176d3bcff651aea2e0c7c38c95f7be884e04648d6fa532c6\": container with ID starting with b77631e8b4a1e4d5176d3bcff651aea2e0c7c38c95f7be884e04648d6fa532c6 not found: ID does not exist" containerID="b77631e8b4a1e4d5176d3bcff651aea2e0c7c38c95f7be884e04648d6fa532c6" Nov 24 13:46:30 crc kubenswrapper[4790]: I1124 13:46:30.823867 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b77631e8b4a1e4d5176d3bcff651aea2e0c7c38c95f7be884e04648d6fa532c6"} err="failed to get container status \"b77631e8b4a1e4d5176d3bcff651aea2e0c7c38c95f7be884e04648d6fa532c6\": rpc error: code = NotFound desc = could not find container \"b77631e8b4a1e4d5176d3bcff651aea2e0c7c38c95f7be884e04648d6fa532c6\": container with ID starting with b77631e8b4a1e4d5176d3bcff651aea2e0c7c38c95f7be884e04648d6fa532c6 not found: ID does not exist" Nov 24 13:46:30 crc kubenswrapper[4790]: I1124 13:46:30.833419 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vvpjq"] Nov 24 13:46:32 crc kubenswrapper[4790]: I1124 13:46:32.328605 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aafebbee-d404-4313-9898-d1baec8e31be" path="/var/lib/kubelet/pods/aafebbee-d404-4313-9898-d1baec8e31be/volumes" Nov 24 13:47:37 crc kubenswrapper[4790]: I1124 13:47:37.435750 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2w4mz"] Nov 24 13:47:37 crc kubenswrapper[4790]: E1124 13:47:37.436665 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aafebbee-d404-4313-9898-d1baec8e31be" containerName="extract-content" Nov 24 13:47:37 crc kubenswrapper[4790]: I1124 13:47:37.436681 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="aafebbee-d404-4313-9898-d1baec8e31be" containerName="extract-content" Nov 24 13:47:37 crc kubenswrapper[4790]: E1124 13:47:37.436690 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aafebbee-d404-4313-9898-d1baec8e31be" containerName="registry-server" Nov 24 13:47:37 crc kubenswrapper[4790]: I1124 13:47:37.436697 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="aafebbee-d404-4313-9898-d1baec8e31be" containerName="registry-server" Nov 24 13:47:37 crc kubenswrapper[4790]: E1124 13:47:37.436715 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aafebbee-d404-4313-9898-d1baec8e31be" containerName="extract-utilities" Nov 24 13:47:37 crc kubenswrapper[4790]: I1124 13:47:37.436722 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="aafebbee-d404-4313-9898-d1baec8e31be" containerName="extract-utilities" Nov 24 13:47:37 crc kubenswrapper[4790]: I1124 13:47:37.436961 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="aafebbee-d404-4313-9898-d1baec8e31be" containerName="registry-server" Nov 24 13:47:37 crc kubenswrapper[4790]: I1124 13:47:37.438246 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2w4mz" Nov 24 13:47:37 crc kubenswrapper[4790]: I1124 13:47:37.449460 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2w4mz"] Nov 24 13:47:37 crc kubenswrapper[4790]: I1124 13:47:37.576833 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/547b3036-e4f0-4071-8661-bf793347a585-utilities\") pod \"redhat-marketplace-2w4mz\" (UID: \"547b3036-e4f0-4071-8661-bf793347a585\") " pod="openshift-marketplace/redhat-marketplace-2w4mz" Nov 24 13:47:37 crc kubenswrapper[4790]: I1124 13:47:37.577166 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/547b3036-e4f0-4071-8661-bf793347a585-catalog-content\") pod \"redhat-marketplace-2w4mz\" (UID: \"547b3036-e4f0-4071-8661-bf793347a585\") " pod="openshift-marketplace/redhat-marketplace-2w4mz" Nov 24 13:47:37 crc kubenswrapper[4790]: I1124 13:47:37.577186 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xr6rq\" (UniqueName: \"kubernetes.io/projected/547b3036-e4f0-4071-8661-bf793347a585-kube-api-access-xr6rq\") pod \"redhat-marketplace-2w4mz\" (UID: \"547b3036-e4f0-4071-8661-bf793347a585\") " pod="openshift-marketplace/redhat-marketplace-2w4mz" Nov 24 13:47:37 crc kubenswrapper[4790]: I1124 13:47:37.678600 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/547b3036-e4f0-4071-8661-bf793347a585-utilities\") pod \"redhat-marketplace-2w4mz\" (UID: \"547b3036-e4f0-4071-8661-bf793347a585\") " pod="openshift-marketplace/redhat-marketplace-2w4mz" Nov 24 13:47:37 crc kubenswrapper[4790]: I1124 13:47:37.678648 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/547b3036-e4f0-4071-8661-bf793347a585-catalog-content\") pod \"redhat-marketplace-2w4mz\" (UID: \"547b3036-e4f0-4071-8661-bf793347a585\") " pod="openshift-marketplace/redhat-marketplace-2w4mz" Nov 24 13:47:37 crc kubenswrapper[4790]: I1124 13:47:37.678671 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xr6rq\" (UniqueName: \"kubernetes.io/projected/547b3036-e4f0-4071-8661-bf793347a585-kube-api-access-xr6rq\") pod \"redhat-marketplace-2w4mz\" (UID: \"547b3036-e4f0-4071-8661-bf793347a585\") " pod="openshift-marketplace/redhat-marketplace-2w4mz" Nov 24 13:47:37 crc kubenswrapper[4790]: I1124 13:47:37.679556 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/547b3036-e4f0-4071-8661-bf793347a585-utilities\") pod \"redhat-marketplace-2w4mz\" (UID: \"547b3036-e4f0-4071-8661-bf793347a585\") " pod="openshift-marketplace/redhat-marketplace-2w4mz" Nov 24 13:47:37 crc kubenswrapper[4790]: I1124 13:47:37.679646 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/547b3036-e4f0-4071-8661-bf793347a585-catalog-content\") pod \"redhat-marketplace-2w4mz\" (UID: \"547b3036-e4f0-4071-8661-bf793347a585\") " pod="openshift-marketplace/redhat-marketplace-2w4mz" Nov 24 13:47:37 crc kubenswrapper[4790]: I1124 13:47:37.700909 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xr6rq\" (UniqueName: \"kubernetes.io/projected/547b3036-e4f0-4071-8661-bf793347a585-kube-api-access-xr6rq\") pod \"redhat-marketplace-2w4mz\" (UID: \"547b3036-e4f0-4071-8661-bf793347a585\") " pod="openshift-marketplace/redhat-marketplace-2w4mz" Nov 24 13:47:37 crc kubenswrapper[4790]: I1124 13:47:37.765930 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2w4mz" Nov 24 13:47:38 crc kubenswrapper[4790]: I1124 13:47:38.216213 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2w4mz"] Nov 24 13:47:38 crc kubenswrapper[4790]: I1124 13:47:38.284603 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2w4mz" event={"ID":"547b3036-e4f0-4071-8661-bf793347a585","Type":"ContainerStarted","Data":"b50829802ba6339b830cda533dcf9216d2da3719cd837079844af7d31543bab1"} Nov 24 13:47:39 crc kubenswrapper[4790]: I1124 13:47:39.292294 4790 generic.go:334] "Generic (PLEG): container finished" podID="547b3036-e4f0-4071-8661-bf793347a585" containerID="7866d7b80d9ddb63e2b6df6defdb23a1d5b68f574eecf764bda6bfe8adb66939" exitCode=0 Nov 24 13:47:39 crc kubenswrapper[4790]: I1124 13:47:39.292412 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2w4mz" event={"ID":"547b3036-e4f0-4071-8661-bf793347a585","Type":"ContainerDied","Data":"7866d7b80d9ddb63e2b6df6defdb23a1d5b68f574eecf764bda6bfe8adb66939"} Nov 24 13:47:40 crc kubenswrapper[4790]: I1124 13:47:40.304832 4790 generic.go:334] "Generic (PLEG): container finished" podID="547b3036-e4f0-4071-8661-bf793347a585" containerID="3709099cf95640d982119c7d36d0e1ed9a8f5aed42694b06b9d7fd926e429c19" exitCode=0 Nov 24 13:47:40 crc kubenswrapper[4790]: I1124 13:47:40.304902 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2w4mz" event={"ID":"547b3036-e4f0-4071-8661-bf793347a585","Type":"ContainerDied","Data":"3709099cf95640d982119c7d36d0e1ed9a8f5aed42694b06b9d7fd926e429c19"} Nov 24 13:47:41 crc kubenswrapper[4790]: I1124 13:47:41.313798 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2w4mz" event={"ID":"547b3036-e4f0-4071-8661-bf793347a585","Type":"ContainerStarted","Data":"ac9326e7a5a578dbd43c12ba357f1c1172a3c0cbfbe6be33b1a37dd7a148ff0e"} Nov 24 13:47:41 crc kubenswrapper[4790]: I1124 13:47:41.335081 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2w4mz" podStartSLOduration=2.942207606 podStartE2EDuration="4.335063104s" podCreationTimestamp="2025-11-24 13:47:37 +0000 UTC" firstStartedPulling="2025-11-24 13:47:39.293748344 +0000 UTC m=+2107.673642006" lastFinishedPulling="2025-11-24 13:47:40.686603842 +0000 UTC m=+2109.066497504" observedRunningTime="2025-11-24 13:47:41.329189026 +0000 UTC m=+2109.709082688" watchObservedRunningTime="2025-11-24 13:47:41.335063104 +0000 UTC m=+2109.714956776" Nov 24 13:47:47 crc kubenswrapper[4790]: I1124 13:47:47.766110 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2w4mz" Nov 24 13:47:47 crc kubenswrapper[4790]: I1124 13:47:47.766617 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2w4mz" Nov 24 13:47:47 crc kubenswrapper[4790]: I1124 13:47:47.807805 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2w4mz" Nov 24 13:47:48 crc kubenswrapper[4790]: I1124 13:47:48.403900 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2w4mz" Nov 24 13:47:48 crc kubenswrapper[4790]: I1124 13:47:48.446617 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2w4mz"] Nov 24 13:47:50 crc kubenswrapper[4790]: I1124 13:47:50.374169 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2w4mz" podUID="547b3036-e4f0-4071-8661-bf793347a585" containerName="registry-server" containerID="cri-o://ac9326e7a5a578dbd43c12ba357f1c1172a3c0cbfbe6be33b1a37dd7a148ff0e" gracePeriod=2 Nov 24 13:47:50 crc kubenswrapper[4790]: I1124 13:47:50.761539 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2w4mz" Nov 24 13:47:50 crc kubenswrapper[4790]: I1124 13:47:50.869704 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/547b3036-e4f0-4071-8661-bf793347a585-utilities\") pod \"547b3036-e4f0-4071-8661-bf793347a585\" (UID: \"547b3036-e4f0-4071-8661-bf793347a585\") " Nov 24 13:47:50 crc kubenswrapper[4790]: I1124 13:47:50.869803 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/547b3036-e4f0-4071-8661-bf793347a585-catalog-content\") pod \"547b3036-e4f0-4071-8661-bf793347a585\" (UID: \"547b3036-e4f0-4071-8661-bf793347a585\") " Nov 24 13:47:50 crc kubenswrapper[4790]: I1124 13:47:50.869855 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xr6rq\" (UniqueName: \"kubernetes.io/projected/547b3036-e4f0-4071-8661-bf793347a585-kube-api-access-xr6rq\") pod \"547b3036-e4f0-4071-8661-bf793347a585\" (UID: \"547b3036-e4f0-4071-8661-bf793347a585\") " Nov 24 13:47:50 crc kubenswrapper[4790]: I1124 13:47:50.871072 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/547b3036-e4f0-4071-8661-bf793347a585-utilities" (OuterVolumeSpecName: "utilities") pod "547b3036-e4f0-4071-8661-bf793347a585" (UID: "547b3036-e4f0-4071-8661-bf793347a585"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:47:50 crc kubenswrapper[4790]: I1124 13:47:50.876151 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/547b3036-e4f0-4071-8661-bf793347a585-kube-api-access-xr6rq" (OuterVolumeSpecName: "kube-api-access-xr6rq") pod "547b3036-e4f0-4071-8661-bf793347a585" (UID: "547b3036-e4f0-4071-8661-bf793347a585"). InnerVolumeSpecName "kube-api-access-xr6rq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:47:50 crc kubenswrapper[4790]: I1124 13:47:50.889939 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/547b3036-e4f0-4071-8661-bf793347a585-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "547b3036-e4f0-4071-8661-bf793347a585" (UID: "547b3036-e4f0-4071-8661-bf793347a585"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:47:50 crc kubenswrapper[4790]: I1124 13:47:50.971779 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/547b3036-e4f0-4071-8661-bf793347a585-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:47:50 crc kubenswrapper[4790]: I1124 13:47:50.971813 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/547b3036-e4f0-4071-8661-bf793347a585-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:47:50 crc kubenswrapper[4790]: I1124 13:47:50.971826 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xr6rq\" (UniqueName: \"kubernetes.io/projected/547b3036-e4f0-4071-8661-bf793347a585-kube-api-access-xr6rq\") on node \"crc\" DevicePath \"\"" Nov 24 13:47:51 crc kubenswrapper[4790]: I1124 13:47:51.387760 4790 generic.go:334] "Generic (PLEG): container finished" podID="547b3036-e4f0-4071-8661-bf793347a585" containerID="ac9326e7a5a578dbd43c12ba357f1c1172a3c0cbfbe6be33b1a37dd7a148ff0e" exitCode=0 Nov 24 13:47:51 crc kubenswrapper[4790]: I1124 13:47:51.387838 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2w4mz" event={"ID":"547b3036-e4f0-4071-8661-bf793347a585","Type":"ContainerDied","Data":"ac9326e7a5a578dbd43c12ba357f1c1172a3c0cbfbe6be33b1a37dd7a148ff0e"} Nov 24 13:47:51 crc kubenswrapper[4790]: I1124 13:47:51.387850 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2w4mz" Nov 24 13:47:51 crc kubenswrapper[4790]: I1124 13:47:51.387873 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2w4mz" event={"ID":"547b3036-e4f0-4071-8661-bf793347a585","Type":"ContainerDied","Data":"b50829802ba6339b830cda533dcf9216d2da3719cd837079844af7d31543bab1"} Nov 24 13:47:51 crc kubenswrapper[4790]: I1124 13:47:51.387917 4790 scope.go:117] "RemoveContainer" containerID="ac9326e7a5a578dbd43c12ba357f1c1172a3c0cbfbe6be33b1a37dd7a148ff0e" Nov 24 13:47:51 crc kubenswrapper[4790]: I1124 13:47:51.418552 4790 scope.go:117] "RemoveContainer" containerID="3709099cf95640d982119c7d36d0e1ed9a8f5aed42694b06b9d7fd926e429c19" Nov 24 13:47:51 crc kubenswrapper[4790]: I1124 13:47:51.448045 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2w4mz"] Nov 24 13:47:51 crc kubenswrapper[4790]: I1124 13:47:51.455516 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2w4mz"] Nov 24 13:47:51 crc kubenswrapper[4790]: I1124 13:47:51.468217 4790 scope.go:117] "RemoveContainer" containerID="7866d7b80d9ddb63e2b6df6defdb23a1d5b68f574eecf764bda6bfe8adb66939" Nov 24 13:47:51 crc kubenswrapper[4790]: I1124 13:47:51.487726 4790 scope.go:117] "RemoveContainer" containerID="ac9326e7a5a578dbd43c12ba357f1c1172a3c0cbfbe6be33b1a37dd7a148ff0e" Nov 24 13:47:51 crc kubenswrapper[4790]: E1124 13:47:51.488192 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac9326e7a5a578dbd43c12ba357f1c1172a3c0cbfbe6be33b1a37dd7a148ff0e\": container with ID starting with ac9326e7a5a578dbd43c12ba357f1c1172a3c0cbfbe6be33b1a37dd7a148ff0e not found: ID does not exist" containerID="ac9326e7a5a578dbd43c12ba357f1c1172a3c0cbfbe6be33b1a37dd7a148ff0e" Nov 24 13:47:51 crc kubenswrapper[4790]: I1124 13:47:51.488235 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac9326e7a5a578dbd43c12ba357f1c1172a3c0cbfbe6be33b1a37dd7a148ff0e"} err="failed to get container status \"ac9326e7a5a578dbd43c12ba357f1c1172a3c0cbfbe6be33b1a37dd7a148ff0e\": rpc error: code = NotFound desc = could not find container \"ac9326e7a5a578dbd43c12ba357f1c1172a3c0cbfbe6be33b1a37dd7a148ff0e\": container with ID starting with ac9326e7a5a578dbd43c12ba357f1c1172a3c0cbfbe6be33b1a37dd7a148ff0e not found: ID does not exist" Nov 24 13:47:51 crc kubenswrapper[4790]: I1124 13:47:51.488265 4790 scope.go:117] "RemoveContainer" containerID="3709099cf95640d982119c7d36d0e1ed9a8f5aed42694b06b9d7fd926e429c19" Nov 24 13:47:51 crc kubenswrapper[4790]: E1124 13:47:51.488822 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3709099cf95640d982119c7d36d0e1ed9a8f5aed42694b06b9d7fd926e429c19\": container with ID starting with 3709099cf95640d982119c7d36d0e1ed9a8f5aed42694b06b9d7fd926e429c19 not found: ID does not exist" containerID="3709099cf95640d982119c7d36d0e1ed9a8f5aed42694b06b9d7fd926e429c19" Nov 24 13:47:51 crc kubenswrapper[4790]: I1124 13:47:51.488852 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3709099cf95640d982119c7d36d0e1ed9a8f5aed42694b06b9d7fd926e429c19"} err="failed to get container status \"3709099cf95640d982119c7d36d0e1ed9a8f5aed42694b06b9d7fd926e429c19\": rpc error: code = NotFound desc = could not find container \"3709099cf95640d982119c7d36d0e1ed9a8f5aed42694b06b9d7fd926e429c19\": container with ID starting with 3709099cf95640d982119c7d36d0e1ed9a8f5aed42694b06b9d7fd926e429c19 not found: ID does not exist" Nov 24 13:47:51 crc kubenswrapper[4790]: I1124 13:47:51.488874 4790 scope.go:117] "RemoveContainer" containerID="7866d7b80d9ddb63e2b6df6defdb23a1d5b68f574eecf764bda6bfe8adb66939" Nov 24 13:47:51 crc kubenswrapper[4790]: E1124 13:47:51.489085 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7866d7b80d9ddb63e2b6df6defdb23a1d5b68f574eecf764bda6bfe8adb66939\": container with ID starting with 7866d7b80d9ddb63e2b6df6defdb23a1d5b68f574eecf764bda6bfe8adb66939 not found: ID does not exist" containerID="7866d7b80d9ddb63e2b6df6defdb23a1d5b68f574eecf764bda6bfe8adb66939" Nov 24 13:47:51 crc kubenswrapper[4790]: I1124 13:47:51.489117 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7866d7b80d9ddb63e2b6df6defdb23a1d5b68f574eecf764bda6bfe8adb66939"} err="failed to get container status \"7866d7b80d9ddb63e2b6df6defdb23a1d5b68f574eecf764bda6bfe8adb66939\": rpc error: code = NotFound desc = could not find container \"7866d7b80d9ddb63e2b6df6defdb23a1d5b68f574eecf764bda6bfe8adb66939\": container with ID starting with 7866d7b80d9ddb63e2b6df6defdb23a1d5b68f574eecf764bda6bfe8adb66939 not found: ID does not exist" Nov 24 13:47:52 crc kubenswrapper[4790]: I1124 13:47:52.326045 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="547b3036-e4f0-4071-8661-bf793347a585" path="/var/lib/kubelet/pods/547b3036-e4f0-4071-8661-bf793347a585/volumes" Nov 24 13:47:55 crc kubenswrapper[4790]: I1124 13:47:55.411021 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4nhlg"] Nov 24 13:47:55 crc kubenswrapper[4790]: E1124 13:47:55.411653 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="547b3036-e4f0-4071-8661-bf793347a585" containerName="extract-utilities" Nov 24 13:47:55 crc kubenswrapper[4790]: I1124 13:47:55.411669 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="547b3036-e4f0-4071-8661-bf793347a585" containerName="extract-utilities" Nov 24 13:47:55 crc kubenswrapper[4790]: E1124 13:47:55.411697 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="547b3036-e4f0-4071-8661-bf793347a585" containerName="registry-server" Nov 24 13:47:55 crc kubenswrapper[4790]: I1124 13:47:55.411705 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="547b3036-e4f0-4071-8661-bf793347a585" containerName="registry-server" Nov 24 13:47:55 crc kubenswrapper[4790]: E1124 13:47:55.411726 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="547b3036-e4f0-4071-8661-bf793347a585" containerName="extract-content" Nov 24 13:47:55 crc kubenswrapper[4790]: I1124 13:47:55.411735 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="547b3036-e4f0-4071-8661-bf793347a585" containerName="extract-content" Nov 24 13:47:55 crc kubenswrapper[4790]: I1124 13:47:55.411956 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="547b3036-e4f0-4071-8661-bf793347a585" containerName="registry-server" Nov 24 13:47:55 crc kubenswrapper[4790]: I1124 13:47:55.413366 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4nhlg" Nov 24 13:47:55 crc kubenswrapper[4790]: I1124 13:47:55.419189 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4nhlg"] Nov 24 13:47:55 crc kubenswrapper[4790]: I1124 13:47:55.540244 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zmk8\" (UniqueName: \"kubernetes.io/projected/54174d67-8e60-4894-94ba-28f535f52266-kube-api-access-7zmk8\") pod \"certified-operators-4nhlg\" (UID: \"54174d67-8e60-4894-94ba-28f535f52266\") " pod="openshift-marketplace/certified-operators-4nhlg" Nov 24 13:47:55 crc kubenswrapper[4790]: I1124 13:47:55.540377 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54174d67-8e60-4894-94ba-28f535f52266-utilities\") pod \"certified-operators-4nhlg\" (UID: \"54174d67-8e60-4894-94ba-28f535f52266\") " pod="openshift-marketplace/certified-operators-4nhlg" Nov 24 13:47:55 crc kubenswrapper[4790]: I1124 13:47:55.540417 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54174d67-8e60-4894-94ba-28f535f52266-catalog-content\") pod \"certified-operators-4nhlg\" (UID: \"54174d67-8e60-4894-94ba-28f535f52266\") " pod="openshift-marketplace/certified-operators-4nhlg" Nov 24 13:47:55 crc kubenswrapper[4790]: I1124 13:47:55.641249 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zmk8\" (UniqueName: \"kubernetes.io/projected/54174d67-8e60-4894-94ba-28f535f52266-kube-api-access-7zmk8\") pod \"certified-operators-4nhlg\" (UID: \"54174d67-8e60-4894-94ba-28f535f52266\") " pod="openshift-marketplace/certified-operators-4nhlg" Nov 24 13:47:55 crc kubenswrapper[4790]: I1124 13:47:55.641356 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54174d67-8e60-4894-94ba-28f535f52266-utilities\") pod \"certified-operators-4nhlg\" (UID: \"54174d67-8e60-4894-94ba-28f535f52266\") " pod="openshift-marketplace/certified-operators-4nhlg" Nov 24 13:47:55 crc kubenswrapper[4790]: I1124 13:47:55.641388 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54174d67-8e60-4894-94ba-28f535f52266-catalog-content\") pod \"certified-operators-4nhlg\" (UID: \"54174d67-8e60-4894-94ba-28f535f52266\") " pod="openshift-marketplace/certified-operators-4nhlg" Nov 24 13:47:55 crc kubenswrapper[4790]: I1124 13:47:55.641922 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54174d67-8e60-4894-94ba-28f535f52266-catalog-content\") pod \"certified-operators-4nhlg\" (UID: \"54174d67-8e60-4894-94ba-28f535f52266\") " pod="openshift-marketplace/certified-operators-4nhlg" Nov 24 13:47:55 crc kubenswrapper[4790]: I1124 13:47:55.641979 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54174d67-8e60-4894-94ba-28f535f52266-utilities\") pod \"certified-operators-4nhlg\" (UID: \"54174d67-8e60-4894-94ba-28f535f52266\") " pod="openshift-marketplace/certified-operators-4nhlg" Nov 24 13:47:55 crc kubenswrapper[4790]: I1124 13:47:55.662920 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zmk8\" (UniqueName: \"kubernetes.io/projected/54174d67-8e60-4894-94ba-28f535f52266-kube-api-access-7zmk8\") pod \"certified-operators-4nhlg\" (UID: \"54174d67-8e60-4894-94ba-28f535f52266\") " pod="openshift-marketplace/certified-operators-4nhlg" Nov 24 13:47:55 crc kubenswrapper[4790]: I1124 13:47:55.730926 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4nhlg" Nov 24 13:47:56 crc kubenswrapper[4790]: I1124 13:47:56.041402 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4nhlg"] Nov 24 13:47:56 crc kubenswrapper[4790]: I1124 13:47:56.434471 4790 generic.go:334] "Generic (PLEG): container finished" podID="54174d67-8e60-4894-94ba-28f535f52266" containerID="05ac3c93831f894b2778ed5900f6940709a60fc799cb9b35d633a53ef2ddd281" exitCode=0 Nov 24 13:47:56 crc kubenswrapper[4790]: I1124 13:47:56.434515 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4nhlg" event={"ID":"54174d67-8e60-4894-94ba-28f535f52266","Type":"ContainerDied","Data":"05ac3c93831f894b2778ed5900f6940709a60fc799cb9b35d633a53ef2ddd281"} Nov 24 13:47:56 crc kubenswrapper[4790]: I1124 13:47:56.434763 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4nhlg" event={"ID":"54174d67-8e60-4894-94ba-28f535f52266","Type":"ContainerStarted","Data":"0c7972d1087c2dc5b066e39bb4c52bff3ceb8151bebec4a8031511b37393451b"} Nov 24 13:47:57 crc kubenswrapper[4790]: I1124 13:47:57.444519 4790 generic.go:334] "Generic (PLEG): container finished" podID="54174d67-8e60-4894-94ba-28f535f52266" containerID="15cfa05ac82da8c1bc1b28eb128fea5f5ab249cb5df5ad8534c0b1a3e20d7b6b" exitCode=0 Nov 24 13:47:57 crc kubenswrapper[4790]: I1124 13:47:57.444568 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4nhlg" event={"ID":"54174d67-8e60-4894-94ba-28f535f52266","Type":"ContainerDied","Data":"15cfa05ac82da8c1bc1b28eb128fea5f5ab249cb5df5ad8534c0b1a3e20d7b6b"} Nov 24 13:47:58 crc kubenswrapper[4790]: I1124 13:47:58.461059 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4nhlg" event={"ID":"54174d67-8e60-4894-94ba-28f535f52266","Type":"ContainerStarted","Data":"2d933fac611b507d714eac65ca584b152423c35d4db939eec02c07041969ff33"} Nov 24 13:47:58 crc kubenswrapper[4790]: I1124 13:47:58.482687 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4nhlg" podStartSLOduration=2.022743502 podStartE2EDuration="3.482673048s" podCreationTimestamp="2025-11-24 13:47:55 +0000 UTC" firstStartedPulling="2025-11-24 13:47:56.435687656 +0000 UTC m=+2124.815581308" lastFinishedPulling="2025-11-24 13:47:57.895617192 +0000 UTC m=+2126.275510854" observedRunningTime="2025-11-24 13:47:58.479825857 +0000 UTC m=+2126.859719519" watchObservedRunningTime="2025-11-24 13:47:58.482673048 +0000 UTC m=+2126.862566700" Nov 24 13:48:05 crc kubenswrapper[4790]: I1124 13:48:05.731499 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4nhlg" Nov 24 13:48:05 crc kubenswrapper[4790]: I1124 13:48:05.732066 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4nhlg" Nov 24 13:48:05 crc kubenswrapper[4790]: I1124 13:48:05.794541 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4nhlg" Nov 24 13:48:06 crc kubenswrapper[4790]: I1124 13:48:06.585940 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4nhlg" Nov 24 13:48:06 crc kubenswrapper[4790]: I1124 13:48:06.656009 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4nhlg"] Nov 24 13:48:08 crc kubenswrapper[4790]: I1124 13:48:08.533419 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4nhlg" podUID="54174d67-8e60-4894-94ba-28f535f52266" containerName="registry-server" containerID="cri-o://2d933fac611b507d714eac65ca584b152423c35d4db939eec02c07041969ff33" gracePeriod=2 Nov 24 13:48:08 crc kubenswrapper[4790]: I1124 13:48:08.988288 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4nhlg" Nov 24 13:48:09 crc kubenswrapper[4790]: I1124 13:48:09.144368 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54174d67-8e60-4894-94ba-28f535f52266-catalog-content\") pod \"54174d67-8e60-4894-94ba-28f535f52266\" (UID: \"54174d67-8e60-4894-94ba-28f535f52266\") " Nov 24 13:48:09 crc kubenswrapper[4790]: I1124 13:48:09.144441 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54174d67-8e60-4894-94ba-28f535f52266-utilities\") pod \"54174d67-8e60-4894-94ba-28f535f52266\" (UID: \"54174d67-8e60-4894-94ba-28f535f52266\") " Nov 24 13:48:09 crc kubenswrapper[4790]: I1124 13:48:09.144532 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zmk8\" (UniqueName: \"kubernetes.io/projected/54174d67-8e60-4894-94ba-28f535f52266-kube-api-access-7zmk8\") pod \"54174d67-8e60-4894-94ba-28f535f52266\" (UID: \"54174d67-8e60-4894-94ba-28f535f52266\") " Nov 24 13:48:09 crc kubenswrapper[4790]: I1124 13:48:09.145404 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54174d67-8e60-4894-94ba-28f535f52266-utilities" (OuterVolumeSpecName: "utilities") pod "54174d67-8e60-4894-94ba-28f535f52266" (UID: "54174d67-8e60-4894-94ba-28f535f52266"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:48:09 crc kubenswrapper[4790]: I1124 13:48:09.149756 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54174d67-8e60-4894-94ba-28f535f52266-kube-api-access-7zmk8" (OuterVolumeSpecName: "kube-api-access-7zmk8") pod "54174d67-8e60-4894-94ba-28f535f52266" (UID: "54174d67-8e60-4894-94ba-28f535f52266"). InnerVolumeSpecName "kube-api-access-7zmk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:48:09 crc kubenswrapper[4790]: I1124 13:48:09.201091 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54174d67-8e60-4894-94ba-28f535f52266-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "54174d67-8e60-4894-94ba-28f535f52266" (UID: "54174d67-8e60-4894-94ba-28f535f52266"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:48:09 crc kubenswrapper[4790]: I1124 13:48:09.246119 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54174d67-8e60-4894-94ba-28f535f52266-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:48:09 crc kubenswrapper[4790]: I1124 13:48:09.246146 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54174d67-8e60-4894-94ba-28f535f52266-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:48:09 crc kubenswrapper[4790]: I1124 13:48:09.246158 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zmk8\" (UniqueName: \"kubernetes.io/projected/54174d67-8e60-4894-94ba-28f535f52266-kube-api-access-7zmk8\") on node \"crc\" DevicePath \"\"" Nov 24 13:48:09 crc kubenswrapper[4790]: I1124 13:48:09.542702 4790 generic.go:334] "Generic (PLEG): container finished" podID="54174d67-8e60-4894-94ba-28f535f52266" containerID="2d933fac611b507d714eac65ca584b152423c35d4db939eec02c07041969ff33" exitCode=0 Nov 24 13:48:09 crc kubenswrapper[4790]: I1124 13:48:09.542746 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4nhlg" event={"ID":"54174d67-8e60-4894-94ba-28f535f52266","Type":"ContainerDied","Data":"2d933fac611b507d714eac65ca584b152423c35d4db939eec02c07041969ff33"} Nov 24 13:48:09 crc kubenswrapper[4790]: I1124 13:48:09.542758 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4nhlg" Nov 24 13:48:09 crc kubenswrapper[4790]: I1124 13:48:09.542774 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4nhlg" event={"ID":"54174d67-8e60-4894-94ba-28f535f52266","Type":"ContainerDied","Data":"0c7972d1087c2dc5b066e39bb4c52bff3ceb8151bebec4a8031511b37393451b"} Nov 24 13:48:09 crc kubenswrapper[4790]: I1124 13:48:09.542789 4790 scope.go:117] "RemoveContainer" containerID="2d933fac611b507d714eac65ca584b152423c35d4db939eec02c07041969ff33" Nov 24 13:48:09 crc kubenswrapper[4790]: I1124 13:48:09.566153 4790 scope.go:117] "RemoveContainer" containerID="15cfa05ac82da8c1bc1b28eb128fea5f5ab249cb5df5ad8534c0b1a3e20d7b6b" Nov 24 13:48:09 crc kubenswrapper[4790]: I1124 13:48:09.582783 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4nhlg"] Nov 24 13:48:09 crc kubenswrapper[4790]: I1124 13:48:09.589438 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4nhlg"] Nov 24 13:48:09 crc kubenswrapper[4790]: I1124 13:48:09.602958 4790 scope.go:117] "RemoveContainer" containerID="05ac3c93831f894b2778ed5900f6940709a60fc799cb9b35d633a53ef2ddd281" Nov 24 13:48:09 crc kubenswrapper[4790]: I1124 13:48:09.620524 4790 scope.go:117] "RemoveContainer" containerID="2d933fac611b507d714eac65ca584b152423c35d4db939eec02c07041969ff33" Nov 24 13:48:09 crc kubenswrapper[4790]: E1124 13:48:09.620960 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d933fac611b507d714eac65ca584b152423c35d4db939eec02c07041969ff33\": container with ID starting with 2d933fac611b507d714eac65ca584b152423c35d4db939eec02c07041969ff33 not found: ID does not exist" containerID="2d933fac611b507d714eac65ca584b152423c35d4db939eec02c07041969ff33" Nov 24 13:48:09 crc kubenswrapper[4790]: I1124 13:48:09.621003 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d933fac611b507d714eac65ca584b152423c35d4db939eec02c07041969ff33"} err="failed to get container status \"2d933fac611b507d714eac65ca584b152423c35d4db939eec02c07041969ff33\": rpc error: code = NotFound desc = could not find container \"2d933fac611b507d714eac65ca584b152423c35d4db939eec02c07041969ff33\": container with ID starting with 2d933fac611b507d714eac65ca584b152423c35d4db939eec02c07041969ff33 not found: ID does not exist" Nov 24 13:48:09 crc kubenswrapper[4790]: I1124 13:48:09.621030 4790 scope.go:117] "RemoveContainer" containerID="15cfa05ac82da8c1bc1b28eb128fea5f5ab249cb5df5ad8534c0b1a3e20d7b6b" Nov 24 13:48:09 crc kubenswrapper[4790]: E1124 13:48:09.621393 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15cfa05ac82da8c1bc1b28eb128fea5f5ab249cb5df5ad8534c0b1a3e20d7b6b\": container with ID starting with 15cfa05ac82da8c1bc1b28eb128fea5f5ab249cb5df5ad8534c0b1a3e20d7b6b not found: ID does not exist" containerID="15cfa05ac82da8c1bc1b28eb128fea5f5ab249cb5df5ad8534c0b1a3e20d7b6b" Nov 24 13:48:09 crc kubenswrapper[4790]: I1124 13:48:09.621424 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15cfa05ac82da8c1bc1b28eb128fea5f5ab249cb5df5ad8534c0b1a3e20d7b6b"} err="failed to get container status \"15cfa05ac82da8c1bc1b28eb128fea5f5ab249cb5df5ad8534c0b1a3e20d7b6b\": rpc error: code = NotFound desc = could not find container \"15cfa05ac82da8c1bc1b28eb128fea5f5ab249cb5df5ad8534c0b1a3e20d7b6b\": container with ID starting with 15cfa05ac82da8c1bc1b28eb128fea5f5ab249cb5df5ad8534c0b1a3e20d7b6b not found: ID does not exist" Nov 24 13:48:09 crc kubenswrapper[4790]: I1124 13:48:09.621481 4790 scope.go:117] "RemoveContainer" containerID="05ac3c93831f894b2778ed5900f6940709a60fc799cb9b35d633a53ef2ddd281" Nov 24 13:48:09 crc kubenswrapper[4790]: E1124 13:48:09.621797 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05ac3c93831f894b2778ed5900f6940709a60fc799cb9b35d633a53ef2ddd281\": container with ID starting with 05ac3c93831f894b2778ed5900f6940709a60fc799cb9b35d633a53ef2ddd281 not found: ID does not exist" containerID="05ac3c93831f894b2778ed5900f6940709a60fc799cb9b35d633a53ef2ddd281" Nov 24 13:48:09 crc kubenswrapper[4790]: I1124 13:48:09.621822 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05ac3c93831f894b2778ed5900f6940709a60fc799cb9b35d633a53ef2ddd281"} err="failed to get container status \"05ac3c93831f894b2778ed5900f6940709a60fc799cb9b35d633a53ef2ddd281\": rpc error: code = NotFound desc = could not find container \"05ac3c93831f894b2778ed5900f6940709a60fc799cb9b35d633a53ef2ddd281\": container with ID starting with 05ac3c93831f894b2778ed5900f6940709a60fc799cb9b35d633a53ef2ddd281 not found: ID does not exist" Nov 24 13:48:10 crc kubenswrapper[4790]: I1124 13:48:10.329327 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54174d67-8e60-4894-94ba-28f535f52266" path="/var/lib/kubelet/pods/54174d67-8e60-4894-94ba-28f535f52266/volumes" Nov 24 13:48:13 crc kubenswrapper[4790]: I1124 13:48:13.938241 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:48:13 crc kubenswrapper[4790]: I1124 13:48:13.939614 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:48:43 crc kubenswrapper[4790]: I1124 13:48:43.938174 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:48:43 crc kubenswrapper[4790]: I1124 13:48:43.938732 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:49:13 crc kubenswrapper[4790]: I1124 13:49:13.939190 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:49:13 crc kubenswrapper[4790]: I1124 13:49:13.939742 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:49:13 crc kubenswrapper[4790]: I1124 13:49:13.939790 4790 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 13:49:13 crc kubenswrapper[4790]: I1124 13:49:13.940349 4790 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d22c23244bf6119211e0ddeceeef48f1aa920868cf60d45e67811b7753e963eb"} pod="openshift-machine-config-operator/machine-config-daemon-xz49t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 13:49:13 crc kubenswrapper[4790]: I1124 13:49:13.940417 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" containerID="cri-o://d22c23244bf6119211e0ddeceeef48f1aa920868cf60d45e67811b7753e963eb" gracePeriod=600 Nov 24 13:49:14 crc kubenswrapper[4790]: I1124 13:49:14.999232 4790 generic.go:334] "Generic (PLEG): container finished" podID="0d73b133-48f0-455f-8f6a-742e633f631a" containerID="d22c23244bf6119211e0ddeceeef48f1aa920868cf60d45e67811b7753e963eb" exitCode=0 Nov 24 13:49:14 crc kubenswrapper[4790]: I1124 13:49:14.999280 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerDied","Data":"d22c23244bf6119211e0ddeceeef48f1aa920868cf60d45e67811b7753e963eb"} Nov 24 13:49:15 crc kubenswrapper[4790]: I1124 13:49:14.999948 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"8310dee54c9b2a0ea195ce3e80b6281248186114131fdd393fbb8cb25d04910a"} Nov 24 13:49:15 crc kubenswrapper[4790]: I1124 13:49:14.999978 4790 scope.go:117] "RemoveContainer" containerID="0e39e707a2686ae4ea3e31dae8bfe7595a875cadfa529bb73dd02722c8084b9e" Nov 24 13:51:43 crc kubenswrapper[4790]: I1124 13:51:43.938641 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:51:43 crc kubenswrapper[4790]: I1124 13:51:43.939301 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:52:13 crc kubenswrapper[4790]: I1124 13:52:13.939100 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:52:13 crc kubenswrapper[4790]: I1124 13:52:13.939644 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:52:43 crc kubenswrapper[4790]: I1124 13:52:43.939064 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:52:43 crc kubenswrapper[4790]: I1124 13:52:43.939692 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:52:43 crc kubenswrapper[4790]: I1124 13:52:43.939753 4790 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 13:52:43 crc kubenswrapper[4790]: I1124 13:52:43.940794 4790 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8310dee54c9b2a0ea195ce3e80b6281248186114131fdd393fbb8cb25d04910a"} pod="openshift-machine-config-operator/machine-config-daemon-xz49t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 13:52:43 crc kubenswrapper[4790]: I1124 13:52:43.941143 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" containerID="cri-o://8310dee54c9b2a0ea195ce3e80b6281248186114131fdd393fbb8cb25d04910a" gracePeriod=600 Nov 24 13:52:44 crc kubenswrapper[4790]: E1124 13:52:44.077282 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:52:44 crc kubenswrapper[4790]: I1124 13:52:44.552814 4790 generic.go:334] "Generic (PLEG): container finished" podID="0d73b133-48f0-455f-8f6a-742e633f631a" containerID="8310dee54c9b2a0ea195ce3e80b6281248186114131fdd393fbb8cb25d04910a" exitCode=0 Nov 24 13:52:44 crc kubenswrapper[4790]: I1124 13:52:44.552921 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerDied","Data":"8310dee54c9b2a0ea195ce3e80b6281248186114131fdd393fbb8cb25d04910a"} Nov 24 13:52:44 crc kubenswrapper[4790]: I1124 13:52:44.552985 4790 scope.go:117] "RemoveContainer" containerID="d22c23244bf6119211e0ddeceeef48f1aa920868cf60d45e67811b7753e963eb" Nov 24 13:52:44 crc kubenswrapper[4790]: I1124 13:52:44.553727 4790 scope.go:117] "RemoveContainer" containerID="8310dee54c9b2a0ea195ce3e80b6281248186114131fdd393fbb8cb25d04910a" Nov 24 13:52:44 crc kubenswrapper[4790]: E1124 13:52:44.554291 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:52:57 crc kubenswrapper[4790]: I1124 13:52:57.314568 4790 scope.go:117] "RemoveContainer" containerID="8310dee54c9b2a0ea195ce3e80b6281248186114131fdd393fbb8cb25d04910a" Nov 24 13:52:57 crc kubenswrapper[4790]: E1124 13:52:57.315288 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:53:09 crc kubenswrapper[4790]: I1124 13:53:09.314727 4790 scope.go:117] "RemoveContainer" containerID="8310dee54c9b2a0ea195ce3e80b6281248186114131fdd393fbb8cb25d04910a" Nov 24 13:53:09 crc kubenswrapper[4790]: E1124 13:53:09.315554 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:53:24 crc kubenswrapper[4790]: I1124 13:53:24.315537 4790 scope.go:117] "RemoveContainer" containerID="8310dee54c9b2a0ea195ce3e80b6281248186114131fdd393fbb8cb25d04910a" Nov 24 13:53:24 crc kubenswrapper[4790]: E1124 13:53:24.316669 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:53:36 crc kubenswrapper[4790]: I1124 13:53:36.750230 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-97gps"] Nov 24 13:53:36 crc kubenswrapper[4790]: E1124 13:53:36.750989 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54174d67-8e60-4894-94ba-28f535f52266" containerName="registry-server" Nov 24 13:53:36 crc kubenswrapper[4790]: I1124 13:53:36.751004 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="54174d67-8e60-4894-94ba-28f535f52266" containerName="registry-server" Nov 24 13:53:36 crc kubenswrapper[4790]: E1124 13:53:36.751023 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54174d67-8e60-4894-94ba-28f535f52266" containerName="extract-content" Nov 24 13:53:36 crc kubenswrapper[4790]: I1124 13:53:36.751031 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="54174d67-8e60-4894-94ba-28f535f52266" containerName="extract-content" Nov 24 13:53:36 crc kubenswrapper[4790]: E1124 13:53:36.751055 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54174d67-8e60-4894-94ba-28f535f52266" containerName="extract-utilities" Nov 24 13:53:36 crc kubenswrapper[4790]: I1124 13:53:36.751062 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="54174d67-8e60-4894-94ba-28f535f52266" containerName="extract-utilities" Nov 24 13:53:36 crc kubenswrapper[4790]: I1124 13:53:36.751181 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="54174d67-8e60-4894-94ba-28f535f52266" containerName="registry-server" Nov 24 13:53:36 crc kubenswrapper[4790]: I1124 13:53:36.752142 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-97gps" Nov 24 13:53:36 crc kubenswrapper[4790]: I1124 13:53:36.764198 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-97gps"] Nov 24 13:53:36 crc kubenswrapper[4790]: I1124 13:53:36.940815 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6vhg\" (UniqueName: \"kubernetes.io/projected/80841c5a-bd46-41a2-bd2c-ac8e59cda2f1-kube-api-access-l6vhg\") pod \"community-operators-97gps\" (UID: \"80841c5a-bd46-41a2-bd2c-ac8e59cda2f1\") " pod="openshift-marketplace/community-operators-97gps" Nov 24 13:53:36 crc kubenswrapper[4790]: I1124 13:53:36.940958 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80841c5a-bd46-41a2-bd2c-ac8e59cda2f1-utilities\") pod \"community-operators-97gps\" (UID: \"80841c5a-bd46-41a2-bd2c-ac8e59cda2f1\") " pod="openshift-marketplace/community-operators-97gps" Nov 24 13:53:36 crc kubenswrapper[4790]: I1124 13:53:36.941043 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80841c5a-bd46-41a2-bd2c-ac8e59cda2f1-catalog-content\") pod \"community-operators-97gps\" (UID: \"80841c5a-bd46-41a2-bd2c-ac8e59cda2f1\") " pod="openshift-marketplace/community-operators-97gps" Nov 24 13:53:37 crc kubenswrapper[4790]: I1124 13:53:37.042518 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6vhg\" (UniqueName: \"kubernetes.io/projected/80841c5a-bd46-41a2-bd2c-ac8e59cda2f1-kube-api-access-l6vhg\") pod \"community-operators-97gps\" (UID: \"80841c5a-bd46-41a2-bd2c-ac8e59cda2f1\") " pod="openshift-marketplace/community-operators-97gps" Nov 24 13:53:37 crc kubenswrapper[4790]: I1124 13:53:37.042649 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80841c5a-bd46-41a2-bd2c-ac8e59cda2f1-utilities\") pod \"community-operators-97gps\" (UID: \"80841c5a-bd46-41a2-bd2c-ac8e59cda2f1\") " pod="openshift-marketplace/community-operators-97gps" Nov 24 13:53:37 crc kubenswrapper[4790]: I1124 13:53:37.042698 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80841c5a-bd46-41a2-bd2c-ac8e59cda2f1-catalog-content\") pod \"community-operators-97gps\" (UID: \"80841c5a-bd46-41a2-bd2c-ac8e59cda2f1\") " pod="openshift-marketplace/community-operators-97gps" Nov 24 13:53:37 crc kubenswrapper[4790]: I1124 13:53:37.043422 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80841c5a-bd46-41a2-bd2c-ac8e59cda2f1-utilities\") pod \"community-operators-97gps\" (UID: \"80841c5a-bd46-41a2-bd2c-ac8e59cda2f1\") " pod="openshift-marketplace/community-operators-97gps" Nov 24 13:53:37 crc kubenswrapper[4790]: I1124 13:53:37.043529 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80841c5a-bd46-41a2-bd2c-ac8e59cda2f1-catalog-content\") pod \"community-operators-97gps\" (UID: \"80841c5a-bd46-41a2-bd2c-ac8e59cda2f1\") " pod="openshift-marketplace/community-operators-97gps" Nov 24 13:53:37 crc kubenswrapper[4790]: I1124 13:53:37.069175 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6vhg\" (UniqueName: \"kubernetes.io/projected/80841c5a-bd46-41a2-bd2c-ac8e59cda2f1-kube-api-access-l6vhg\") pod \"community-operators-97gps\" (UID: \"80841c5a-bd46-41a2-bd2c-ac8e59cda2f1\") " pod="openshift-marketplace/community-operators-97gps" Nov 24 13:53:37 crc kubenswrapper[4790]: I1124 13:53:37.073484 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-97gps" Nov 24 13:53:37 crc kubenswrapper[4790]: I1124 13:53:37.315918 4790 scope.go:117] "RemoveContainer" containerID="8310dee54c9b2a0ea195ce3e80b6281248186114131fdd393fbb8cb25d04910a" Nov 24 13:53:37 crc kubenswrapper[4790]: E1124 13:53:37.316375 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:53:37 crc kubenswrapper[4790]: I1124 13:53:37.344200 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-97gps"] Nov 24 13:53:37 crc kubenswrapper[4790]: I1124 13:53:37.971430 4790 generic.go:334] "Generic (PLEG): container finished" podID="80841c5a-bd46-41a2-bd2c-ac8e59cda2f1" containerID="987fbbf926b7439356d34bcc5d313aad754bafd7f54b5938a9fd88ff0d7a5e13" exitCode=0 Nov 24 13:53:37 crc kubenswrapper[4790]: I1124 13:53:37.971541 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-97gps" event={"ID":"80841c5a-bd46-41a2-bd2c-ac8e59cda2f1","Type":"ContainerDied","Data":"987fbbf926b7439356d34bcc5d313aad754bafd7f54b5938a9fd88ff0d7a5e13"} Nov 24 13:53:37 crc kubenswrapper[4790]: I1124 13:53:37.971766 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-97gps" event={"ID":"80841c5a-bd46-41a2-bd2c-ac8e59cda2f1","Type":"ContainerStarted","Data":"eb9e5ce20461ff26d6a43d34457b3ad7402be6b66346be22ba26001b80bb2109"} Nov 24 13:53:37 crc kubenswrapper[4790]: I1124 13:53:37.973681 4790 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 13:53:42 crc kubenswrapper[4790]: I1124 13:53:42.001705 4790 generic.go:334] "Generic (PLEG): container finished" podID="80841c5a-bd46-41a2-bd2c-ac8e59cda2f1" containerID="b964fc41bf324f22d7e6dd7aa6bde80943d294254f8a3ddf88ab216e3698d6d8" exitCode=0 Nov 24 13:53:42 crc kubenswrapper[4790]: I1124 13:53:42.001755 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-97gps" event={"ID":"80841c5a-bd46-41a2-bd2c-ac8e59cda2f1","Type":"ContainerDied","Data":"b964fc41bf324f22d7e6dd7aa6bde80943d294254f8a3ddf88ab216e3698d6d8"} Nov 24 13:53:43 crc kubenswrapper[4790]: I1124 13:53:43.009674 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-97gps" event={"ID":"80841c5a-bd46-41a2-bd2c-ac8e59cda2f1","Type":"ContainerStarted","Data":"1dff8f79eb33c86b294dc2a50b6034dd725baf71597c195ab8e49daa47d7ca87"} Nov 24 13:53:47 crc kubenswrapper[4790]: I1124 13:53:47.074115 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-97gps" Nov 24 13:53:47 crc kubenswrapper[4790]: I1124 13:53:47.074713 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-97gps" Nov 24 13:53:47 crc kubenswrapper[4790]: I1124 13:53:47.147281 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-97gps" Nov 24 13:53:47 crc kubenswrapper[4790]: I1124 13:53:47.180264 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-97gps" podStartSLOduration=6.508221984 podStartE2EDuration="11.180242774s" podCreationTimestamp="2025-11-24 13:53:36 +0000 UTC" firstStartedPulling="2025-11-24 13:53:37.973450187 +0000 UTC m=+2466.353343839" lastFinishedPulling="2025-11-24 13:53:42.645470967 +0000 UTC m=+2471.025364629" observedRunningTime="2025-11-24 13:53:43.030463908 +0000 UTC m=+2471.410357590" watchObservedRunningTime="2025-11-24 13:53:47.180242774 +0000 UTC m=+2475.560136446" Nov 24 13:53:48 crc kubenswrapper[4790]: I1124 13:53:48.106154 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-97gps" Nov 24 13:53:48 crc kubenswrapper[4790]: I1124 13:53:48.188770 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-97gps"] Nov 24 13:53:48 crc kubenswrapper[4790]: I1124 13:53:48.229798 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r2j97"] Nov 24 13:53:48 crc kubenswrapper[4790]: I1124 13:53:48.230094 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-r2j97" podUID="b790d0d0-0038-4153-8508-b5f974dda04c" containerName="registry-server" containerID="cri-o://1009438344b82ad72e3d2bc97fd6bf27af2d1b40ec336f7a8a508c0e939c478e" gracePeriod=2 Nov 24 13:53:49 crc kubenswrapper[4790]: I1124 13:53:49.058639 4790 generic.go:334] "Generic (PLEG): container finished" podID="b790d0d0-0038-4153-8508-b5f974dda04c" containerID="1009438344b82ad72e3d2bc97fd6bf27af2d1b40ec336f7a8a508c0e939c478e" exitCode=0 Nov 24 13:53:49 crc kubenswrapper[4790]: I1124 13:53:49.065188 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r2j97" event={"ID":"b790d0d0-0038-4153-8508-b5f974dda04c","Type":"ContainerDied","Data":"1009438344b82ad72e3d2bc97fd6bf27af2d1b40ec336f7a8a508c0e939c478e"} Nov 24 13:53:49 crc kubenswrapper[4790]: I1124 13:53:49.198114 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r2j97" Nov 24 13:53:49 crc kubenswrapper[4790]: I1124 13:53:49.315470 4790 scope.go:117] "RemoveContainer" containerID="8310dee54c9b2a0ea195ce3e80b6281248186114131fdd393fbb8cb25d04910a" Nov 24 13:53:49 crc kubenswrapper[4790]: E1124 13:53:49.315966 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:53:49 crc kubenswrapper[4790]: I1124 13:53:49.319654 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmn7x\" (UniqueName: \"kubernetes.io/projected/b790d0d0-0038-4153-8508-b5f974dda04c-kube-api-access-cmn7x\") pod \"b790d0d0-0038-4153-8508-b5f974dda04c\" (UID: \"b790d0d0-0038-4153-8508-b5f974dda04c\") " Nov 24 13:53:49 crc kubenswrapper[4790]: I1124 13:53:49.319773 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b790d0d0-0038-4153-8508-b5f974dda04c-utilities\") pod \"b790d0d0-0038-4153-8508-b5f974dda04c\" (UID: \"b790d0d0-0038-4153-8508-b5f974dda04c\") " Nov 24 13:53:49 crc kubenswrapper[4790]: I1124 13:53:49.320312 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b790d0d0-0038-4153-8508-b5f974dda04c-catalog-content\") pod \"b790d0d0-0038-4153-8508-b5f974dda04c\" (UID: \"b790d0d0-0038-4153-8508-b5f974dda04c\") " Nov 24 13:53:49 crc kubenswrapper[4790]: I1124 13:53:49.320588 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b790d0d0-0038-4153-8508-b5f974dda04c-utilities" (OuterVolumeSpecName: "utilities") pod "b790d0d0-0038-4153-8508-b5f974dda04c" (UID: "b790d0d0-0038-4153-8508-b5f974dda04c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:53:49 crc kubenswrapper[4790]: I1124 13:53:49.327480 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b790d0d0-0038-4153-8508-b5f974dda04c-kube-api-access-cmn7x" (OuterVolumeSpecName: "kube-api-access-cmn7x") pod "b790d0d0-0038-4153-8508-b5f974dda04c" (UID: "b790d0d0-0038-4153-8508-b5f974dda04c"). InnerVolumeSpecName "kube-api-access-cmn7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:53:49 crc kubenswrapper[4790]: I1124 13:53:49.361438 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b790d0d0-0038-4153-8508-b5f974dda04c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b790d0d0-0038-4153-8508-b5f974dda04c" (UID: "b790d0d0-0038-4153-8508-b5f974dda04c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:53:49 crc kubenswrapper[4790]: I1124 13:53:49.422513 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b790d0d0-0038-4153-8508-b5f974dda04c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:53:49 crc kubenswrapper[4790]: I1124 13:53:49.422813 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmn7x\" (UniqueName: \"kubernetes.io/projected/b790d0d0-0038-4153-8508-b5f974dda04c-kube-api-access-cmn7x\") on node \"crc\" DevicePath \"\"" Nov 24 13:53:49 crc kubenswrapper[4790]: I1124 13:53:49.422902 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b790d0d0-0038-4153-8508-b5f974dda04c-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:53:50 crc kubenswrapper[4790]: I1124 13:53:50.068412 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r2j97" event={"ID":"b790d0d0-0038-4153-8508-b5f974dda04c","Type":"ContainerDied","Data":"2b998ed567a370390fc27eb766a29f0317efd7587263e94e797e501fca823d98"} Nov 24 13:53:50 crc kubenswrapper[4790]: I1124 13:53:50.068465 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r2j97" Nov 24 13:53:50 crc kubenswrapper[4790]: I1124 13:53:50.068480 4790 scope.go:117] "RemoveContainer" containerID="1009438344b82ad72e3d2bc97fd6bf27af2d1b40ec336f7a8a508c0e939c478e" Nov 24 13:53:50 crc kubenswrapper[4790]: I1124 13:53:50.098119 4790 scope.go:117] "RemoveContainer" containerID="671625f2690507a78224d63267e12dd69a2e222abb0770bd6a718f7a7ff7d773" Nov 24 13:53:50 crc kubenswrapper[4790]: I1124 13:53:50.102371 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r2j97"] Nov 24 13:53:50 crc kubenswrapper[4790]: I1124 13:53:50.110603 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-r2j97"] Nov 24 13:53:50 crc kubenswrapper[4790]: I1124 13:53:50.118753 4790 scope.go:117] "RemoveContainer" containerID="8c635a22d6853ab2b43811d154d38b26922332371ba5f476e3809cdba17fd607" Nov 24 13:53:50 crc kubenswrapper[4790]: I1124 13:53:50.323780 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b790d0d0-0038-4153-8508-b5f974dda04c" path="/var/lib/kubelet/pods/b790d0d0-0038-4153-8508-b5f974dda04c/volumes" Nov 24 13:54:00 crc kubenswrapper[4790]: I1124 13:54:00.314596 4790 scope.go:117] "RemoveContainer" containerID="8310dee54c9b2a0ea195ce3e80b6281248186114131fdd393fbb8cb25d04910a" Nov 24 13:54:00 crc kubenswrapper[4790]: E1124 13:54:00.315243 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:54:12 crc kubenswrapper[4790]: I1124 13:54:12.318928 4790 scope.go:117] "RemoveContainer" containerID="8310dee54c9b2a0ea195ce3e80b6281248186114131fdd393fbb8cb25d04910a" Nov 24 13:54:12 crc kubenswrapper[4790]: E1124 13:54:12.319750 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:54:27 crc kubenswrapper[4790]: I1124 13:54:27.314179 4790 scope.go:117] "RemoveContainer" containerID="8310dee54c9b2a0ea195ce3e80b6281248186114131fdd393fbb8cb25d04910a" Nov 24 13:54:27 crc kubenswrapper[4790]: E1124 13:54:27.314986 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:54:42 crc kubenswrapper[4790]: I1124 13:54:42.320189 4790 scope.go:117] "RemoveContainer" containerID="8310dee54c9b2a0ea195ce3e80b6281248186114131fdd393fbb8cb25d04910a" Nov 24 13:54:42 crc kubenswrapper[4790]: E1124 13:54:42.320973 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:54:57 crc kubenswrapper[4790]: I1124 13:54:57.314510 4790 scope.go:117] "RemoveContainer" containerID="8310dee54c9b2a0ea195ce3e80b6281248186114131fdd393fbb8cb25d04910a" Nov 24 13:54:57 crc kubenswrapper[4790]: E1124 13:54:57.315126 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:55:12 crc kubenswrapper[4790]: I1124 13:55:12.322684 4790 scope.go:117] "RemoveContainer" containerID="8310dee54c9b2a0ea195ce3e80b6281248186114131fdd393fbb8cb25d04910a" Nov 24 13:55:12 crc kubenswrapper[4790]: E1124 13:55:12.323906 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:55:27 crc kubenswrapper[4790]: I1124 13:55:27.314787 4790 scope.go:117] "RemoveContainer" containerID="8310dee54c9b2a0ea195ce3e80b6281248186114131fdd393fbb8cb25d04910a" Nov 24 13:55:27 crc kubenswrapper[4790]: E1124 13:55:27.315870 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:55:39 crc kubenswrapper[4790]: I1124 13:55:39.315080 4790 scope.go:117] "RemoveContainer" containerID="8310dee54c9b2a0ea195ce3e80b6281248186114131fdd393fbb8cb25d04910a" Nov 24 13:55:39 crc kubenswrapper[4790]: E1124 13:55:39.316184 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:55:50 crc kubenswrapper[4790]: I1124 13:55:50.314338 4790 scope.go:117] "RemoveContainer" containerID="8310dee54c9b2a0ea195ce3e80b6281248186114131fdd393fbb8cb25d04910a" Nov 24 13:55:50 crc kubenswrapper[4790]: E1124 13:55:50.315112 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:56:01 crc kubenswrapper[4790]: I1124 13:56:01.314997 4790 scope.go:117] "RemoveContainer" containerID="8310dee54c9b2a0ea195ce3e80b6281248186114131fdd393fbb8cb25d04910a" Nov 24 13:56:01 crc kubenswrapper[4790]: E1124 13:56:01.315636 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:56:16 crc kubenswrapper[4790]: I1124 13:56:16.315561 4790 scope.go:117] "RemoveContainer" containerID="8310dee54c9b2a0ea195ce3e80b6281248186114131fdd393fbb8cb25d04910a" Nov 24 13:56:16 crc kubenswrapper[4790]: E1124 13:56:16.316675 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:56:30 crc kubenswrapper[4790]: I1124 13:56:30.314274 4790 scope.go:117] "RemoveContainer" containerID="8310dee54c9b2a0ea195ce3e80b6281248186114131fdd393fbb8cb25d04910a" Nov 24 13:56:30 crc kubenswrapper[4790]: E1124 13:56:30.314946 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:56:42 crc kubenswrapper[4790]: I1124 13:56:42.318788 4790 scope.go:117] "RemoveContainer" containerID="8310dee54c9b2a0ea195ce3e80b6281248186114131fdd393fbb8cb25d04910a" Nov 24 13:56:42 crc kubenswrapper[4790]: E1124 13:56:42.319534 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:56:55 crc kubenswrapper[4790]: I1124 13:56:55.314731 4790 scope.go:117] "RemoveContainer" containerID="8310dee54c9b2a0ea195ce3e80b6281248186114131fdd393fbb8cb25d04910a" Nov 24 13:56:55 crc kubenswrapper[4790]: E1124 13:56:55.316392 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:57:08 crc kubenswrapper[4790]: I1124 13:57:08.314808 4790 scope.go:117] "RemoveContainer" containerID="8310dee54c9b2a0ea195ce3e80b6281248186114131fdd393fbb8cb25d04910a" Nov 24 13:57:08 crc kubenswrapper[4790]: E1124 13:57:08.315596 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:57:21 crc kubenswrapper[4790]: I1124 13:57:21.314421 4790 scope.go:117] "RemoveContainer" containerID="8310dee54c9b2a0ea195ce3e80b6281248186114131fdd393fbb8cb25d04910a" Nov 24 13:57:21 crc kubenswrapper[4790]: E1124 13:57:21.315287 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:57:34 crc kubenswrapper[4790]: I1124 13:57:34.314958 4790 scope.go:117] "RemoveContainer" containerID="8310dee54c9b2a0ea195ce3e80b6281248186114131fdd393fbb8cb25d04910a" Nov 24 13:57:34 crc kubenswrapper[4790]: E1124 13:57:34.316682 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 13:57:46 crc kubenswrapper[4790]: I1124 13:57:46.315171 4790 scope.go:117] "RemoveContainer" containerID="8310dee54c9b2a0ea195ce3e80b6281248186114131fdd393fbb8cb25d04910a" Nov 24 13:57:46 crc kubenswrapper[4790]: I1124 13:57:46.832959 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"d5298328afc25b5319c1cc5ddb054410b112d29084866f4e201ce5c5fa1d883c"} Nov 24 13:58:00 crc kubenswrapper[4790]: I1124 13:58:00.653813 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-j7rl8"] Nov 24 13:58:00 crc kubenswrapper[4790]: E1124 13:58:00.654617 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b790d0d0-0038-4153-8508-b5f974dda04c" containerName="registry-server" Nov 24 13:58:00 crc kubenswrapper[4790]: I1124 13:58:00.654633 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="b790d0d0-0038-4153-8508-b5f974dda04c" containerName="registry-server" Nov 24 13:58:00 crc kubenswrapper[4790]: E1124 13:58:00.654647 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b790d0d0-0038-4153-8508-b5f974dda04c" containerName="extract-utilities" Nov 24 13:58:00 crc kubenswrapper[4790]: I1124 13:58:00.654655 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="b790d0d0-0038-4153-8508-b5f974dda04c" containerName="extract-utilities" Nov 24 13:58:00 crc kubenswrapper[4790]: E1124 13:58:00.654689 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b790d0d0-0038-4153-8508-b5f974dda04c" containerName="extract-content" Nov 24 13:58:00 crc kubenswrapper[4790]: I1124 13:58:00.654699 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="b790d0d0-0038-4153-8508-b5f974dda04c" containerName="extract-content" Nov 24 13:58:00 crc kubenswrapper[4790]: I1124 13:58:00.654871 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="b790d0d0-0038-4153-8508-b5f974dda04c" containerName="registry-server" Nov 24 13:58:00 crc kubenswrapper[4790]: I1124 13:58:00.656087 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j7rl8" Nov 24 13:58:00 crc kubenswrapper[4790]: I1124 13:58:00.673655 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j7rl8"] Nov 24 13:58:00 crc kubenswrapper[4790]: I1124 13:58:00.780807 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7468685e-a517-4dcc-af7d-f58e54eda771-catalog-content\") pod \"certified-operators-j7rl8\" (UID: \"7468685e-a517-4dcc-af7d-f58e54eda771\") " pod="openshift-marketplace/certified-operators-j7rl8" Nov 24 13:58:00 crc kubenswrapper[4790]: I1124 13:58:00.781453 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7468685e-a517-4dcc-af7d-f58e54eda771-utilities\") pod \"certified-operators-j7rl8\" (UID: \"7468685e-a517-4dcc-af7d-f58e54eda771\") " pod="openshift-marketplace/certified-operators-j7rl8" Nov 24 13:58:00 crc kubenswrapper[4790]: I1124 13:58:00.781803 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5lxp\" (UniqueName: \"kubernetes.io/projected/7468685e-a517-4dcc-af7d-f58e54eda771-kube-api-access-x5lxp\") pod \"certified-operators-j7rl8\" (UID: \"7468685e-a517-4dcc-af7d-f58e54eda771\") " pod="openshift-marketplace/certified-operators-j7rl8" Nov 24 13:58:00 crc kubenswrapper[4790]: I1124 13:58:00.882772 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7468685e-a517-4dcc-af7d-f58e54eda771-catalog-content\") pod \"certified-operators-j7rl8\" (UID: \"7468685e-a517-4dcc-af7d-f58e54eda771\") " pod="openshift-marketplace/certified-operators-j7rl8" Nov 24 13:58:00 crc kubenswrapper[4790]: I1124 13:58:00.882829 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7468685e-a517-4dcc-af7d-f58e54eda771-utilities\") pod \"certified-operators-j7rl8\" (UID: \"7468685e-a517-4dcc-af7d-f58e54eda771\") " pod="openshift-marketplace/certified-operators-j7rl8" Nov 24 13:58:00 crc kubenswrapper[4790]: I1124 13:58:00.882920 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5lxp\" (UniqueName: \"kubernetes.io/projected/7468685e-a517-4dcc-af7d-f58e54eda771-kube-api-access-x5lxp\") pod \"certified-operators-j7rl8\" (UID: \"7468685e-a517-4dcc-af7d-f58e54eda771\") " pod="openshift-marketplace/certified-operators-j7rl8" Nov 24 13:58:00 crc kubenswrapper[4790]: I1124 13:58:00.883280 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7468685e-a517-4dcc-af7d-f58e54eda771-catalog-content\") pod \"certified-operators-j7rl8\" (UID: \"7468685e-a517-4dcc-af7d-f58e54eda771\") " pod="openshift-marketplace/certified-operators-j7rl8" Nov 24 13:58:00 crc kubenswrapper[4790]: I1124 13:58:00.883661 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7468685e-a517-4dcc-af7d-f58e54eda771-utilities\") pod \"certified-operators-j7rl8\" (UID: \"7468685e-a517-4dcc-af7d-f58e54eda771\") " pod="openshift-marketplace/certified-operators-j7rl8" Nov 24 13:58:00 crc kubenswrapper[4790]: I1124 13:58:00.908728 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5lxp\" (UniqueName: \"kubernetes.io/projected/7468685e-a517-4dcc-af7d-f58e54eda771-kube-api-access-x5lxp\") pod \"certified-operators-j7rl8\" (UID: \"7468685e-a517-4dcc-af7d-f58e54eda771\") " pod="openshift-marketplace/certified-operators-j7rl8" Nov 24 13:58:00 crc kubenswrapper[4790]: I1124 13:58:00.988694 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j7rl8" Nov 24 13:58:01 crc kubenswrapper[4790]: I1124 13:58:01.453945 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j7rl8"] Nov 24 13:58:01 crc kubenswrapper[4790]: I1124 13:58:01.941116 4790 generic.go:334] "Generic (PLEG): container finished" podID="7468685e-a517-4dcc-af7d-f58e54eda771" containerID="de4e13c8dd9cba8fcd780ad6e6ac65fea34119b15b1cae8837e6bb88a66eb3d7" exitCode=0 Nov 24 13:58:01 crc kubenswrapper[4790]: I1124 13:58:01.941180 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j7rl8" event={"ID":"7468685e-a517-4dcc-af7d-f58e54eda771","Type":"ContainerDied","Data":"de4e13c8dd9cba8fcd780ad6e6ac65fea34119b15b1cae8837e6bb88a66eb3d7"} Nov 24 13:58:01 crc kubenswrapper[4790]: I1124 13:58:01.941218 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j7rl8" event={"ID":"7468685e-a517-4dcc-af7d-f58e54eda771","Type":"ContainerStarted","Data":"c8269000bcfd66e6af89fa80618313417e8fb28b46d907ab09aeb8753b5a5de0"} Nov 24 13:58:02 crc kubenswrapper[4790]: I1124 13:58:02.950112 4790 generic.go:334] "Generic (PLEG): container finished" podID="7468685e-a517-4dcc-af7d-f58e54eda771" containerID="f57ef45117c66d2f102c4f7e3aeda448670b7a4f83c25c5b680364457c54d58c" exitCode=0 Nov 24 13:58:02 crc kubenswrapper[4790]: I1124 13:58:02.950221 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j7rl8" event={"ID":"7468685e-a517-4dcc-af7d-f58e54eda771","Type":"ContainerDied","Data":"f57ef45117c66d2f102c4f7e3aeda448670b7a4f83c25c5b680364457c54d58c"} Nov 24 13:58:03 crc kubenswrapper[4790]: I1124 13:58:03.960404 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j7rl8" event={"ID":"7468685e-a517-4dcc-af7d-f58e54eda771","Type":"ContainerStarted","Data":"edc4dfae7239809312c3c2ff1e35796b0fd63d130639ee58fa58a214ee1bb34c"} Nov 24 13:58:03 crc kubenswrapper[4790]: I1124 13:58:03.979015 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-j7rl8" podStartSLOduration=2.551085451 podStartE2EDuration="3.978992251s" podCreationTimestamp="2025-11-24 13:58:00 +0000 UTC" firstStartedPulling="2025-11-24 13:58:01.94274264 +0000 UTC m=+2730.322636322" lastFinishedPulling="2025-11-24 13:58:03.37064946 +0000 UTC m=+2731.750543122" observedRunningTime="2025-11-24 13:58:03.976357765 +0000 UTC m=+2732.356251437" watchObservedRunningTime="2025-11-24 13:58:03.978992251 +0000 UTC m=+2732.358885913" Nov 24 13:58:04 crc kubenswrapper[4790]: I1124 13:58:04.646619 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-j6hzv"] Nov 24 13:58:04 crc kubenswrapper[4790]: I1124 13:58:04.648555 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j6hzv" Nov 24 13:58:04 crc kubenswrapper[4790]: I1124 13:58:04.666236 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j6hzv"] Nov 24 13:58:04 crc kubenswrapper[4790]: I1124 13:58:04.762993 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0980c6b4-8327-4046-9f89-ceb79b90786d-utilities\") pod \"redhat-marketplace-j6hzv\" (UID: \"0980c6b4-8327-4046-9f89-ceb79b90786d\") " pod="openshift-marketplace/redhat-marketplace-j6hzv" Nov 24 13:58:04 crc kubenswrapper[4790]: I1124 13:58:04.763632 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55jbd\" (UniqueName: \"kubernetes.io/projected/0980c6b4-8327-4046-9f89-ceb79b90786d-kube-api-access-55jbd\") pod \"redhat-marketplace-j6hzv\" (UID: \"0980c6b4-8327-4046-9f89-ceb79b90786d\") " pod="openshift-marketplace/redhat-marketplace-j6hzv" Nov 24 13:58:04 crc kubenswrapper[4790]: I1124 13:58:04.763737 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0980c6b4-8327-4046-9f89-ceb79b90786d-catalog-content\") pod \"redhat-marketplace-j6hzv\" (UID: \"0980c6b4-8327-4046-9f89-ceb79b90786d\") " pod="openshift-marketplace/redhat-marketplace-j6hzv" Nov 24 13:58:04 crc kubenswrapper[4790]: I1124 13:58:04.865763 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0980c6b4-8327-4046-9f89-ceb79b90786d-utilities\") pod \"redhat-marketplace-j6hzv\" (UID: \"0980c6b4-8327-4046-9f89-ceb79b90786d\") " pod="openshift-marketplace/redhat-marketplace-j6hzv" Nov 24 13:58:04 crc kubenswrapper[4790]: I1124 13:58:04.865847 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55jbd\" (UniqueName: \"kubernetes.io/projected/0980c6b4-8327-4046-9f89-ceb79b90786d-kube-api-access-55jbd\") pod \"redhat-marketplace-j6hzv\" (UID: \"0980c6b4-8327-4046-9f89-ceb79b90786d\") " pod="openshift-marketplace/redhat-marketplace-j6hzv" Nov 24 13:58:04 crc kubenswrapper[4790]: I1124 13:58:04.865865 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0980c6b4-8327-4046-9f89-ceb79b90786d-catalog-content\") pod \"redhat-marketplace-j6hzv\" (UID: \"0980c6b4-8327-4046-9f89-ceb79b90786d\") " pod="openshift-marketplace/redhat-marketplace-j6hzv" Nov 24 13:58:04 crc kubenswrapper[4790]: I1124 13:58:04.866326 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0980c6b4-8327-4046-9f89-ceb79b90786d-catalog-content\") pod \"redhat-marketplace-j6hzv\" (UID: \"0980c6b4-8327-4046-9f89-ceb79b90786d\") " pod="openshift-marketplace/redhat-marketplace-j6hzv" Nov 24 13:58:04 crc kubenswrapper[4790]: I1124 13:58:04.866527 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0980c6b4-8327-4046-9f89-ceb79b90786d-utilities\") pod \"redhat-marketplace-j6hzv\" (UID: \"0980c6b4-8327-4046-9f89-ceb79b90786d\") " pod="openshift-marketplace/redhat-marketplace-j6hzv" Nov 24 13:58:04 crc kubenswrapper[4790]: I1124 13:58:04.886125 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55jbd\" (UniqueName: \"kubernetes.io/projected/0980c6b4-8327-4046-9f89-ceb79b90786d-kube-api-access-55jbd\") pod \"redhat-marketplace-j6hzv\" (UID: \"0980c6b4-8327-4046-9f89-ceb79b90786d\") " pod="openshift-marketplace/redhat-marketplace-j6hzv" Nov 24 13:58:04 crc kubenswrapper[4790]: I1124 13:58:04.969194 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j6hzv" Nov 24 13:58:05 crc kubenswrapper[4790]: I1124 13:58:05.386468 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j6hzv"] Nov 24 13:58:05 crc kubenswrapper[4790]: I1124 13:58:05.651460 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-96c9c"] Nov 24 13:58:05 crc kubenswrapper[4790]: I1124 13:58:05.653142 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-96c9c" Nov 24 13:58:05 crc kubenswrapper[4790]: I1124 13:58:05.679915 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-96c9c"] Nov 24 13:58:05 crc kubenswrapper[4790]: I1124 13:58:05.782585 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90d19b1e-1040-456c-be95-32e15e32a013-catalog-content\") pod \"redhat-operators-96c9c\" (UID: \"90d19b1e-1040-456c-be95-32e15e32a013\") " pod="openshift-marketplace/redhat-operators-96c9c" Nov 24 13:58:05 crc kubenswrapper[4790]: I1124 13:58:05.782636 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvv7x\" (UniqueName: \"kubernetes.io/projected/90d19b1e-1040-456c-be95-32e15e32a013-kube-api-access-fvv7x\") pod \"redhat-operators-96c9c\" (UID: \"90d19b1e-1040-456c-be95-32e15e32a013\") " pod="openshift-marketplace/redhat-operators-96c9c" Nov 24 13:58:05 crc kubenswrapper[4790]: I1124 13:58:05.782668 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90d19b1e-1040-456c-be95-32e15e32a013-utilities\") pod \"redhat-operators-96c9c\" (UID: \"90d19b1e-1040-456c-be95-32e15e32a013\") " pod="openshift-marketplace/redhat-operators-96c9c" Nov 24 13:58:05 crc kubenswrapper[4790]: I1124 13:58:05.884014 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90d19b1e-1040-456c-be95-32e15e32a013-catalog-content\") pod \"redhat-operators-96c9c\" (UID: \"90d19b1e-1040-456c-be95-32e15e32a013\") " pod="openshift-marketplace/redhat-operators-96c9c" Nov 24 13:58:05 crc kubenswrapper[4790]: I1124 13:58:05.884091 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvv7x\" (UniqueName: \"kubernetes.io/projected/90d19b1e-1040-456c-be95-32e15e32a013-kube-api-access-fvv7x\") pod \"redhat-operators-96c9c\" (UID: \"90d19b1e-1040-456c-be95-32e15e32a013\") " pod="openshift-marketplace/redhat-operators-96c9c" Nov 24 13:58:05 crc kubenswrapper[4790]: I1124 13:58:05.884138 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90d19b1e-1040-456c-be95-32e15e32a013-utilities\") pod \"redhat-operators-96c9c\" (UID: \"90d19b1e-1040-456c-be95-32e15e32a013\") " pod="openshift-marketplace/redhat-operators-96c9c" Nov 24 13:58:05 crc kubenswrapper[4790]: I1124 13:58:05.884584 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90d19b1e-1040-456c-be95-32e15e32a013-catalog-content\") pod \"redhat-operators-96c9c\" (UID: \"90d19b1e-1040-456c-be95-32e15e32a013\") " pod="openshift-marketplace/redhat-operators-96c9c" Nov 24 13:58:05 crc kubenswrapper[4790]: I1124 13:58:05.884668 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90d19b1e-1040-456c-be95-32e15e32a013-utilities\") pod \"redhat-operators-96c9c\" (UID: \"90d19b1e-1040-456c-be95-32e15e32a013\") " pod="openshift-marketplace/redhat-operators-96c9c" Nov 24 13:58:05 crc kubenswrapper[4790]: I1124 13:58:05.903464 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvv7x\" (UniqueName: \"kubernetes.io/projected/90d19b1e-1040-456c-be95-32e15e32a013-kube-api-access-fvv7x\") pod \"redhat-operators-96c9c\" (UID: \"90d19b1e-1040-456c-be95-32e15e32a013\") " pod="openshift-marketplace/redhat-operators-96c9c" Nov 24 13:58:05 crc kubenswrapper[4790]: I1124 13:58:05.974922 4790 generic.go:334] "Generic (PLEG): container finished" podID="0980c6b4-8327-4046-9f89-ceb79b90786d" containerID="aa2af5c99319fa7e0465591dd209487cc5300292d117ec2f2bcfcd04eef048b9" exitCode=0 Nov 24 13:58:05 crc kubenswrapper[4790]: I1124 13:58:05.975008 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j6hzv" event={"ID":"0980c6b4-8327-4046-9f89-ceb79b90786d","Type":"ContainerDied","Data":"aa2af5c99319fa7e0465591dd209487cc5300292d117ec2f2bcfcd04eef048b9"} Nov 24 13:58:05 crc kubenswrapper[4790]: I1124 13:58:05.975242 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j6hzv" event={"ID":"0980c6b4-8327-4046-9f89-ceb79b90786d","Type":"ContainerStarted","Data":"f9796f76f19486c71a6d65a889d0538f071b6ec388375636b7769db0c427f369"} Nov 24 13:58:05 crc kubenswrapper[4790]: I1124 13:58:05.985824 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-96c9c" Nov 24 13:58:06 crc kubenswrapper[4790]: I1124 13:58:06.450717 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-96c9c"] Nov 24 13:58:06 crc kubenswrapper[4790]: W1124 13:58:06.455140 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod90d19b1e_1040_456c_be95_32e15e32a013.slice/crio-4ab30ec3802395732f52b6089bc34c6ba50fe8ef92c512cfaea226979e33a948 WatchSource:0}: Error finding container 4ab30ec3802395732f52b6089bc34c6ba50fe8ef92c512cfaea226979e33a948: Status 404 returned error can't find the container with id 4ab30ec3802395732f52b6089bc34c6ba50fe8ef92c512cfaea226979e33a948 Nov 24 13:58:06 crc kubenswrapper[4790]: I1124 13:58:06.982276 4790 generic.go:334] "Generic (PLEG): container finished" podID="0980c6b4-8327-4046-9f89-ceb79b90786d" containerID="2b2d098b161fe0d91cec193801a4857b8ffb0ce25e3381d82e3df0ac58743b5b" exitCode=0 Nov 24 13:58:06 crc kubenswrapper[4790]: I1124 13:58:06.982332 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j6hzv" event={"ID":"0980c6b4-8327-4046-9f89-ceb79b90786d","Type":"ContainerDied","Data":"2b2d098b161fe0d91cec193801a4857b8ffb0ce25e3381d82e3df0ac58743b5b"} Nov 24 13:58:06 crc kubenswrapper[4790]: I1124 13:58:06.984162 4790 generic.go:334] "Generic (PLEG): container finished" podID="90d19b1e-1040-456c-be95-32e15e32a013" containerID="7ff1828fb9b063ab7d9298790fc0fec064882d7429407dea6c805ebfc16c6313" exitCode=0 Nov 24 13:58:06 crc kubenswrapper[4790]: I1124 13:58:06.984188 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-96c9c" event={"ID":"90d19b1e-1040-456c-be95-32e15e32a013","Type":"ContainerDied","Data":"7ff1828fb9b063ab7d9298790fc0fec064882d7429407dea6c805ebfc16c6313"} Nov 24 13:58:06 crc kubenswrapper[4790]: I1124 13:58:06.984211 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-96c9c" event={"ID":"90d19b1e-1040-456c-be95-32e15e32a013","Type":"ContainerStarted","Data":"4ab30ec3802395732f52b6089bc34c6ba50fe8ef92c512cfaea226979e33a948"} Nov 24 13:58:07 crc kubenswrapper[4790]: I1124 13:58:07.993235 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j6hzv" event={"ID":"0980c6b4-8327-4046-9f89-ceb79b90786d","Type":"ContainerStarted","Data":"2c251051738bb3da0d9648c2b090526aa6e42b1d05f00be680110b74593b796f"} Nov 24 13:58:08 crc kubenswrapper[4790]: I1124 13:58:08.011775 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-j6hzv" podStartSLOduration=2.626073104 podStartE2EDuration="4.011751746s" podCreationTimestamp="2025-11-24 13:58:04 +0000 UTC" firstStartedPulling="2025-11-24 13:58:05.97673418 +0000 UTC m=+2734.356627842" lastFinishedPulling="2025-11-24 13:58:07.362412822 +0000 UTC m=+2735.742306484" observedRunningTime="2025-11-24 13:58:08.00945871 +0000 UTC m=+2736.389352372" watchObservedRunningTime="2025-11-24 13:58:08.011751746 +0000 UTC m=+2736.391645408" Nov 24 13:58:10 crc kubenswrapper[4790]: I1124 13:58:10.989458 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-j7rl8" Nov 24 13:58:10 crc kubenswrapper[4790]: I1124 13:58:10.989751 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-j7rl8" Nov 24 13:58:11 crc kubenswrapper[4790]: I1124 13:58:11.034625 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-j7rl8" Nov 24 13:58:11 crc kubenswrapper[4790]: I1124 13:58:11.078169 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-j7rl8" Nov 24 13:58:14 crc kubenswrapper[4790]: I1124 13:58:14.036865 4790 generic.go:334] "Generic (PLEG): container finished" podID="90d19b1e-1040-456c-be95-32e15e32a013" containerID="56452aac30a9719549303efe43baa6aff815459a7d2666c6e2bbeee31943252c" exitCode=0 Nov 24 13:58:14 crc kubenswrapper[4790]: I1124 13:58:14.036928 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-96c9c" event={"ID":"90d19b1e-1040-456c-be95-32e15e32a013","Type":"ContainerDied","Data":"56452aac30a9719549303efe43baa6aff815459a7d2666c6e2bbeee31943252c"} Nov 24 13:58:14 crc kubenswrapper[4790]: I1124 13:58:14.970050 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-j6hzv" Nov 24 13:58:14 crc kubenswrapper[4790]: I1124 13:58:14.970114 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-j6hzv" Nov 24 13:58:15 crc kubenswrapper[4790]: I1124 13:58:15.028536 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-j6hzv" Nov 24 13:58:15 crc kubenswrapper[4790]: I1124 13:58:15.044842 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-96c9c" event={"ID":"90d19b1e-1040-456c-be95-32e15e32a013","Type":"ContainerStarted","Data":"caec78013ff1d9d5db7fa09e36e0b2f3ef3901a99d4b732afc888710af5a6a20"} Nov 24 13:58:15 crc kubenswrapper[4790]: I1124 13:58:15.067024 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-96c9c" podStartSLOduration=2.601412765 podStartE2EDuration="10.067008888s" podCreationTimestamp="2025-11-24 13:58:05 +0000 UTC" firstStartedPulling="2025-11-24 13:58:06.985168279 +0000 UTC m=+2735.365061941" lastFinishedPulling="2025-11-24 13:58:14.450764402 +0000 UTC m=+2742.830658064" observedRunningTime="2025-11-24 13:58:15.061788178 +0000 UTC m=+2743.441681840" watchObservedRunningTime="2025-11-24 13:58:15.067008888 +0000 UTC m=+2743.446902550" Nov 24 13:58:15 crc kubenswrapper[4790]: I1124 13:58:15.108071 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-j6hzv" Nov 24 13:58:15 crc kubenswrapper[4790]: I1124 13:58:15.986140 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-96c9c" Nov 24 13:58:15 crc kubenswrapper[4790]: I1124 13:58:15.986489 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-96c9c" Nov 24 13:58:16 crc kubenswrapper[4790]: I1124 13:58:16.641034 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j7rl8"] Nov 24 13:58:16 crc kubenswrapper[4790]: I1124 13:58:16.641303 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-j7rl8" podUID="7468685e-a517-4dcc-af7d-f58e54eda771" containerName="registry-server" containerID="cri-o://edc4dfae7239809312c3c2ff1e35796b0fd63d130639ee58fa58a214ee1bb34c" gracePeriod=2 Nov 24 13:58:17 crc kubenswrapper[4790]: I1124 13:58:17.033742 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-96c9c" podUID="90d19b1e-1040-456c-be95-32e15e32a013" containerName="registry-server" probeResult="failure" output=< Nov 24 13:58:17 crc kubenswrapper[4790]: timeout: failed to connect service ":50051" within 1s Nov 24 13:58:17 crc kubenswrapper[4790]: > Nov 24 13:58:17 crc kubenswrapper[4790]: I1124 13:58:17.062313 4790 generic.go:334] "Generic (PLEG): container finished" podID="7468685e-a517-4dcc-af7d-f58e54eda771" containerID="edc4dfae7239809312c3c2ff1e35796b0fd63d130639ee58fa58a214ee1bb34c" exitCode=0 Nov 24 13:58:17 crc kubenswrapper[4790]: I1124 13:58:17.062403 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j7rl8" event={"ID":"7468685e-a517-4dcc-af7d-f58e54eda771","Type":"ContainerDied","Data":"edc4dfae7239809312c3c2ff1e35796b0fd63d130639ee58fa58a214ee1bb34c"} Nov 24 13:58:18 crc kubenswrapper[4790]: I1124 13:58:18.422683 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j7rl8" Nov 24 13:58:18 crc kubenswrapper[4790]: I1124 13:58:18.468416 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7468685e-a517-4dcc-af7d-f58e54eda771-catalog-content\") pod \"7468685e-a517-4dcc-af7d-f58e54eda771\" (UID: \"7468685e-a517-4dcc-af7d-f58e54eda771\") " Nov 24 13:58:18 crc kubenswrapper[4790]: I1124 13:58:18.468566 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5lxp\" (UniqueName: \"kubernetes.io/projected/7468685e-a517-4dcc-af7d-f58e54eda771-kube-api-access-x5lxp\") pod \"7468685e-a517-4dcc-af7d-f58e54eda771\" (UID: \"7468685e-a517-4dcc-af7d-f58e54eda771\") " Nov 24 13:58:18 crc kubenswrapper[4790]: I1124 13:58:18.468674 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7468685e-a517-4dcc-af7d-f58e54eda771-utilities\") pod \"7468685e-a517-4dcc-af7d-f58e54eda771\" (UID: \"7468685e-a517-4dcc-af7d-f58e54eda771\") " Nov 24 13:58:18 crc kubenswrapper[4790]: I1124 13:58:18.470587 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7468685e-a517-4dcc-af7d-f58e54eda771-utilities" (OuterVolumeSpecName: "utilities") pod "7468685e-a517-4dcc-af7d-f58e54eda771" (UID: "7468685e-a517-4dcc-af7d-f58e54eda771"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:58:18 crc kubenswrapper[4790]: I1124 13:58:18.474670 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7468685e-a517-4dcc-af7d-f58e54eda771-kube-api-access-x5lxp" (OuterVolumeSpecName: "kube-api-access-x5lxp") pod "7468685e-a517-4dcc-af7d-f58e54eda771" (UID: "7468685e-a517-4dcc-af7d-f58e54eda771"). InnerVolumeSpecName "kube-api-access-x5lxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:58:18 crc kubenswrapper[4790]: I1124 13:58:18.543074 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7468685e-a517-4dcc-af7d-f58e54eda771-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7468685e-a517-4dcc-af7d-f58e54eda771" (UID: "7468685e-a517-4dcc-af7d-f58e54eda771"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:58:18 crc kubenswrapper[4790]: I1124 13:58:18.571920 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7468685e-a517-4dcc-af7d-f58e54eda771-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:58:18 crc kubenswrapper[4790]: I1124 13:58:18.571963 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5lxp\" (UniqueName: \"kubernetes.io/projected/7468685e-a517-4dcc-af7d-f58e54eda771-kube-api-access-x5lxp\") on node \"crc\" DevicePath \"\"" Nov 24 13:58:18 crc kubenswrapper[4790]: I1124 13:58:18.571976 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7468685e-a517-4dcc-af7d-f58e54eda771-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:58:19 crc kubenswrapper[4790]: I1124 13:58:19.078011 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j7rl8" event={"ID":"7468685e-a517-4dcc-af7d-f58e54eda771","Type":"ContainerDied","Data":"c8269000bcfd66e6af89fa80618313417e8fb28b46d907ab09aeb8753b5a5de0"} Nov 24 13:58:19 crc kubenswrapper[4790]: I1124 13:58:19.078071 4790 scope.go:117] "RemoveContainer" containerID="edc4dfae7239809312c3c2ff1e35796b0fd63d130639ee58fa58a214ee1bb34c" Nov 24 13:58:19 crc kubenswrapper[4790]: I1124 13:58:19.078118 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j7rl8" Nov 24 13:58:19 crc kubenswrapper[4790]: I1124 13:58:19.101000 4790 scope.go:117] "RemoveContainer" containerID="f57ef45117c66d2f102c4f7e3aeda448670b7a4f83c25c5b680364457c54d58c" Nov 24 13:58:19 crc kubenswrapper[4790]: I1124 13:58:19.111036 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j7rl8"] Nov 24 13:58:19 crc kubenswrapper[4790]: I1124 13:58:19.123349 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-j7rl8"] Nov 24 13:58:19 crc kubenswrapper[4790]: I1124 13:58:19.149122 4790 scope.go:117] "RemoveContainer" containerID="de4e13c8dd9cba8fcd780ad6e6ac65fea34119b15b1cae8837e6bb88a66eb3d7" Nov 24 13:58:20 crc kubenswrapper[4790]: I1124 13:58:20.042889 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j6hzv"] Nov 24 13:58:20 crc kubenswrapper[4790]: I1124 13:58:20.043121 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-j6hzv" podUID="0980c6b4-8327-4046-9f89-ceb79b90786d" containerName="registry-server" containerID="cri-o://2c251051738bb3da0d9648c2b090526aa6e42b1d05f00be680110b74593b796f" gracePeriod=2 Nov 24 13:58:20 crc kubenswrapper[4790]: E1124 13:58:20.121667 4790 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0980c6b4_8327_4046_9f89_ceb79b90786d.slice/crio-2c251051738bb3da0d9648c2b090526aa6e42b1d05f00be680110b74593b796f.scope\": RecentStats: unable to find data in memory cache]" Nov 24 13:58:20 crc kubenswrapper[4790]: I1124 13:58:20.323911 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7468685e-a517-4dcc-af7d-f58e54eda771" path="/var/lib/kubelet/pods/7468685e-a517-4dcc-af7d-f58e54eda771/volumes" Nov 24 13:58:20 crc kubenswrapper[4790]: I1124 13:58:20.947140 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j6hzv" Nov 24 13:58:21 crc kubenswrapper[4790]: I1124 13:58:21.005297 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0980c6b4-8327-4046-9f89-ceb79b90786d-catalog-content\") pod \"0980c6b4-8327-4046-9f89-ceb79b90786d\" (UID: \"0980c6b4-8327-4046-9f89-ceb79b90786d\") " Nov 24 13:58:21 crc kubenswrapper[4790]: I1124 13:58:21.005346 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55jbd\" (UniqueName: \"kubernetes.io/projected/0980c6b4-8327-4046-9f89-ceb79b90786d-kube-api-access-55jbd\") pod \"0980c6b4-8327-4046-9f89-ceb79b90786d\" (UID: \"0980c6b4-8327-4046-9f89-ceb79b90786d\") " Nov 24 13:58:21 crc kubenswrapper[4790]: I1124 13:58:21.005380 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0980c6b4-8327-4046-9f89-ceb79b90786d-utilities\") pod \"0980c6b4-8327-4046-9f89-ceb79b90786d\" (UID: \"0980c6b4-8327-4046-9f89-ceb79b90786d\") " Nov 24 13:58:21 crc kubenswrapper[4790]: I1124 13:58:21.006455 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0980c6b4-8327-4046-9f89-ceb79b90786d-utilities" (OuterVolumeSpecName: "utilities") pod "0980c6b4-8327-4046-9f89-ceb79b90786d" (UID: "0980c6b4-8327-4046-9f89-ceb79b90786d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:58:21 crc kubenswrapper[4790]: I1124 13:58:21.011065 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0980c6b4-8327-4046-9f89-ceb79b90786d-kube-api-access-55jbd" (OuterVolumeSpecName: "kube-api-access-55jbd") pod "0980c6b4-8327-4046-9f89-ceb79b90786d" (UID: "0980c6b4-8327-4046-9f89-ceb79b90786d"). InnerVolumeSpecName "kube-api-access-55jbd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:58:21 crc kubenswrapper[4790]: I1124 13:58:21.021847 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0980c6b4-8327-4046-9f89-ceb79b90786d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0980c6b4-8327-4046-9f89-ceb79b90786d" (UID: "0980c6b4-8327-4046-9f89-ceb79b90786d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:58:21 crc kubenswrapper[4790]: I1124 13:58:21.095319 4790 generic.go:334] "Generic (PLEG): container finished" podID="0980c6b4-8327-4046-9f89-ceb79b90786d" containerID="2c251051738bb3da0d9648c2b090526aa6e42b1d05f00be680110b74593b796f" exitCode=0 Nov 24 13:58:21 crc kubenswrapper[4790]: I1124 13:58:21.095368 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j6hzv" event={"ID":"0980c6b4-8327-4046-9f89-ceb79b90786d","Type":"ContainerDied","Data":"2c251051738bb3da0d9648c2b090526aa6e42b1d05f00be680110b74593b796f"} Nov 24 13:58:21 crc kubenswrapper[4790]: I1124 13:58:21.095403 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j6hzv" event={"ID":"0980c6b4-8327-4046-9f89-ceb79b90786d","Type":"ContainerDied","Data":"f9796f76f19486c71a6d65a889d0538f071b6ec388375636b7769db0c427f369"} Nov 24 13:58:21 crc kubenswrapper[4790]: I1124 13:58:21.095424 4790 scope.go:117] "RemoveContainer" containerID="2c251051738bb3da0d9648c2b090526aa6e42b1d05f00be680110b74593b796f" Nov 24 13:58:21 crc kubenswrapper[4790]: I1124 13:58:21.095454 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j6hzv" Nov 24 13:58:21 crc kubenswrapper[4790]: I1124 13:58:21.106831 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0980c6b4-8327-4046-9f89-ceb79b90786d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:58:21 crc kubenswrapper[4790]: I1124 13:58:21.106858 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55jbd\" (UniqueName: \"kubernetes.io/projected/0980c6b4-8327-4046-9f89-ceb79b90786d-kube-api-access-55jbd\") on node \"crc\" DevicePath \"\"" Nov 24 13:58:21 crc kubenswrapper[4790]: I1124 13:58:21.106868 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0980c6b4-8327-4046-9f89-ceb79b90786d-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:58:21 crc kubenswrapper[4790]: I1124 13:58:21.117224 4790 scope.go:117] "RemoveContainer" containerID="2b2d098b161fe0d91cec193801a4857b8ffb0ce25e3381d82e3df0ac58743b5b" Nov 24 13:58:21 crc kubenswrapper[4790]: I1124 13:58:21.131848 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j6hzv"] Nov 24 13:58:21 crc kubenswrapper[4790]: I1124 13:58:21.137693 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-j6hzv"] Nov 24 13:58:21 crc kubenswrapper[4790]: I1124 13:58:21.156137 4790 scope.go:117] "RemoveContainer" containerID="aa2af5c99319fa7e0465591dd209487cc5300292d117ec2f2bcfcd04eef048b9" Nov 24 13:58:21 crc kubenswrapper[4790]: I1124 13:58:21.173611 4790 scope.go:117] "RemoveContainer" containerID="2c251051738bb3da0d9648c2b090526aa6e42b1d05f00be680110b74593b796f" Nov 24 13:58:21 crc kubenswrapper[4790]: E1124 13:58:21.174031 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c251051738bb3da0d9648c2b090526aa6e42b1d05f00be680110b74593b796f\": container with ID starting with 2c251051738bb3da0d9648c2b090526aa6e42b1d05f00be680110b74593b796f not found: ID does not exist" containerID="2c251051738bb3da0d9648c2b090526aa6e42b1d05f00be680110b74593b796f" Nov 24 13:58:21 crc kubenswrapper[4790]: I1124 13:58:21.174073 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c251051738bb3da0d9648c2b090526aa6e42b1d05f00be680110b74593b796f"} err="failed to get container status \"2c251051738bb3da0d9648c2b090526aa6e42b1d05f00be680110b74593b796f\": rpc error: code = NotFound desc = could not find container \"2c251051738bb3da0d9648c2b090526aa6e42b1d05f00be680110b74593b796f\": container with ID starting with 2c251051738bb3da0d9648c2b090526aa6e42b1d05f00be680110b74593b796f not found: ID does not exist" Nov 24 13:58:21 crc kubenswrapper[4790]: I1124 13:58:21.174102 4790 scope.go:117] "RemoveContainer" containerID="2b2d098b161fe0d91cec193801a4857b8ffb0ce25e3381d82e3df0ac58743b5b" Nov 24 13:58:21 crc kubenswrapper[4790]: E1124 13:58:21.174384 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b2d098b161fe0d91cec193801a4857b8ffb0ce25e3381d82e3df0ac58743b5b\": container with ID starting with 2b2d098b161fe0d91cec193801a4857b8ffb0ce25e3381d82e3df0ac58743b5b not found: ID does not exist" containerID="2b2d098b161fe0d91cec193801a4857b8ffb0ce25e3381d82e3df0ac58743b5b" Nov 24 13:58:21 crc kubenswrapper[4790]: I1124 13:58:21.174413 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b2d098b161fe0d91cec193801a4857b8ffb0ce25e3381d82e3df0ac58743b5b"} err="failed to get container status \"2b2d098b161fe0d91cec193801a4857b8ffb0ce25e3381d82e3df0ac58743b5b\": rpc error: code = NotFound desc = could not find container \"2b2d098b161fe0d91cec193801a4857b8ffb0ce25e3381d82e3df0ac58743b5b\": container with ID starting with 2b2d098b161fe0d91cec193801a4857b8ffb0ce25e3381d82e3df0ac58743b5b not found: ID does not exist" Nov 24 13:58:21 crc kubenswrapper[4790]: I1124 13:58:21.174436 4790 scope.go:117] "RemoveContainer" containerID="aa2af5c99319fa7e0465591dd209487cc5300292d117ec2f2bcfcd04eef048b9" Nov 24 13:58:21 crc kubenswrapper[4790]: E1124 13:58:21.174631 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa2af5c99319fa7e0465591dd209487cc5300292d117ec2f2bcfcd04eef048b9\": container with ID starting with aa2af5c99319fa7e0465591dd209487cc5300292d117ec2f2bcfcd04eef048b9 not found: ID does not exist" containerID="aa2af5c99319fa7e0465591dd209487cc5300292d117ec2f2bcfcd04eef048b9" Nov 24 13:58:21 crc kubenswrapper[4790]: I1124 13:58:21.174656 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa2af5c99319fa7e0465591dd209487cc5300292d117ec2f2bcfcd04eef048b9"} err="failed to get container status \"aa2af5c99319fa7e0465591dd209487cc5300292d117ec2f2bcfcd04eef048b9\": rpc error: code = NotFound desc = could not find container \"aa2af5c99319fa7e0465591dd209487cc5300292d117ec2f2bcfcd04eef048b9\": container with ID starting with aa2af5c99319fa7e0465591dd209487cc5300292d117ec2f2bcfcd04eef048b9 not found: ID does not exist" Nov 24 13:58:22 crc kubenswrapper[4790]: I1124 13:58:22.328042 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0980c6b4-8327-4046-9f89-ceb79b90786d" path="/var/lib/kubelet/pods/0980c6b4-8327-4046-9f89-ceb79b90786d/volumes" Nov 24 13:58:26 crc kubenswrapper[4790]: I1124 13:58:26.036722 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-96c9c" Nov 24 13:58:26 crc kubenswrapper[4790]: I1124 13:58:26.093273 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-96c9c" Nov 24 13:58:26 crc kubenswrapper[4790]: I1124 13:58:26.261791 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-96c9c"] Nov 24 13:58:26 crc kubenswrapper[4790]: I1124 13:58:26.440940 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j69qv"] Nov 24 13:58:26 crc kubenswrapper[4790]: I1124 13:58:26.441206 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-j69qv" podUID="324e1511-0e35-441f-a819-6faf957fe189" containerName="registry-server" containerID="cri-o://e208ec35513fd0ab410faec2c705e25f136f5d5626eddaf8636794dd56ddef2b" gracePeriod=2 Nov 24 13:58:26 crc kubenswrapper[4790]: I1124 13:58:26.824685 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j69qv" Nov 24 13:58:26 crc kubenswrapper[4790]: I1124 13:58:26.990598 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/324e1511-0e35-441f-a819-6faf957fe189-catalog-content\") pod \"324e1511-0e35-441f-a819-6faf957fe189\" (UID: \"324e1511-0e35-441f-a819-6faf957fe189\") " Nov 24 13:58:26 crc kubenswrapper[4790]: I1124 13:58:26.990671 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crb5n\" (UniqueName: \"kubernetes.io/projected/324e1511-0e35-441f-a819-6faf957fe189-kube-api-access-crb5n\") pod \"324e1511-0e35-441f-a819-6faf957fe189\" (UID: \"324e1511-0e35-441f-a819-6faf957fe189\") " Nov 24 13:58:26 crc kubenswrapper[4790]: I1124 13:58:26.990771 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/324e1511-0e35-441f-a819-6faf957fe189-utilities\") pod \"324e1511-0e35-441f-a819-6faf957fe189\" (UID: \"324e1511-0e35-441f-a819-6faf957fe189\") " Nov 24 13:58:26 crc kubenswrapper[4790]: I1124 13:58:26.991450 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/324e1511-0e35-441f-a819-6faf957fe189-utilities" (OuterVolumeSpecName: "utilities") pod "324e1511-0e35-441f-a819-6faf957fe189" (UID: "324e1511-0e35-441f-a819-6faf957fe189"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:58:26 crc kubenswrapper[4790]: I1124 13:58:26.997334 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/324e1511-0e35-441f-a819-6faf957fe189-kube-api-access-crb5n" (OuterVolumeSpecName: "kube-api-access-crb5n") pod "324e1511-0e35-441f-a819-6faf957fe189" (UID: "324e1511-0e35-441f-a819-6faf957fe189"). InnerVolumeSpecName "kube-api-access-crb5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:58:27 crc kubenswrapper[4790]: I1124 13:58:27.066746 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/324e1511-0e35-441f-a819-6faf957fe189-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "324e1511-0e35-441f-a819-6faf957fe189" (UID: "324e1511-0e35-441f-a819-6faf957fe189"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:58:27 crc kubenswrapper[4790]: I1124 13:58:27.092166 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/324e1511-0e35-441f-a819-6faf957fe189-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:58:27 crc kubenswrapper[4790]: I1124 13:58:27.092515 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crb5n\" (UniqueName: \"kubernetes.io/projected/324e1511-0e35-441f-a819-6faf957fe189-kube-api-access-crb5n\") on node \"crc\" DevicePath \"\"" Nov 24 13:58:27 crc kubenswrapper[4790]: I1124 13:58:27.092533 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/324e1511-0e35-441f-a819-6faf957fe189-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:58:27 crc kubenswrapper[4790]: I1124 13:58:27.135918 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j69qv" event={"ID":"324e1511-0e35-441f-a819-6faf957fe189","Type":"ContainerDied","Data":"e208ec35513fd0ab410faec2c705e25f136f5d5626eddaf8636794dd56ddef2b"} Nov 24 13:58:27 crc kubenswrapper[4790]: I1124 13:58:27.135973 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j69qv" Nov 24 13:58:27 crc kubenswrapper[4790]: I1124 13:58:27.135922 4790 generic.go:334] "Generic (PLEG): container finished" podID="324e1511-0e35-441f-a819-6faf957fe189" containerID="e208ec35513fd0ab410faec2c705e25f136f5d5626eddaf8636794dd56ddef2b" exitCode=0 Nov 24 13:58:27 crc kubenswrapper[4790]: I1124 13:58:27.135982 4790 scope.go:117] "RemoveContainer" containerID="e208ec35513fd0ab410faec2c705e25f136f5d5626eddaf8636794dd56ddef2b" Nov 24 13:58:27 crc kubenswrapper[4790]: I1124 13:58:27.136224 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j69qv" event={"ID":"324e1511-0e35-441f-a819-6faf957fe189","Type":"ContainerDied","Data":"44e31064f8c48b15d799d6a7f3f5b1d4e94f1702466c412bbcdb4e7de8676b54"} Nov 24 13:58:27 crc kubenswrapper[4790]: I1124 13:58:27.166177 4790 scope.go:117] "RemoveContainer" containerID="c5ed873b1002b1b75dadfe4f6037daa841206c87c98fb9df269ab7a2d7c4f465" Nov 24 13:58:27 crc kubenswrapper[4790]: I1124 13:58:27.181642 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j69qv"] Nov 24 13:58:27 crc kubenswrapper[4790]: I1124 13:58:27.189028 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-j69qv"] Nov 24 13:58:27 crc kubenswrapper[4790]: I1124 13:58:27.196092 4790 scope.go:117] "RemoveContainer" containerID="bccf83f5efa4bb596c5a1856f780440b6fa155b46a95ebaf00a5d582f20392e1" Nov 24 13:58:27 crc kubenswrapper[4790]: I1124 13:58:27.213038 4790 scope.go:117] "RemoveContainer" containerID="e208ec35513fd0ab410faec2c705e25f136f5d5626eddaf8636794dd56ddef2b" Nov 24 13:58:27 crc kubenswrapper[4790]: E1124 13:58:27.213489 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e208ec35513fd0ab410faec2c705e25f136f5d5626eddaf8636794dd56ddef2b\": container with ID starting with e208ec35513fd0ab410faec2c705e25f136f5d5626eddaf8636794dd56ddef2b not found: ID does not exist" containerID="e208ec35513fd0ab410faec2c705e25f136f5d5626eddaf8636794dd56ddef2b" Nov 24 13:58:27 crc kubenswrapper[4790]: I1124 13:58:27.213515 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e208ec35513fd0ab410faec2c705e25f136f5d5626eddaf8636794dd56ddef2b"} err="failed to get container status \"e208ec35513fd0ab410faec2c705e25f136f5d5626eddaf8636794dd56ddef2b\": rpc error: code = NotFound desc = could not find container \"e208ec35513fd0ab410faec2c705e25f136f5d5626eddaf8636794dd56ddef2b\": container with ID starting with e208ec35513fd0ab410faec2c705e25f136f5d5626eddaf8636794dd56ddef2b not found: ID does not exist" Nov 24 13:58:27 crc kubenswrapper[4790]: I1124 13:58:27.213534 4790 scope.go:117] "RemoveContainer" containerID="c5ed873b1002b1b75dadfe4f6037daa841206c87c98fb9df269ab7a2d7c4f465" Nov 24 13:58:27 crc kubenswrapper[4790]: E1124 13:58:27.213963 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5ed873b1002b1b75dadfe4f6037daa841206c87c98fb9df269ab7a2d7c4f465\": container with ID starting with c5ed873b1002b1b75dadfe4f6037daa841206c87c98fb9df269ab7a2d7c4f465 not found: ID does not exist" containerID="c5ed873b1002b1b75dadfe4f6037daa841206c87c98fb9df269ab7a2d7c4f465" Nov 24 13:58:27 crc kubenswrapper[4790]: I1124 13:58:27.213990 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5ed873b1002b1b75dadfe4f6037daa841206c87c98fb9df269ab7a2d7c4f465"} err="failed to get container status \"c5ed873b1002b1b75dadfe4f6037daa841206c87c98fb9df269ab7a2d7c4f465\": rpc error: code = NotFound desc = could not find container \"c5ed873b1002b1b75dadfe4f6037daa841206c87c98fb9df269ab7a2d7c4f465\": container with ID starting with c5ed873b1002b1b75dadfe4f6037daa841206c87c98fb9df269ab7a2d7c4f465 not found: ID does not exist" Nov 24 13:58:27 crc kubenswrapper[4790]: I1124 13:58:27.214022 4790 scope.go:117] "RemoveContainer" containerID="bccf83f5efa4bb596c5a1856f780440b6fa155b46a95ebaf00a5d582f20392e1" Nov 24 13:58:27 crc kubenswrapper[4790]: E1124 13:58:27.214376 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bccf83f5efa4bb596c5a1856f780440b6fa155b46a95ebaf00a5d582f20392e1\": container with ID starting with bccf83f5efa4bb596c5a1856f780440b6fa155b46a95ebaf00a5d582f20392e1 not found: ID does not exist" containerID="bccf83f5efa4bb596c5a1856f780440b6fa155b46a95ebaf00a5d582f20392e1" Nov 24 13:58:27 crc kubenswrapper[4790]: I1124 13:58:27.214396 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bccf83f5efa4bb596c5a1856f780440b6fa155b46a95ebaf00a5d582f20392e1"} err="failed to get container status \"bccf83f5efa4bb596c5a1856f780440b6fa155b46a95ebaf00a5d582f20392e1\": rpc error: code = NotFound desc = could not find container \"bccf83f5efa4bb596c5a1856f780440b6fa155b46a95ebaf00a5d582f20392e1\": container with ID starting with bccf83f5efa4bb596c5a1856f780440b6fa155b46a95ebaf00a5d582f20392e1 not found: ID does not exist" Nov 24 13:58:28 crc kubenswrapper[4790]: I1124 13:58:28.323029 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="324e1511-0e35-441f-a819-6faf957fe189" path="/var/lib/kubelet/pods/324e1511-0e35-441f-a819-6faf957fe189/volumes" Nov 24 14:00:00 crc kubenswrapper[4790]: I1124 14:00:00.143662 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399880-6cbwh"] Nov 24 14:00:00 crc kubenswrapper[4790]: E1124 14:00:00.144570 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7468685e-a517-4dcc-af7d-f58e54eda771" containerName="extract-content" Nov 24 14:00:00 crc kubenswrapper[4790]: I1124 14:00:00.144585 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="7468685e-a517-4dcc-af7d-f58e54eda771" containerName="extract-content" Nov 24 14:00:00 crc kubenswrapper[4790]: E1124 14:00:00.144604 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7468685e-a517-4dcc-af7d-f58e54eda771" containerName="extract-utilities" Nov 24 14:00:00 crc kubenswrapper[4790]: I1124 14:00:00.144612 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="7468685e-a517-4dcc-af7d-f58e54eda771" containerName="extract-utilities" Nov 24 14:00:00 crc kubenswrapper[4790]: E1124 14:00:00.144625 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="324e1511-0e35-441f-a819-6faf957fe189" containerName="extract-content" Nov 24 14:00:00 crc kubenswrapper[4790]: I1124 14:00:00.144632 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="324e1511-0e35-441f-a819-6faf957fe189" containerName="extract-content" Nov 24 14:00:00 crc kubenswrapper[4790]: E1124 14:00:00.144643 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="324e1511-0e35-441f-a819-6faf957fe189" containerName="extract-utilities" Nov 24 14:00:00 crc kubenswrapper[4790]: I1124 14:00:00.144648 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="324e1511-0e35-441f-a819-6faf957fe189" containerName="extract-utilities" Nov 24 14:00:00 crc kubenswrapper[4790]: E1124 14:00:00.144657 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7468685e-a517-4dcc-af7d-f58e54eda771" containerName="registry-server" Nov 24 14:00:00 crc kubenswrapper[4790]: I1124 14:00:00.144663 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="7468685e-a517-4dcc-af7d-f58e54eda771" containerName="registry-server" Nov 24 14:00:00 crc kubenswrapper[4790]: E1124 14:00:00.144674 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="324e1511-0e35-441f-a819-6faf957fe189" containerName="registry-server" Nov 24 14:00:00 crc kubenswrapper[4790]: I1124 14:00:00.144679 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="324e1511-0e35-441f-a819-6faf957fe189" containerName="registry-server" Nov 24 14:00:00 crc kubenswrapper[4790]: E1124 14:00:00.144694 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0980c6b4-8327-4046-9f89-ceb79b90786d" containerName="registry-server" Nov 24 14:00:00 crc kubenswrapper[4790]: I1124 14:00:00.144700 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="0980c6b4-8327-4046-9f89-ceb79b90786d" containerName="registry-server" Nov 24 14:00:00 crc kubenswrapper[4790]: E1124 14:00:00.144713 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0980c6b4-8327-4046-9f89-ceb79b90786d" containerName="extract-content" Nov 24 14:00:00 crc kubenswrapper[4790]: I1124 14:00:00.144719 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="0980c6b4-8327-4046-9f89-ceb79b90786d" containerName="extract-content" Nov 24 14:00:00 crc kubenswrapper[4790]: E1124 14:00:00.144734 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0980c6b4-8327-4046-9f89-ceb79b90786d" containerName="extract-utilities" Nov 24 14:00:00 crc kubenswrapper[4790]: I1124 14:00:00.144741 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="0980c6b4-8327-4046-9f89-ceb79b90786d" containerName="extract-utilities" Nov 24 14:00:00 crc kubenswrapper[4790]: I1124 14:00:00.144880 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="324e1511-0e35-441f-a819-6faf957fe189" containerName="registry-server" Nov 24 14:00:00 crc kubenswrapper[4790]: I1124 14:00:00.144912 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="0980c6b4-8327-4046-9f89-ceb79b90786d" containerName="registry-server" Nov 24 14:00:00 crc kubenswrapper[4790]: I1124 14:00:00.144918 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="7468685e-a517-4dcc-af7d-f58e54eda771" containerName="registry-server" Nov 24 14:00:00 crc kubenswrapper[4790]: I1124 14:00:00.145415 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-6cbwh" Nov 24 14:00:00 crc kubenswrapper[4790]: I1124 14:00:00.147647 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 14:00:00 crc kubenswrapper[4790]: I1124 14:00:00.154863 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399880-6cbwh"] Nov 24 14:00:00 crc kubenswrapper[4790]: I1124 14:00:00.155908 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 14:00:00 crc kubenswrapper[4790]: I1124 14:00:00.179911 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f84w8\" (UniqueName: \"kubernetes.io/projected/b64e3a97-00c5-4c24-9d42-faec03c7709a-kube-api-access-f84w8\") pod \"collect-profiles-29399880-6cbwh\" (UID: \"b64e3a97-00c5-4c24-9d42-faec03c7709a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-6cbwh" Nov 24 14:00:00 crc kubenswrapper[4790]: I1124 14:00:00.180657 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b64e3a97-00c5-4c24-9d42-faec03c7709a-config-volume\") pod \"collect-profiles-29399880-6cbwh\" (UID: \"b64e3a97-00c5-4c24-9d42-faec03c7709a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-6cbwh" Nov 24 14:00:00 crc kubenswrapper[4790]: I1124 14:00:00.180729 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b64e3a97-00c5-4c24-9d42-faec03c7709a-secret-volume\") pod \"collect-profiles-29399880-6cbwh\" (UID: \"b64e3a97-00c5-4c24-9d42-faec03c7709a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-6cbwh" Nov 24 14:00:00 crc kubenswrapper[4790]: I1124 14:00:00.281330 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b64e3a97-00c5-4c24-9d42-faec03c7709a-config-volume\") pod \"collect-profiles-29399880-6cbwh\" (UID: \"b64e3a97-00c5-4c24-9d42-faec03c7709a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-6cbwh" Nov 24 14:00:00 crc kubenswrapper[4790]: I1124 14:00:00.281378 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b64e3a97-00c5-4c24-9d42-faec03c7709a-secret-volume\") pod \"collect-profiles-29399880-6cbwh\" (UID: \"b64e3a97-00c5-4c24-9d42-faec03c7709a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-6cbwh" Nov 24 14:00:00 crc kubenswrapper[4790]: I1124 14:00:00.281469 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f84w8\" (UniqueName: \"kubernetes.io/projected/b64e3a97-00c5-4c24-9d42-faec03c7709a-kube-api-access-f84w8\") pod \"collect-profiles-29399880-6cbwh\" (UID: \"b64e3a97-00c5-4c24-9d42-faec03c7709a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-6cbwh" Nov 24 14:00:00 crc kubenswrapper[4790]: I1124 14:00:00.282509 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b64e3a97-00c5-4c24-9d42-faec03c7709a-config-volume\") pod \"collect-profiles-29399880-6cbwh\" (UID: \"b64e3a97-00c5-4c24-9d42-faec03c7709a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-6cbwh" Nov 24 14:00:00 crc kubenswrapper[4790]: I1124 14:00:00.288393 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b64e3a97-00c5-4c24-9d42-faec03c7709a-secret-volume\") pod \"collect-profiles-29399880-6cbwh\" (UID: \"b64e3a97-00c5-4c24-9d42-faec03c7709a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-6cbwh" Nov 24 14:00:00 crc kubenswrapper[4790]: I1124 14:00:00.301670 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f84w8\" (UniqueName: \"kubernetes.io/projected/b64e3a97-00c5-4c24-9d42-faec03c7709a-kube-api-access-f84w8\") pod \"collect-profiles-29399880-6cbwh\" (UID: \"b64e3a97-00c5-4c24-9d42-faec03c7709a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-6cbwh" Nov 24 14:00:00 crc kubenswrapper[4790]: I1124 14:00:00.465865 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-6cbwh" Nov 24 14:00:00 crc kubenswrapper[4790]: I1124 14:00:00.703990 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399880-6cbwh"] Nov 24 14:00:00 crc kubenswrapper[4790]: I1124 14:00:00.821803 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-6cbwh" event={"ID":"b64e3a97-00c5-4c24-9d42-faec03c7709a","Type":"ContainerStarted","Data":"c50fa80755c421a8ac52884da942e0c06d004302a20119c559f7afdaf90170ba"} Nov 24 14:00:01 crc kubenswrapper[4790]: I1124 14:00:01.831477 4790 generic.go:334] "Generic (PLEG): container finished" podID="b64e3a97-00c5-4c24-9d42-faec03c7709a" containerID="9ec50958969b0ff5c3ef1791470136baa532899ab450dc32dcd2c2677cf16c67" exitCode=0 Nov 24 14:00:01 crc kubenswrapper[4790]: I1124 14:00:01.831517 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-6cbwh" event={"ID":"b64e3a97-00c5-4c24-9d42-faec03c7709a","Type":"ContainerDied","Data":"9ec50958969b0ff5c3ef1791470136baa532899ab450dc32dcd2c2677cf16c67"} Nov 24 14:00:03 crc kubenswrapper[4790]: I1124 14:00:03.108514 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-6cbwh" Nov 24 14:00:03 crc kubenswrapper[4790]: I1124 14:00:03.223712 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b64e3a97-00c5-4c24-9d42-faec03c7709a-config-volume\") pod \"b64e3a97-00c5-4c24-9d42-faec03c7709a\" (UID: \"b64e3a97-00c5-4c24-9d42-faec03c7709a\") " Nov 24 14:00:03 crc kubenswrapper[4790]: I1124 14:00:03.224076 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b64e3a97-00c5-4c24-9d42-faec03c7709a-secret-volume\") pod \"b64e3a97-00c5-4c24-9d42-faec03c7709a\" (UID: \"b64e3a97-00c5-4c24-9d42-faec03c7709a\") " Nov 24 14:00:03 crc kubenswrapper[4790]: I1124 14:00:03.224105 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f84w8\" (UniqueName: \"kubernetes.io/projected/b64e3a97-00c5-4c24-9d42-faec03c7709a-kube-api-access-f84w8\") pod \"b64e3a97-00c5-4c24-9d42-faec03c7709a\" (UID: \"b64e3a97-00c5-4c24-9d42-faec03c7709a\") " Nov 24 14:00:03 crc kubenswrapper[4790]: I1124 14:00:03.224558 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b64e3a97-00c5-4c24-9d42-faec03c7709a-config-volume" (OuterVolumeSpecName: "config-volume") pod "b64e3a97-00c5-4c24-9d42-faec03c7709a" (UID: "b64e3a97-00c5-4c24-9d42-faec03c7709a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:00:03 crc kubenswrapper[4790]: I1124 14:00:03.229280 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b64e3a97-00c5-4c24-9d42-faec03c7709a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b64e3a97-00c5-4c24-9d42-faec03c7709a" (UID: "b64e3a97-00c5-4c24-9d42-faec03c7709a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:00:03 crc kubenswrapper[4790]: I1124 14:00:03.229644 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b64e3a97-00c5-4c24-9d42-faec03c7709a-kube-api-access-f84w8" (OuterVolumeSpecName: "kube-api-access-f84w8") pod "b64e3a97-00c5-4c24-9d42-faec03c7709a" (UID: "b64e3a97-00c5-4c24-9d42-faec03c7709a"). InnerVolumeSpecName "kube-api-access-f84w8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:00:03 crc kubenswrapper[4790]: I1124 14:00:03.325919 4790 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b64e3a97-00c5-4c24-9d42-faec03c7709a-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 14:00:03 crc kubenswrapper[4790]: I1124 14:00:03.325951 4790 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b64e3a97-00c5-4c24-9d42-faec03c7709a-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 14:00:03 crc kubenswrapper[4790]: I1124 14:00:03.325961 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f84w8\" (UniqueName: \"kubernetes.io/projected/b64e3a97-00c5-4c24-9d42-faec03c7709a-kube-api-access-f84w8\") on node \"crc\" DevicePath \"\"" Nov 24 14:00:03 crc kubenswrapper[4790]: I1124 14:00:03.846236 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-6cbwh" event={"ID":"b64e3a97-00c5-4c24-9d42-faec03c7709a","Type":"ContainerDied","Data":"c50fa80755c421a8ac52884da942e0c06d004302a20119c559f7afdaf90170ba"} Nov 24 14:00:03 crc kubenswrapper[4790]: I1124 14:00:03.846531 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c50fa80755c421a8ac52884da942e0c06d004302a20119c559f7afdaf90170ba" Nov 24 14:00:03 crc kubenswrapper[4790]: I1124 14:00:03.846297 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-6cbwh" Nov 24 14:00:04 crc kubenswrapper[4790]: I1124 14:00:04.180835 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399835-8lrlt"] Nov 24 14:00:04 crc kubenswrapper[4790]: I1124 14:00:04.185863 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399835-8lrlt"] Nov 24 14:00:04 crc kubenswrapper[4790]: I1124 14:00:04.330314 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="620bb83f-d131-4cce-b059-63ac16d078db" path="/var/lib/kubelet/pods/620bb83f-d131-4cce-b059-63ac16d078db/volumes" Nov 24 14:00:13 crc kubenswrapper[4790]: I1124 14:00:13.939282 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:00:13 crc kubenswrapper[4790]: I1124 14:00:13.940023 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:00:43 crc kubenswrapper[4790]: I1124 14:00:43.938994 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:00:43 crc kubenswrapper[4790]: I1124 14:00:43.939449 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:00:45 crc kubenswrapper[4790]: I1124 14:00:45.859599 4790 scope.go:117] "RemoveContainer" containerID="35193b2e078bbc0405057f2ab163fa82c6b85b36d9d0b00ae061fc71843346c1" Nov 24 14:01:13 crc kubenswrapper[4790]: I1124 14:01:13.938979 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:01:13 crc kubenswrapper[4790]: I1124 14:01:13.939675 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:01:13 crc kubenswrapper[4790]: I1124 14:01:13.939738 4790 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 14:01:13 crc kubenswrapper[4790]: I1124 14:01:13.940570 4790 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d5298328afc25b5319c1cc5ddb054410b112d29084866f4e201ce5c5fa1d883c"} pod="openshift-machine-config-operator/machine-config-daemon-xz49t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:01:13 crc kubenswrapper[4790]: I1124 14:01:13.940942 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" containerID="cri-o://d5298328afc25b5319c1cc5ddb054410b112d29084866f4e201ce5c5fa1d883c" gracePeriod=600 Nov 24 14:01:14 crc kubenswrapper[4790]: I1124 14:01:14.345401 4790 generic.go:334] "Generic (PLEG): container finished" podID="0d73b133-48f0-455f-8f6a-742e633f631a" containerID="d5298328afc25b5319c1cc5ddb054410b112d29084866f4e201ce5c5fa1d883c" exitCode=0 Nov 24 14:01:14 crc kubenswrapper[4790]: I1124 14:01:14.345484 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerDied","Data":"d5298328afc25b5319c1cc5ddb054410b112d29084866f4e201ce5c5fa1d883c"} Nov 24 14:01:14 crc kubenswrapper[4790]: I1124 14:01:14.345769 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"451fe5d9ca399689fc2ebac539dfde44079a1a33f2d2b78ba4e87e453a78b4b6"} Nov 24 14:01:14 crc kubenswrapper[4790]: I1124 14:01:14.345792 4790 scope.go:117] "RemoveContainer" containerID="8310dee54c9b2a0ea195ce3e80b6281248186114131fdd393fbb8cb25d04910a" Nov 24 14:03:43 crc kubenswrapper[4790]: I1124 14:03:43.938921 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:03:43 crc kubenswrapper[4790]: I1124 14:03:43.939994 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:03:58 crc kubenswrapper[4790]: I1124 14:03:58.676714 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-k6h9s"] Nov 24 14:03:58 crc kubenswrapper[4790]: E1124 14:03:58.677682 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b64e3a97-00c5-4c24-9d42-faec03c7709a" containerName="collect-profiles" Nov 24 14:03:58 crc kubenswrapper[4790]: I1124 14:03:58.677700 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="b64e3a97-00c5-4c24-9d42-faec03c7709a" containerName="collect-profiles" Nov 24 14:03:58 crc kubenswrapper[4790]: I1124 14:03:58.677907 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="b64e3a97-00c5-4c24-9d42-faec03c7709a" containerName="collect-profiles" Nov 24 14:03:58 crc kubenswrapper[4790]: I1124 14:03:58.679035 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k6h9s" Nov 24 14:03:58 crc kubenswrapper[4790]: I1124 14:03:58.690100 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k6h9s"] Nov 24 14:03:58 crc kubenswrapper[4790]: I1124 14:03:58.772611 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/790a857e-5d11-4a30-b9d2-78ecfa4ea8cc-catalog-content\") pod \"community-operators-k6h9s\" (UID: \"790a857e-5d11-4a30-b9d2-78ecfa4ea8cc\") " pod="openshift-marketplace/community-operators-k6h9s" Nov 24 14:03:58 crc kubenswrapper[4790]: I1124 14:03:58.772706 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/790a857e-5d11-4a30-b9d2-78ecfa4ea8cc-utilities\") pod \"community-operators-k6h9s\" (UID: \"790a857e-5d11-4a30-b9d2-78ecfa4ea8cc\") " pod="openshift-marketplace/community-operators-k6h9s" Nov 24 14:03:58 crc kubenswrapper[4790]: I1124 14:03:58.772757 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7h48\" (UniqueName: \"kubernetes.io/projected/790a857e-5d11-4a30-b9d2-78ecfa4ea8cc-kube-api-access-b7h48\") pod \"community-operators-k6h9s\" (UID: \"790a857e-5d11-4a30-b9d2-78ecfa4ea8cc\") " pod="openshift-marketplace/community-operators-k6h9s" Nov 24 14:03:58 crc kubenswrapper[4790]: I1124 14:03:58.874132 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/790a857e-5d11-4a30-b9d2-78ecfa4ea8cc-utilities\") pod \"community-operators-k6h9s\" (UID: \"790a857e-5d11-4a30-b9d2-78ecfa4ea8cc\") " pod="openshift-marketplace/community-operators-k6h9s" Nov 24 14:03:58 crc kubenswrapper[4790]: I1124 14:03:58.874405 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7h48\" (UniqueName: \"kubernetes.io/projected/790a857e-5d11-4a30-b9d2-78ecfa4ea8cc-kube-api-access-b7h48\") pod \"community-operators-k6h9s\" (UID: \"790a857e-5d11-4a30-b9d2-78ecfa4ea8cc\") " pod="openshift-marketplace/community-operators-k6h9s" Nov 24 14:03:58 crc kubenswrapper[4790]: I1124 14:03:58.874472 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/790a857e-5d11-4a30-b9d2-78ecfa4ea8cc-catalog-content\") pod \"community-operators-k6h9s\" (UID: \"790a857e-5d11-4a30-b9d2-78ecfa4ea8cc\") " pod="openshift-marketplace/community-operators-k6h9s" Nov 24 14:03:58 crc kubenswrapper[4790]: I1124 14:03:58.874577 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/790a857e-5d11-4a30-b9d2-78ecfa4ea8cc-utilities\") pod \"community-operators-k6h9s\" (UID: \"790a857e-5d11-4a30-b9d2-78ecfa4ea8cc\") " pod="openshift-marketplace/community-operators-k6h9s" Nov 24 14:03:58 crc kubenswrapper[4790]: I1124 14:03:58.874735 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/790a857e-5d11-4a30-b9d2-78ecfa4ea8cc-catalog-content\") pod \"community-operators-k6h9s\" (UID: \"790a857e-5d11-4a30-b9d2-78ecfa4ea8cc\") " pod="openshift-marketplace/community-operators-k6h9s" Nov 24 14:03:58 crc kubenswrapper[4790]: I1124 14:03:58.893436 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7h48\" (UniqueName: \"kubernetes.io/projected/790a857e-5d11-4a30-b9d2-78ecfa4ea8cc-kube-api-access-b7h48\") pod \"community-operators-k6h9s\" (UID: \"790a857e-5d11-4a30-b9d2-78ecfa4ea8cc\") " pod="openshift-marketplace/community-operators-k6h9s" Nov 24 14:03:58 crc kubenswrapper[4790]: I1124 14:03:58.998784 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k6h9s" Nov 24 14:03:59 crc kubenswrapper[4790]: I1124 14:03:59.515074 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k6h9s"] Nov 24 14:03:59 crc kubenswrapper[4790]: W1124 14:03:59.524205 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod790a857e_5d11_4a30_b9d2_78ecfa4ea8cc.slice/crio-041e094114ecdfdba82c462081eb42ecc0908a04f7a81098e330967d100f5cbd WatchSource:0}: Error finding container 041e094114ecdfdba82c462081eb42ecc0908a04f7a81098e330967d100f5cbd: Status 404 returned error can't find the container with id 041e094114ecdfdba82c462081eb42ecc0908a04f7a81098e330967d100f5cbd Nov 24 14:03:59 crc kubenswrapper[4790]: I1124 14:03:59.691193 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k6h9s" event={"ID":"790a857e-5d11-4a30-b9d2-78ecfa4ea8cc","Type":"ContainerStarted","Data":"041e094114ecdfdba82c462081eb42ecc0908a04f7a81098e330967d100f5cbd"} Nov 24 14:04:00 crc kubenswrapper[4790]: I1124 14:04:00.700835 4790 generic.go:334] "Generic (PLEG): container finished" podID="790a857e-5d11-4a30-b9d2-78ecfa4ea8cc" containerID="d65a9612f2fb735dd8872148f1ee6ccf681bd62d9d07dc83e5912b981ba26c05" exitCode=0 Nov 24 14:04:00 crc kubenswrapper[4790]: I1124 14:04:00.700921 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k6h9s" event={"ID":"790a857e-5d11-4a30-b9d2-78ecfa4ea8cc","Type":"ContainerDied","Data":"d65a9612f2fb735dd8872148f1ee6ccf681bd62d9d07dc83e5912b981ba26c05"} Nov 24 14:04:00 crc kubenswrapper[4790]: I1124 14:04:00.704594 4790 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 14:04:01 crc kubenswrapper[4790]: I1124 14:04:01.709813 4790 generic.go:334] "Generic (PLEG): container finished" podID="790a857e-5d11-4a30-b9d2-78ecfa4ea8cc" containerID="35b1e6f6e44b1eb0b44e4501f03d0c5008f7581efd3f3dbbbdac677c45cc1a9e" exitCode=0 Nov 24 14:04:01 crc kubenswrapper[4790]: I1124 14:04:01.709859 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k6h9s" event={"ID":"790a857e-5d11-4a30-b9d2-78ecfa4ea8cc","Type":"ContainerDied","Data":"35b1e6f6e44b1eb0b44e4501f03d0c5008f7581efd3f3dbbbdac677c45cc1a9e"} Nov 24 14:04:02 crc kubenswrapper[4790]: I1124 14:04:02.718871 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k6h9s" event={"ID":"790a857e-5d11-4a30-b9d2-78ecfa4ea8cc","Type":"ContainerStarted","Data":"c4708a73916614f15064a10e072ee605645d98da50dd3fee4403a500e01a46b6"} Nov 24 14:04:02 crc kubenswrapper[4790]: I1124 14:04:02.741095 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-k6h9s" podStartSLOduration=3.335250739 podStartE2EDuration="4.741070578s" podCreationTimestamp="2025-11-24 14:03:58 +0000 UTC" firstStartedPulling="2025-11-24 14:04:00.704275597 +0000 UTC m=+3089.084169259" lastFinishedPulling="2025-11-24 14:04:02.110095426 +0000 UTC m=+3090.489989098" observedRunningTime="2025-11-24 14:04:02.734692826 +0000 UTC m=+3091.114586498" watchObservedRunningTime="2025-11-24 14:04:02.741070578 +0000 UTC m=+3091.120964250" Nov 24 14:04:08 crc kubenswrapper[4790]: I1124 14:04:08.999622 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-k6h9s" Nov 24 14:04:08 crc kubenswrapper[4790]: I1124 14:04:08.999949 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-k6h9s" Nov 24 14:04:09 crc kubenswrapper[4790]: I1124 14:04:09.044944 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-k6h9s" Nov 24 14:04:09 crc kubenswrapper[4790]: I1124 14:04:09.849955 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-k6h9s" Nov 24 14:04:09 crc kubenswrapper[4790]: I1124 14:04:09.896646 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k6h9s"] Nov 24 14:04:11 crc kubenswrapper[4790]: I1124 14:04:11.787835 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-k6h9s" podUID="790a857e-5d11-4a30-b9d2-78ecfa4ea8cc" containerName="registry-server" containerID="cri-o://c4708a73916614f15064a10e072ee605645d98da50dd3fee4403a500e01a46b6" gracePeriod=2 Nov 24 14:04:12 crc kubenswrapper[4790]: I1124 14:04:12.190605 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k6h9s" Nov 24 14:04:12 crc kubenswrapper[4790]: I1124 14:04:12.264360 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7h48\" (UniqueName: \"kubernetes.io/projected/790a857e-5d11-4a30-b9d2-78ecfa4ea8cc-kube-api-access-b7h48\") pod \"790a857e-5d11-4a30-b9d2-78ecfa4ea8cc\" (UID: \"790a857e-5d11-4a30-b9d2-78ecfa4ea8cc\") " Nov 24 14:04:12 crc kubenswrapper[4790]: I1124 14:04:12.264526 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/790a857e-5d11-4a30-b9d2-78ecfa4ea8cc-catalog-content\") pod \"790a857e-5d11-4a30-b9d2-78ecfa4ea8cc\" (UID: \"790a857e-5d11-4a30-b9d2-78ecfa4ea8cc\") " Nov 24 14:04:12 crc kubenswrapper[4790]: I1124 14:04:12.264966 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/790a857e-5d11-4a30-b9d2-78ecfa4ea8cc-utilities\") pod \"790a857e-5d11-4a30-b9d2-78ecfa4ea8cc\" (UID: \"790a857e-5d11-4a30-b9d2-78ecfa4ea8cc\") " Nov 24 14:04:12 crc kubenswrapper[4790]: I1124 14:04:12.266636 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/790a857e-5d11-4a30-b9d2-78ecfa4ea8cc-utilities" (OuterVolumeSpecName: "utilities") pod "790a857e-5d11-4a30-b9d2-78ecfa4ea8cc" (UID: "790a857e-5d11-4a30-b9d2-78ecfa4ea8cc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:04:12 crc kubenswrapper[4790]: I1124 14:04:12.275522 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/790a857e-5d11-4a30-b9d2-78ecfa4ea8cc-kube-api-access-b7h48" (OuterVolumeSpecName: "kube-api-access-b7h48") pod "790a857e-5d11-4a30-b9d2-78ecfa4ea8cc" (UID: "790a857e-5d11-4a30-b9d2-78ecfa4ea8cc"). InnerVolumeSpecName "kube-api-access-b7h48". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:04:12 crc kubenswrapper[4790]: I1124 14:04:12.326041 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/790a857e-5d11-4a30-b9d2-78ecfa4ea8cc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "790a857e-5d11-4a30-b9d2-78ecfa4ea8cc" (UID: "790a857e-5d11-4a30-b9d2-78ecfa4ea8cc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:04:12 crc kubenswrapper[4790]: I1124 14:04:12.367821 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7h48\" (UniqueName: \"kubernetes.io/projected/790a857e-5d11-4a30-b9d2-78ecfa4ea8cc-kube-api-access-b7h48\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:12 crc kubenswrapper[4790]: I1124 14:04:12.367866 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/790a857e-5d11-4a30-b9d2-78ecfa4ea8cc-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:12 crc kubenswrapper[4790]: I1124 14:04:12.367882 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/790a857e-5d11-4a30-b9d2-78ecfa4ea8cc-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:12 crc kubenswrapper[4790]: I1124 14:04:12.800684 4790 generic.go:334] "Generic (PLEG): container finished" podID="790a857e-5d11-4a30-b9d2-78ecfa4ea8cc" containerID="c4708a73916614f15064a10e072ee605645d98da50dd3fee4403a500e01a46b6" exitCode=0 Nov 24 14:04:12 crc kubenswrapper[4790]: I1124 14:04:12.800777 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k6h9s" event={"ID":"790a857e-5d11-4a30-b9d2-78ecfa4ea8cc","Type":"ContainerDied","Data":"c4708a73916614f15064a10e072ee605645d98da50dd3fee4403a500e01a46b6"} Nov 24 14:04:12 crc kubenswrapper[4790]: I1124 14:04:12.800827 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k6h9s" event={"ID":"790a857e-5d11-4a30-b9d2-78ecfa4ea8cc","Type":"ContainerDied","Data":"041e094114ecdfdba82c462081eb42ecc0908a04f7a81098e330967d100f5cbd"} Nov 24 14:04:12 crc kubenswrapper[4790]: I1124 14:04:12.800868 4790 scope.go:117] "RemoveContainer" containerID="c4708a73916614f15064a10e072ee605645d98da50dd3fee4403a500e01a46b6" Nov 24 14:04:12 crc kubenswrapper[4790]: I1124 14:04:12.801112 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k6h9s" Nov 24 14:04:12 crc kubenswrapper[4790]: I1124 14:04:12.827408 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k6h9s"] Nov 24 14:04:12 crc kubenswrapper[4790]: I1124 14:04:12.833519 4790 scope.go:117] "RemoveContainer" containerID="35b1e6f6e44b1eb0b44e4501f03d0c5008f7581efd3f3dbbbdac677c45cc1a9e" Nov 24 14:04:12 crc kubenswrapper[4790]: I1124 14:04:12.838777 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-k6h9s"] Nov 24 14:04:12 crc kubenswrapper[4790]: I1124 14:04:12.857369 4790 scope.go:117] "RemoveContainer" containerID="d65a9612f2fb735dd8872148f1ee6ccf681bd62d9d07dc83e5912b981ba26c05" Nov 24 14:04:12 crc kubenswrapper[4790]: I1124 14:04:12.888385 4790 scope.go:117] "RemoveContainer" containerID="c4708a73916614f15064a10e072ee605645d98da50dd3fee4403a500e01a46b6" Nov 24 14:04:12 crc kubenswrapper[4790]: E1124 14:04:12.888964 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4708a73916614f15064a10e072ee605645d98da50dd3fee4403a500e01a46b6\": container with ID starting with c4708a73916614f15064a10e072ee605645d98da50dd3fee4403a500e01a46b6 not found: ID does not exist" containerID="c4708a73916614f15064a10e072ee605645d98da50dd3fee4403a500e01a46b6" Nov 24 14:04:12 crc kubenswrapper[4790]: I1124 14:04:12.889052 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4708a73916614f15064a10e072ee605645d98da50dd3fee4403a500e01a46b6"} err="failed to get container status \"c4708a73916614f15064a10e072ee605645d98da50dd3fee4403a500e01a46b6\": rpc error: code = NotFound desc = could not find container \"c4708a73916614f15064a10e072ee605645d98da50dd3fee4403a500e01a46b6\": container with ID starting with c4708a73916614f15064a10e072ee605645d98da50dd3fee4403a500e01a46b6 not found: ID does not exist" Nov 24 14:04:12 crc kubenswrapper[4790]: I1124 14:04:12.889128 4790 scope.go:117] "RemoveContainer" containerID="35b1e6f6e44b1eb0b44e4501f03d0c5008f7581efd3f3dbbbdac677c45cc1a9e" Nov 24 14:04:12 crc kubenswrapper[4790]: E1124 14:04:12.889562 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35b1e6f6e44b1eb0b44e4501f03d0c5008f7581efd3f3dbbbdac677c45cc1a9e\": container with ID starting with 35b1e6f6e44b1eb0b44e4501f03d0c5008f7581efd3f3dbbbdac677c45cc1a9e not found: ID does not exist" containerID="35b1e6f6e44b1eb0b44e4501f03d0c5008f7581efd3f3dbbbdac677c45cc1a9e" Nov 24 14:04:12 crc kubenswrapper[4790]: I1124 14:04:12.889605 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35b1e6f6e44b1eb0b44e4501f03d0c5008f7581efd3f3dbbbdac677c45cc1a9e"} err="failed to get container status \"35b1e6f6e44b1eb0b44e4501f03d0c5008f7581efd3f3dbbbdac677c45cc1a9e\": rpc error: code = NotFound desc = could not find container \"35b1e6f6e44b1eb0b44e4501f03d0c5008f7581efd3f3dbbbdac677c45cc1a9e\": container with ID starting with 35b1e6f6e44b1eb0b44e4501f03d0c5008f7581efd3f3dbbbdac677c45cc1a9e not found: ID does not exist" Nov 24 14:04:12 crc kubenswrapper[4790]: I1124 14:04:12.889632 4790 scope.go:117] "RemoveContainer" containerID="d65a9612f2fb735dd8872148f1ee6ccf681bd62d9d07dc83e5912b981ba26c05" Nov 24 14:04:12 crc kubenswrapper[4790]: E1124 14:04:12.890108 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d65a9612f2fb735dd8872148f1ee6ccf681bd62d9d07dc83e5912b981ba26c05\": container with ID starting with d65a9612f2fb735dd8872148f1ee6ccf681bd62d9d07dc83e5912b981ba26c05 not found: ID does not exist" containerID="d65a9612f2fb735dd8872148f1ee6ccf681bd62d9d07dc83e5912b981ba26c05" Nov 24 14:04:12 crc kubenswrapper[4790]: I1124 14:04:12.890197 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d65a9612f2fb735dd8872148f1ee6ccf681bd62d9d07dc83e5912b981ba26c05"} err="failed to get container status \"d65a9612f2fb735dd8872148f1ee6ccf681bd62d9d07dc83e5912b981ba26c05\": rpc error: code = NotFound desc = could not find container \"d65a9612f2fb735dd8872148f1ee6ccf681bd62d9d07dc83e5912b981ba26c05\": container with ID starting with d65a9612f2fb735dd8872148f1ee6ccf681bd62d9d07dc83e5912b981ba26c05 not found: ID does not exist" Nov 24 14:04:13 crc kubenswrapper[4790]: I1124 14:04:13.938501 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:04:13 crc kubenswrapper[4790]: I1124 14:04:13.939160 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:04:14 crc kubenswrapper[4790]: I1124 14:04:14.323511 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="790a857e-5d11-4a30-b9d2-78ecfa4ea8cc" path="/var/lib/kubelet/pods/790a857e-5d11-4a30-b9d2-78ecfa4ea8cc/volumes" Nov 24 14:04:43 crc kubenswrapper[4790]: I1124 14:04:43.938649 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:04:43 crc kubenswrapper[4790]: I1124 14:04:43.939223 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:04:43 crc kubenswrapper[4790]: I1124 14:04:43.939274 4790 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 14:04:43 crc kubenswrapper[4790]: I1124 14:04:43.940074 4790 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"451fe5d9ca399689fc2ebac539dfde44079a1a33f2d2b78ba4e87e453a78b4b6"} pod="openshift-machine-config-operator/machine-config-daemon-xz49t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:04:43 crc kubenswrapper[4790]: I1124 14:04:43.940139 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" containerID="cri-o://451fe5d9ca399689fc2ebac539dfde44079a1a33f2d2b78ba4e87e453a78b4b6" gracePeriod=600 Nov 24 14:04:44 crc kubenswrapper[4790]: E1124 14:04:44.065635 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:04:45 crc kubenswrapper[4790]: I1124 14:04:45.047132 4790 generic.go:334] "Generic (PLEG): container finished" podID="0d73b133-48f0-455f-8f6a-742e633f631a" containerID="451fe5d9ca399689fc2ebac539dfde44079a1a33f2d2b78ba4e87e453a78b4b6" exitCode=0 Nov 24 14:04:45 crc kubenswrapper[4790]: I1124 14:04:45.047219 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerDied","Data":"451fe5d9ca399689fc2ebac539dfde44079a1a33f2d2b78ba4e87e453a78b4b6"} Nov 24 14:04:45 crc kubenswrapper[4790]: I1124 14:04:45.047285 4790 scope.go:117] "RemoveContainer" containerID="d5298328afc25b5319c1cc5ddb054410b112d29084866f4e201ce5c5fa1d883c" Nov 24 14:04:45 crc kubenswrapper[4790]: I1124 14:04:45.049583 4790 scope.go:117] "RemoveContainer" containerID="451fe5d9ca399689fc2ebac539dfde44079a1a33f2d2b78ba4e87e453a78b4b6" Nov 24 14:04:45 crc kubenswrapper[4790]: E1124 14:04:45.050452 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:04:59 crc kubenswrapper[4790]: I1124 14:04:59.314403 4790 scope.go:117] "RemoveContainer" containerID="451fe5d9ca399689fc2ebac539dfde44079a1a33f2d2b78ba4e87e453a78b4b6" Nov 24 14:04:59 crc kubenswrapper[4790]: E1124 14:04:59.315100 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:05:14 crc kubenswrapper[4790]: I1124 14:05:14.314581 4790 scope.go:117] "RemoveContainer" containerID="451fe5d9ca399689fc2ebac539dfde44079a1a33f2d2b78ba4e87e453a78b4b6" Nov 24 14:05:14 crc kubenswrapper[4790]: E1124 14:05:14.315423 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:05:27 crc kubenswrapper[4790]: I1124 14:05:27.314224 4790 scope.go:117] "RemoveContainer" containerID="451fe5d9ca399689fc2ebac539dfde44079a1a33f2d2b78ba4e87e453a78b4b6" Nov 24 14:05:27 crc kubenswrapper[4790]: E1124 14:05:27.315021 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:05:42 crc kubenswrapper[4790]: I1124 14:05:42.322683 4790 scope.go:117] "RemoveContainer" containerID="451fe5d9ca399689fc2ebac539dfde44079a1a33f2d2b78ba4e87e453a78b4b6" Nov 24 14:05:42 crc kubenswrapper[4790]: E1124 14:05:42.323706 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:05:56 crc kubenswrapper[4790]: I1124 14:05:56.314630 4790 scope.go:117] "RemoveContainer" containerID="451fe5d9ca399689fc2ebac539dfde44079a1a33f2d2b78ba4e87e453a78b4b6" Nov 24 14:05:56 crc kubenswrapper[4790]: E1124 14:05:56.315417 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:06:11 crc kubenswrapper[4790]: I1124 14:06:11.315951 4790 scope.go:117] "RemoveContainer" containerID="451fe5d9ca399689fc2ebac539dfde44079a1a33f2d2b78ba4e87e453a78b4b6" Nov 24 14:06:11 crc kubenswrapper[4790]: E1124 14:06:11.316843 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:06:24 crc kubenswrapper[4790]: I1124 14:06:24.314979 4790 scope.go:117] "RemoveContainer" containerID="451fe5d9ca399689fc2ebac539dfde44079a1a33f2d2b78ba4e87e453a78b4b6" Nov 24 14:06:24 crc kubenswrapper[4790]: E1124 14:06:24.315919 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:06:36 crc kubenswrapper[4790]: I1124 14:06:36.315625 4790 scope.go:117] "RemoveContainer" containerID="451fe5d9ca399689fc2ebac539dfde44079a1a33f2d2b78ba4e87e453a78b4b6" Nov 24 14:06:36 crc kubenswrapper[4790]: E1124 14:06:36.317410 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:06:48 crc kubenswrapper[4790]: I1124 14:06:48.314182 4790 scope.go:117] "RemoveContainer" containerID="451fe5d9ca399689fc2ebac539dfde44079a1a33f2d2b78ba4e87e453a78b4b6" Nov 24 14:06:48 crc kubenswrapper[4790]: E1124 14:06:48.314989 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:06:59 crc kubenswrapper[4790]: I1124 14:06:59.315835 4790 scope.go:117] "RemoveContainer" containerID="451fe5d9ca399689fc2ebac539dfde44079a1a33f2d2b78ba4e87e453a78b4b6" Nov 24 14:06:59 crc kubenswrapper[4790]: E1124 14:06:59.316980 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:07:10 crc kubenswrapper[4790]: I1124 14:07:10.315066 4790 scope.go:117] "RemoveContainer" containerID="451fe5d9ca399689fc2ebac539dfde44079a1a33f2d2b78ba4e87e453a78b4b6" Nov 24 14:07:10 crc kubenswrapper[4790]: E1124 14:07:10.315777 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:07:25 crc kubenswrapper[4790]: I1124 14:07:25.314459 4790 scope.go:117] "RemoveContainer" containerID="451fe5d9ca399689fc2ebac539dfde44079a1a33f2d2b78ba4e87e453a78b4b6" Nov 24 14:07:25 crc kubenswrapper[4790]: E1124 14:07:25.315088 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:07:39 crc kubenswrapper[4790]: I1124 14:07:39.314406 4790 scope.go:117] "RemoveContainer" containerID="451fe5d9ca399689fc2ebac539dfde44079a1a33f2d2b78ba4e87e453a78b4b6" Nov 24 14:07:39 crc kubenswrapper[4790]: E1124 14:07:39.315200 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:07:54 crc kubenswrapper[4790]: I1124 14:07:54.314862 4790 scope.go:117] "RemoveContainer" containerID="451fe5d9ca399689fc2ebac539dfde44079a1a33f2d2b78ba4e87e453a78b4b6" Nov 24 14:07:54 crc kubenswrapper[4790]: E1124 14:07:54.315670 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:08:09 crc kubenswrapper[4790]: I1124 14:08:09.314363 4790 scope.go:117] "RemoveContainer" containerID="451fe5d9ca399689fc2ebac539dfde44079a1a33f2d2b78ba4e87e453a78b4b6" Nov 24 14:08:09 crc kubenswrapper[4790]: E1124 14:08:09.315164 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:08:14 crc kubenswrapper[4790]: I1124 14:08:14.506844 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gnk76"] Nov 24 14:08:14 crc kubenswrapper[4790]: E1124 14:08:14.507913 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="790a857e-5d11-4a30-b9d2-78ecfa4ea8cc" containerName="extract-content" Nov 24 14:08:14 crc kubenswrapper[4790]: I1124 14:08:14.507938 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="790a857e-5d11-4a30-b9d2-78ecfa4ea8cc" containerName="extract-content" Nov 24 14:08:14 crc kubenswrapper[4790]: E1124 14:08:14.507969 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="790a857e-5d11-4a30-b9d2-78ecfa4ea8cc" containerName="registry-server" Nov 24 14:08:14 crc kubenswrapper[4790]: I1124 14:08:14.507980 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="790a857e-5d11-4a30-b9d2-78ecfa4ea8cc" containerName="registry-server" Nov 24 14:08:14 crc kubenswrapper[4790]: E1124 14:08:14.508020 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="790a857e-5d11-4a30-b9d2-78ecfa4ea8cc" containerName="extract-utilities" Nov 24 14:08:14 crc kubenswrapper[4790]: I1124 14:08:14.508032 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="790a857e-5d11-4a30-b9d2-78ecfa4ea8cc" containerName="extract-utilities" Nov 24 14:08:14 crc kubenswrapper[4790]: I1124 14:08:14.508286 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="790a857e-5d11-4a30-b9d2-78ecfa4ea8cc" containerName="registry-server" Nov 24 14:08:14 crc kubenswrapper[4790]: I1124 14:08:14.509936 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gnk76" Nov 24 14:08:14 crc kubenswrapper[4790]: I1124 14:08:14.519470 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gnk76"] Nov 24 14:08:14 crc kubenswrapper[4790]: I1124 14:08:14.628397 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbm55\" (UniqueName: \"kubernetes.io/projected/f280252b-a411-460b-9fe0-843261c02836-kube-api-access-rbm55\") pod \"certified-operators-gnk76\" (UID: \"f280252b-a411-460b-9fe0-843261c02836\") " pod="openshift-marketplace/certified-operators-gnk76" Nov 24 14:08:14 crc kubenswrapper[4790]: I1124 14:08:14.628467 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f280252b-a411-460b-9fe0-843261c02836-catalog-content\") pod \"certified-operators-gnk76\" (UID: \"f280252b-a411-460b-9fe0-843261c02836\") " pod="openshift-marketplace/certified-operators-gnk76" Nov 24 14:08:14 crc kubenswrapper[4790]: I1124 14:08:14.628486 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f280252b-a411-460b-9fe0-843261c02836-utilities\") pod \"certified-operators-gnk76\" (UID: \"f280252b-a411-460b-9fe0-843261c02836\") " pod="openshift-marketplace/certified-operators-gnk76" Nov 24 14:08:14 crc kubenswrapper[4790]: I1124 14:08:14.729434 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f280252b-a411-460b-9fe0-843261c02836-catalog-content\") pod \"certified-operators-gnk76\" (UID: \"f280252b-a411-460b-9fe0-843261c02836\") " pod="openshift-marketplace/certified-operators-gnk76" Nov 24 14:08:14 crc kubenswrapper[4790]: I1124 14:08:14.729489 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f280252b-a411-460b-9fe0-843261c02836-utilities\") pod \"certified-operators-gnk76\" (UID: \"f280252b-a411-460b-9fe0-843261c02836\") " pod="openshift-marketplace/certified-operators-gnk76" Nov 24 14:08:14 crc kubenswrapper[4790]: I1124 14:08:14.729578 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbm55\" (UniqueName: \"kubernetes.io/projected/f280252b-a411-460b-9fe0-843261c02836-kube-api-access-rbm55\") pod \"certified-operators-gnk76\" (UID: \"f280252b-a411-460b-9fe0-843261c02836\") " pod="openshift-marketplace/certified-operators-gnk76" Nov 24 14:08:14 crc kubenswrapper[4790]: I1124 14:08:14.730010 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f280252b-a411-460b-9fe0-843261c02836-catalog-content\") pod \"certified-operators-gnk76\" (UID: \"f280252b-a411-460b-9fe0-843261c02836\") " pod="openshift-marketplace/certified-operators-gnk76" Nov 24 14:08:14 crc kubenswrapper[4790]: I1124 14:08:14.730066 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f280252b-a411-460b-9fe0-843261c02836-utilities\") pod \"certified-operators-gnk76\" (UID: \"f280252b-a411-460b-9fe0-843261c02836\") " pod="openshift-marketplace/certified-operators-gnk76" Nov 24 14:08:14 crc kubenswrapper[4790]: I1124 14:08:14.749726 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbm55\" (UniqueName: \"kubernetes.io/projected/f280252b-a411-460b-9fe0-843261c02836-kube-api-access-rbm55\") pod \"certified-operators-gnk76\" (UID: \"f280252b-a411-460b-9fe0-843261c02836\") " pod="openshift-marketplace/certified-operators-gnk76" Nov 24 14:08:14 crc kubenswrapper[4790]: I1124 14:08:14.841212 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gnk76" Nov 24 14:08:15 crc kubenswrapper[4790]: I1124 14:08:15.335739 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gnk76"] Nov 24 14:08:16 crc kubenswrapper[4790]: I1124 14:08:16.162704 4790 generic.go:334] "Generic (PLEG): container finished" podID="f280252b-a411-460b-9fe0-843261c02836" containerID="82ac48f9b645a50d9059a1a74a91767ed99a57dd24b36e0c203e450e46463321" exitCode=0 Nov 24 14:08:16 crc kubenswrapper[4790]: I1124 14:08:16.162783 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gnk76" event={"ID":"f280252b-a411-460b-9fe0-843261c02836","Type":"ContainerDied","Data":"82ac48f9b645a50d9059a1a74a91767ed99a57dd24b36e0c203e450e46463321"} Nov 24 14:08:16 crc kubenswrapper[4790]: I1124 14:08:16.163066 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gnk76" event={"ID":"f280252b-a411-460b-9fe0-843261c02836","Type":"ContainerStarted","Data":"a29265ab5f23fd95d4182bad1e6eb1afcebae93e1cbadb8840258226233cdf06"} Nov 24 14:08:17 crc kubenswrapper[4790]: I1124 14:08:17.172029 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gnk76" event={"ID":"f280252b-a411-460b-9fe0-843261c02836","Type":"ContainerStarted","Data":"1637d635fb06fe031c5c99910c348c31bb9fb4132746c58f0023ec63e8b734f9"} Nov 24 14:08:18 crc kubenswrapper[4790]: I1124 14:08:18.185014 4790 generic.go:334] "Generic (PLEG): container finished" podID="f280252b-a411-460b-9fe0-843261c02836" containerID="1637d635fb06fe031c5c99910c348c31bb9fb4132746c58f0023ec63e8b734f9" exitCode=0 Nov 24 14:08:18 crc kubenswrapper[4790]: I1124 14:08:18.185117 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gnk76" event={"ID":"f280252b-a411-460b-9fe0-843261c02836","Type":"ContainerDied","Data":"1637d635fb06fe031c5c99910c348c31bb9fb4132746c58f0023ec63e8b734f9"} Nov 24 14:08:19 crc kubenswrapper[4790]: I1124 14:08:19.194336 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gnk76" event={"ID":"f280252b-a411-460b-9fe0-843261c02836","Type":"ContainerStarted","Data":"02644981edab7d70417fd5e4f986dbb820889a28c689ed676619698d3427e6a0"} Nov 24 14:08:19 crc kubenswrapper[4790]: I1124 14:08:19.209487 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gnk76" podStartSLOduration=2.750603385 podStartE2EDuration="5.209467033s" podCreationTimestamp="2025-11-24 14:08:14 +0000 UTC" firstStartedPulling="2025-11-24 14:08:16.165727868 +0000 UTC m=+3344.545621530" lastFinishedPulling="2025-11-24 14:08:18.624591516 +0000 UTC m=+3347.004485178" observedRunningTime="2025-11-24 14:08:19.209104503 +0000 UTC m=+3347.588998165" watchObservedRunningTime="2025-11-24 14:08:19.209467033 +0000 UTC m=+3347.589360695" Nov 24 14:08:21 crc kubenswrapper[4790]: I1124 14:08:21.314443 4790 scope.go:117] "RemoveContainer" containerID="451fe5d9ca399689fc2ebac539dfde44079a1a33f2d2b78ba4e87e453a78b4b6" Nov 24 14:08:21 crc kubenswrapper[4790]: E1124 14:08:21.315052 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:08:24 crc kubenswrapper[4790]: I1124 14:08:24.842615 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gnk76" Nov 24 14:08:24 crc kubenswrapper[4790]: I1124 14:08:24.843036 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gnk76" Nov 24 14:08:24 crc kubenswrapper[4790]: I1124 14:08:24.890228 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gnk76" Nov 24 14:08:25 crc kubenswrapper[4790]: I1124 14:08:25.280199 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gnk76" Nov 24 14:08:25 crc kubenswrapper[4790]: I1124 14:08:25.347771 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gnk76"] Nov 24 14:08:27 crc kubenswrapper[4790]: I1124 14:08:27.247568 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gnk76" podUID="f280252b-a411-460b-9fe0-843261c02836" containerName="registry-server" containerID="cri-o://02644981edab7d70417fd5e4f986dbb820889a28c689ed676619698d3427e6a0" gracePeriod=2 Nov 24 14:08:27 crc kubenswrapper[4790]: I1124 14:08:27.623036 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gnk76" Nov 24 14:08:27 crc kubenswrapper[4790]: I1124 14:08:27.709176 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbm55\" (UniqueName: \"kubernetes.io/projected/f280252b-a411-460b-9fe0-843261c02836-kube-api-access-rbm55\") pod \"f280252b-a411-460b-9fe0-843261c02836\" (UID: \"f280252b-a411-460b-9fe0-843261c02836\") " Nov 24 14:08:27 crc kubenswrapper[4790]: I1124 14:08:27.709517 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f280252b-a411-460b-9fe0-843261c02836-catalog-content\") pod \"f280252b-a411-460b-9fe0-843261c02836\" (UID: \"f280252b-a411-460b-9fe0-843261c02836\") " Nov 24 14:08:27 crc kubenswrapper[4790]: I1124 14:08:27.709738 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f280252b-a411-460b-9fe0-843261c02836-utilities\") pod \"f280252b-a411-460b-9fe0-843261c02836\" (UID: \"f280252b-a411-460b-9fe0-843261c02836\") " Nov 24 14:08:27 crc kubenswrapper[4790]: I1124 14:08:27.710478 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f280252b-a411-460b-9fe0-843261c02836-utilities" (OuterVolumeSpecName: "utilities") pod "f280252b-a411-460b-9fe0-843261c02836" (UID: "f280252b-a411-460b-9fe0-843261c02836"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:08:27 crc kubenswrapper[4790]: I1124 14:08:27.718292 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f280252b-a411-460b-9fe0-843261c02836-kube-api-access-rbm55" (OuterVolumeSpecName: "kube-api-access-rbm55") pod "f280252b-a411-460b-9fe0-843261c02836" (UID: "f280252b-a411-460b-9fe0-843261c02836"). InnerVolumeSpecName "kube-api-access-rbm55". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:08:27 crc kubenswrapper[4790]: I1124 14:08:27.762119 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f280252b-a411-460b-9fe0-843261c02836-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f280252b-a411-460b-9fe0-843261c02836" (UID: "f280252b-a411-460b-9fe0-843261c02836"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:08:27 crc kubenswrapper[4790]: I1124 14:08:27.811171 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f280252b-a411-460b-9fe0-843261c02836-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:08:27 crc kubenswrapper[4790]: I1124 14:08:27.811462 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f280252b-a411-460b-9fe0-843261c02836-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:08:27 crc kubenswrapper[4790]: I1124 14:08:27.811529 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbm55\" (UniqueName: \"kubernetes.io/projected/f280252b-a411-460b-9fe0-843261c02836-kube-api-access-rbm55\") on node \"crc\" DevicePath \"\"" Nov 24 14:08:28 crc kubenswrapper[4790]: I1124 14:08:28.254923 4790 generic.go:334] "Generic (PLEG): container finished" podID="f280252b-a411-460b-9fe0-843261c02836" containerID="02644981edab7d70417fd5e4f986dbb820889a28c689ed676619698d3427e6a0" exitCode=0 Nov 24 14:08:28 crc kubenswrapper[4790]: I1124 14:08:28.254973 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gnk76" event={"ID":"f280252b-a411-460b-9fe0-843261c02836","Type":"ContainerDied","Data":"02644981edab7d70417fd5e4f986dbb820889a28c689ed676619698d3427e6a0"} Nov 24 14:08:28 crc kubenswrapper[4790]: I1124 14:08:28.254996 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gnk76" Nov 24 14:08:28 crc kubenswrapper[4790]: I1124 14:08:28.255603 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gnk76" event={"ID":"f280252b-a411-460b-9fe0-843261c02836","Type":"ContainerDied","Data":"a29265ab5f23fd95d4182bad1e6eb1afcebae93e1cbadb8840258226233cdf06"} Nov 24 14:08:28 crc kubenswrapper[4790]: I1124 14:08:28.255753 4790 scope.go:117] "RemoveContainer" containerID="02644981edab7d70417fd5e4f986dbb820889a28c689ed676619698d3427e6a0" Nov 24 14:08:28 crc kubenswrapper[4790]: I1124 14:08:28.283107 4790 scope.go:117] "RemoveContainer" containerID="1637d635fb06fe031c5c99910c348c31bb9fb4132746c58f0023ec63e8b734f9" Nov 24 14:08:28 crc kubenswrapper[4790]: I1124 14:08:28.290299 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gnk76"] Nov 24 14:08:28 crc kubenswrapper[4790]: I1124 14:08:28.297430 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gnk76"] Nov 24 14:08:28 crc kubenswrapper[4790]: I1124 14:08:28.305285 4790 scope.go:117] "RemoveContainer" containerID="82ac48f9b645a50d9059a1a74a91767ed99a57dd24b36e0c203e450e46463321" Nov 24 14:08:28 crc kubenswrapper[4790]: I1124 14:08:28.323566 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f280252b-a411-460b-9fe0-843261c02836" path="/var/lib/kubelet/pods/f280252b-a411-460b-9fe0-843261c02836/volumes" Nov 24 14:08:28 crc kubenswrapper[4790]: I1124 14:08:28.323675 4790 scope.go:117] "RemoveContainer" containerID="02644981edab7d70417fd5e4f986dbb820889a28c689ed676619698d3427e6a0" Nov 24 14:08:28 crc kubenswrapper[4790]: E1124 14:08:28.324137 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02644981edab7d70417fd5e4f986dbb820889a28c689ed676619698d3427e6a0\": container with ID starting with 02644981edab7d70417fd5e4f986dbb820889a28c689ed676619698d3427e6a0 not found: ID does not exist" containerID="02644981edab7d70417fd5e4f986dbb820889a28c689ed676619698d3427e6a0" Nov 24 14:08:28 crc kubenswrapper[4790]: I1124 14:08:28.324174 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02644981edab7d70417fd5e4f986dbb820889a28c689ed676619698d3427e6a0"} err="failed to get container status \"02644981edab7d70417fd5e4f986dbb820889a28c689ed676619698d3427e6a0\": rpc error: code = NotFound desc = could not find container \"02644981edab7d70417fd5e4f986dbb820889a28c689ed676619698d3427e6a0\": container with ID starting with 02644981edab7d70417fd5e4f986dbb820889a28c689ed676619698d3427e6a0 not found: ID does not exist" Nov 24 14:08:28 crc kubenswrapper[4790]: I1124 14:08:28.324202 4790 scope.go:117] "RemoveContainer" containerID="1637d635fb06fe031c5c99910c348c31bb9fb4132746c58f0023ec63e8b734f9" Nov 24 14:08:28 crc kubenswrapper[4790]: E1124 14:08:28.324763 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1637d635fb06fe031c5c99910c348c31bb9fb4132746c58f0023ec63e8b734f9\": container with ID starting with 1637d635fb06fe031c5c99910c348c31bb9fb4132746c58f0023ec63e8b734f9 not found: ID does not exist" containerID="1637d635fb06fe031c5c99910c348c31bb9fb4132746c58f0023ec63e8b734f9" Nov 24 14:08:28 crc kubenswrapper[4790]: I1124 14:08:28.324795 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1637d635fb06fe031c5c99910c348c31bb9fb4132746c58f0023ec63e8b734f9"} err="failed to get container status \"1637d635fb06fe031c5c99910c348c31bb9fb4132746c58f0023ec63e8b734f9\": rpc error: code = NotFound desc = could not find container \"1637d635fb06fe031c5c99910c348c31bb9fb4132746c58f0023ec63e8b734f9\": container with ID starting with 1637d635fb06fe031c5c99910c348c31bb9fb4132746c58f0023ec63e8b734f9 not found: ID does not exist" Nov 24 14:08:28 crc kubenswrapper[4790]: I1124 14:08:28.324813 4790 scope.go:117] "RemoveContainer" containerID="82ac48f9b645a50d9059a1a74a91767ed99a57dd24b36e0c203e450e46463321" Nov 24 14:08:28 crc kubenswrapper[4790]: E1124 14:08:28.325104 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82ac48f9b645a50d9059a1a74a91767ed99a57dd24b36e0c203e450e46463321\": container with ID starting with 82ac48f9b645a50d9059a1a74a91767ed99a57dd24b36e0c203e450e46463321 not found: ID does not exist" containerID="82ac48f9b645a50d9059a1a74a91767ed99a57dd24b36e0c203e450e46463321" Nov 24 14:08:28 crc kubenswrapper[4790]: I1124 14:08:28.325189 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82ac48f9b645a50d9059a1a74a91767ed99a57dd24b36e0c203e450e46463321"} err="failed to get container status \"82ac48f9b645a50d9059a1a74a91767ed99a57dd24b36e0c203e450e46463321\": rpc error: code = NotFound desc = could not find container \"82ac48f9b645a50d9059a1a74a91767ed99a57dd24b36e0c203e450e46463321\": container with ID starting with 82ac48f9b645a50d9059a1a74a91767ed99a57dd24b36e0c203e450e46463321 not found: ID does not exist" Nov 24 14:08:33 crc kubenswrapper[4790]: I1124 14:08:33.314053 4790 scope.go:117] "RemoveContainer" containerID="451fe5d9ca399689fc2ebac539dfde44079a1a33f2d2b78ba4e87e453a78b4b6" Nov 24 14:08:33 crc kubenswrapper[4790]: E1124 14:08:33.314730 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:08:45 crc kubenswrapper[4790]: I1124 14:08:45.315387 4790 scope.go:117] "RemoveContainer" containerID="451fe5d9ca399689fc2ebac539dfde44079a1a33f2d2b78ba4e87e453a78b4b6" Nov 24 14:08:45 crc kubenswrapper[4790]: E1124 14:08:45.316091 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:09:00 crc kubenswrapper[4790]: I1124 14:09:00.315572 4790 scope.go:117] "RemoveContainer" containerID="451fe5d9ca399689fc2ebac539dfde44079a1a33f2d2b78ba4e87e453a78b4b6" Nov 24 14:09:00 crc kubenswrapper[4790]: E1124 14:09:00.316388 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:09:14 crc kubenswrapper[4790]: I1124 14:09:14.314672 4790 scope.go:117] "RemoveContainer" containerID="451fe5d9ca399689fc2ebac539dfde44079a1a33f2d2b78ba4e87e453a78b4b6" Nov 24 14:09:14 crc kubenswrapper[4790]: E1124 14:09:14.315813 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:09:28 crc kubenswrapper[4790]: I1124 14:09:28.314446 4790 scope.go:117] "RemoveContainer" containerID="451fe5d9ca399689fc2ebac539dfde44079a1a33f2d2b78ba4e87e453a78b4b6" Nov 24 14:09:28 crc kubenswrapper[4790]: E1124 14:09:28.315099 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:09:40 crc kubenswrapper[4790]: I1124 14:09:40.314331 4790 scope.go:117] "RemoveContainer" containerID="451fe5d9ca399689fc2ebac539dfde44079a1a33f2d2b78ba4e87e453a78b4b6" Nov 24 14:09:40 crc kubenswrapper[4790]: E1124 14:09:40.315010 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:09:54 crc kubenswrapper[4790]: I1124 14:09:54.314066 4790 scope.go:117] "RemoveContainer" containerID="451fe5d9ca399689fc2ebac539dfde44079a1a33f2d2b78ba4e87e453a78b4b6" Nov 24 14:09:54 crc kubenswrapper[4790]: I1124 14:09:54.855830 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"d126dc358baf6e31345e3b42df23e22cb22fe7a763b1b6d8958121963d242656"} Nov 24 14:12:13 crc kubenswrapper[4790]: I1124 14:12:13.938678 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:12:13 crc kubenswrapper[4790]: I1124 14:12:13.939228 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:12:20 crc kubenswrapper[4790]: I1124 14:12:20.034132 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-r4vz8"] Nov 24 14:12:20 crc kubenswrapper[4790]: E1124 14:12:20.034682 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f280252b-a411-460b-9fe0-843261c02836" containerName="registry-server" Nov 24 14:12:20 crc kubenswrapper[4790]: I1124 14:12:20.034698 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="f280252b-a411-460b-9fe0-843261c02836" containerName="registry-server" Nov 24 14:12:20 crc kubenswrapper[4790]: E1124 14:12:20.034721 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f280252b-a411-460b-9fe0-843261c02836" containerName="extract-utilities" Nov 24 14:12:20 crc kubenswrapper[4790]: I1124 14:12:20.034729 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="f280252b-a411-460b-9fe0-843261c02836" containerName="extract-utilities" Nov 24 14:12:20 crc kubenswrapper[4790]: E1124 14:12:20.034747 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f280252b-a411-460b-9fe0-843261c02836" containerName="extract-content" Nov 24 14:12:20 crc kubenswrapper[4790]: I1124 14:12:20.034755 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="f280252b-a411-460b-9fe0-843261c02836" containerName="extract-content" Nov 24 14:12:20 crc kubenswrapper[4790]: I1124 14:12:20.034976 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="f280252b-a411-460b-9fe0-843261c02836" containerName="registry-server" Nov 24 14:12:20 crc kubenswrapper[4790]: I1124 14:12:20.036555 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r4vz8" Nov 24 14:12:20 crc kubenswrapper[4790]: I1124 14:12:20.042419 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r4vz8"] Nov 24 14:12:20 crc kubenswrapper[4790]: I1124 14:12:20.205959 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec04e11d-eb48-4e98-aa3b-11ea094a5924-utilities\") pod \"redhat-operators-r4vz8\" (UID: \"ec04e11d-eb48-4e98-aa3b-11ea094a5924\") " pod="openshift-marketplace/redhat-operators-r4vz8" Nov 24 14:12:20 crc kubenswrapper[4790]: I1124 14:12:20.206025 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec04e11d-eb48-4e98-aa3b-11ea094a5924-catalog-content\") pod \"redhat-operators-r4vz8\" (UID: \"ec04e11d-eb48-4e98-aa3b-11ea094a5924\") " pod="openshift-marketplace/redhat-operators-r4vz8" Nov 24 14:12:20 crc kubenswrapper[4790]: I1124 14:12:20.206063 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nz85d\" (UniqueName: \"kubernetes.io/projected/ec04e11d-eb48-4e98-aa3b-11ea094a5924-kube-api-access-nz85d\") pod \"redhat-operators-r4vz8\" (UID: \"ec04e11d-eb48-4e98-aa3b-11ea094a5924\") " pod="openshift-marketplace/redhat-operators-r4vz8" Nov 24 14:12:20 crc kubenswrapper[4790]: I1124 14:12:20.307796 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec04e11d-eb48-4e98-aa3b-11ea094a5924-utilities\") pod \"redhat-operators-r4vz8\" (UID: \"ec04e11d-eb48-4e98-aa3b-11ea094a5924\") " pod="openshift-marketplace/redhat-operators-r4vz8" Nov 24 14:12:20 crc kubenswrapper[4790]: I1124 14:12:20.307855 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec04e11d-eb48-4e98-aa3b-11ea094a5924-catalog-content\") pod \"redhat-operators-r4vz8\" (UID: \"ec04e11d-eb48-4e98-aa3b-11ea094a5924\") " pod="openshift-marketplace/redhat-operators-r4vz8" Nov 24 14:12:20 crc kubenswrapper[4790]: I1124 14:12:20.307897 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nz85d\" (UniqueName: \"kubernetes.io/projected/ec04e11d-eb48-4e98-aa3b-11ea094a5924-kube-api-access-nz85d\") pod \"redhat-operators-r4vz8\" (UID: \"ec04e11d-eb48-4e98-aa3b-11ea094a5924\") " pod="openshift-marketplace/redhat-operators-r4vz8" Nov 24 14:12:20 crc kubenswrapper[4790]: I1124 14:12:20.308291 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec04e11d-eb48-4e98-aa3b-11ea094a5924-utilities\") pod \"redhat-operators-r4vz8\" (UID: \"ec04e11d-eb48-4e98-aa3b-11ea094a5924\") " pod="openshift-marketplace/redhat-operators-r4vz8" Nov 24 14:12:20 crc kubenswrapper[4790]: I1124 14:12:20.308311 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec04e11d-eb48-4e98-aa3b-11ea094a5924-catalog-content\") pod \"redhat-operators-r4vz8\" (UID: \"ec04e11d-eb48-4e98-aa3b-11ea094a5924\") " pod="openshift-marketplace/redhat-operators-r4vz8" Nov 24 14:12:20 crc kubenswrapper[4790]: I1124 14:12:20.326754 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nz85d\" (UniqueName: \"kubernetes.io/projected/ec04e11d-eb48-4e98-aa3b-11ea094a5924-kube-api-access-nz85d\") pod \"redhat-operators-r4vz8\" (UID: \"ec04e11d-eb48-4e98-aa3b-11ea094a5924\") " pod="openshift-marketplace/redhat-operators-r4vz8" Nov 24 14:12:20 crc kubenswrapper[4790]: I1124 14:12:20.372393 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r4vz8" Nov 24 14:12:20 crc kubenswrapper[4790]: I1124 14:12:20.786962 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r4vz8"] Nov 24 14:12:20 crc kubenswrapper[4790]: I1124 14:12:20.863282 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r4vz8" event={"ID":"ec04e11d-eb48-4e98-aa3b-11ea094a5924","Type":"ContainerStarted","Data":"7b25b3ec3bc9578d6af2afcbde3cfc595f58a3433c9059099794cc76bac44141"} Nov 24 14:12:21 crc kubenswrapper[4790]: I1124 14:12:21.872846 4790 generic.go:334] "Generic (PLEG): container finished" podID="ec04e11d-eb48-4e98-aa3b-11ea094a5924" containerID="9e73b5b5a2037b6dc73b05d2a768206326f0b6e7dd53ed5e545ca410a39f50e0" exitCode=0 Nov 24 14:12:21 crc kubenswrapper[4790]: I1124 14:12:21.872923 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r4vz8" event={"ID":"ec04e11d-eb48-4e98-aa3b-11ea094a5924","Type":"ContainerDied","Data":"9e73b5b5a2037b6dc73b05d2a768206326f0b6e7dd53ed5e545ca410a39f50e0"} Nov 24 14:12:21 crc kubenswrapper[4790]: I1124 14:12:21.876049 4790 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 14:12:22 crc kubenswrapper[4790]: I1124 14:12:22.880284 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r4vz8" event={"ID":"ec04e11d-eb48-4e98-aa3b-11ea094a5924","Type":"ContainerStarted","Data":"df8da810c824df26b7cd3da333c5c28d20b91f3f7a99c7471f5d971ee48af589"} Nov 24 14:12:23 crc kubenswrapper[4790]: I1124 14:12:23.889647 4790 generic.go:334] "Generic (PLEG): container finished" podID="ec04e11d-eb48-4e98-aa3b-11ea094a5924" containerID="df8da810c824df26b7cd3da333c5c28d20b91f3f7a99c7471f5d971ee48af589" exitCode=0 Nov 24 14:12:23 crc kubenswrapper[4790]: I1124 14:12:23.889697 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r4vz8" event={"ID":"ec04e11d-eb48-4e98-aa3b-11ea094a5924","Type":"ContainerDied","Data":"df8da810c824df26b7cd3da333c5c28d20b91f3f7a99c7471f5d971ee48af589"} Nov 24 14:12:24 crc kubenswrapper[4790]: I1124 14:12:24.900163 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r4vz8" event={"ID":"ec04e11d-eb48-4e98-aa3b-11ea094a5924","Type":"ContainerStarted","Data":"914bb44cf2487ae1ed22c62f607464307cf808d47fc33fb6c2a922757348278b"} Nov 24 14:12:24 crc kubenswrapper[4790]: I1124 14:12:24.919419 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-r4vz8" podStartSLOduration=2.430813678 podStartE2EDuration="4.919399659s" podCreationTimestamp="2025-11-24 14:12:20 +0000 UTC" firstStartedPulling="2025-11-24 14:12:21.875727985 +0000 UTC m=+3590.255621647" lastFinishedPulling="2025-11-24 14:12:24.364313956 +0000 UTC m=+3592.744207628" observedRunningTime="2025-11-24 14:12:24.918117423 +0000 UTC m=+3593.298011095" watchObservedRunningTime="2025-11-24 14:12:24.919399659 +0000 UTC m=+3593.299293321" Nov 24 14:12:25 crc kubenswrapper[4790]: I1124 14:12:25.635821 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gc8zh"] Nov 24 14:12:25 crc kubenswrapper[4790]: I1124 14:12:25.638218 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gc8zh" Nov 24 14:12:25 crc kubenswrapper[4790]: I1124 14:12:25.650918 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gc8zh"] Nov 24 14:12:25 crc kubenswrapper[4790]: I1124 14:12:25.788984 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwmkz\" (UniqueName: \"kubernetes.io/projected/b0120341-ad67-4000-975a-5cb7186514da-kube-api-access-lwmkz\") pod \"redhat-marketplace-gc8zh\" (UID: \"b0120341-ad67-4000-975a-5cb7186514da\") " pod="openshift-marketplace/redhat-marketplace-gc8zh" Nov 24 14:12:25 crc kubenswrapper[4790]: I1124 14:12:25.789205 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0120341-ad67-4000-975a-5cb7186514da-utilities\") pod \"redhat-marketplace-gc8zh\" (UID: \"b0120341-ad67-4000-975a-5cb7186514da\") " pod="openshift-marketplace/redhat-marketplace-gc8zh" Nov 24 14:12:25 crc kubenswrapper[4790]: I1124 14:12:25.789352 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0120341-ad67-4000-975a-5cb7186514da-catalog-content\") pod \"redhat-marketplace-gc8zh\" (UID: \"b0120341-ad67-4000-975a-5cb7186514da\") " pod="openshift-marketplace/redhat-marketplace-gc8zh" Nov 24 14:12:25 crc kubenswrapper[4790]: I1124 14:12:25.890438 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0120341-ad67-4000-975a-5cb7186514da-catalog-content\") pod \"redhat-marketplace-gc8zh\" (UID: \"b0120341-ad67-4000-975a-5cb7186514da\") " pod="openshift-marketplace/redhat-marketplace-gc8zh" Nov 24 14:12:25 crc kubenswrapper[4790]: I1124 14:12:25.890518 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwmkz\" (UniqueName: \"kubernetes.io/projected/b0120341-ad67-4000-975a-5cb7186514da-kube-api-access-lwmkz\") pod \"redhat-marketplace-gc8zh\" (UID: \"b0120341-ad67-4000-975a-5cb7186514da\") " pod="openshift-marketplace/redhat-marketplace-gc8zh" Nov 24 14:12:25 crc kubenswrapper[4790]: I1124 14:12:25.890589 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0120341-ad67-4000-975a-5cb7186514da-utilities\") pod \"redhat-marketplace-gc8zh\" (UID: \"b0120341-ad67-4000-975a-5cb7186514da\") " pod="openshift-marketplace/redhat-marketplace-gc8zh" Nov 24 14:12:25 crc kubenswrapper[4790]: I1124 14:12:25.891023 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0120341-ad67-4000-975a-5cb7186514da-catalog-content\") pod \"redhat-marketplace-gc8zh\" (UID: \"b0120341-ad67-4000-975a-5cb7186514da\") " pod="openshift-marketplace/redhat-marketplace-gc8zh" Nov 24 14:12:25 crc kubenswrapper[4790]: I1124 14:12:25.891048 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0120341-ad67-4000-975a-5cb7186514da-utilities\") pod \"redhat-marketplace-gc8zh\" (UID: \"b0120341-ad67-4000-975a-5cb7186514da\") " pod="openshift-marketplace/redhat-marketplace-gc8zh" Nov 24 14:12:25 crc kubenswrapper[4790]: I1124 14:12:25.912760 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwmkz\" (UniqueName: \"kubernetes.io/projected/b0120341-ad67-4000-975a-5cb7186514da-kube-api-access-lwmkz\") pod \"redhat-marketplace-gc8zh\" (UID: \"b0120341-ad67-4000-975a-5cb7186514da\") " pod="openshift-marketplace/redhat-marketplace-gc8zh" Nov 24 14:12:25 crc kubenswrapper[4790]: I1124 14:12:25.959069 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gc8zh" Nov 24 14:12:26 crc kubenswrapper[4790]: I1124 14:12:26.395087 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gc8zh"] Nov 24 14:12:26 crc kubenswrapper[4790]: I1124 14:12:26.914151 4790 generic.go:334] "Generic (PLEG): container finished" podID="b0120341-ad67-4000-975a-5cb7186514da" containerID="3e1b078ac2673c5182657ccdc833bcb2203090eb58d7117c41a2a2e680ab65de" exitCode=0 Nov 24 14:12:26 crc kubenswrapper[4790]: I1124 14:12:26.914216 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gc8zh" event={"ID":"b0120341-ad67-4000-975a-5cb7186514da","Type":"ContainerDied","Data":"3e1b078ac2673c5182657ccdc833bcb2203090eb58d7117c41a2a2e680ab65de"} Nov 24 14:12:26 crc kubenswrapper[4790]: I1124 14:12:26.914241 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gc8zh" event={"ID":"b0120341-ad67-4000-975a-5cb7186514da","Type":"ContainerStarted","Data":"8a90ea8675edbbb19d1e52cb8127727f2c96b86672867daeb8b0f942ca3572a7"} Nov 24 14:12:27 crc kubenswrapper[4790]: I1124 14:12:27.923691 4790 generic.go:334] "Generic (PLEG): container finished" podID="b0120341-ad67-4000-975a-5cb7186514da" containerID="3a253dfa7dd8376b939cb5841ca623ba9ad6db13b796d1b9be69b5dbd1464f51" exitCode=0 Nov 24 14:12:27 crc kubenswrapper[4790]: I1124 14:12:27.923748 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gc8zh" event={"ID":"b0120341-ad67-4000-975a-5cb7186514da","Type":"ContainerDied","Data":"3a253dfa7dd8376b939cb5841ca623ba9ad6db13b796d1b9be69b5dbd1464f51"} Nov 24 14:12:28 crc kubenswrapper[4790]: I1124 14:12:28.932933 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gc8zh" event={"ID":"b0120341-ad67-4000-975a-5cb7186514da","Type":"ContainerStarted","Data":"9fe3267b2fa74d507261b83410e37c4f64deefc8f6adc7d8b01a28d18cd74e50"} Nov 24 14:12:28 crc kubenswrapper[4790]: I1124 14:12:28.953144 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gc8zh" podStartSLOduration=2.309732717 podStartE2EDuration="3.953126641s" podCreationTimestamp="2025-11-24 14:12:25 +0000 UTC" firstStartedPulling="2025-11-24 14:12:26.915716176 +0000 UTC m=+3595.295609838" lastFinishedPulling="2025-11-24 14:12:28.5591101 +0000 UTC m=+3596.939003762" observedRunningTime="2025-11-24 14:12:28.947921852 +0000 UTC m=+3597.327815524" watchObservedRunningTime="2025-11-24 14:12:28.953126641 +0000 UTC m=+3597.333020303" Nov 24 14:12:30 crc kubenswrapper[4790]: I1124 14:12:30.373627 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-r4vz8" Nov 24 14:12:30 crc kubenswrapper[4790]: I1124 14:12:30.373822 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-r4vz8" Nov 24 14:12:30 crc kubenswrapper[4790]: I1124 14:12:30.420288 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-r4vz8" Nov 24 14:12:30 crc kubenswrapper[4790]: I1124 14:12:30.987328 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-r4vz8" Nov 24 14:12:32 crc kubenswrapper[4790]: I1124 14:12:32.226704 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r4vz8"] Nov 24 14:12:32 crc kubenswrapper[4790]: I1124 14:12:32.971680 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-r4vz8" podUID="ec04e11d-eb48-4e98-aa3b-11ea094a5924" containerName="registry-server" containerID="cri-o://914bb44cf2487ae1ed22c62f607464307cf808d47fc33fb6c2a922757348278b" gracePeriod=2 Nov 24 14:12:33 crc kubenswrapper[4790]: I1124 14:12:33.373628 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r4vz8" Nov 24 14:12:33 crc kubenswrapper[4790]: I1124 14:12:33.495579 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec04e11d-eb48-4e98-aa3b-11ea094a5924-utilities\") pod \"ec04e11d-eb48-4e98-aa3b-11ea094a5924\" (UID: \"ec04e11d-eb48-4e98-aa3b-11ea094a5924\") " Nov 24 14:12:33 crc kubenswrapper[4790]: I1124 14:12:33.495948 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nz85d\" (UniqueName: \"kubernetes.io/projected/ec04e11d-eb48-4e98-aa3b-11ea094a5924-kube-api-access-nz85d\") pod \"ec04e11d-eb48-4e98-aa3b-11ea094a5924\" (UID: \"ec04e11d-eb48-4e98-aa3b-11ea094a5924\") " Nov 24 14:12:33 crc kubenswrapper[4790]: I1124 14:12:33.496145 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec04e11d-eb48-4e98-aa3b-11ea094a5924-catalog-content\") pod \"ec04e11d-eb48-4e98-aa3b-11ea094a5924\" (UID: \"ec04e11d-eb48-4e98-aa3b-11ea094a5924\") " Nov 24 14:12:33 crc kubenswrapper[4790]: I1124 14:12:33.496585 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec04e11d-eb48-4e98-aa3b-11ea094a5924-utilities" (OuterVolumeSpecName: "utilities") pod "ec04e11d-eb48-4e98-aa3b-11ea094a5924" (UID: "ec04e11d-eb48-4e98-aa3b-11ea094a5924"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:12:33 crc kubenswrapper[4790]: I1124 14:12:33.503394 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec04e11d-eb48-4e98-aa3b-11ea094a5924-kube-api-access-nz85d" (OuterVolumeSpecName: "kube-api-access-nz85d") pod "ec04e11d-eb48-4e98-aa3b-11ea094a5924" (UID: "ec04e11d-eb48-4e98-aa3b-11ea094a5924"). InnerVolumeSpecName "kube-api-access-nz85d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:12:33 crc kubenswrapper[4790]: I1124 14:12:33.597994 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec04e11d-eb48-4e98-aa3b-11ea094a5924-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:12:33 crc kubenswrapper[4790]: I1124 14:12:33.598022 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nz85d\" (UniqueName: \"kubernetes.io/projected/ec04e11d-eb48-4e98-aa3b-11ea094a5924-kube-api-access-nz85d\") on node \"crc\" DevicePath \"\"" Nov 24 14:12:33 crc kubenswrapper[4790]: I1124 14:12:33.980477 4790 generic.go:334] "Generic (PLEG): container finished" podID="ec04e11d-eb48-4e98-aa3b-11ea094a5924" containerID="914bb44cf2487ae1ed22c62f607464307cf808d47fc33fb6c2a922757348278b" exitCode=0 Nov 24 14:12:33 crc kubenswrapper[4790]: I1124 14:12:33.980712 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r4vz8" event={"ID":"ec04e11d-eb48-4e98-aa3b-11ea094a5924","Type":"ContainerDied","Data":"914bb44cf2487ae1ed22c62f607464307cf808d47fc33fb6c2a922757348278b"} Nov 24 14:12:33 crc kubenswrapper[4790]: I1124 14:12:33.980791 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r4vz8" event={"ID":"ec04e11d-eb48-4e98-aa3b-11ea094a5924","Type":"ContainerDied","Data":"7b25b3ec3bc9578d6af2afcbde3cfc595f58a3433c9059099794cc76bac44141"} Nov 24 14:12:33 crc kubenswrapper[4790]: I1124 14:12:33.980916 4790 scope.go:117] "RemoveContainer" containerID="914bb44cf2487ae1ed22c62f607464307cf808d47fc33fb6c2a922757348278b" Nov 24 14:12:33 crc kubenswrapper[4790]: I1124 14:12:33.981057 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r4vz8" Nov 24 14:12:34 crc kubenswrapper[4790]: I1124 14:12:34.001848 4790 scope.go:117] "RemoveContainer" containerID="df8da810c824df26b7cd3da333c5c28d20b91f3f7a99c7471f5d971ee48af589" Nov 24 14:12:34 crc kubenswrapper[4790]: I1124 14:12:34.023686 4790 scope.go:117] "RemoveContainer" containerID="9e73b5b5a2037b6dc73b05d2a768206326f0b6e7dd53ed5e545ca410a39f50e0" Nov 24 14:12:34 crc kubenswrapper[4790]: I1124 14:12:34.046547 4790 scope.go:117] "RemoveContainer" containerID="914bb44cf2487ae1ed22c62f607464307cf808d47fc33fb6c2a922757348278b" Nov 24 14:12:34 crc kubenswrapper[4790]: E1124 14:12:34.046979 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"914bb44cf2487ae1ed22c62f607464307cf808d47fc33fb6c2a922757348278b\": container with ID starting with 914bb44cf2487ae1ed22c62f607464307cf808d47fc33fb6c2a922757348278b not found: ID does not exist" containerID="914bb44cf2487ae1ed22c62f607464307cf808d47fc33fb6c2a922757348278b" Nov 24 14:12:34 crc kubenswrapper[4790]: I1124 14:12:34.047006 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"914bb44cf2487ae1ed22c62f607464307cf808d47fc33fb6c2a922757348278b"} err="failed to get container status \"914bb44cf2487ae1ed22c62f607464307cf808d47fc33fb6c2a922757348278b\": rpc error: code = NotFound desc = could not find container \"914bb44cf2487ae1ed22c62f607464307cf808d47fc33fb6c2a922757348278b\": container with ID starting with 914bb44cf2487ae1ed22c62f607464307cf808d47fc33fb6c2a922757348278b not found: ID does not exist" Nov 24 14:12:34 crc kubenswrapper[4790]: I1124 14:12:34.047027 4790 scope.go:117] "RemoveContainer" containerID="df8da810c824df26b7cd3da333c5c28d20b91f3f7a99c7471f5d971ee48af589" Nov 24 14:12:34 crc kubenswrapper[4790]: E1124 14:12:34.047273 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df8da810c824df26b7cd3da333c5c28d20b91f3f7a99c7471f5d971ee48af589\": container with ID starting with df8da810c824df26b7cd3da333c5c28d20b91f3f7a99c7471f5d971ee48af589 not found: ID does not exist" containerID="df8da810c824df26b7cd3da333c5c28d20b91f3f7a99c7471f5d971ee48af589" Nov 24 14:12:34 crc kubenswrapper[4790]: I1124 14:12:34.047302 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df8da810c824df26b7cd3da333c5c28d20b91f3f7a99c7471f5d971ee48af589"} err="failed to get container status \"df8da810c824df26b7cd3da333c5c28d20b91f3f7a99c7471f5d971ee48af589\": rpc error: code = NotFound desc = could not find container \"df8da810c824df26b7cd3da333c5c28d20b91f3f7a99c7471f5d971ee48af589\": container with ID starting with df8da810c824df26b7cd3da333c5c28d20b91f3f7a99c7471f5d971ee48af589 not found: ID does not exist" Nov 24 14:12:34 crc kubenswrapper[4790]: I1124 14:12:34.047321 4790 scope.go:117] "RemoveContainer" containerID="9e73b5b5a2037b6dc73b05d2a768206326f0b6e7dd53ed5e545ca410a39f50e0" Nov 24 14:12:34 crc kubenswrapper[4790]: E1124 14:12:34.047573 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e73b5b5a2037b6dc73b05d2a768206326f0b6e7dd53ed5e545ca410a39f50e0\": container with ID starting with 9e73b5b5a2037b6dc73b05d2a768206326f0b6e7dd53ed5e545ca410a39f50e0 not found: ID does not exist" containerID="9e73b5b5a2037b6dc73b05d2a768206326f0b6e7dd53ed5e545ca410a39f50e0" Nov 24 14:12:34 crc kubenswrapper[4790]: I1124 14:12:34.047592 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e73b5b5a2037b6dc73b05d2a768206326f0b6e7dd53ed5e545ca410a39f50e0"} err="failed to get container status \"9e73b5b5a2037b6dc73b05d2a768206326f0b6e7dd53ed5e545ca410a39f50e0\": rpc error: code = NotFound desc = could not find container \"9e73b5b5a2037b6dc73b05d2a768206326f0b6e7dd53ed5e545ca410a39f50e0\": container with ID starting with 9e73b5b5a2037b6dc73b05d2a768206326f0b6e7dd53ed5e545ca410a39f50e0 not found: ID does not exist" Nov 24 14:12:35 crc kubenswrapper[4790]: I1124 14:12:35.092276 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec04e11d-eb48-4e98-aa3b-11ea094a5924-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ec04e11d-eb48-4e98-aa3b-11ea094a5924" (UID: "ec04e11d-eb48-4e98-aa3b-11ea094a5924"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:12:35 crc kubenswrapper[4790]: I1124 14:12:35.117545 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec04e11d-eb48-4e98-aa3b-11ea094a5924-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:12:35 crc kubenswrapper[4790]: I1124 14:12:35.214787 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r4vz8"] Nov 24 14:12:35 crc kubenswrapper[4790]: I1124 14:12:35.221590 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-r4vz8"] Nov 24 14:12:35 crc kubenswrapper[4790]: I1124 14:12:35.960151 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gc8zh" Nov 24 14:12:35 crc kubenswrapper[4790]: I1124 14:12:35.960726 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gc8zh" Nov 24 14:12:36 crc kubenswrapper[4790]: I1124 14:12:36.007216 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gc8zh" Nov 24 14:12:36 crc kubenswrapper[4790]: I1124 14:12:36.329179 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec04e11d-eb48-4e98-aa3b-11ea094a5924" path="/var/lib/kubelet/pods/ec04e11d-eb48-4e98-aa3b-11ea094a5924/volumes" Nov 24 14:12:37 crc kubenswrapper[4790]: I1124 14:12:37.043435 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gc8zh" Nov 24 14:12:37 crc kubenswrapper[4790]: I1124 14:12:37.622953 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gc8zh"] Nov 24 14:12:39 crc kubenswrapper[4790]: I1124 14:12:39.016631 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gc8zh" podUID="b0120341-ad67-4000-975a-5cb7186514da" containerName="registry-server" containerID="cri-o://9fe3267b2fa74d507261b83410e37c4f64deefc8f6adc7d8b01a28d18cd74e50" gracePeriod=2 Nov 24 14:12:39 crc kubenswrapper[4790]: I1124 14:12:39.369814 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gc8zh" Nov 24 14:12:39 crc kubenswrapper[4790]: I1124 14:12:39.483782 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwmkz\" (UniqueName: \"kubernetes.io/projected/b0120341-ad67-4000-975a-5cb7186514da-kube-api-access-lwmkz\") pod \"b0120341-ad67-4000-975a-5cb7186514da\" (UID: \"b0120341-ad67-4000-975a-5cb7186514da\") " Nov 24 14:12:39 crc kubenswrapper[4790]: I1124 14:12:39.483973 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0120341-ad67-4000-975a-5cb7186514da-utilities\") pod \"b0120341-ad67-4000-975a-5cb7186514da\" (UID: \"b0120341-ad67-4000-975a-5cb7186514da\") " Nov 24 14:12:39 crc kubenswrapper[4790]: I1124 14:12:39.484023 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0120341-ad67-4000-975a-5cb7186514da-catalog-content\") pod \"b0120341-ad67-4000-975a-5cb7186514da\" (UID: \"b0120341-ad67-4000-975a-5cb7186514da\") " Nov 24 14:12:39 crc kubenswrapper[4790]: I1124 14:12:39.485144 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0120341-ad67-4000-975a-5cb7186514da-utilities" (OuterVolumeSpecName: "utilities") pod "b0120341-ad67-4000-975a-5cb7186514da" (UID: "b0120341-ad67-4000-975a-5cb7186514da"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:12:39 crc kubenswrapper[4790]: I1124 14:12:39.490321 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0120341-ad67-4000-975a-5cb7186514da-kube-api-access-lwmkz" (OuterVolumeSpecName: "kube-api-access-lwmkz") pod "b0120341-ad67-4000-975a-5cb7186514da" (UID: "b0120341-ad67-4000-975a-5cb7186514da"). InnerVolumeSpecName "kube-api-access-lwmkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:12:39 crc kubenswrapper[4790]: I1124 14:12:39.508372 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0120341-ad67-4000-975a-5cb7186514da-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b0120341-ad67-4000-975a-5cb7186514da" (UID: "b0120341-ad67-4000-975a-5cb7186514da"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:12:39 crc kubenswrapper[4790]: I1124 14:12:39.585709 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0120341-ad67-4000-975a-5cb7186514da-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:12:39 crc kubenswrapper[4790]: I1124 14:12:39.585759 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0120341-ad67-4000-975a-5cb7186514da-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:12:39 crc kubenswrapper[4790]: I1124 14:12:39.585777 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwmkz\" (UniqueName: \"kubernetes.io/projected/b0120341-ad67-4000-975a-5cb7186514da-kube-api-access-lwmkz\") on node \"crc\" DevicePath \"\"" Nov 24 14:12:40 crc kubenswrapper[4790]: I1124 14:12:40.034313 4790 generic.go:334] "Generic (PLEG): container finished" podID="b0120341-ad67-4000-975a-5cb7186514da" containerID="9fe3267b2fa74d507261b83410e37c4f64deefc8f6adc7d8b01a28d18cd74e50" exitCode=0 Nov 24 14:12:40 crc kubenswrapper[4790]: I1124 14:12:40.034394 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gc8zh" event={"ID":"b0120341-ad67-4000-975a-5cb7186514da","Type":"ContainerDied","Data":"9fe3267b2fa74d507261b83410e37c4f64deefc8f6adc7d8b01a28d18cd74e50"} Nov 24 14:12:40 crc kubenswrapper[4790]: I1124 14:12:40.034434 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gc8zh" event={"ID":"b0120341-ad67-4000-975a-5cb7186514da","Type":"ContainerDied","Data":"8a90ea8675edbbb19d1e52cb8127727f2c96b86672867daeb8b0f942ca3572a7"} Nov 24 14:12:40 crc kubenswrapper[4790]: I1124 14:12:40.034459 4790 scope.go:117] "RemoveContainer" containerID="9fe3267b2fa74d507261b83410e37c4f64deefc8f6adc7d8b01a28d18cd74e50" Nov 24 14:12:40 crc kubenswrapper[4790]: I1124 14:12:40.034704 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gc8zh" Nov 24 14:12:40 crc kubenswrapper[4790]: I1124 14:12:40.068381 4790 scope.go:117] "RemoveContainer" containerID="3a253dfa7dd8376b939cb5841ca623ba9ad6db13b796d1b9be69b5dbd1464f51" Nov 24 14:12:40 crc kubenswrapper[4790]: I1124 14:12:40.076518 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gc8zh"] Nov 24 14:12:40 crc kubenswrapper[4790]: I1124 14:12:40.082155 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gc8zh"] Nov 24 14:12:40 crc kubenswrapper[4790]: I1124 14:12:40.090122 4790 scope.go:117] "RemoveContainer" containerID="3e1b078ac2673c5182657ccdc833bcb2203090eb58d7117c41a2a2e680ab65de" Nov 24 14:12:40 crc kubenswrapper[4790]: I1124 14:12:40.123121 4790 scope.go:117] "RemoveContainer" containerID="9fe3267b2fa74d507261b83410e37c4f64deefc8f6adc7d8b01a28d18cd74e50" Nov 24 14:12:40 crc kubenswrapper[4790]: E1124 14:12:40.123721 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fe3267b2fa74d507261b83410e37c4f64deefc8f6adc7d8b01a28d18cd74e50\": container with ID starting with 9fe3267b2fa74d507261b83410e37c4f64deefc8f6adc7d8b01a28d18cd74e50 not found: ID does not exist" containerID="9fe3267b2fa74d507261b83410e37c4f64deefc8f6adc7d8b01a28d18cd74e50" Nov 24 14:12:40 crc kubenswrapper[4790]: I1124 14:12:40.123776 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fe3267b2fa74d507261b83410e37c4f64deefc8f6adc7d8b01a28d18cd74e50"} err="failed to get container status \"9fe3267b2fa74d507261b83410e37c4f64deefc8f6adc7d8b01a28d18cd74e50\": rpc error: code = NotFound desc = could not find container \"9fe3267b2fa74d507261b83410e37c4f64deefc8f6adc7d8b01a28d18cd74e50\": container with ID starting with 9fe3267b2fa74d507261b83410e37c4f64deefc8f6adc7d8b01a28d18cd74e50 not found: ID does not exist" Nov 24 14:12:40 crc kubenswrapper[4790]: I1124 14:12:40.123814 4790 scope.go:117] "RemoveContainer" containerID="3a253dfa7dd8376b939cb5841ca623ba9ad6db13b796d1b9be69b5dbd1464f51" Nov 24 14:12:40 crc kubenswrapper[4790]: E1124 14:12:40.124215 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a253dfa7dd8376b939cb5841ca623ba9ad6db13b796d1b9be69b5dbd1464f51\": container with ID starting with 3a253dfa7dd8376b939cb5841ca623ba9ad6db13b796d1b9be69b5dbd1464f51 not found: ID does not exist" containerID="3a253dfa7dd8376b939cb5841ca623ba9ad6db13b796d1b9be69b5dbd1464f51" Nov 24 14:12:40 crc kubenswrapper[4790]: I1124 14:12:40.124257 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a253dfa7dd8376b939cb5841ca623ba9ad6db13b796d1b9be69b5dbd1464f51"} err="failed to get container status \"3a253dfa7dd8376b939cb5841ca623ba9ad6db13b796d1b9be69b5dbd1464f51\": rpc error: code = NotFound desc = could not find container \"3a253dfa7dd8376b939cb5841ca623ba9ad6db13b796d1b9be69b5dbd1464f51\": container with ID starting with 3a253dfa7dd8376b939cb5841ca623ba9ad6db13b796d1b9be69b5dbd1464f51 not found: ID does not exist" Nov 24 14:12:40 crc kubenswrapper[4790]: I1124 14:12:40.124313 4790 scope.go:117] "RemoveContainer" containerID="3e1b078ac2673c5182657ccdc833bcb2203090eb58d7117c41a2a2e680ab65de" Nov 24 14:12:40 crc kubenswrapper[4790]: E1124 14:12:40.124562 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e1b078ac2673c5182657ccdc833bcb2203090eb58d7117c41a2a2e680ab65de\": container with ID starting with 3e1b078ac2673c5182657ccdc833bcb2203090eb58d7117c41a2a2e680ab65de not found: ID does not exist" containerID="3e1b078ac2673c5182657ccdc833bcb2203090eb58d7117c41a2a2e680ab65de" Nov 24 14:12:40 crc kubenswrapper[4790]: I1124 14:12:40.124617 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e1b078ac2673c5182657ccdc833bcb2203090eb58d7117c41a2a2e680ab65de"} err="failed to get container status \"3e1b078ac2673c5182657ccdc833bcb2203090eb58d7117c41a2a2e680ab65de\": rpc error: code = NotFound desc = could not find container \"3e1b078ac2673c5182657ccdc833bcb2203090eb58d7117c41a2a2e680ab65de\": container with ID starting with 3e1b078ac2673c5182657ccdc833bcb2203090eb58d7117c41a2a2e680ab65de not found: ID does not exist" Nov 24 14:12:40 crc kubenswrapper[4790]: I1124 14:12:40.324952 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0120341-ad67-4000-975a-5cb7186514da" path="/var/lib/kubelet/pods/b0120341-ad67-4000-975a-5cb7186514da/volumes" Nov 24 14:12:43 crc kubenswrapper[4790]: I1124 14:12:43.938492 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:12:43 crc kubenswrapper[4790]: I1124 14:12:43.938800 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:13:13 crc kubenswrapper[4790]: I1124 14:13:13.938792 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:13:13 crc kubenswrapper[4790]: I1124 14:13:13.940039 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:13:13 crc kubenswrapper[4790]: I1124 14:13:13.940108 4790 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 14:13:13 crc kubenswrapper[4790]: I1124 14:13:13.940767 4790 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d126dc358baf6e31345e3b42df23e22cb22fe7a763b1b6d8958121963d242656"} pod="openshift-machine-config-operator/machine-config-daemon-xz49t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:13:13 crc kubenswrapper[4790]: I1124 14:13:13.940896 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" containerID="cri-o://d126dc358baf6e31345e3b42df23e22cb22fe7a763b1b6d8958121963d242656" gracePeriod=600 Nov 24 14:13:14 crc kubenswrapper[4790]: I1124 14:13:14.269474 4790 generic.go:334] "Generic (PLEG): container finished" podID="0d73b133-48f0-455f-8f6a-742e633f631a" containerID="d126dc358baf6e31345e3b42df23e22cb22fe7a763b1b6d8958121963d242656" exitCode=0 Nov 24 14:13:14 crc kubenswrapper[4790]: I1124 14:13:14.269543 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerDied","Data":"d126dc358baf6e31345e3b42df23e22cb22fe7a763b1b6d8958121963d242656"} Nov 24 14:13:14 crc kubenswrapper[4790]: I1124 14:13:14.269783 4790 scope.go:117] "RemoveContainer" containerID="451fe5d9ca399689fc2ebac539dfde44079a1a33f2d2b78ba4e87e453a78b4b6" Nov 24 14:13:15 crc kubenswrapper[4790]: I1124 14:13:15.278692 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"9d160b5f0ea013bde38ba30b4fd4ea41f13ba76ae92c0901dbd7ffee1185aac0"} Nov 24 14:15:00 crc kubenswrapper[4790]: I1124 14:15:00.207276 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399895-t6g7z"] Nov 24 14:15:00 crc kubenswrapper[4790]: E1124 14:15:00.208201 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0120341-ad67-4000-975a-5cb7186514da" containerName="extract-utilities" Nov 24 14:15:00 crc kubenswrapper[4790]: I1124 14:15:00.208221 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0120341-ad67-4000-975a-5cb7186514da" containerName="extract-utilities" Nov 24 14:15:00 crc kubenswrapper[4790]: E1124 14:15:00.208242 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec04e11d-eb48-4e98-aa3b-11ea094a5924" containerName="extract-content" Nov 24 14:15:00 crc kubenswrapper[4790]: I1124 14:15:00.208252 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec04e11d-eb48-4e98-aa3b-11ea094a5924" containerName="extract-content" Nov 24 14:15:00 crc kubenswrapper[4790]: E1124 14:15:00.208267 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0120341-ad67-4000-975a-5cb7186514da" containerName="registry-server" Nov 24 14:15:00 crc kubenswrapper[4790]: I1124 14:15:00.208276 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0120341-ad67-4000-975a-5cb7186514da" containerName="registry-server" Nov 24 14:15:00 crc kubenswrapper[4790]: E1124 14:15:00.208296 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec04e11d-eb48-4e98-aa3b-11ea094a5924" containerName="extract-utilities" Nov 24 14:15:00 crc kubenswrapper[4790]: I1124 14:15:00.208304 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec04e11d-eb48-4e98-aa3b-11ea094a5924" containerName="extract-utilities" Nov 24 14:15:00 crc kubenswrapper[4790]: E1124 14:15:00.208318 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec04e11d-eb48-4e98-aa3b-11ea094a5924" containerName="registry-server" Nov 24 14:15:00 crc kubenswrapper[4790]: I1124 14:15:00.208326 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec04e11d-eb48-4e98-aa3b-11ea094a5924" containerName="registry-server" Nov 24 14:15:00 crc kubenswrapper[4790]: E1124 14:15:00.208339 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0120341-ad67-4000-975a-5cb7186514da" containerName="extract-content" Nov 24 14:15:00 crc kubenswrapper[4790]: I1124 14:15:00.208346 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0120341-ad67-4000-975a-5cb7186514da" containerName="extract-content" Nov 24 14:15:00 crc kubenswrapper[4790]: I1124 14:15:00.208527 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0120341-ad67-4000-975a-5cb7186514da" containerName="registry-server" Nov 24 14:15:00 crc kubenswrapper[4790]: I1124 14:15:00.208540 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec04e11d-eb48-4e98-aa3b-11ea094a5924" containerName="registry-server" Nov 24 14:15:00 crc kubenswrapper[4790]: I1124 14:15:00.209119 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-t6g7z" Nov 24 14:15:00 crc kubenswrapper[4790]: I1124 14:15:00.228486 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 14:15:00 crc kubenswrapper[4790]: I1124 14:15:00.229105 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399895-t6g7z"] Nov 24 14:15:00 crc kubenswrapper[4790]: I1124 14:15:00.230954 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 14:15:00 crc kubenswrapper[4790]: I1124 14:15:00.363783 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2bd45d25-7193-44bd-9a32-06329ca3fac1-secret-volume\") pod \"collect-profiles-29399895-t6g7z\" (UID: \"2bd45d25-7193-44bd-9a32-06329ca3fac1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-t6g7z" Nov 24 14:15:00 crc kubenswrapper[4790]: I1124 14:15:00.363872 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9plmn\" (UniqueName: \"kubernetes.io/projected/2bd45d25-7193-44bd-9a32-06329ca3fac1-kube-api-access-9plmn\") pod \"collect-profiles-29399895-t6g7z\" (UID: \"2bd45d25-7193-44bd-9a32-06329ca3fac1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-t6g7z" Nov 24 14:15:00 crc kubenswrapper[4790]: I1124 14:15:00.363962 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2bd45d25-7193-44bd-9a32-06329ca3fac1-config-volume\") pod \"collect-profiles-29399895-t6g7z\" (UID: \"2bd45d25-7193-44bd-9a32-06329ca3fac1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-t6g7z" Nov 24 14:15:00 crc kubenswrapper[4790]: I1124 14:15:00.465351 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2bd45d25-7193-44bd-9a32-06329ca3fac1-secret-volume\") pod \"collect-profiles-29399895-t6g7z\" (UID: \"2bd45d25-7193-44bd-9a32-06329ca3fac1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-t6g7z" Nov 24 14:15:00 crc kubenswrapper[4790]: I1124 14:15:00.465455 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9plmn\" (UniqueName: \"kubernetes.io/projected/2bd45d25-7193-44bd-9a32-06329ca3fac1-kube-api-access-9plmn\") pod \"collect-profiles-29399895-t6g7z\" (UID: \"2bd45d25-7193-44bd-9a32-06329ca3fac1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-t6g7z" Nov 24 14:15:00 crc kubenswrapper[4790]: I1124 14:15:00.465526 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2bd45d25-7193-44bd-9a32-06329ca3fac1-config-volume\") pod \"collect-profiles-29399895-t6g7z\" (UID: \"2bd45d25-7193-44bd-9a32-06329ca3fac1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-t6g7z" Nov 24 14:15:00 crc kubenswrapper[4790]: I1124 14:15:00.466577 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2bd45d25-7193-44bd-9a32-06329ca3fac1-config-volume\") pod \"collect-profiles-29399895-t6g7z\" (UID: \"2bd45d25-7193-44bd-9a32-06329ca3fac1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-t6g7z" Nov 24 14:15:00 crc kubenswrapper[4790]: I1124 14:15:00.475419 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2bd45d25-7193-44bd-9a32-06329ca3fac1-secret-volume\") pod \"collect-profiles-29399895-t6g7z\" (UID: \"2bd45d25-7193-44bd-9a32-06329ca3fac1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-t6g7z" Nov 24 14:15:00 crc kubenswrapper[4790]: I1124 14:15:00.491732 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9plmn\" (UniqueName: \"kubernetes.io/projected/2bd45d25-7193-44bd-9a32-06329ca3fac1-kube-api-access-9plmn\") pod \"collect-profiles-29399895-t6g7z\" (UID: \"2bd45d25-7193-44bd-9a32-06329ca3fac1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-t6g7z" Nov 24 14:15:00 crc kubenswrapper[4790]: I1124 14:15:00.532386 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-t6g7z" Nov 24 14:15:00 crc kubenswrapper[4790]: I1124 14:15:00.996312 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399895-t6g7z"] Nov 24 14:15:02 crc kubenswrapper[4790]: I1124 14:15:02.025194 4790 generic.go:334] "Generic (PLEG): container finished" podID="2bd45d25-7193-44bd-9a32-06329ca3fac1" containerID="f836dfded5e80709de981a9ca934d8ffdd578779daa2106a025e0792c6a7b839" exitCode=0 Nov 24 14:15:02 crc kubenswrapper[4790]: I1124 14:15:02.025299 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-t6g7z" event={"ID":"2bd45d25-7193-44bd-9a32-06329ca3fac1","Type":"ContainerDied","Data":"f836dfded5e80709de981a9ca934d8ffdd578779daa2106a025e0792c6a7b839"} Nov 24 14:15:02 crc kubenswrapper[4790]: I1124 14:15:02.025509 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-t6g7z" event={"ID":"2bd45d25-7193-44bd-9a32-06329ca3fac1","Type":"ContainerStarted","Data":"331191b826dab1ea25ca11215c41098f059464dbb2e2d28f09a3cbdf366f36e7"} Nov 24 14:15:03 crc kubenswrapper[4790]: I1124 14:15:03.306842 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-t6g7z" Nov 24 14:15:03 crc kubenswrapper[4790]: I1124 14:15:03.410848 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2bd45d25-7193-44bd-9a32-06329ca3fac1-secret-volume\") pod \"2bd45d25-7193-44bd-9a32-06329ca3fac1\" (UID: \"2bd45d25-7193-44bd-9a32-06329ca3fac1\") " Nov 24 14:15:03 crc kubenswrapper[4790]: I1124 14:15:03.410957 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2bd45d25-7193-44bd-9a32-06329ca3fac1-config-volume\") pod \"2bd45d25-7193-44bd-9a32-06329ca3fac1\" (UID: \"2bd45d25-7193-44bd-9a32-06329ca3fac1\") " Nov 24 14:15:03 crc kubenswrapper[4790]: I1124 14:15:03.410996 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9plmn\" (UniqueName: \"kubernetes.io/projected/2bd45d25-7193-44bd-9a32-06329ca3fac1-kube-api-access-9plmn\") pod \"2bd45d25-7193-44bd-9a32-06329ca3fac1\" (UID: \"2bd45d25-7193-44bd-9a32-06329ca3fac1\") " Nov 24 14:15:03 crc kubenswrapper[4790]: I1124 14:15:03.412474 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bd45d25-7193-44bd-9a32-06329ca3fac1-config-volume" (OuterVolumeSpecName: "config-volume") pod "2bd45d25-7193-44bd-9a32-06329ca3fac1" (UID: "2bd45d25-7193-44bd-9a32-06329ca3fac1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:15:03 crc kubenswrapper[4790]: I1124 14:15:03.417108 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bd45d25-7193-44bd-9a32-06329ca3fac1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2bd45d25-7193-44bd-9a32-06329ca3fac1" (UID: "2bd45d25-7193-44bd-9a32-06329ca3fac1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:15:03 crc kubenswrapper[4790]: I1124 14:15:03.417995 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bd45d25-7193-44bd-9a32-06329ca3fac1-kube-api-access-9plmn" (OuterVolumeSpecName: "kube-api-access-9plmn") pod "2bd45d25-7193-44bd-9a32-06329ca3fac1" (UID: "2bd45d25-7193-44bd-9a32-06329ca3fac1"). InnerVolumeSpecName "kube-api-access-9plmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:15:03 crc kubenswrapper[4790]: I1124 14:15:03.513198 4790 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2bd45d25-7193-44bd-9a32-06329ca3fac1-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 14:15:03 crc kubenswrapper[4790]: I1124 14:15:03.513230 4790 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2bd45d25-7193-44bd-9a32-06329ca3fac1-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 14:15:03 crc kubenswrapper[4790]: I1124 14:15:03.513242 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9plmn\" (UniqueName: \"kubernetes.io/projected/2bd45d25-7193-44bd-9a32-06329ca3fac1-kube-api-access-9plmn\") on node \"crc\" DevicePath \"\"" Nov 24 14:15:04 crc kubenswrapper[4790]: I1124 14:15:04.041037 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-t6g7z" event={"ID":"2bd45d25-7193-44bd-9a32-06329ca3fac1","Type":"ContainerDied","Data":"331191b826dab1ea25ca11215c41098f059464dbb2e2d28f09a3cbdf366f36e7"} Nov 24 14:15:04 crc kubenswrapper[4790]: I1124 14:15:04.041082 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="331191b826dab1ea25ca11215c41098f059464dbb2e2d28f09a3cbdf366f36e7" Nov 24 14:15:04 crc kubenswrapper[4790]: I1124 14:15:04.041089 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-t6g7z" Nov 24 14:15:04 crc kubenswrapper[4790]: I1124 14:15:04.391552 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399850-dhxbs"] Nov 24 14:15:04 crc kubenswrapper[4790]: I1124 14:15:04.395939 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399850-dhxbs"] Nov 24 14:15:06 crc kubenswrapper[4790]: I1124 14:15:06.322659 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ed9d0d9-aa9f-44ec-918a-59231503a4ba" path="/var/lib/kubelet/pods/9ed9d0d9-aa9f-44ec-918a-59231503a4ba/volumes" Nov 24 14:15:07 crc kubenswrapper[4790]: I1124 14:15:07.942795 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pdbm7"] Nov 24 14:15:07 crc kubenswrapper[4790]: E1124 14:15:07.943197 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bd45d25-7193-44bd-9a32-06329ca3fac1" containerName="collect-profiles" Nov 24 14:15:07 crc kubenswrapper[4790]: I1124 14:15:07.943211 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bd45d25-7193-44bd-9a32-06329ca3fac1" containerName="collect-profiles" Nov 24 14:15:07 crc kubenswrapper[4790]: I1124 14:15:07.943360 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bd45d25-7193-44bd-9a32-06329ca3fac1" containerName="collect-profiles" Nov 24 14:15:07 crc kubenswrapper[4790]: I1124 14:15:07.944396 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pdbm7" Nov 24 14:15:07 crc kubenswrapper[4790]: I1124 14:15:07.959350 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pdbm7"] Nov 24 14:15:08 crc kubenswrapper[4790]: I1124 14:15:08.073418 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a415403-8f42-4dd2-9387-30b9c2af07b3-catalog-content\") pod \"community-operators-pdbm7\" (UID: \"1a415403-8f42-4dd2-9387-30b9c2af07b3\") " pod="openshift-marketplace/community-operators-pdbm7" Nov 24 14:15:08 crc kubenswrapper[4790]: I1124 14:15:08.073811 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a415403-8f42-4dd2-9387-30b9c2af07b3-utilities\") pod \"community-operators-pdbm7\" (UID: \"1a415403-8f42-4dd2-9387-30b9c2af07b3\") " pod="openshift-marketplace/community-operators-pdbm7" Nov 24 14:15:08 crc kubenswrapper[4790]: I1124 14:15:08.073898 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4lnb\" (UniqueName: \"kubernetes.io/projected/1a415403-8f42-4dd2-9387-30b9c2af07b3-kube-api-access-q4lnb\") pod \"community-operators-pdbm7\" (UID: \"1a415403-8f42-4dd2-9387-30b9c2af07b3\") " pod="openshift-marketplace/community-operators-pdbm7" Nov 24 14:15:08 crc kubenswrapper[4790]: I1124 14:15:08.175178 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a415403-8f42-4dd2-9387-30b9c2af07b3-catalog-content\") pod \"community-operators-pdbm7\" (UID: \"1a415403-8f42-4dd2-9387-30b9c2af07b3\") " pod="openshift-marketplace/community-operators-pdbm7" Nov 24 14:15:08 crc kubenswrapper[4790]: I1124 14:15:08.175254 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a415403-8f42-4dd2-9387-30b9c2af07b3-utilities\") pod \"community-operators-pdbm7\" (UID: \"1a415403-8f42-4dd2-9387-30b9c2af07b3\") " pod="openshift-marketplace/community-operators-pdbm7" Nov 24 14:15:08 crc kubenswrapper[4790]: I1124 14:15:08.175307 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4lnb\" (UniqueName: \"kubernetes.io/projected/1a415403-8f42-4dd2-9387-30b9c2af07b3-kube-api-access-q4lnb\") pod \"community-operators-pdbm7\" (UID: \"1a415403-8f42-4dd2-9387-30b9c2af07b3\") " pod="openshift-marketplace/community-operators-pdbm7" Nov 24 14:15:08 crc kubenswrapper[4790]: I1124 14:15:08.175780 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a415403-8f42-4dd2-9387-30b9c2af07b3-catalog-content\") pod \"community-operators-pdbm7\" (UID: \"1a415403-8f42-4dd2-9387-30b9c2af07b3\") " pod="openshift-marketplace/community-operators-pdbm7" Nov 24 14:15:08 crc kubenswrapper[4790]: I1124 14:15:08.176108 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a415403-8f42-4dd2-9387-30b9c2af07b3-utilities\") pod \"community-operators-pdbm7\" (UID: \"1a415403-8f42-4dd2-9387-30b9c2af07b3\") " pod="openshift-marketplace/community-operators-pdbm7" Nov 24 14:15:08 crc kubenswrapper[4790]: I1124 14:15:08.199659 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4lnb\" (UniqueName: \"kubernetes.io/projected/1a415403-8f42-4dd2-9387-30b9c2af07b3-kube-api-access-q4lnb\") pod \"community-operators-pdbm7\" (UID: \"1a415403-8f42-4dd2-9387-30b9c2af07b3\") " pod="openshift-marketplace/community-operators-pdbm7" Nov 24 14:15:08 crc kubenswrapper[4790]: I1124 14:15:08.281698 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pdbm7" Nov 24 14:15:08 crc kubenswrapper[4790]: I1124 14:15:08.762733 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pdbm7"] Nov 24 14:15:09 crc kubenswrapper[4790]: I1124 14:15:09.072923 4790 generic.go:334] "Generic (PLEG): container finished" podID="1a415403-8f42-4dd2-9387-30b9c2af07b3" containerID="8d5064d024f41a27cd92a7b60a8d0f8704660ecde1ea290c63f128a03b91c586" exitCode=0 Nov 24 14:15:09 crc kubenswrapper[4790]: I1124 14:15:09.072970 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pdbm7" event={"ID":"1a415403-8f42-4dd2-9387-30b9c2af07b3","Type":"ContainerDied","Data":"8d5064d024f41a27cd92a7b60a8d0f8704660ecde1ea290c63f128a03b91c586"} Nov 24 14:15:09 crc kubenswrapper[4790]: I1124 14:15:09.073000 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pdbm7" event={"ID":"1a415403-8f42-4dd2-9387-30b9c2af07b3","Type":"ContainerStarted","Data":"79b36d8dc97d19c7bec48f96ab057cc114dec4e3c1b2506301f1ed2fb1b73246"} Nov 24 14:15:13 crc kubenswrapper[4790]: I1124 14:15:13.099317 4790 generic.go:334] "Generic (PLEG): container finished" podID="1a415403-8f42-4dd2-9387-30b9c2af07b3" containerID="546b8f3b04dd9052f1c204a7eb7dd6b20cd306ecb9342e506d76e403d6f73bb8" exitCode=0 Nov 24 14:15:13 crc kubenswrapper[4790]: I1124 14:15:13.099382 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pdbm7" event={"ID":"1a415403-8f42-4dd2-9387-30b9c2af07b3","Type":"ContainerDied","Data":"546b8f3b04dd9052f1c204a7eb7dd6b20cd306ecb9342e506d76e403d6f73bb8"} Nov 24 14:15:14 crc kubenswrapper[4790]: I1124 14:15:14.107639 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pdbm7" event={"ID":"1a415403-8f42-4dd2-9387-30b9c2af07b3","Type":"ContainerStarted","Data":"bb4a95744b54bbbdb68dbd8cd4a650b7a66aa6135408708c5cfc9b0014b04512"} Nov 24 14:15:14 crc kubenswrapper[4790]: I1124 14:15:14.123206 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pdbm7" podStartSLOduration=2.5040112900000002 podStartE2EDuration="7.123190364s" podCreationTimestamp="2025-11-24 14:15:07 +0000 UTC" firstStartedPulling="2025-11-24 14:15:09.074727519 +0000 UTC m=+3757.454621181" lastFinishedPulling="2025-11-24 14:15:13.693906593 +0000 UTC m=+3762.073800255" observedRunningTime="2025-11-24 14:15:14.121013102 +0000 UTC m=+3762.500906794" watchObservedRunningTime="2025-11-24 14:15:14.123190364 +0000 UTC m=+3762.503084026" Nov 24 14:15:18 crc kubenswrapper[4790]: I1124 14:15:18.282978 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pdbm7" Nov 24 14:15:18 crc kubenswrapper[4790]: I1124 14:15:18.284184 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pdbm7" Nov 24 14:15:18 crc kubenswrapper[4790]: I1124 14:15:18.332803 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pdbm7" Nov 24 14:15:19 crc kubenswrapper[4790]: I1124 14:15:19.195131 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pdbm7" Nov 24 14:15:19 crc kubenswrapper[4790]: I1124 14:15:19.273707 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pdbm7"] Nov 24 14:15:19 crc kubenswrapper[4790]: I1124 14:15:19.313599 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-97gps"] Nov 24 14:15:19 crc kubenswrapper[4790]: I1124 14:15:19.314012 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-97gps" podUID="80841c5a-bd46-41a2-bd2c-ac8e59cda2f1" containerName="registry-server" containerID="cri-o://1dff8f79eb33c86b294dc2a50b6034dd725baf71597c195ab8e49daa47d7ca87" gracePeriod=2 Nov 24 14:15:19 crc kubenswrapper[4790]: E1124 14:15:19.421098 4790 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod80841c5a_bd46_41a2_bd2c_ac8e59cda2f1.slice/crio-1dff8f79eb33c86b294dc2a50b6034dd725baf71597c195ab8e49daa47d7ca87.scope\": RecentStats: unable to find data in memory cache]" Nov 24 14:15:19 crc kubenswrapper[4790]: I1124 14:15:19.707995 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-97gps" Nov 24 14:15:19 crc kubenswrapper[4790]: I1124 14:15:19.842636 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6vhg\" (UniqueName: \"kubernetes.io/projected/80841c5a-bd46-41a2-bd2c-ac8e59cda2f1-kube-api-access-l6vhg\") pod \"80841c5a-bd46-41a2-bd2c-ac8e59cda2f1\" (UID: \"80841c5a-bd46-41a2-bd2c-ac8e59cda2f1\") " Nov 24 14:15:19 crc kubenswrapper[4790]: I1124 14:15:19.842788 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80841c5a-bd46-41a2-bd2c-ac8e59cda2f1-catalog-content\") pod \"80841c5a-bd46-41a2-bd2c-ac8e59cda2f1\" (UID: \"80841c5a-bd46-41a2-bd2c-ac8e59cda2f1\") " Nov 24 14:15:19 crc kubenswrapper[4790]: I1124 14:15:19.842985 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80841c5a-bd46-41a2-bd2c-ac8e59cda2f1-utilities\") pod \"80841c5a-bd46-41a2-bd2c-ac8e59cda2f1\" (UID: \"80841c5a-bd46-41a2-bd2c-ac8e59cda2f1\") " Nov 24 14:15:19 crc kubenswrapper[4790]: I1124 14:15:19.843683 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80841c5a-bd46-41a2-bd2c-ac8e59cda2f1-utilities" (OuterVolumeSpecName: "utilities") pod "80841c5a-bd46-41a2-bd2c-ac8e59cda2f1" (UID: "80841c5a-bd46-41a2-bd2c-ac8e59cda2f1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:15:19 crc kubenswrapper[4790]: I1124 14:15:19.850091 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80841c5a-bd46-41a2-bd2c-ac8e59cda2f1-kube-api-access-l6vhg" (OuterVolumeSpecName: "kube-api-access-l6vhg") pod "80841c5a-bd46-41a2-bd2c-ac8e59cda2f1" (UID: "80841c5a-bd46-41a2-bd2c-ac8e59cda2f1"). InnerVolumeSpecName "kube-api-access-l6vhg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:15:19 crc kubenswrapper[4790]: I1124 14:15:19.913357 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80841c5a-bd46-41a2-bd2c-ac8e59cda2f1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "80841c5a-bd46-41a2-bd2c-ac8e59cda2f1" (UID: "80841c5a-bd46-41a2-bd2c-ac8e59cda2f1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:15:19 crc kubenswrapper[4790]: I1124 14:15:19.945465 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80841c5a-bd46-41a2-bd2c-ac8e59cda2f1-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:15:19 crc kubenswrapper[4790]: I1124 14:15:19.945513 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80841c5a-bd46-41a2-bd2c-ac8e59cda2f1-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:15:19 crc kubenswrapper[4790]: I1124 14:15:19.945527 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6vhg\" (UniqueName: \"kubernetes.io/projected/80841c5a-bd46-41a2-bd2c-ac8e59cda2f1-kube-api-access-l6vhg\") on node \"crc\" DevicePath \"\"" Nov 24 14:15:20 crc kubenswrapper[4790]: I1124 14:15:20.149015 4790 generic.go:334] "Generic (PLEG): container finished" podID="80841c5a-bd46-41a2-bd2c-ac8e59cda2f1" containerID="1dff8f79eb33c86b294dc2a50b6034dd725baf71597c195ab8e49daa47d7ca87" exitCode=0 Nov 24 14:15:20 crc kubenswrapper[4790]: I1124 14:15:20.149104 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-97gps" Nov 24 14:15:20 crc kubenswrapper[4790]: I1124 14:15:20.149110 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-97gps" event={"ID":"80841c5a-bd46-41a2-bd2c-ac8e59cda2f1","Type":"ContainerDied","Data":"1dff8f79eb33c86b294dc2a50b6034dd725baf71597c195ab8e49daa47d7ca87"} Nov 24 14:15:20 crc kubenswrapper[4790]: I1124 14:15:20.149172 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-97gps" event={"ID":"80841c5a-bd46-41a2-bd2c-ac8e59cda2f1","Type":"ContainerDied","Data":"eb9e5ce20461ff26d6a43d34457b3ad7402be6b66346be22ba26001b80bb2109"} Nov 24 14:15:20 crc kubenswrapper[4790]: I1124 14:15:20.149193 4790 scope.go:117] "RemoveContainer" containerID="1dff8f79eb33c86b294dc2a50b6034dd725baf71597c195ab8e49daa47d7ca87" Nov 24 14:15:20 crc kubenswrapper[4790]: I1124 14:15:20.168273 4790 scope.go:117] "RemoveContainer" containerID="b964fc41bf324f22d7e6dd7aa6bde80943d294254f8a3ddf88ab216e3698d6d8" Nov 24 14:15:20 crc kubenswrapper[4790]: I1124 14:15:20.179833 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-97gps"] Nov 24 14:15:20 crc kubenswrapper[4790]: I1124 14:15:20.193520 4790 scope.go:117] "RemoveContainer" containerID="987fbbf926b7439356d34bcc5d313aad754bafd7f54b5938a9fd88ff0d7a5e13" Nov 24 14:15:20 crc kubenswrapper[4790]: I1124 14:15:20.202224 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-97gps"] Nov 24 14:15:20 crc kubenswrapper[4790]: I1124 14:15:20.235624 4790 scope.go:117] "RemoveContainer" containerID="1dff8f79eb33c86b294dc2a50b6034dd725baf71597c195ab8e49daa47d7ca87" Nov 24 14:15:20 crc kubenswrapper[4790]: E1124 14:15:20.236160 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1dff8f79eb33c86b294dc2a50b6034dd725baf71597c195ab8e49daa47d7ca87\": container with ID starting with 1dff8f79eb33c86b294dc2a50b6034dd725baf71597c195ab8e49daa47d7ca87 not found: ID does not exist" containerID="1dff8f79eb33c86b294dc2a50b6034dd725baf71597c195ab8e49daa47d7ca87" Nov 24 14:15:20 crc kubenswrapper[4790]: I1124 14:15:20.236199 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dff8f79eb33c86b294dc2a50b6034dd725baf71597c195ab8e49daa47d7ca87"} err="failed to get container status \"1dff8f79eb33c86b294dc2a50b6034dd725baf71597c195ab8e49daa47d7ca87\": rpc error: code = NotFound desc = could not find container \"1dff8f79eb33c86b294dc2a50b6034dd725baf71597c195ab8e49daa47d7ca87\": container with ID starting with 1dff8f79eb33c86b294dc2a50b6034dd725baf71597c195ab8e49daa47d7ca87 not found: ID does not exist" Nov 24 14:15:20 crc kubenswrapper[4790]: I1124 14:15:20.236224 4790 scope.go:117] "RemoveContainer" containerID="b964fc41bf324f22d7e6dd7aa6bde80943d294254f8a3ddf88ab216e3698d6d8" Nov 24 14:15:20 crc kubenswrapper[4790]: E1124 14:15:20.236747 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b964fc41bf324f22d7e6dd7aa6bde80943d294254f8a3ddf88ab216e3698d6d8\": container with ID starting with b964fc41bf324f22d7e6dd7aa6bde80943d294254f8a3ddf88ab216e3698d6d8 not found: ID does not exist" containerID="b964fc41bf324f22d7e6dd7aa6bde80943d294254f8a3ddf88ab216e3698d6d8" Nov 24 14:15:20 crc kubenswrapper[4790]: I1124 14:15:20.236778 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b964fc41bf324f22d7e6dd7aa6bde80943d294254f8a3ddf88ab216e3698d6d8"} err="failed to get container status \"b964fc41bf324f22d7e6dd7aa6bde80943d294254f8a3ddf88ab216e3698d6d8\": rpc error: code = NotFound desc = could not find container \"b964fc41bf324f22d7e6dd7aa6bde80943d294254f8a3ddf88ab216e3698d6d8\": container with ID starting with b964fc41bf324f22d7e6dd7aa6bde80943d294254f8a3ddf88ab216e3698d6d8 not found: ID does not exist" Nov 24 14:15:20 crc kubenswrapper[4790]: I1124 14:15:20.236797 4790 scope.go:117] "RemoveContainer" containerID="987fbbf926b7439356d34bcc5d313aad754bafd7f54b5938a9fd88ff0d7a5e13" Nov 24 14:15:20 crc kubenswrapper[4790]: E1124 14:15:20.237335 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"987fbbf926b7439356d34bcc5d313aad754bafd7f54b5938a9fd88ff0d7a5e13\": container with ID starting with 987fbbf926b7439356d34bcc5d313aad754bafd7f54b5938a9fd88ff0d7a5e13 not found: ID does not exist" containerID="987fbbf926b7439356d34bcc5d313aad754bafd7f54b5938a9fd88ff0d7a5e13" Nov 24 14:15:20 crc kubenswrapper[4790]: I1124 14:15:20.237392 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"987fbbf926b7439356d34bcc5d313aad754bafd7f54b5938a9fd88ff0d7a5e13"} err="failed to get container status \"987fbbf926b7439356d34bcc5d313aad754bafd7f54b5938a9fd88ff0d7a5e13\": rpc error: code = NotFound desc = could not find container \"987fbbf926b7439356d34bcc5d313aad754bafd7f54b5938a9fd88ff0d7a5e13\": container with ID starting with 987fbbf926b7439356d34bcc5d313aad754bafd7f54b5938a9fd88ff0d7a5e13 not found: ID does not exist" Nov 24 14:15:20 crc kubenswrapper[4790]: I1124 14:15:20.323897 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80841c5a-bd46-41a2-bd2c-ac8e59cda2f1" path="/var/lib/kubelet/pods/80841c5a-bd46-41a2-bd2c-ac8e59cda2f1/volumes" Nov 24 14:15:43 crc kubenswrapper[4790]: I1124 14:15:43.938349 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:15:43 crc kubenswrapper[4790]: I1124 14:15:43.938927 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:15:46 crc kubenswrapper[4790]: I1124 14:15:46.207383 4790 scope.go:117] "RemoveContainer" containerID="b53df9baaaa61a1ffd04ee843ecc8ca5fc1c09e5b75dae56c5c6b1bb8b96c4b5" Nov 24 14:16:13 crc kubenswrapper[4790]: I1124 14:16:13.938594 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:16:13 crc kubenswrapper[4790]: I1124 14:16:13.939122 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:16:43 crc kubenswrapper[4790]: I1124 14:16:43.938159 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:16:43 crc kubenswrapper[4790]: I1124 14:16:43.938684 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:16:43 crc kubenswrapper[4790]: I1124 14:16:43.938726 4790 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 14:16:43 crc kubenswrapper[4790]: I1124 14:16:43.939318 4790 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9d160b5f0ea013bde38ba30b4fd4ea41f13ba76ae92c0901dbd7ffee1185aac0"} pod="openshift-machine-config-operator/machine-config-daemon-xz49t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:16:43 crc kubenswrapper[4790]: I1124 14:16:43.939370 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" containerID="cri-o://9d160b5f0ea013bde38ba30b4fd4ea41f13ba76ae92c0901dbd7ffee1185aac0" gracePeriod=600 Nov 24 14:16:44 crc kubenswrapper[4790]: E1124 14:16:44.724332 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:16:44 crc kubenswrapper[4790]: I1124 14:16:44.739403 4790 generic.go:334] "Generic (PLEG): container finished" podID="0d73b133-48f0-455f-8f6a-742e633f631a" containerID="9d160b5f0ea013bde38ba30b4fd4ea41f13ba76ae92c0901dbd7ffee1185aac0" exitCode=0 Nov 24 14:16:44 crc kubenswrapper[4790]: I1124 14:16:44.739475 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerDied","Data":"9d160b5f0ea013bde38ba30b4fd4ea41f13ba76ae92c0901dbd7ffee1185aac0"} Nov 24 14:16:44 crc kubenswrapper[4790]: I1124 14:16:44.739555 4790 scope.go:117] "RemoveContainer" containerID="d126dc358baf6e31345e3b42df23e22cb22fe7a763b1b6d8958121963d242656" Nov 24 14:16:44 crc kubenswrapper[4790]: I1124 14:16:44.740158 4790 scope.go:117] "RemoveContainer" containerID="9d160b5f0ea013bde38ba30b4fd4ea41f13ba76ae92c0901dbd7ffee1185aac0" Nov 24 14:16:44 crc kubenswrapper[4790]: E1124 14:16:44.740465 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:17:00 crc kubenswrapper[4790]: I1124 14:17:00.314773 4790 scope.go:117] "RemoveContainer" containerID="9d160b5f0ea013bde38ba30b4fd4ea41f13ba76ae92c0901dbd7ffee1185aac0" Nov 24 14:17:00 crc kubenswrapper[4790]: E1124 14:17:00.315521 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:17:12 crc kubenswrapper[4790]: I1124 14:17:12.320675 4790 scope.go:117] "RemoveContainer" containerID="9d160b5f0ea013bde38ba30b4fd4ea41f13ba76ae92c0901dbd7ffee1185aac0" Nov 24 14:17:12 crc kubenswrapper[4790]: E1124 14:17:12.321576 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:17:24 crc kubenswrapper[4790]: I1124 14:17:24.314798 4790 scope.go:117] "RemoveContainer" containerID="9d160b5f0ea013bde38ba30b4fd4ea41f13ba76ae92c0901dbd7ffee1185aac0" Nov 24 14:17:24 crc kubenswrapper[4790]: E1124 14:17:24.315477 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:17:35 crc kubenswrapper[4790]: I1124 14:17:35.315112 4790 scope.go:117] "RemoveContainer" containerID="9d160b5f0ea013bde38ba30b4fd4ea41f13ba76ae92c0901dbd7ffee1185aac0" Nov 24 14:17:35 crc kubenswrapper[4790]: E1124 14:17:35.315966 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:17:47 crc kubenswrapper[4790]: I1124 14:17:47.314991 4790 scope.go:117] "RemoveContainer" containerID="9d160b5f0ea013bde38ba30b4fd4ea41f13ba76ae92c0901dbd7ffee1185aac0" Nov 24 14:17:47 crc kubenswrapper[4790]: E1124 14:17:47.315579 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:18:02 crc kubenswrapper[4790]: I1124 14:18:02.319552 4790 scope.go:117] "RemoveContainer" containerID="9d160b5f0ea013bde38ba30b4fd4ea41f13ba76ae92c0901dbd7ffee1185aac0" Nov 24 14:18:02 crc kubenswrapper[4790]: E1124 14:18:02.320265 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:18:13 crc kubenswrapper[4790]: I1124 14:18:13.315455 4790 scope.go:117] "RemoveContainer" containerID="9d160b5f0ea013bde38ba30b4fd4ea41f13ba76ae92c0901dbd7ffee1185aac0" Nov 24 14:18:13 crc kubenswrapper[4790]: E1124 14:18:13.317649 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:18:27 crc kubenswrapper[4790]: I1124 14:18:27.314170 4790 scope.go:117] "RemoveContainer" containerID="9d160b5f0ea013bde38ba30b4fd4ea41f13ba76ae92c0901dbd7ffee1185aac0" Nov 24 14:18:27 crc kubenswrapper[4790]: E1124 14:18:27.315017 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:18:30 crc kubenswrapper[4790]: I1124 14:18:30.949445 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-97ffx"] Nov 24 14:18:30 crc kubenswrapper[4790]: E1124 14:18:30.950135 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80841c5a-bd46-41a2-bd2c-ac8e59cda2f1" containerName="registry-server" Nov 24 14:18:30 crc kubenswrapper[4790]: I1124 14:18:30.950166 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="80841c5a-bd46-41a2-bd2c-ac8e59cda2f1" containerName="registry-server" Nov 24 14:18:30 crc kubenswrapper[4790]: E1124 14:18:30.950197 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80841c5a-bd46-41a2-bd2c-ac8e59cda2f1" containerName="extract-utilities" Nov 24 14:18:30 crc kubenswrapper[4790]: I1124 14:18:30.950206 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="80841c5a-bd46-41a2-bd2c-ac8e59cda2f1" containerName="extract-utilities" Nov 24 14:18:30 crc kubenswrapper[4790]: E1124 14:18:30.950218 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80841c5a-bd46-41a2-bd2c-ac8e59cda2f1" containerName="extract-content" Nov 24 14:18:30 crc kubenswrapper[4790]: I1124 14:18:30.950226 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="80841c5a-bd46-41a2-bd2c-ac8e59cda2f1" containerName="extract-content" Nov 24 14:18:30 crc kubenswrapper[4790]: I1124 14:18:30.950451 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="80841c5a-bd46-41a2-bd2c-ac8e59cda2f1" containerName="registry-server" Nov 24 14:18:30 crc kubenswrapper[4790]: I1124 14:18:30.951700 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-97ffx" Nov 24 14:18:30 crc kubenswrapper[4790]: I1124 14:18:30.964304 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-97ffx"] Nov 24 14:18:31 crc kubenswrapper[4790]: I1124 14:18:31.125486 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d703ae8e-4aaf-4c29-b1fe-27212fec7239-utilities\") pod \"certified-operators-97ffx\" (UID: \"d703ae8e-4aaf-4c29-b1fe-27212fec7239\") " pod="openshift-marketplace/certified-operators-97ffx" Nov 24 14:18:31 crc kubenswrapper[4790]: I1124 14:18:31.125543 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d703ae8e-4aaf-4c29-b1fe-27212fec7239-catalog-content\") pod \"certified-operators-97ffx\" (UID: \"d703ae8e-4aaf-4c29-b1fe-27212fec7239\") " pod="openshift-marketplace/certified-operators-97ffx" Nov 24 14:18:31 crc kubenswrapper[4790]: I1124 14:18:31.125635 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zdcz\" (UniqueName: \"kubernetes.io/projected/d703ae8e-4aaf-4c29-b1fe-27212fec7239-kube-api-access-8zdcz\") pod \"certified-operators-97ffx\" (UID: \"d703ae8e-4aaf-4c29-b1fe-27212fec7239\") " pod="openshift-marketplace/certified-operators-97ffx" Nov 24 14:18:31 crc kubenswrapper[4790]: I1124 14:18:31.227048 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zdcz\" (UniqueName: \"kubernetes.io/projected/d703ae8e-4aaf-4c29-b1fe-27212fec7239-kube-api-access-8zdcz\") pod \"certified-operators-97ffx\" (UID: \"d703ae8e-4aaf-4c29-b1fe-27212fec7239\") " pod="openshift-marketplace/certified-operators-97ffx" Nov 24 14:18:31 crc kubenswrapper[4790]: I1124 14:18:31.227112 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d703ae8e-4aaf-4c29-b1fe-27212fec7239-utilities\") pod \"certified-operators-97ffx\" (UID: \"d703ae8e-4aaf-4c29-b1fe-27212fec7239\") " pod="openshift-marketplace/certified-operators-97ffx" Nov 24 14:18:31 crc kubenswrapper[4790]: I1124 14:18:31.227147 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d703ae8e-4aaf-4c29-b1fe-27212fec7239-catalog-content\") pod \"certified-operators-97ffx\" (UID: \"d703ae8e-4aaf-4c29-b1fe-27212fec7239\") " pod="openshift-marketplace/certified-operators-97ffx" Nov 24 14:18:31 crc kubenswrapper[4790]: I1124 14:18:31.228154 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d703ae8e-4aaf-4c29-b1fe-27212fec7239-catalog-content\") pod \"certified-operators-97ffx\" (UID: \"d703ae8e-4aaf-4c29-b1fe-27212fec7239\") " pod="openshift-marketplace/certified-operators-97ffx" Nov 24 14:18:31 crc kubenswrapper[4790]: I1124 14:18:31.228596 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d703ae8e-4aaf-4c29-b1fe-27212fec7239-utilities\") pod \"certified-operators-97ffx\" (UID: \"d703ae8e-4aaf-4c29-b1fe-27212fec7239\") " pod="openshift-marketplace/certified-operators-97ffx" Nov 24 14:18:31 crc kubenswrapper[4790]: I1124 14:18:31.252901 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zdcz\" (UniqueName: \"kubernetes.io/projected/d703ae8e-4aaf-4c29-b1fe-27212fec7239-kube-api-access-8zdcz\") pod \"certified-operators-97ffx\" (UID: \"d703ae8e-4aaf-4c29-b1fe-27212fec7239\") " pod="openshift-marketplace/certified-operators-97ffx" Nov 24 14:18:31 crc kubenswrapper[4790]: I1124 14:18:31.271648 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-97ffx" Nov 24 14:18:31 crc kubenswrapper[4790]: I1124 14:18:31.828042 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-97ffx"] Nov 24 14:18:32 crc kubenswrapper[4790]: I1124 14:18:32.476362 4790 generic.go:334] "Generic (PLEG): container finished" podID="d703ae8e-4aaf-4c29-b1fe-27212fec7239" containerID="3449d30fdc10cb4f64fa8812a39873c564645acc2060132db95cf35641c46f4d" exitCode=0 Nov 24 14:18:32 crc kubenswrapper[4790]: I1124 14:18:32.476694 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-97ffx" event={"ID":"d703ae8e-4aaf-4c29-b1fe-27212fec7239","Type":"ContainerDied","Data":"3449d30fdc10cb4f64fa8812a39873c564645acc2060132db95cf35641c46f4d"} Nov 24 14:18:32 crc kubenswrapper[4790]: I1124 14:18:32.476724 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-97ffx" event={"ID":"d703ae8e-4aaf-4c29-b1fe-27212fec7239","Type":"ContainerStarted","Data":"0a1e577b2933cef8b5b3bff6dc0fa6372876620c4dcb54c0f203174f4c01afe4"} Nov 24 14:18:32 crc kubenswrapper[4790]: I1124 14:18:32.478553 4790 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 14:18:33 crc kubenswrapper[4790]: I1124 14:18:33.485171 4790 generic.go:334] "Generic (PLEG): container finished" podID="d703ae8e-4aaf-4c29-b1fe-27212fec7239" containerID="928a171abecbc1f843df72e9a5dd9bff1b0c3914ad93d6c9f5b51f00ecaa9d73" exitCode=0 Nov 24 14:18:33 crc kubenswrapper[4790]: I1124 14:18:33.485216 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-97ffx" event={"ID":"d703ae8e-4aaf-4c29-b1fe-27212fec7239","Type":"ContainerDied","Data":"928a171abecbc1f843df72e9a5dd9bff1b0c3914ad93d6c9f5b51f00ecaa9d73"} Nov 24 14:18:34 crc kubenswrapper[4790]: I1124 14:18:34.493130 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-97ffx" event={"ID":"d703ae8e-4aaf-4c29-b1fe-27212fec7239","Type":"ContainerStarted","Data":"aecdb66ea2bea94dc8bdb3c553f8eb7a8c9d8a511e7b71aeeff121459f34fff5"} Nov 24 14:18:34 crc kubenswrapper[4790]: I1124 14:18:34.508357 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-97ffx" podStartSLOduration=3.080220358 podStartE2EDuration="4.508341958s" podCreationTimestamp="2025-11-24 14:18:30 +0000 UTC" firstStartedPulling="2025-11-24 14:18:32.478311164 +0000 UTC m=+3960.858204826" lastFinishedPulling="2025-11-24 14:18:33.906432764 +0000 UTC m=+3962.286326426" observedRunningTime="2025-11-24 14:18:34.5070134 +0000 UTC m=+3962.886907072" watchObservedRunningTime="2025-11-24 14:18:34.508341958 +0000 UTC m=+3962.888235620" Nov 24 14:18:39 crc kubenswrapper[4790]: I1124 14:18:39.315197 4790 scope.go:117] "RemoveContainer" containerID="9d160b5f0ea013bde38ba30b4fd4ea41f13ba76ae92c0901dbd7ffee1185aac0" Nov 24 14:18:39 crc kubenswrapper[4790]: E1124 14:18:39.315789 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:18:41 crc kubenswrapper[4790]: I1124 14:18:41.272002 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-97ffx" Nov 24 14:18:41 crc kubenswrapper[4790]: I1124 14:18:41.272299 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-97ffx" Nov 24 14:18:41 crc kubenswrapper[4790]: I1124 14:18:41.312170 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-97ffx" Nov 24 14:18:41 crc kubenswrapper[4790]: I1124 14:18:41.607261 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-97ffx" Nov 24 14:18:41 crc kubenswrapper[4790]: I1124 14:18:41.650958 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-97ffx"] Nov 24 14:18:43 crc kubenswrapper[4790]: I1124 14:18:43.568670 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-97ffx" podUID="d703ae8e-4aaf-4c29-b1fe-27212fec7239" containerName="registry-server" containerID="cri-o://aecdb66ea2bea94dc8bdb3c553f8eb7a8c9d8a511e7b71aeeff121459f34fff5" gracePeriod=2 Nov 24 14:18:45 crc kubenswrapper[4790]: I1124 14:18:45.453008 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-97ffx" Nov 24 14:18:45 crc kubenswrapper[4790]: I1124 14:18:45.582870 4790 generic.go:334] "Generic (PLEG): container finished" podID="d703ae8e-4aaf-4c29-b1fe-27212fec7239" containerID="aecdb66ea2bea94dc8bdb3c553f8eb7a8c9d8a511e7b71aeeff121459f34fff5" exitCode=0 Nov 24 14:18:45 crc kubenswrapper[4790]: I1124 14:18:45.582943 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-97ffx" event={"ID":"d703ae8e-4aaf-4c29-b1fe-27212fec7239","Type":"ContainerDied","Data":"aecdb66ea2bea94dc8bdb3c553f8eb7a8c9d8a511e7b71aeeff121459f34fff5"} Nov 24 14:18:45 crc kubenswrapper[4790]: I1124 14:18:45.582969 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-97ffx" Nov 24 14:18:45 crc kubenswrapper[4790]: I1124 14:18:45.582977 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-97ffx" event={"ID":"d703ae8e-4aaf-4c29-b1fe-27212fec7239","Type":"ContainerDied","Data":"0a1e577b2933cef8b5b3bff6dc0fa6372876620c4dcb54c0f203174f4c01afe4"} Nov 24 14:18:45 crc kubenswrapper[4790]: I1124 14:18:45.582995 4790 scope.go:117] "RemoveContainer" containerID="aecdb66ea2bea94dc8bdb3c553f8eb7a8c9d8a511e7b71aeeff121459f34fff5" Nov 24 14:18:45 crc kubenswrapper[4790]: I1124 14:18:45.604560 4790 scope.go:117] "RemoveContainer" containerID="928a171abecbc1f843df72e9a5dd9bff1b0c3914ad93d6c9f5b51f00ecaa9d73" Nov 24 14:18:45 crc kubenswrapper[4790]: I1124 14:18:45.622092 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d703ae8e-4aaf-4c29-b1fe-27212fec7239-catalog-content\") pod \"d703ae8e-4aaf-4c29-b1fe-27212fec7239\" (UID: \"d703ae8e-4aaf-4c29-b1fe-27212fec7239\") " Nov 24 14:18:45 crc kubenswrapper[4790]: I1124 14:18:45.622222 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d703ae8e-4aaf-4c29-b1fe-27212fec7239-utilities\") pod \"d703ae8e-4aaf-4c29-b1fe-27212fec7239\" (UID: \"d703ae8e-4aaf-4c29-b1fe-27212fec7239\") " Nov 24 14:18:45 crc kubenswrapper[4790]: I1124 14:18:45.622321 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zdcz\" (UniqueName: \"kubernetes.io/projected/d703ae8e-4aaf-4c29-b1fe-27212fec7239-kube-api-access-8zdcz\") pod \"d703ae8e-4aaf-4c29-b1fe-27212fec7239\" (UID: \"d703ae8e-4aaf-4c29-b1fe-27212fec7239\") " Nov 24 14:18:45 crc kubenswrapper[4790]: I1124 14:18:45.623262 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d703ae8e-4aaf-4c29-b1fe-27212fec7239-utilities" (OuterVolumeSpecName: "utilities") pod "d703ae8e-4aaf-4c29-b1fe-27212fec7239" (UID: "d703ae8e-4aaf-4c29-b1fe-27212fec7239"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:18:45 crc kubenswrapper[4790]: I1124 14:18:45.627580 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d703ae8e-4aaf-4c29-b1fe-27212fec7239-kube-api-access-8zdcz" (OuterVolumeSpecName: "kube-api-access-8zdcz") pod "d703ae8e-4aaf-4c29-b1fe-27212fec7239" (UID: "d703ae8e-4aaf-4c29-b1fe-27212fec7239"). InnerVolumeSpecName "kube-api-access-8zdcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:18:45 crc kubenswrapper[4790]: I1124 14:18:45.646322 4790 scope.go:117] "RemoveContainer" containerID="3449d30fdc10cb4f64fa8812a39873c564645acc2060132db95cf35641c46f4d" Nov 24 14:18:45 crc kubenswrapper[4790]: I1124 14:18:45.677795 4790 scope.go:117] "RemoveContainer" containerID="aecdb66ea2bea94dc8bdb3c553f8eb7a8c9d8a511e7b71aeeff121459f34fff5" Nov 24 14:18:45 crc kubenswrapper[4790]: E1124 14:18:45.678469 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aecdb66ea2bea94dc8bdb3c553f8eb7a8c9d8a511e7b71aeeff121459f34fff5\": container with ID starting with aecdb66ea2bea94dc8bdb3c553f8eb7a8c9d8a511e7b71aeeff121459f34fff5 not found: ID does not exist" containerID="aecdb66ea2bea94dc8bdb3c553f8eb7a8c9d8a511e7b71aeeff121459f34fff5" Nov 24 14:18:45 crc kubenswrapper[4790]: I1124 14:18:45.678545 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aecdb66ea2bea94dc8bdb3c553f8eb7a8c9d8a511e7b71aeeff121459f34fff5"} err="failed to get container status \"aecdb66ea2bea94dc8bdb3c553f8eb7a8c9d8a511e7b71aeeff121459f34fff5\": rpc error: code = NotFound desc = could not find container \"aecdb66ea2bea94dc8bdb3c553f8eb7a8c9d8a511e7b71aeeff121459f34fff5\": container with ID starting with aecdb66ea2bea94dc8bdb3c553f8eb7a8c9d8a511e7b71aeeff121459f34fff5 not found: ID does not exist" Nov 24 14:18:45 crc kubenswrapper[4790]: I1124 14:18:45.678579 4790 scope.go:117] "RemoveContainer" containerID="928a171abecbc1f843df72e9a5dd9bff1b0c3914ad93d6c9f5b51f00ecaa9d73" Nov 24 14:18:45 crc kubenswrapper[4790]: E1124 14:18:45.679184 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"928a171abecbc1f843df72e9a5dd9bff1b0c3914ad93d6c9f5b51f00ecaa9d73\": container with ID starting with 928a171abecbc1f843df72e9a5dd9bff1b0c3914ad93d6c9f5b51f00ecaa9d73 not found: ID does not exist" containerID="928a171abecbc1f843df72e9a5dd9bff1b0c3914ad93d6c9f5b51f00ecaa9d73" Nov 24 14:18:45 crc kubenswrapper[4790]: I1124 14:18:45.679232 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"928a171abecbc1f843df72e9a5dd9bff1b0c3914ad93d6c9f5b51f00ecaa9d73"} err="failed to get container status \"928a171abecbc1f843df72e9a5dd9bff1b0c3914ad93d6c9f5b51f00ecaa9d73\": rpc error: code = NotFound desc = could not find container \"928a171abecbc1f843df72e9a5dd9bff1b0c3914ad93d6c9f5b51f00ecaa9d73\": container with ID starting with 928a171abecbc1f843df72e9a5dd9bff1b0c3914ad93d6c9f5b51f00ecaa9d73 not found: ID does not exist" Nov 24 14:18:45 crc kubenswrapper[4790]: I1124 14:18:45.679267 4790 scope.go:117] "RemoveContainer" containerID="3449d30fdc10cb4f64fa8812a39873c564645acc2060132db95cf35641c46f4d" Nov 24 14:18:45 crc kubenswrapper[4790]: E1124 14:18:45.679623 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3449d30fdc10cb4f64fa8812a39873c564645acc2060132db95cf35641c46f4d\": container with ID starting with 3449d30fdc10cb4f64fa8812a39873c564645acc2060132db95cf35641c46f4d not found: ID does not exist" containerID="3449d30fdc10cb4f64fa8812a39873c564645acc2060132db95cf35641c46f4d" Nov 24 14:18:45 crc kubenswrapper[4790]: I1124 14:18:45.679657 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3449d30fdc10cb4f64fa8812a39873c564645acc2060132db95cf35641c46f4d"} err="failed to get container status \"3449d30fdc10cb4f64fa8812a39873c564645acc2060132db95cf35641c46f4d\": rpc error: code = NotFound desc = could not find container \"3449d30fdc10cb4f64fa8812a39873c564645acc2060132db95cf35641c46f4d\": container with ID starting with 3449d30fdc10cb4f64fa8812a39873c564645acc2060132db95cf35641c46f4d not found: ID does not exist" Nov 24 14:18:45 crc kubenswrapper[4790]: I1124 14:18:45.698111 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d703ae8e-4aaf-4c29-b1fe-27212fec7239-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d703ae8e-4aaf-4c29-b1fe-27212fec7239" (UID: "d703ae8e-4aaf-4c29-b1fe-27212fec7239"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:18:45 crc kubenswrapper[4790]: I1124 14:18:45.724641 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d703ae8e-4aaf-4c29-b1fe-27212fec7239-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:18:45 crc kubenswrapper[4790]: I1124 14:18:45.724693 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d703ae8e-4aaf-4c29-b1fe-27212fec7239-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:18:45 crc kubenswrapper[4790]: I1124 14:18:45.724707 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zdcz\" (UniqueName: \"kubernetes.io/projected/d703ae8e-4aaf-4c29-b1fe-27212fec7239-kube-api-access-8zdcz\") on node \"crc\" DevicePath \"\"" Nov 24 14:18:45 crc kubenswrapper[4790]: I1124 14:18:45.909572 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-97ffx"] Nov 24 14:18:45 crc kubenswrapper[4790]: I1124 14:18:45.914046 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-97ffx"] Nov 24 14:18:46 crc kubenswrapper[4790]: I1124 14:18:46.322647 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d703ae8e-4aaf-4c29-b1fe-27212fec7239" path="/var/lib/kubelet/pods/d703ae8e-4aaf-4c29-b1fe-27212fec7239/volumes" Nov 24 14:18:52 crc kubenswrapper[4790]: I1124 14:18:52.319780 4790 scope.go:117] "RemoveContainer" containerID="9d160b5f0ea013bde38ba30b4fd4ea41f13ba76ae92c0901dbd7ffee1185aac0" Nov 24 14:18:52 crc kubenswrapper[4790]: E1124 14:18:52.320362 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:19:04 crc kubenswrapper[4790]: I1124 14:19:04.315072 4790 scope.go:117] "RemoveContainer" containerID="9d160b5f0ea013bde38ba30b4fd4ea41f13ba76ae92c0901dbd7ffee1185aac0" Nov 24 14:19:04 crc kubenswrapper[4790]: E1124 14:19:04.315866 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:19:15 crc kubenswrapper[4790]: I1124 14:19:15.314503 4790 scope.go:117] "RemoveContainer" containerID="9d160b5f0ea013bde38ba30b4fd4ea41f13ba76ae92c0901dbd7ffee1185aac0" Nov 24 14:19:15 crc kubenswrapper[4790]: E1124 14:19:15.314973 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:19:26 crc kubenswrapper[4790]: I1124 14:19:26.314404 4790 scope.go:117] "RemoveContainer" containerID="9d160b5f0ea013bde38ba30b4fd4ea41f13ba76ae92c0901dbd7ffee1185aac0" Nov 24 14:19:26 crc kubenswrapper[4790]: E1124 14:19:26.315105 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:19:40 crc kubenswrapper[4790]: I1124 14:19:40.315036 4790 scope.go:117] "RemoveContainer" containerID="9d160b5f0ea013bde38ba30b4fd4ea41f13ba76ae92c0901dbd7ffee1185aac0" Nov 24 14:19:40 crc kubenswrapper[4790]: E1124 14:19:40.315777 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:19:55 crc kubenswrapper[4790]: I1124 14:19:55.315382 4790 scope.go:117] "RemoveContainer" containerID="9d160b5f0ea013bde38ba30b4fd4ea41f13ba76ae92c0901dbd7ffee1185aac0" Nov 24 14:19:55 crc kubenswrapper[4790]: E1124 14:19:55.316585 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:20:07 crc kubenswrapper[4790]: I1124 14:20:07.314773 4790 scope.go:117] "RemoveContainer" containerID="9d160b5f0ea013bde38ba30b4fd4ea41f13ba76ae92c0901dbd7ffee1185aac0" Nov 24 14:20:07 crc kubenswrapper[4790]: E1124 14:20:07.316040 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:20:22 crc kubenswrapper[4790]: I1124 14:20:22.317635 4790 scope.go:117] "RemoveContainer" containerID="9d160b5f0ea013bde38ba30b4fd4ea41f13ba76ae92c0901dbd7ffee1185aac0" Nov 24 14:20:22 crc kubenswrapper[4790]: E1124 14:20:22.318329 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:20:34 crc kubenswrapper[4790]: I1124 14:20:34.314248 4790 scope.go:117] "RemoveContainer" containerID="9d160b5f0ea013bde38ba30b4fd4ea41f13ba76ae92c0901dbd7ffee1185aac0" Nov 24 14:20:34 crc kubenswrapper[4790]: E1124 14:20:34.315008 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:20:49 crc kubenswrapper[4790]: I1124 14:20:49.315096 4790 scope.go:117] "RemoveContainer" containerID="9d160b5f0ea013bde38ba30b4fd4ea41f13ba76ae92c0901dbd7ffee1185aac0" Nov 24 14:20:49 crc kubenswrapper[4790]: E1124 14:20:49.316352 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:21:04 crc kubenswrapper[4790]: I1124 14:21:04.314239 4790 scope.go:117] "RemoveContainer" containerID="9d160b5f0ea013bde38ba30b4fd4ea41f13ba76ae92c0901dbd7ffee1185aac0" Nov 24 14:21:04 crc kubenswrapper[4790]: E1124 14:21:04.314936 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:21:16 crc kubenswrapper[4790]: I1124 14:21:16.314831 4790 scope.go:117] "RemoveContainer" containerID="9d160b5f0ea013bde38ba30b4fd4ea41f13ba76ae92c0901dbd7ffee1185aac0" Nov 24 14:21:16 crc kubenswrapper[4790]: E1124 14:21:16.317341 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:21:29 crc kubenswrapper[4790]: I1124 14:21:29.314025 4790 scope.go:117] "RemoveContainer" containerID="9d160b5f0ea013bde38ba30b4fd4ea41f13ba76ae92c0901dbd7ffee1185aac0" Nov 24 14:21:29 crc kubenswrapper[4790]: E1124 14:21:29.314709 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:21:43 crc kubenswrapper[4790]: I1124 14:21:43.314399 4790 scope.go:117] "RemoveContainer" containerID="9d160b5f0ea013bde38ba30b4fd4ea41f13ba76ae92c0901dbd7ffee1185aac0" Nov 24 14:21:43 crc kubenswrapper[4790]: E1124 14:21:43.315144 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:21:54 crc kubenswrapper[4790]: I1124 14:21:54.314026 4790 scope.go:117] "RemoveContainer" containerID="9d160b5f0ea013bde38ba30b4fd4ea41f13ba76ae92c0901dbd7ffee1185aac0" Nov 24 14:21:54 crc kubenswrapper[4790]: I1124 14:21:54.917577 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"b25648b43123ad262f034d5f38126a0ec7bbda6c1db0e5dea770deb92fb3c707"} Nov 24 14:23:03 crc kubenswrapper[4790]: I1124 14:23:03.757453 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kzd9f"] Nov 24 14:23:03 crc kubenswrapper[4790]: E1124 14:23:03.758221 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d703ae8e-4aaf-4c29-b1fe-27212fec7239" containerName="extract-content" Nov 24 14:23:03 crc kubenswrapper[4790]: I1124 14:23:03.758234 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="d703ae8e-4aaf-4c29-b1fe-27212fec7239" containerName="extract-content" Nov 24 14:23:03 crc kubenswrapper[4790]: E1124 14:23:03.758250 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d703ae8e-4aaf-4c29-b1fe-27212fec7239" containerName="extract-utilities" Nov 24 14:23:03 crc kubenswrapper[4790]: I1124 14:23:03.758257 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="d703ae8e-4aaf-4c29-b1fe-27212fec7239" containerName="extract-utilities" Nov 24 14:23:03 crc kubenswrapper[4790]: E1124 14:23:03.758269 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d703ae8e-4aaf-4c29-b1fe-27212fec7239" containerName="registry-server" Nov 24 14:23:03 crc kubenswrapper[4790]: I1124 14:23:03.758275 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="d703ae8e-4aaf-4c29-b1fe-27212fec7239" containerName="registry-server" Nov 24 14:23:03 crc kubenswrapper[4790]: I1124 14:23:03.758446 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="d703ae8e-4aaf-4c29-b1fe-27212fec7239" containerName="registry-server" Nov 24 14:23:03 crc kubenswrapper[4790]: I1124 14:23:03.759599 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kzd9f" Nov 24 14:23:03 crc kubenswrapper[4790]: I1124 14:23:03.774127 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kzd9f"] Nov 24 14:23:03 crc kubenswrapper[4790]: I1124 14:23:03.917735 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c5br\" (UniqueName: \"kubernetes.io/projected/44add7b4-c478-4cde-84b7-c5fa5aed6bca-kube-api-access-8c5br\") pod \"redhat-marketplace-kzd9f\" (UID: \"44add7b4-c478-4cde-84b7-c5fa5aed6bca\") " pod="openshift-marketplace/redhat-marketplace-kzd9f" Nov 24 14:23:03 crc kubenswrapper[4790]: I1124 14:23:03.918099 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44add7b4-c478-4cde-84b7-c5fa5aed6bca-utilities\") pod \"redhat-marketplace-kzd9f\" (UID: \"44add7b4-c478-4cde-84b7-c5fa5aed6bca\") " pod="openshift-marketplace/redhat-marketplace-kzd9f" Nov 24 14:23:03 crc kubenswrapper[4790]: I1124 14:23:03.918587 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44add7b4-c478-4cde-84b7-c5fa5aed6bca-catalog-content\") pod \"redhat-marketplace-kzd9f\" (UID: \"44add7b4-c478-4cde-84b7-c5fa5aed6bca\") " pod="openshift-marketplace/redhat-marketplace-kzd9f" Nov 24 14:23:04 crc kubenswrapper[4790]: I1124 14:23:04.020370 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44add7b4-c478-4cde-84b7-c5fa5aed6bca-catalog-content\") pod \"redhat-marketplace-kzd9f\" (UID: \"44add7b4-c478-4cde-84b7-c5fa5aed6bca\") " pod="openshift-marketplace/redhat-marketplace-kzd9f" Nov 24 14:23:04 crc kubenswrapper[4790]: I1124 14:23:04.020455 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8c5br\" (UniqueName: \"kubernetes.io/projected/44add7b4-c478-4cde-84b7-c5fa5aed6bca-kube-api-access-8c5br\") pod \"redhat-marketplace-kzd9f\" (UID: \"44add7b4-c478-4cde-84b7-c5fa5aed6bca\") " pod="openshift-marketplace/redhat-marketplace-kzd9f" Nov 24 14:23:04 crc kubenswrapper[4790]: I1124 14:23:04.020479 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44add7b4-c478-4cde-84b7-c5fa5aed6bca-utilities\") pod \"redhat-marketplace-kzd9f\" (UID: \"44add7b4-c478-4cde-84b7-c5fa5aed6bca\") " pod="openshift-marketplace/redhat-marketplace-kzd9f" Nov 24 14:23:04 crc kubenswrapper[4790]: I1124 14:23:04.021416 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44add7b4-c478-4cde-84b7-c5fa5aed6bca-utilities\") pod \"redhat-marketplace-kzd9f\" (UID: \"44add7b4-c478-4cde-84b7-c5fa5aed6bca\") " pod="openshift-marketplace/redhat-marketplace-kzd9f" Nov 24 14:23:04 crc kubenswrapper[4790]: I1124 14:23:04.021479 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44add7b4-c478-4cde-84b7-c5fa5aed6bca-catalog-content\") pod \"redhat-marketplace-kzd9f\" (UID: \"44add7b4-c478-4cde-84b7-c5fa5aed6bca\") " pod="openshift-marketplace/redhat-marketplace-kzd9f" Nov 24 14:23:04 crc kubenswrapper[4790]: I1124 14:23:04.046138 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8c5br\" (UniqueName: \"kubernetes.io/projected/44add7b4-c478-4cde-84b7-c5fa5aed6bca-kube-api-access-8c5br\") pod \"redhat-marketplace-kzd9f\" (UID: \"44add7b4-c478-4cde-84b7-c5fa5aed6bca\") " pod="openshift-marketplace/redhat-marketplace-kzd9f" Nov 24 14:23:04 crc kubenswrapper[4790]: I1124 14:23:04.086267 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kzd9f" Nov 24 14:23:04 crc kubenswrapper[4790]: I1124 14:23:04.338140 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kzd9f"] Nov 24 14:23:04 crc kubenswrapper[4790]: I1124 14:23:04.409459 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kzd9f" event={"ID":"44add7b4-c478-4cde-84b7-c5fa5aed6bca","Type":"ContainerStarted","Data":"009b98378eebd0d51664105650f979f82d51e1cbf29b650b45381e4dee7f0823"} Nov 24 14:23:05 crc kubenswrapper[4790]: I1124 14:23:05.424502 4790 generic.go:334] "Generic (PLEG): container finished" podID="44add7b4-c478-4cde-84b7-c5fa5aed6bca" containerID="9563da6033dbb895b9a46f761ebc8b83974cdea95e37ef746c56ffab0cbd7a61" exitCode=0 Nov 24 14:23:05 crc kubenswrapper[4790]: I1124 14:23:05.424583 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kzd9f" event={"ID":"44add7b4-c478-4cde-84b7-c5fa5aed6bca","Type":"ContainerDied","Data":"9563da6033dbb895b9a46f761ebc8b83974cdea95e37ef746c56ffab0cbd7a61"} Nov 24 14:23:07 crc kubenswrapper[4790]: I1124 14:23:07.441122 4790 generic.go:334] "Generic (PLEG): container finished" podID="44add7b4-c478-4cde-84b7-c5fa5aed6bca" containerID="d0030509a1b0eedaf8cc0145a1b67b56dc5219fd283636772e0e1f1dd64fcc55" exitCode=0 Nov 24 14:23:07 crc kubenswrapper[4790]: I1124 14:23:07.441173 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kzd9f" event={"ID":"44add7b4-c478-4cde-84b7-c5fa5aed6bca","Type":"ContainerDied","Data":"d0030509a1b0eedaf8cc0145a1b67b56dc5219fd283636772e0e1f1dd64fcc55"} Nov 24 14:23:08 crc kubenswrapper[4790]: I1124 14:23:08.449522 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kzd9f" event={"ID":"44add7b4-c478-4cde-84b7-c5fa5aed6bca","Type":"ContainerStarted","Data":"4e63fced43a940aa8a367b8caa8c8735ea637897e31a4aa1acafc784caaddf93"} Nov 24 14:23:08 crc kubenswrapper[4790]: I1124 14:23:08.471995 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kzd9f" podStartSLOduration=2.938176597 podStartE2EDuration="5.471979843s" podCreationTimestamp="2025-11-24 14:23:03 +0000 UTC" firstStartedPulling="2025-11-24 14:23:05.427556027 +0000 UTC m=+4233.807449689" lastFinishedPulling="2025-11-24 14:23:07.961359263 +0000 UTC m=+4236.341252935" observedRunningTime="2025-11-24 14:23:08.46631476 +0000 UTC m=+4236.846208422" watchObservedRunningTime="2025-11-24 14:23:08.471979843 +0000 UTC m=+4236.851873505" Nov 24 14:23:14 crc kubenswrapper[4790]: I1124 14:23:14.086729 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kzd9f" Nov 24 14:23:14 crc kubenswrapper[4790]: I1124 14:23:14.087270 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kzd9f" Nov 24 14:23:14 crc kubenswrapper[4790]: I1124 14:23:14.124151 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kzd9f" Nov 24 14:23:14 crc kubenswrapper[4790]: I1124 14:23:14.553934 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kzd9f" Nov 24 14:23:14 crc kubenswrapper[4790]: I1124 14:23:14.598291 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kzd9f"] Nov 24 14:23:16 crc kubenswrapper[4790]: I1124 14:23:16.504783 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kzd9f" podUID="44add7b4-c478-4cde-84b7-c5fa5aed6bca" containerName="registry-server" containerID="cri-o://4e63fced43a940aa8a367b8caa8c8735ea637897e31a4aa1acafc784caaddf93" gracePeriod=2 Nov 24 14:23:16 crc kubenswrapper[4790]: I1124 14:23:16.879157 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kzd9f" Nov 24 14:23:17 crc kubenswrapper[4790]: I1124 14:23:17.004930 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44add7b4-c478-4cde-84b7-c5fa5aed6bca-catalog-content\") pod \"44add7b4-c478-4cde-84b7-c5fa5aed6bca\" (UID: \"44add7b4-c478-4cde-84b7-c5fa5aed6bca\") " Nov 24 14:23:17 crc kubenswrapper[4790]: I1124 14:23:17.004984 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44add7b4-c478-4cde-84b7-c5fa5aed6bca-utilities\") pod \"44add7b4-c478-4cde-84b7-c5fa5aed6bca\" (UID: \"44add7b4-c478-4cde-84b7-c5fa5aed6bca\") " Nov 24 14:23:17 crc kubenswrapper[4790]: I1124 14:23:17.005027 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8c5br\" (UniqueName: \"kubernetes.io/projected/44add7b4-c478-4cde-84b7-c5fa5aed6bca-kube-api-access-8c5br\") pod \"44add7b4-c478-4cde-84b7-c5fa5aed6bca\" (UID: \"44add7b4-c478-4cde-84b7-c5fa5aed6bca\") " Nov 24 14:23:17 crc kubenswrapper[4790]: I1124 14:23:17.006374 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44add7b4-c478-4cde-84b7-c5fa5aed6bca-utilities" (OuterVolumeSpecName: "utilities") pod "44add7b4-c478-4cde-84b7-c5fa5aed6bca" (UID: "44add7b4-c478-4cde-84b7-c5fa5aed6bca"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:23:17 crc kubenswrapper[4790]: I1124 14:23:17.010672 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44add7b4-c478-4cde-84b7-c5fa5aed6bca-kube-api-access-8c5br" (OuterVolumeSpecName: "kube-api-access-8c5br") pod "44add7b4-c478-4cde-84b7-c5fa5aed6bca" (UID: "44add7b4-c478-4cde-84b7-c5fa5aed6bca"). InnerVolumeSpecName "kube-api-access-8c5br". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:23:17 crc kubenswrapper[4790]: I1124 14:23:17.024369 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44add7b4-c478-4cde-84b7-c5fa5aed6bca-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "44add7b4-c478-4cde-84b7-c5fa5aed6bca" (UID: "44add7b4-c478-4cde-84b7-c5fa5aed6bca"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:23:17 crc kubenswrapper[4790]: I1124 14:23:17.106175 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44add7b4-c478-4cde-84b7-c5fa5aed6bca-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:17 crc kubenswrapper[4790]: I1124 14:23:17.106203 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44add7b4-c478-4cde-84b7-c5fa5aed6bca-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:17 crc kubenswrapper[4790]: I1124 14:23:17.106221 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8c5br\" (UniqueName: \"kubernetes.io/projected/44add7b4-c478-4cde-84b7-c5fa5aed6bca-kube-api-access-8c5br\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:17 crc kubenswrapper[4790]: I1124 14:23:17.520150 4790 generic.go:334] "Generic (PLEG): container finished" podID="44add7b4-c478-4cde-84b7-c5fa5aed6bca" containerID="4e63fced43a940aa8a367b8caa8c8735ea637897e31a4aa1acafc784caaddf93" exitCode=0 Nov 24 14:23:17 crc kubenswrapper[4790]: I1124 14:23:17.520215 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kzd9f" event={"ID":"44add7b4-c478-4cde-84b7-c5fa5aed6bca","Type":"ContainerDied","Data":"4e63fced43a940aa8a367b8caa8c8735ea637897e31a4aa1acafc784caaddf93"} Nov 24 14:23:17 crc kubenswrapper[4790]: I1124 14:23:17.520279 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kzd9f" event={"ID":"44add7b4-c478-4cde-84b7-c5fa5aed6bca","Type":"ContainerDied","Data":"009b98378eebd0d51664105650f979f82d51e1cbf29b650b45381e4dee7f0823"} Nov 24 14:23:17 crc kubenswrapper[4790]: I1124 14:23:17.520307 4790 scope.go:117] "RemoveContainer" containerID="4e63fced43a940aa8a367b8caa8c8735ea637897e31a4aa1acafc784caaddf93" Nov 24 14:23:17 crc kubenswrapper[4790]: I1124 14:23:17.521110 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kzd9f" Nov 24 14:23:17 crc kubenswrapper[4790]: I1124 14:23:17.539508 4790 scope.go:117] "RemoveContainer" containerID="d0030509a1b0eedaf8cc0145a1b67b56dc5219fd283636772e0e1f1dd64fcc55" Nov 24 14:23:17 crc kubenswrapper[4790]: I1124 14:23:17.564732 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kzd9f"] Nov 24 14:23:17 crc kubenswrapper[4790]: I1124 14:23:17.571549 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kzd9f"] Nov 24 14:23:17 crc kubenswrapper[4790]: I1124 14:23:17.582498 4790 scope.go:117] "RemoveContainer" containerID="9563da6033dbb895b9a46f761ebc8b83974cdea95e37ef746c56ffab0cbd7a61" Nov 24 14:23:17 crc kubenswrapper[4790]: I1124 14:23:17.600836 4790 scope.go:117] "RemoveContainer" containerID="4e63fced43a940aa8a367b8caa8c8735ea637897e31a4aa1acafc784caaddf93" Nov 24 14:23:17 crc kubenswrapper[4790]: E1124 14:23:17.601305 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e63fced43a940aa8a367b8caa8c8735ea637897e31a4aa1acafc784caaddf93\": container with ID starting with 4e63fced43a940aa8a367b8caa8c8735ea637897e31a4aa1acafc784caaddf93 not found: ID does not exist" containerID="4e63fced43a940aa8a367b8caa8c8735ea637897e31a4aa1acafc784caaddf93" Nov 24 14:23:17 crc kubenswrapper[4790]: I1124 14:23:17.601337 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e63fced43a940aa8a367b8caa8c8735ea637897e31a4aa1acafc784caaddf93"} err="failed to get container status \"4e63fced43a940aa8a367b8caa8c8735ea637897e31a4aa1acafc784caaddf93\": rpc error: code = NotFound desc = could not find container \"4e63fced43a940aa8a367b8caa8c8735ea637897e31a4aa1acafc784caaddf93\": container with ID starting with 4e63fced43a940aa8a367b8caa8c8735ea637897e31a4aa1acafc784caaddf93 not found: ID does not exist" Nov 24 14:23:17 crc kubenswrapper[4790]: I1124 14:23:17.601358 4790 scope.go:117] "RemoveContainer" containerID="d0030509a1b0eedaf8cc0145a1b67b56dc5219fd283636772e0e1f1dd64fcc55" Nov 24 14:23:17 crc kubenswrapper[4790]: E1124 14:23:17.601642 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0030509a1b0eedaf8cc0145a1b67b56dc5219fd283636772e0e1f1dd64fcc55\": container with ID starting with d0030509a1b0eedaf8cc0145a1b67b56dc5219fd283636772e0e1f1dd64fcc55 not found: ID does not exist" containerID="d0030509a1b0eedaf8cc0145a1b67b56dc5219fd283636772e0e1f1dd64fcc55" Nov 24 14:23:17 crc kubenswrapper[4790]: I1124 14:23:17.601657 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0030509a1b0eedaf8cc0145a1b67b56dc5219fd283636772e0e1f1dd64fcc55"} err="failed to get container status \"d0030509a1b0eedaf8cc0145a1b67b56dc5219fd283636772e0e1f1dd64fcc55\": rpc error: code = NotFound desc = could not find container \"d0030509a1b0eedaf8cc0145a1b67b56dc5219fd283636772e0e1f1dd64fcc55\": container with ID starting with d0030509a1b0eedaf8cc0145a1b67b56dc5219fd283636772e0e1f1dd64fcc55 not found: ID does not exist" Nov 24 14:23:17 crc kubenswrapper[4790]: I1124 14:23:17.601669 4790 scope.go:117] "RemoveContainer" containerID="9563da6033dbb895b9a46f761ebc8b83974cdea95e37ef746c56ffab0cbd7a61" Nov 24 14:23:17 crc kubenswrapper[4790]: E1124 14:23:17.601971 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9563da6033dbb895b9a46f761ebc8b83974cdea95e37ef746c56ffab0cbd7a61\": container with ID starting with 9563da6033dbb895b9a46f761ebc8b83974cdea95e37ef746c56ffab0cbd7a61 not found: ID does not exist" containerID="9563da6033dbb895b9a46f761ebc8b83974cdea95e37ef746c56ffab0cbd7a61" Nov 24 14:23:17 crc kubenswrapper[4790]: I1124 14:23:17.602055 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9563da6033dbb895b9a46f761ebc8b83974cdea95e37ef746c56ffab0cbd7a61"} err="failed to get container status \"9563da6033dbb895b9a46f761ebc8b83974cdea95e37ef746c56ffab0cbd7a61\": rpc error: code = NotFound desc = could not find container \"9563da6033dbb895b9a46f761ebc8b83974cdea95e37ef746c56ffab0cbd7a61\": container with ID starting with 9563da6033dbb895b9a46f761ebc8b83974cdea95e37ef746c56ffab0cbd7a61 not found: ID does not exist" Nov 24 14:23:18 crc kubenswrapper[4790]: I1124 14:23:18.323858 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44add7b4-c478-4cde-84b7-c5fa5aed6bca" path="/var/lib/kubelet/pods/44add7b4-c478-4cde-84b7-c5fa5aed6bca/volumes" Nov 24 14:23:28 crc kubenswrapper[4790]: I1124 14:23:28.704357 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-m5lwm"] Nov 24 14:23:28 crc kubenswrapper[4790]: E1124 14:23:28.705267 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44add7b4-c478-4cde-84b7-c5fa5aed6bca" containerName="extract-content" Nov 24 14:23:28 crc kubenswrapper[4790]: I1124 14:23:28.705294 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="44add7b4-c478-4cde-84b7-c5fa5aed6bca" containerName="extract-content" Nov 24 14:23:28 crc kubenswrapper[4790]: E1124 14:23:28.705319 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44add7b4-c478-4cde-84b7-c5fa5aed6bca" containerName="registry-server" Nov 24 14:23:28 crc kubenswrapper[4790]: I1124 14:23:28.705328 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="44add7b4-c478-4cde-84b7-c5fa5aed6bca" containerName="registry-server" Nov 24 14:23:28 crc kubenswrapper[4790]: E1124 14:23:28.705353 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44add7b4-c478-4cde-84b7-c5fa5aed6bca" containerName="extract-utilities" Nov 24 14:23:28 crc kubenswrapper[4790]: I1124 14:23:28.705362 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="44add7b4-c478-4cde-84b7-c5fa5aed6bca" containerName="extract-utilities" Nov 24 14:23:28 crc kubenswrapper[4790]: I1124 14:23:28.705515 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="44add7b4-c478-4cde-84b7-c5fa5aed6bca" containerName="registry-server" Nov 24 14:23:28 crc kubenswrapper[4790]: I1124 14:23:28.706756 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m5lwm" Nov 24 14:23:28 crc kubenswrapper[4790]: I1124 14:23:28.716503 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-m5lwm"] Nov 24 14:23:28 crc kubenswrapper[4790]: I1124 14:23:28.878173 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f4eaeea-ed9a-4cfb-a09f-e5ed26fb312c-utilities\") pod \"redhat-operators-m5lwm\" (UID: \"8f4eaeea-ed9a-4cfb-a09f-e5ed26fb312c\") " pod="openshift-marketplace/redhat-operators-m5lwm" Nov 24 14:23:28 crc kubenswrapper[4790]: I1124 14:23:28.878567 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f4eaeea-ed9a-4cfb-a09f-e5ed26fb312c-catalog-content\") pod \"redhat-operators-m5lwm\" (UID: \"8f4eaeea-ed9a-4cfb-a09f-e5ed26fb312c\") " pod="openshift-marketplace/redhat-operators-m5lwm" Nov 24 14:23:28 crc kubenswrapper[4790]: I1124 14:23:28.878605 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctrbr\" (UniqueName: \"kubernetes.io/projected/8f4eaeea-ed9a-4cfb-a09f-e5ed26fb312c-kube-api-access-ctrbr\") pod \"redhat-operators-m5lwm\" (UID: \"8f4eaeea-ed9a-4cfb-a09f-e5ed26fb312c\") " pod="openshift-marketplace/redhat-operators-m5lwm" Nov 24 14:23:28 crc kubenswrapper[4790]: I1124 14:23:28.979736 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f4eaeea-ed9a-4cfb-a09f-e5ed26fb312c-utilities\") pod \"redhat-operators-m5lwm\" (UID: \"8f4eaeea-ed9a-4cfb-a09f-e5ed26fb312c\") " pod="openshift-marketplace/redhat-operators-m5lwm" Nov 24 14:23:28 crc kubenswrapper[4790]: I1124 14:23:28.979812 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f4eaeea-ed9a-4cfb-a09f-e5ed26fb312c-catalog-content\") pod \"redhat-operators-m5lwm\" (UID: \"8f4eaeea-ed9a-4cfb-a09f-e5ed26fb312c\") " pod="openshift-marketplace/redhat-operators-m5lwm" Nov 24 14:23:28 crc kubenswrapper[4790]: I1124 14:23:28.979843 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctrbr\" (UniqueName: \"kubernetes.io/projected/8f4eaeea-ed9a-4cfb-a09f-e5ed26fb312c-kube-api-access-ctrbr\") pod \"redhat-operators-m5lwm\" (UID: \"8f4eaeea-ed9a-4cfb-a09f-e5ed26fb312c\") " pod="openshift-marketplace/redhat-operators-m5lwm" Nov 24 14:23:28 crc kubenswrapper[4790]: I1124 14:23:28.980326 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f4eaeea-ed9a-4cfb-a09f-e5ed26fb312c-utilities\") pod \"redhat-operators-m5lwm\" (UID: \"8f4eaeea-ed9a-4cfb-a09f-e5ed26fb312c\") " pod="openshift-marketplace/redhat-operators-m5lwm" Nov 24 14:23:28 crc kubenswrapper[4790]: I1124 14:23:28.980440 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f4eaeea-ed9a-4cfb-a09f-e5ed26fb312c-catalog-content\") pod \"redhat-operators-m5lwm\" (UID: \"8f4eaeea-ed9a-4cfb-a09f-e5ed26fb312c\") " pod="openshift-marketplace/redhat-operators-m5lwm" Nov 24 14:23:29 crc kubenswrapper[4790]: I1124 14:23:29.008819 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctrbr\" (UniqueName: \"kubernetes.io/projected/8f4eaeea-ed9a-4cfb-a09f-e5ed26fb312c-kube-api-access-ctrbr\") pod \"redhat-operators-m5lwm\" (UID: \"8f4eaeea-ed9a-4cfb-a09f-e5ed26fb312c\") " pod="openshift-marketplace/redhat-operators-m5lwm" Nov 24 14:23:29 crc kubenswrapper[4790]: I1124 14:23:29.037635 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m5lwm" Nov 24 14:23:29 crc kubenswrapper[4790]: I1124 14:23:29.251608 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-m5lwm"] Nov 24 14:23:29 crc kubenswrapper[4790]: I1124 14:23:29.612374 4790 generic.go:334] "Generic (PLEG): container finished" podID="8f4eaeea-ed9a-4cfb-a09f-e5ed26fb312c" containerID="c1c0ca049017b99c5c5472d978edb853a4bd06b8053353f8ba22a22557f08469" exitCode=0 Nov 24 14:23:29 crc kubenswrapper[4790]: I1124 14:23:29.612412 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m5lwm" event={"ID":"8f4eaeea-ed9a-4cfb-a09f-e5ed26fb312c","Type":"ContainerDied","Data":"c1c0ca049017b99c5c5472d978edb853a4bd06b8053353f8ba22a22557f08469"} Nov 24 14:23:29 crc kubenswrapper[4790]: I1124 14:23:29.612437 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m5lwm" event={"ID":"8f4eaeea-ed9a-4cfb-a09f-e5ed26fb312c","Type":"ContainerStarted","Data":"cbc6391b68fc8116c1f1092e4f8e396175b0ac022b68340b9bf01ad45b085785"} Nov 24 14:23:35 crc kubenswrapper[4790]: I1124 14:23:35.652397 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m5lwm" event={"ID":"8f4eaeea-ed9a-4cfb-a09f-e5ed26fb312c","Type":"ContainerStarted","Data":"f96155bbb3e762b61bd40cb2df1502bb043b9f6cb6a1c58db65951a4de524c8d"} Nov 24 14:23:36 crc kubenswrapper[4790]: I1124 14:23:36.660771 4790 generic.go:334] "Generic (PLEG): container finished" podID="8f4eaeea-ed9a-4cfb-a09f-e5ed26fb312c" containerID="f96155bbb3e762b61bd40cb2df1502bb043b9f6cb6a1c58db65951a4de524c8d" exitCode=0 Nov 24 14:23:36 crc kubenswrapper[4790]: I1124 14:23:36.660854 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m5lwm" event={"ID":"8f4eaeea-ed9a-4cfb-a09f-e5ed26fb312c","Type":"ContainerDied","Data":"f96155bbb3e762b61bd40cb2df1502bb043b9f6cb6a1c58db65951a4de524c8d"} Nov 24 14:23:36 crc kubenswrapper[4790]: I1124 14:23:36.662968 4790 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 14:23:37 crc kubenswrapper[4790]: I1124 14:23:37.669170 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m5lwm" event={"ID":"8f4eaeea-ed9a-4cfb-a09f-e5ed26fb312c","Type":"ContainerStarted","Data":"18897c9416ab6437d856b5527375b8210ab7bb4c3aa1fab4695aafac4bb9818f"} Nov 24 14:23:37 crc kubenswrapper[4790]: I1124 14:23:37.689001 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-m5lwm" podStartSLOduration=2.160129581 podStartE2EDuration="9.688978053s" podCreationTimestamp="2025-11-24 14:23:28 +0000 UTC" firstStartedPulling="2025-11-24 14:23:29.613943653 +0000 UTC m=+4257.993837325" lastFinishedPulling="2025-11-24 14:23:37.142792125 +0000 UTC m=+4265.522685797" observedRunningTime="2025-11-24 14:23:37.685063811 +0000 UTC m=+4266.064957483" watchObservedRunningTime="2025-11-24 14:23:37.688978053 +0000 UTC m=+4266.068871715" Nov 24 14:23:39 crc kubenswrapper[4790]: I1124 14:23:39.038264 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-m5lwm" Nov 24 14:23:39 crc kubenswrapper[4790]: I1124 14:23:39.038580 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-m5lwm" Nov 24 14:23:40 crc kubenswrapper[4790]: I1124 14:23:40.084151 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-m5lwm" podUID="8f4eaeea-ed9a-4cfb-a09f-e5ed26fb312c" containerName="registry-server" probeResult="failure" output=< Nov 24 14:23:40 crc kubenswrapper[4790]: timeout: failed to connect service ":50051" within 1s Nov 24 14:23:40 crc kubenswrapper[4790]: > Nov 24 14:23:49 crc kubenswrapper[4790]: I1124 14:23:49.086758 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-m5lwm" Nov 24 14:23:49 crc kubenswrapper[4790]: I1124 14:23:49.132279 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-m5lwm" Nov 24 14:23:49 crc kubenswrapper[4790]: I1124 14:23:49.185187 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-m5lwm"] Nov 24 14:23:49 crc kubenswrapper[4790]: I1124 14:23:49.322446 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-96c9c"] Nov 24 14:23:49 crc kubenswrapper[4790]: I1124 14:23:49.322754 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-96c9c" podUID="90d19b1e-1040-456c-be95-32e15e32a013" containerName="registry-server" containerID="cri-o://caec78013ff1d9d5db7fa09e36e0b2f3ef3901a99d4b732afc888710af5a6a20" gracePeriod=2 Nov 24 14:23:49 crc kubenswrapper[4790]: I1124 14:23:49.688630 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-96c9c" Nov 24 14:23:49 crc kubenswrapper[4790]: I1124 14:23:49.759119 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90d19b1e-1040-456c-be95-32e15e32a013-utilities\") pod \"90d19b1e-1040-456c-be95-32e15e32a013\" (UID: \"90d19b1e-1040-456c-be95-32e15e32a013\") " Nov 24 14:23:49 crc kubenswrapper[4790]: I1124 14:23:49.759230 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90d19b1e-1040-456c-be95-32e15e32a013-catalog-content\") pod \"90d19b1e-1040-456c-be95-32e15e32a013\" (UID: \"90d19b1e-1040-456c-be95-32e15e32a013\") " Nov 24 14:23:49 crc kubenswrapper[4790]: I1124 14:23:49.759255 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvv7x\" (UniqueName: \"kubernetes.io/projected/90d19b1e-1040-456c-be95-32e15e32a013-kube-api-access-fvv7x\") pod \"90d19b1e-1040-456c-be95-32e15e32a013\" (UID: \"90d19b1e-1040-456c-be95-32e15e32a013\") " Nov 24 14:23:49 crc kubenswrapper[4790]: I1124 14:23:49.762991 4790 generic.go:334] "Generic (PLEG): container finished" podID="90d19b1e-1040-456c-be95-32e15e32a013" containerID="caec78013ff1d9d5db7fa09e36e0b2f3ef3901a99d4b732afc888710af5a6a20" exitCode=0 Nov 24 14:23:49 crc kubenswrapper[4790]: I1124 14:23:49.763573 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90d19b1e-1040-456c-be95-32e15e32a013-utilities" (OuterVolumeSpecName: "utilities") pod "90d19b1e-1040-456c-be95-32e15e32a013" (UID: "90d19b1e-1040-456c-be95-32e15e32a013"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:23:49 crc kubenswrapper[4790]: I1124 14:23:49.763683 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-96c9c" Nov 24 14:23:49 crc kubenswrapper[4790]: I1124 14:23:49.764041 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-96c9c" event={"ID":"90d19b1e-1040-456c-be95-32e15e32a013","Type":"ContainerDied","Data":"caec78013ff1d9d5db7fa09e36e0b2f3ef3901a99d4b732afc888710af5a6a20"} Nov 24 14:23:49 crc kubenswrapper[4790]: I1124 14:23:49.764069 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-96c9c" event={"ID":"90d19b1e-1040-456c-be95-32e15e32a013","Type":"ContainerDied","Data":"4ab30ec3802395732f52b6089bc34c6ba50fe8ef92c512cfaea226979e33a948"} Nov 24 14:23:49 crc kubenswrapper[4790]: I1124 14:23:49.764086 4790 scope.go:117] "RemoveContainer" containerID="caec78013ff1d9d5db7fa09e36e0b2f3ef3901a99d4b732afc888710af5a6a20" Nov 24 14:23:49 crc kubenswrapper[4790]: I1124 14:23:49.774950 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90d19b1e-1040-456c-be95-32e15e32a013-kube-api-access-fvv7x" (OuterVolumeSpecName: "kube-api-access-fvv7x") pod "90d19b1e-1040-456c-be95-32e15e32a013" (UID: "90d19b1e-1040-456c-be95-32e15e32a013"). InnerVolumeSpecName "kube-api-access-fvv7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:23:49 crc kubenswrapper[4790]: I1124 14:23:49.805962 4790 scope.go:117] "RemoveContainer" containerID="56452aac30a9719549303efe43baa6aff815459a7d2666c6e2bbeee31943252c" Nov 24 14:23:49 crc kubenswrapper[4790]: I1124 14:23:49.825485 4790 scope.go:117] "RemoveContainer" containerID="7ff1828fb9b063ab7d9298790fc0fec064882d7429407dea6c805ebfc16c6313" Nov 24 14:23:49 crc kubenswrapper[4790]: I1124 14:23:49.847997 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90d19b1e-1040-456c-be95-32e15e32a013-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "90d19b1e-1040-456c-be95-32e15e32a013" (UID: "90d19b1e-1040-456c-be95-32e15e32a013"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:23:49 crc kubenswrapper[4790]: I1124 14:23:49.849407 4790 scope.go:117] "RemoveContainer" containerID="caec78013ff1d9d5db7fa09e36e0b2f3ef3901a99d4b732afc888710af5a6a20" Nov 24 14:23:49 crc kubenswrapper[4790]: E1124 14:23:49.850340 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"caec78013ff1d9d5db7fa09e36e0b2f3ef3901a99d4b732afc888710af5a6a20\": container with ID starting with caec78013ff1d9d5db7fa09e36e0b2f3ef3901a99d4b732afc888710af5a6a20 not found: ID does not exist" containerID="caec78013ff1d9d5db7fa09e36e0b2f3ef3901a99d4b732afc888710af5a6a20" Nov 24 14:23:49 crc kubenswrapper[4790]: I1124 14:23:49.850403 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"caec78013ff1d9d5db7fa09e36e0b2f3ef3901a99d4b732afc888710af5a6a20"} err="failed to get container status \"caec78013ff1d9d5db7fa09e36e0b2f3ef3901a99d4b732afc888710af5a6a20\": rpc error: code = NotFound desc = could not find container \"caec78013ff1d9d5db7fa09e36e0b2f3ef3901a99d4b732afc888710af5a6a20\": container with ID starting with caec78013ff1d9d5db7fa09e36e0b2f3ef3901a99d4b732afc888710af5a6a20 not found: ID does not exist" Nov 24 14:23:49 crc kubenswrapper[4790]: I1124 14:23:49.850436 4790 scope.go:117] "RemoveContainer" containerID="56452aac30a9719549303efe43baa6aff815459a7d2666c6e2bbeee31943252c" Nov 24 14:23:49 crc kubenswrapper[4790]: E1124 14:23:49.852949 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56452aac30a9719549303efe43baa6aff815459a7d2666c6e2bbeee31943252c\": container with ID starting with 56452aac30a9719549303efe43baa6aff815459a7d2666c6e2bbeee31943252c not found: ID does not exist" containerID="56452aac30a9719549303efe43baa6aff815459a7d2666c6e2bbeee31943252c" Nov 24 14:23:49 crc kubenswrapper[4790]: I1124 14:23:49.852990 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56452aac30a9719549303efe43baa6aff815459a7d2666c6e2bbeee31943252c"} err="failed to get container status \"56452aac30a9719549303efe43baa6aff815459a7d2666c6e2bbeee31943252c\": rpc error: code = NotFound desc = could not find container \"56452aac30a9719549303efe43baa6aff815459a7d2666c6e2bbeee31943252c\": container with ID starting with 56452aac30a9719549303efe43baa6aff815459a7d2666c6e2bbeee31943252c not found: ID does not exist" Nov 24 14:23:49 crc kubenswrapper[4790]: I1124 14:23:49.853011 4790 scope.go:117] "RemoveContainer" containerID="7ff1828fb9b063ab7d9298790fc0fec064882d7429407dea6c805ebfc16c6313" Nov 24 14:23:49 crc kubenswrapper[4790]: E1124 14:23:49.853520 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ff1828fb9b063ab7d9298790fc0fec064882d7429407dea6c805ebfc16c6313\": container with ID starting with 7ff1828fb9b063ab7d9298790fc0fec064882d7429407dea6c805ebfc16c6313 not found: ID does not exist" containerID="7ff1828fb9b063ab7d9298790fc0fec064882d7429407dea6c805ebfc16c6313" Nov 24 14:23:49 crc kubenswrapper[4790]: I1124 14:23:49.853604 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ff1828fb9b063ab7d9298790fc0fec064882d7429407dea6c805ebfc16c6313"} err="failed to get container status \"7ff1828fb9b063ab7d9298790fc0fec064882d7429407dea6c805ebfc16c6313\": rpc error: code = NotFound desc = could not find container \"7ff1828fb9b063ab7d9298790fc0fec064882d7429407dea6c805ebfc16c6313\": container with ID starting with 7ff1828fb9b063ab7d9298790fc0fec064882d7429407dea6c805ebfc16c6313 not found: ID does not exist" Nov 24 14:23:49 crc kubenswrapper[4790]: I1124 14:23:49.860817 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90d19b1e-1040-456c-be95-32e15e32a013-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:49 crc kubenswrapper[4790]: I1124 14:23:49.860853 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90d19b1e-1040-456c-be95-32e15e32a013-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:49 crc kubenswrapper[4790]: I1124 14:23:49.860866 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvv7x\" (UniqueName: \"kubernetes.io/projected/90d19b1e-1040-456c-be95-32e15e32a013-kube-api-access-fvv7x\") on node \"crc\" DevicePath \"\"" Nov 24 14:23:50 crc kubenswrapper[4790]: I1124 14:23:50.093092 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-96c9c"] Nov 24 14:23:50 crc kubenswrapper[4790]: I1124 14:23:50.100731 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-96c9c"] Nov 24 14:23:50 crc kubenswrapper[4790]: I1124 14:23:50.323495 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90d19b1e-1040-456c-be95-32e15e32a013" path="/var/lib/kubelet/pods/90d19b1e-1040-456c-be95-32e15e32a013/volumes" Nov 24 14:24:13 crc kubenswrapper[4790]: I1124 14:24:13.939226 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:24:13 crc kubenswrapper[4790]: I1124 14:24:13.939823 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:24:43 crc kubenswrapper[4790]: I1124 14:24:43.938185 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:24:43 crc kubenswrapper[4790]: I1124 14:24:43.938779 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:25:13 crc kubenswrapper[4790]: I1124 14:25:13.938494 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:25:13 crc kubenswrapper[4790]: I1124 14:25:13.939161 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:25:13 crc kubenswrapper[4790]: I1124 14:25:13.939223 4790 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 14:25:13 crc kubenswrapper[4790]: I1124 14:25:13.939971 4790 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b25648b43123ad262f034d5f38126a0ec7bbda6c1db0e5dea770deb92fb3c707"} pod="openshift-machine-config-operator/machine-config-daemon-xz49t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:25:13 crc kubenswrapper[4790]: I1124 14:25:13.940062 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" containerID="cri-o://b25648b43123ad262f034d5f38126a0ec7bbda6c1db0e5dea770deb92fb3c707" gracePeriod=600 Nov 24 14:25:14 crc kubenswrapper[4790]: I1124 14:25:14.887345 4790 generic.go:334] "Generic (PLEG): container finished" podID="0d73b133-48f0-455f-8f6a-742e633f631a" containerID="b25648b43123ad262f034d5f38126a0ec7bbda6c1db0e5dea770deb92fb3c707" exitCode=0 Nov 24 14:25:14 crc kubenswrapper[4790]: I1124 14:25:14.887411 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerDied","Data":"b25648b43123ad262f034d5f38126a0ec7bbda6c1db0e5dea770deb92fb3c707"} Nov 24 14:25:14 crc kubenswrapper[4790]: I1124 14:25:14.888267 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"571bd0c27e6ad4756f7116ed8b5a40d06301f3d85436cea369594fdb322154a7"} Nov 24 14:25:14 crc kubenswrapper[4790]: I1124 14:25:14.888363 4790 scope.go:117] "RemoveContainer" containerID="9d160b5f0ea013bde38ba30b4fd4ea41f13ba76ae92c0901dbd7ffee1185aac0" Nov 24 14:25:41 crc kubenswrapper[4790]: I1124 14:25:41.104519 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hqzm4"] Nov 24 14:25:41 crc kubenswrapper[4790]: E1124 14:25:41.106073 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90d19b1e-1040-456c-be95-32e15e32a013" containerName="registry-server" Nov 24 14:25:41 crc kubenswrapper[4790]: I1124 14:25:41.106104 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="90d19b1e-1040-456c-be95-32e15e32a013" containerName="registry-server" Nov 24 14:25:41 crc kubenswrapper[4790]: E1124 14:25:41.106174 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90d19b1e-1040-456c-be95-32e15e32a013" containerName="extract-content" Nov 24 14:25:41 crc kubenswrapper[4790]: I1124 14:25:41.106192 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="90d19b1e-1040-456c-be95-32e15e32a013" containerName="extract-content" Nov 24 14:25:41 crc kubenswrapper[4790]: E1124 14:25:41.106259 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90d19b1e-1040-456c-be95-32e15e32a013" containerName="extract-utilities" Nov 24 14:25:41 crc kubenswrapper[4790]: I1124 14:25:41.106278 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="90d19b1e-1040-456c-be95-32e15e32a013" containerName="extract-utilities" Nov 24 14:25:41 crc kubenswrapper[4790]: I1124 14:25:41.106616 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="90d19b1e-1040-456c-be95-32e15e32a013" containerName="registry-server" Nov 24 14:25:41 crc kubenswrapper[4790]: I1124 14:25:41.109287 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hqzm4" Nov 24 14:25:41 crc kubenswrapper[4790]: I1124 14:25:41.126172 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hqzm4"] Nov 24 14:25:41 crc kubenswrapper[4790]: I1124 14:25:41.133699 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khpp6\" (UniqueName: \"kubernetes.io/projected/ce5b3273-5304-4d77-8d49-398174e348de-kube-api-access-khpp6\") pod \"community-operators-hqzm4\" (UID: \"ce5b3273-5304-4d77-8d49-398174e348de\") " pod="openshift-marketplace/community-operators-hqzm4" Nov 24 14:25:41 crc kubenswrapper[4790]: I1124 14:25:41.134091 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce5b3273-5304-4d77-8d49-398174e348de-catalog-content\") pod \"community-operators-hqzm4\" (UID: \"ce5b3273-5304-4d77-8d49-398174e348de\") " pod="openshift-marketplace/community-operators-hqzm4" Nov 24 14:25:41 crc kubenswrapper[4790]: I1124 14:25:41.134188 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce5b3273-5304-4d77-8d49-398174e348de-utilities\") pod \"community-operators-hqzm4\" (UID: \"ce5b3273-5304-4d77-8d49-398174e348de\") " pod="openshift-marketplace/community-operators-hqzm4" Nov 24 14:25:41 crc kubenswrapper[4790]: I1124 14:25:41.234851 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce5b3273-5304-4d77-8d49-398174e348de-catalog-content\") pod \"community-operators-hqzm4\" (UID: \"ce5b3273-5304-4d77-8d49-398174e348de\") " pod="openshift-marketplace/community-operators-hqzm4" Nov 24 14:25:41 crc kubenswrapper[4790]: I1124 14:25:41.235050 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce5b3273-5304-4d77-8d49-398174e348de-utilities\") pod \"community-operators-hqzm4\" (UID: \"ce5b3273-5304-4d77-8d49-398174e348de\") " pod="openshift-marketplace/community-operators-hqzm4" Nov 24 14:25:41 crc kubenswrapper[4790]: I1124 14:25:41.235175 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khpp6\" (UniqueName: \"kubernetes.io/projected/ce5b3273-5304-4d77-8d49-398174e348de-kube-api-access-khpp6\") pod \"community-operators-hqzm4\" (UID: \"ce5b3273-5304-4d77-8d49-398174e348de\") " pod="openshift-marketplace/community-operators-hqzm4" Nov 24 14:25:41 crc kubenswrapper[4790]: I1124 14:25:41.235455 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce5b3273-5304-4d77-8d49-398174e348de-catalog-content\") pod \"community-operators-hqzm4\" (UID: \"ce5b3273-5304-4d77-8d49-398174e348de\") " pod="openshift-marketplace/community-operators-hqzm4" Nov 24 14:25:41 crc kubenswrapper[4790]: I1124 14:25:41.235538 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce5b3273-5304-4d77-8d49-398174e348de-utilities\") pod \"community-operators-hqzm4\" (UID: \"ce5b3273-5304-4d77-8d49-398174e348de\") " pod="openshift-marketplace/community-operators-hqzm4" Nov 24 14:25:41 crc kubenswrapper[4790]: I1124 14:25:41.263312 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khpp6\" (UniqueName: \"kubernetes.io/projected/ce5b3273-5304-4d77-8d49-398174e348de-kube-api-access-khpp6\") pod \"community-operators-hqzm4\" (UID: \"ce5b3273-5304-4d77-8d49-398174e348de\") " pod="openshift-marketplace/community-operators-hqzm4" Nov 24 14:25:41 crc kubenswrapper[4790]: I1124 14:25:41.433597 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hqzm4" Nov 24 14:25:41 crc kubenswrapper[4790]: I1124 14:25:41.964567 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hqzm4"] Nov 24 14:25:42 crc kubenswrapper[4790]: I1124 14:25:42.138306 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hqzm4" event={"ID":"ce5b3273-5304-4d77-8d49-398174e348de","Type":"ContainerStarted","Data":"883f2f9ddbaca2e671a332cb63fff3796c3ba745e5243a384d7a4404cf9b1ad3"} Nov 24 14:25:43 crc kubenswrapper[4790]: I1124 14:25:43.150767 4790 generic.go:334] "Generic (PLEG): container finished" podID="ce5b3273-5304-4d77-8d49-398174e348de" containerID="a4905a1abd2ec0423542afd505c1f529460c4f4dd1e044264ef953a400628c38" exitCode=0 Nov 24 14:25:43 crc kubenswrapper[4790]: I1124 14:25:43.151053 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hqzm4" event={"ID":"ce5b3273-5304-4d77-8d49-398174e348de","Type":"ContainerDied","Data":"a4905a1abd2ec0423542afd505c1f529460c4f4dd1e044264ef953a400628c38"} Nov 24 14:25:45 crc kubenswrapper[4790]: I1124 14:25:45.180726 4790 generic.go:334] "Generic (PLEG): container finished" podID="ce5b3273-5304-4d77-8d49-398174e348de" containerID="4ed864e95c4bfbb17015091f17d2b9b766838cf6487b00c48dc5c4ae4123a5bd" exitCode=0 Nov 24 14:25:45 crc kubenswrapper[4790]: I1124 14:25:45.181035 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hqzm4" event={"ID":"ce5b3273-5304-4d77-8d49-398174e348de","Type":"ContainerDied","Data":"4ed864e95c4bfbb17015091f17d2b9b766838cf6487b00c48dc5c4ae4123a5bd"} Nov 24 14:25:47 crc kubenswrapper[4790]: I1124 14:25:47.197713 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hqzm4" event={"ID":"ce5b3273-5304-4d77-8d49-398174e348de","Type":"ContainerStarted","Data":"30b3885384f9d9637dd99a093f600339bc7618e17f47d19d8f1130016aca5f5a"} Nov 24 14:25:47 crc kubenswrapper[4790]: I1124 14:25:47.221203 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hqzm4" podStartSLOduration=3.506635668 podStartE2EDuration="6.221176669s" podCreationTimestamp="2025-11-24 14:25:41 +0000 UTC" firstStartedPulling="2025-11-24 14:25:43.154493764 +0000 UTC m=+4391.534387436" lastFinishedPulling="2025-11-24 14:25:45.869034765 +0000 UTC m=+4394.248928437" observedRunningTime="2025-11-24 14:25:47.214457467 +0000 UTC m=+4395.594351139" watchObservedRunningTime="2025-11-24 14:25:47.221176669 +0000 UTC m=+4395.601070351" Nov 24 14:25:51 crc kubenswrapper[4790]: I1124 14:25:51.434253 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hqzm4" Nov 24 14:25:51 crc kubenswrapper[4790]: I1124 14:25:51.435106 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hqzm4" Nov 24 14:25:51 crc kubenswrapper[4790]: I1124 14:25:51.516908 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hqzm4" Nov 24 14:25:52 crc kubenswrapper[4790]: I1124 14:25:52.329670 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hqzm4" Nov 24 14:25:52 crc kubenswrapper[4790]: I1124 14:25:52.385193 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hqzm4"] Nov 24 14:25:54 crc kubenswrapper[4790]: I1124 14:25:54.259587 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hqzm4" podUID="ce5b3273-5304-4d77-8d49-398174e348de" containerName="registry-server" containerID="cri-o://30b3885384f9d9637dd99a093f600339bc7618e17f47d19d8f1130016aca5f5a" gracePeriod=2 Nov 24 14:25:54 crc kubenswrapper[4790]: I1124 14:25:54.900430 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hqzm4" Nov 24 14:25:54 crc kubenswrapper[4790]: I1124 14:25:54.938960 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce5b3273-5304-4d77-8d49-398174e348de-utilities\") pod \"ce5b3273-5304-4d77-8d49-398174e348de\" (UID: \"ce5b3273-5304-4d77-8d49-398174e348de\") " Nov 24 14:25:54 crc kubenswrapper[4790]: I1124 14:25:54.939030 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khpp6\" (UniqueName: \"kubernetes.io/projected/ce5b3273-5304-4d77-8d49-398174e348de-kube-api-access-khpp6\") pod \"ce5b3273-5304-4d77-8d49-398174e348de\" (UID: \"ce5b3273-5304-4d77-8d49-398174e348de\") " Nov 24 14:25:54 crc kubenswrapper[4790]: I1124 14:25:54.939056 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce5b3273-5304-4d77-8d49-398174e348de-catalog-content\") pod \"ce5b3273-5304-4d77-8d49-398174e348de\" (UID: \"ce5b3273-5304-4d77-8d49-398174e348de\") " Nov 24 14:25:54 crc kubenswrapper[4790]: I1124 14:25:54.940114 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce5b3273-5304-4d77-8d49-398174e348de-utilities" (OuterVolumeSpecName: "utilities") pod "ce5b3273-5304-4d77-8d49-398174e348de" (UID: "ce5b3273-5304-4d77-8d49-398174e348de"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:25:54 crc kubenswrapper[4790]: I1124 14:25:54.944653 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce5b3273-5304-4d77-8d49-398174e348de-kube-api-access-khpp6" (OuterVolumeSpecName: "kube-api-access-khpp6") pod "ce5b3273-5304-4d77-8d49-398174e348de" (UID: "ce5b3273-5304-4d77-8d49-398174e348de"). InnerVolumeSpecName "kube-api-access-khpp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:25:54 crc kubenswrapper[4790]: I1124 14:25:54.999274 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce5b3273-5304-4d77-8d49-398174e348de-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ce5b3273-5304-4d77-8d49-398174e348de" (UID: "ce5b3273-5304-4d77-8d49-398174e348de"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:25:55 crc kubenswrapper[4790]: I1124 14:25:55.039964 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce5b3273-5304-4d77-8d49-398174e348de-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:25:55 crc kubenswrapper[4790]: I1124 14:25:55.040043 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khpp6\" (UniqueName: \"kubernetes.io/projected/ce5b3273-5304-4d77-8d49-398174e348de-kube-api-access-khpp6\") on node \"crc\" DevicePath \"\"" Nov 24 14:25:55 crc kubenswrapper[4790]: I1124 14:25:55.040066 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce5b3273-5304-4d77-8d49-398174e348de-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:25:55 crc kubenswrapper[4790]: I1124 14:25:55.268639 4790 generic.go:334] "Generic (PLEG): container finished" podID="ce5b3273-5304-4d77-8d49-398174e348de" containerID="30b3885384f9d9637dd99a093f600339bc7618e17f47d19d8f1130016aca5f5a" exitCode=0 Nov 24 14:25:55 crc kubenswrapper[4790]: I1124 14:25:55.268697 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hqzm4" event={"ID":"ce5b3273-5304-4d77-8d49-398174e348de","Type":"ContainerDied","Data":"30b3885384f9d9637dd99a093f600339bc7618e17f47d19d8f1130016aca5f5a"} Nov 24 14:25:55 crc kubenswrapper[4790]: I1124 14:25:55.268746 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hqzm4" event={"ID":"ce5b3273-5304-4d77-8d49-398174e348de","Type":"ContainerDied","Data":"883f2f9ddbaca2e671a332cb63fff3796c3ba745e5243a384d7a4404cf9b1ad3"} Nov 24 14:25:55 crc kubenswrapper[4790]: I1124 14:25:55.268789 4790 scope.go:117] "RemoveContainer" containerID="30b3885384f9d9637dd99a093f600339bc7618e17f47d19d8f1130016aca5f5a" Nov 24 14:25:55 crc kubenswrapper[4790]: I1124 14:25:55.269297 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hqzm4" Nov 24 14:25:55 crc kubenswrapper[4790]: I1124 14:25:55.285426 4790 scope.go:117] "RemoveContainer" containerID="4ed864e95c4bfbb17015091f17d2b9b766838cf6487b00c48dc5c4ae4123a5bd" Nov 24 14:25:55 crc kubenswrapper[4790]: I1124 14:25:55.299970 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hqzm4"] Nov 24 14:25:55 crc kubenswrapper[4790]: I1124 14:25:55.309138 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hqzm4"] Nov 24 14:25:55 crc kubenswrapper[4790]: I1124 14:25:55.323149 4790 scope.go:117] "RemoveContainer" containerID="a4905a1abd2ec0423542afd505c1f529460c4f4dd1e044264ef953a400628c38" Nov 24 14:25:55 crc kubenswrapper[4790]: I1124 14:25:55.339835 4790 scope.go:117] "RemoveContainer" containerID="30b3885384f9d9637dd99a093f600339bc7618e17f47d19d8f1130016aca5f5a" Nov 24 14:25:55 crc kubenswrapper[4790]: E1124 14:25:55.340296 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30b3885384f9d9637dd99a093f600339bc7618e17f47d19d8f1130016aca5f5a\": container with ID starting with 30b3885384f9d9637dd99a093f600339bc7618e17f47d19d8f1130016aca5f5a not found: ID does not exist" containerID="30b3885384f9d9637dd99a093f600339bc7618e17f47d19d8f1130016aca5f5a" Nov 24 14:25:55 crc kubenswrapper[4790]: I1124 14:25:55.340339 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30b3885384f9d9637dd99a093f600339bc7618e17f47d19d8f1130016aca5f5a"} err="failed to get container status \"30b3885384f9d9637dd99a093f600339bc7618e17f47d19d8f1130016aca5f5a\": rpc error: code = NotFound desc = could not find container \"30b3885384f9d9637dd99a093f600339bc7618e17f47d19d8f1130016aca5f5a\": container with ID starting with 30b3885384f9d9637dd99a093f600339bc7618e17f47d19d8f1130016aca5f5a not found: ID does not exist" Nov 24 14:25:55 crc kubenswrapper[4790]: I1124 14:25:55.340365 4790 scope.go:117] "RemoveContainer" containerID="4ed864e95c4bfbb17015091f17d2b9b766838cf6487b00c48dc5c4ae4123a5bd" Nov 24 14:25:55 crc kubenswrapper[4790]: E1124 14:25:55.340674 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ed864e95c4bfbb17015091f17d2b9b766838cf6487b00c48dc5c4ae4123a5bd\": container with ID starting with 4ed864e95c4bfbb17015091f17d2b9b766838cf6487b00c48dc5c4ae4123a5bd not found: ID does not exist" containerID="4ed864e95c4bfbb17015091f17d2b9b766838cf6487b00c48dc5c4ae4123a5bd" Nov 24 14:25:55 crc kubenswrapper[4790]: I1124 14:25:55.340694 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ed864e95c4bfbb17015091f17d2b9b766838cf6487b00c48dc5c4ae4123a5bd"} err="failed to get container status \"4ed864e95c4bfbb17015091f17d2b9b766838cf6487b00c48dc5c4ae4123a5bd\": rpc error: code = NotFound desc = could not find container \"4ed864e95c4bfbb17015091f17d2b9b766838cf6487b00c48dc5c4ae4123a5bd\": container with ID starting with 4ed864e95c4bfbb17015091f17d2b9b766838cf6487b00c48dc5c4ae4123a5bd not found: ID does not exist" Nov 24 14:25:55 crc kubenswrapper[4790]: I1124 14:25:55.340707 4790 scope.go:117] "RemoveContainer" containerID="a4905a1abd2ec0423542afd505c1f529460c4f4dd1e044264ef953a400628c38" Nov 24 14:25:55 crc kubenswrapper[4790]: E1124 14:25:55.341120 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4905a1abd2ec0423542afd505c1f529460c4f4dd1e044264ef953a400628c38\": container with ID starting with a4905a1abd2ec0423542afd505c1f529460c4f4dd1e044264ef953a400628c38 not found: ID does not exist" containerID="a4905a1abd2ec0423542afd505c1f529460c4f4dd1e044264ef953a400628c38" Nov 24 14:25:55 crc kubenswrapper[4790]: I1124 14:25:55.341184 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4905a1abd2ec0423542afd505c1f529460c4f4dd1e044264ef953a400628c38"} err="failed to get container status \"a4905a1abd2ec0423542afd505c1f529460c4f4dd1e044264ef953a400628c38\": rpc error: code = NotFound desc = could not find container \"a4905a1abd2ec0423542afd505c1f529460c4f4dd1e044264ef953a400628c38\": container with ID starting with a4905a1abd2ec0423542afd505c1f529460c4f4dd1e044264ef953a400628c38 not found: ID does not exist" Nov 24 14:25:56 crc kubenswrapper[4790]: I1124 14:25:56.327258 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce5b3273-5304-4d77-8d49-398174e348de" path="/var/lib/kubelet/pods/ce5b3273-5304-4d77-8d49-398174e348de/volumes" Nov 24 14:27:30 crc kubenswrapper[4790]: I1124 14:27:30.404083 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-kpg6g"] Nov 24 14:27:30 crc kubenswrapper[4790]: I1124 14:27:30.408365 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-kpg6g"] Nov 24 14:27:30 crc kubenswrapper[4790]: I1124 14:27:30.532500 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-6dgkt"] Nov 24 14:27:30 crc kubenswrapper[4790]: E1124 14:27:30.532869 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce5b3273-5304-4d77-8d49-398174e348de" containerName="extract-utilities" Nov 24 14:27:30 crc kubenswrapper[4790]: I1124 14:27:30.532907 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce5b3273-5304-4d77-8d49-398174e348de" containerName="extract-utilities" Nov 24 14:27:30 crc kubenswrapper[4790]: E1124 14:27:30.532940 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce5b3273-5304-4d77-8d49-398174e348de" containerName="registry-server" Nov 24 14:27:30 crc kubenswrapper[4790]: I1124 14:27:30.532951 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce5b3273-5304-4d77-8d49-398174e348de" containerName="registry-server" Nov 24 14:27:30 crc kubenswrapper[4790]: E1124 14:27:30.532964 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce5b3273-5304-4d77-8d49-398174e348de" containerName="extract-content" Nov 24 14:27:30 crc kubenswrapper[4790]: I1124 14:27:30.532971 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce5b3273-5304-4d77-8d49-398174e348de" containerName="extract-content" Nov 24 14:27:30 crc kubenswrapper[4790]: I1124 14:27:30.533147 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce5b3273-5304-4d77-8d49-398174e348de" containerName="registry-server" Nov 24 14:27:30 crc kubenswrapper[4790]: I1124 14:27:30.533707 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-6dgkt" Nov 24 14:27:30 crc kubenswrapper[4790]: I1124 14:27:30.537428 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Nov 24 14:27:30 crc kubenswrapper[4790]: I1124 14:27:30.538003 4790 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-xfpzp" Nov 24 14:27:30 crc kubenswrapper[4790]: I1124 14:27:30.540051 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Nov 24 14:27:30 crc kubenswrapper[4790]: I1124 14:27:30.540202 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Nov 24 14:27:30 crc kubenswrapper[4790]: I1124 14:27:30.542923 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-6dgkt"] Nov 24 14:27:30 crc kubenswrapper[4790]: I1124 14:27:30.673527 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/793f055c-5f6e-453c-9e46-1afc107747cc-node-mnt\") pod \"crc-storage-crc-6dgkt\" (UID: \"793f055c-5f6e-453c-9e46-1afc107747cc\") " pod="crc-storage/crc-storage-crc-6dgkt" Nov 24 14:27:30 crc kubenswrapper[4790]: I1124 14:27:30.673583 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/793f055c-5f6e-453c-9e46-1afc107747cc-crc-storage\") pod \"crc-storage-crc-6dgkt\" (UID: \"793f055c-5f6e-453c-9e46-1afc107747cc\") " pod="crc-storage/crc-storage-crc-6dgkt" Nov 24 14:27:30 crc kubenswrapper[4790]: I1124 14:27:30.673605 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ts67f\" (UniqueName: \"kubernetes.io/projected/793f055c-5f6e-453c-9e46-1afc107747cc-kube-api-access-ts67f\") pod \"crc-storage-crc-6dgkt\" (UID: \"793f055c-5f6e-453c-9e46-1afc107747cc\") " pod="crc-storage/crc-storage-crc-6dgkt" Nov 24 14:27:30 crc kubenswrapper[4790]: I1124 14:27:30.775134 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/793f055c-5f6e-453c-9e46-1afc107747cc-node-mnt\") pod \"crc-storage-crc-6dgkt\" (UID: \"793f055c-5f6e-453c-9e46-1afc107747cc\") " pod="crc-storage/crc-storage-crc-6dgkt" Nov 24 14:27:30 crc kubenswrapper[4790]: I1124 14:27:30.775246 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/793f055c-5f6e-453c-9e46-1afc107747cc-crc-storage\") pod \"crc-storage-crc-6dgkt\" (UID: \"793f055c-5f6e-453c-9e46-1afc107747cc\") " pod="crc-storage/crc-storage-crc-6dgkt" Nov 24 14:27:30 crc kubenswrapper[4790]: I1124 14:27:30.775299 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ts67f\" (UniqueName: \"kubernetes.io/projected/793f055c-5f6e-453c-9e46-1afc107747cc-kube-api-access-ts67f\") pod \"crc-storage-crc-6dgkt\" (UID: \"793f055c-5f6e-453c-9e46-1afc107747cc\") " pod="crc-storage/crc-storage-crc-6dgkt" Nov 24 14:27:30 crc kubenswrapper[4790]: I1124 14:27:30.775469 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/793f055c-5f6e-453c-9e46-1afc107747cc-node-mnt\") pod \"crc-storage-crc-6dgkt\" (UID: \"793f055c-5f6e-453c-9e46-1afc107747cc\") " pod="crc-storage/crc-storage-crc-6dgkt" Nov 24 14:27:30 crc kubenswrapper[4790]: I1124 14:27:30.776576 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/793f055c-5f6e-453c-9e46-1afc107747cc-crc-storage\") pod \"crc-storage-crc-6dgkt\" (UID: \"793f055c-5f6e-453c-9e46-1afc107747cc\") " pod="crc-storage/crc-storage-crc-6dgkt" Nov 24 14:27:30 crc kubenswrapper[4790]: I1124 14:27:30.802238 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ts67f\" (UniqueName: \"kubernetes.io/projected/793f055c-5f6e-453c-9e46-1afc107747cc-kube-api-access-ts67f\") pod \"crc-storage-crc-6dgkt\" (UID: \"793f055c-5f6e-453c-9e46-1afc107747cc\") " pod="crc-storage/crc-storage-crc-6dgkt" Nov 24 14:27:30 crc kubenswrapper[4790]: I1124 14:27:30.855571 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-6dgkt" Nov 24 14:27:31 crc kubenswrapper[4790]: I1124 14:27:31.102115 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-6dgkt"] Nov 24 14:27:32 crc kubenswrapper[4790]: I1124 14:27:32.071916 4790 generic.go:334] "Generic (PLEG): container finished" podID="793f055c-5f6e-453c-9e46-1afc107747cc" containerID="327852f78d1c40248ad2a4e37cb155535377c520bc90cb223080cf976bf095fa" exitCode=0 Nov 24 14:27:32 crc kubenswrapper[4790]: I1124 14:27:32.072092 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-6dgkt" event={"ID":"793f055c-5f6e-453c-9e46-1afc107747cc","Type":"ContainerDied","Data":"327852f78d1c40248ad2a4e37cb155535377c520bc90cb223080cf976bf095fa"} Nov 24 14:27:32 crc kubenswrapper[4790]: I1124 14:27:32.072193 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-6dgkt" event={"ID":"793f055c-5f6e-453c-9e46-1afc107747cc","Type":"ContainerStarted","Data":"fcd34f45b8ff3475745e5d8bc0d94a4054545e645de8a39af7974e7b47233878"} Nov 24 14:27:32 crc kubenswrapper[4790]: I1124 14:27:32.327707 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5" path="/var/lib/kubelet/pods/412bae7a-3f92-47ce-bd17-5f0e5fd5c1e5/volumes" Nov 24 14:27:33 crc kubenswrapper[4790]: I1124 14:27:33.401785 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-6dgkt" Nov 24 14:27:33 crc kubenswrapper[4790]: I1124 14:27:33.515394 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/793f055c-5f6e-453c-9e46-1afc107747cc-node-mnt\") pod \"793f055c-5f6e-453c-9e46-1afc107747cc\" (UID: \"793f055c-5f6e-453c-9e46-1afc107747cc\") " Nov 24 14:27:33 crc kubenswrapper[4790]: I1124 14:27:33.515487 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ts67f\" (UniqueName: \"kubernetes.io/projected/793f055c-5f6e-453c-9e46-1afc107747cc-kube-api-access-ts67f\") pod \"793f055c-5f6e-453c-9e46-1afc107747cc\" (UID: \"793f055c-5f6e-453c-9e46-1afc107747cc\") " Nov 24 14:27:33 crc kubenswrapper[4790]: I1124 14:27:33.515523 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/793f055c-5f6e-453c-9e46-1afc107747cc-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "793f055c-5f6e-453c-9e46-1afc107747cc" (UID: "793f055c-5f6e-453c-9e46-1afc107747cc"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:27:33 crc kubenswrapper[4790]: I1124 14:27:33.515551 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/793f055c-5f6e-453c-9e46-1afc107747cc-crc-storage\") pod \"793f055c-5f6e-453c-9e46-1afc107747cc\" (UID: \"793f055c-5f6e-453c-9e46-1afc107747cc\") " Nov 24 14:27:33 crc kubenswrapper[4790]: I1124 14:27:33.516085 4790 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/793f055c-5f6e-453c-9e46-1afc107747cc-node-mnt\") on node \"crc\" DevicePath \"\"" Nov 24 14:27:33 crc kubenswrapper[4790]: I1124 14:27:33.522866 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/793f055c-5f6e-453c-9e46-1afc107747cc-kube-api-access-ts67f" (OuterVolumeSpecName: "kube-api-access-ts67f") pod "793f055c-5f6e-453c-9e46-1afc107747cc" (UID: "793f055c-5f6e-453c-9e46-1afc107747cc"). InnerVolumeSpecName "kube-api-access-ts67f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:27:33 crc kubenswrapper[4790]: I1124 14:27:33.539783 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/793f055c-5f6e-453c-9e46-1afc107747cc-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "793f055c-5f6e-453c-9e46-1afc107747cc" (UID: "793f055c-5f6e-453c-9e46-1afc107747cc"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:27:33 crc kubenswrapper[4790]: I1124 14:27:33.617747 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ts67f\" (UniqueName: \"kubernetes.io/projected/793f055c-5f6e-453c-9e46-1afc107747cc-kube-api-access-ts67f\") on node \"crc\" DevicePath \"\"" Nov 24 14:27:33 crc kubenswrapper[4790]: I1124 14:27:33.617809 4790 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/793f055c-5f6e-453c-9e46-1afc107747cc-crc-storage\") on node \"crc\" DevicePath \"\"" Nov 24 14:27:34 crc kubenswrapper[4790]: I1124 14:27:34.092285 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-6dgkt" event={"ID":"793f055c-5f6e-453c-9e46-1afc107747cc","Type":"ContainerDied","Data":"fcd34f45b8ff3475745e5d8bc0d94a4054545e645de8a39af7974e7b47233878"} Nov 24 14:27:34 crc kubenswrapper[4790]: I1124 14:27:34.092329 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fcd34f45b8ff3475745e5d8bc0d94a4054545e645de8a39af7974e7b47233878" Nov 24 14:27:34 crc kubenswrapper[4790]: I1124 14:27:34.092345 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-6dgkt" Nov 24 14:27:35 crc kubenswrapper[4790]: I1124 14:27:35.508928 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-6dgkt"] Nov 24 14:27:35 crc kubenswrapper[4790]: I1124 14:27:35.515361 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-6dgkt"] Nov 24 14:27:35 crc kubenswrapper[4790]: I1124 14:27:35.639160 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-9bmbz"] Nov 24 14:27:35 crc kubenswrapper[4790]: E1124 14:27:35.639484 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="793f055c-5f6e-453c-9e46-1afc107747cc" containerName="storage" Nov 24 14:27:35 crc kubenswrapper[4790]: I1124 14:27:35.639498 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="793f055c-5f6e-453c-9e46-1afc107747cc" containerName="storage" Nov 24 14:27:35 crc kubenswrapper[4790]: I1124 14:27:35.640049 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="793f055c-5f6e-453c-9e46-1afc107747cc" containerName="storage" Nov 24 14:27:35 crc kubenswrapper[4790]: I1124 14:27:35.640988 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-9bmbz" Nov 24 14:27:35 crc kubenswrapper[4790]: I1124 14:27:35.645929 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Nov 24 14:27:35 crc kubenswrapper[4790]: I1124 14:27:35.646161 4790 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-xfpzp" Nov 24 14:27:35 crc kubenswrapper[4790]: I1124 14:27:35.646280 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Nov 24 14:27:35 crc kubenswrapper[4790]: I1124 14:27:35.646315 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Nov 24 14:27:35 crc kubenswrapper[4790]: I1124 14:27:35.652372 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtr7h\" (UniqueName: \"kubernetes.io/projected/01503096-6d72-49fc-9a8f-efa422e9d3b0-kube-api-access-vtr7h\") pod \"crc-storage-crc-9bmbz\" (UID: \"01503096-6d72-49fc-9a8f-efa422e9d3b0\") " pod="crc-storage/crc-storage-crc-9bmbz" Nov 24 14:27:35 crc kubenswrapper[4790]: I1124 14:27:35.652415 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/01503096-6d72-49fc-9a8f-efa422e9d3b0-crc-storage\") pod \"crc-storage-crc-9bmbz\" (UID: \"01503096-6d72-49fc-9a8f-efa422e9d3b0\") " pod="crc-storage/crc-storage-crc-9bmbz" Nov 24 14:27:35 crc kubenswrapper[4790]: I1124 14:27:35.652492 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/01503096-6d72-49fc-9a8f-efa422e9d3b0-node-mnt\") pod \"crc-storage-crc-9bmbz\" (UID: \"01503096-6d72-49fc-9a8f-efa422e9d3b0\") " pod="crc-storage/crc-storage-crc-9bmbz" Nov 24 14:27:35 crc kubenswrapper[4790]: I1124 14:27:35.658109 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-9bmbz"] Nov 24 14:27:35 crc kubenswrapper[4790]: I1124 14:27:35.753756 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtr7h\" (UniqueName: \"kubernetes.io/projected/01503096-6d72-49fc-9a8f-efa422e9d3b0-kube-api-access-vtr7h\") pod \"crc-storage-crc-9bmbz\" (UID: \"01503096-6d72-49fc-9a8f-efa422e9d3b0\") " pod="crc-storage/crc-storage-crc-9bmbz" Nov 24 14:27:35 crc kubenswrapper[4790]: I1124 14:27:35.753816 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/01503096-6d72-49fc-9a8f-efa422e9d3b0-crc-storage\") pod \"crc-storage-crc-9bmbz\" (UID: \"01503096-6d72-49fc-9a8f-efa422e9d3b0\") " pod="crc-storage/crc-storage-crc-9bmbz" Nov 24 14:27:35 crc kubenswrapper[4790]: I1124 14:27:35.753990 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/01503096-6d72-49fc-9a8f-efa422e9d3b0-node-mnt\") pod \"crc-storage-crc-9bmbz\" (UID: \"01503096-6d72-49fc-9a8f-efa422e9d3b0\") " pod="crc-storage/crc-storage-crc-9bmbz" Nov 24 14:27:35 crc kubenswrapper[4790]: I1124 14:27:35.754370 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/01503096-6d72-49fc-9a8f-efa422e9d3b0-node-mnt\") pod \"crc-storage-crc-9bmbz\" (UID: \"01503096-6d72-49fc-9a8f-efa422e9d3b0\") " pod="crc-storage/crc-storage-crc-9bmbz" Nov 24 14:27:35 crc kubenswrapper[4790]: I1124 14:27:35.754814 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/01503096-6d72-49fc-9a8f-efa422e9d3b0-crc-storage\") pod \"crc-storage-crc-9bmbz\" (UID: \"01503096-6d72-49fc-9a8f-efa422e9d3b0\") " pod="crc-storage/crc-storage-crc-9bmbz" Nov 24 14:27:35 crc kubenswrapper[4790]: I1124 14:27:35.776158 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtr7h\" (UniqueName: \"kubernetes.io/projected/01503096-6d72-49fc-9a8f-efa422e9d3b0-kube-api-access-vtr7h\") pod \"crc-storage-crc-9bmbz\" (UID: \"01503096-6d72-49fc-9a8f-efa422e9d3b0\") " pod="crc-storage/crc-storage-crc-9bmbz" Nov 24 14:27:35 crc kubenswrapper[4790]: I1124 14:27:35.971167 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-9bmbz" Nov 24 14:27:36 crc kubenswrapper[4790]: I1124 14:27:36.218023 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-9bmbz"] Nov 24 14:27:36 crc kubenswrapper[4790]: W1124 14:27:36.222338 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01503096_6d72_49fc_9a8f_efa422e9d3b0.slice/crio-d2d002c50d460057f018dffdfea6a26da12ba75ab6392bac5fba1429ccf518c5 WatchSource:0}: Error finding container d2d002c50d460057f018dffdfea6a26da12ba75ab6392bac5fba1429ccf518c5: Status 404 returned error can't find the container with id d2d002c50d460057f018dffdfea6a26da12ba75ab6392bac5fba1429ccf518c5 Nov 24 14:27:36 crc kubenswrapper[4790]: I1124 14:27:36.323366 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="793f055c-5f6e-453c-9e46-1afc107747cc" path="/var/lib/kubelet/pods/793f055c-5f6e-453c-9e46-1afc107747cc/volumes" Nov 24 14:27:37 crc kubenswrapper[4790]: I1124 14:27:37.125867 4790 generic.go:334] "Generic (PLEG): container finished" podID="01503096-6d72-49fc-9a8f-efa422e9d3b0" containerID="3ab8380803bdd95d99465d160209dcc6a52d44e274ae0892f63b6fdbc051383c" exitCode=0 Nov 24 14:27:37 crc kubenswrapper[4790]: I1124 14:27:37.125985 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-9bmbz" event={"ID":"01503096-6d72-49fc-9a8f-efa422e9d3b0","Type":"ContainerDied","Data":"3ab8380803bdd95d99465d160209dcc6a52d44e274ae0892f63b6fdbc051383c"} Nov 24 14:27:37 crc kubenswrapper[4790]: I1124 14:27:37.126878 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-9bmbz" event={"ID":"01503096-6d72-49fc-9a8f-efa422e9d3b0","Type":"ContainerStarted","Data":"d2d002c50d460057f018dffdfea6a26da12ba75ab6392bac5fba1429ccf518c5"} Nov 24 14:27:38 crc kubenswrapper[4790]: I1124 14:27:38.524301 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-9bmbz" Nov 24 14:27:38 crc kubenswrapper[4790]: I1124 14:27:38.697986 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtr7h\" (UniqueName: \"kubernetes.io/projected/01503096-6d72-49fc-9a8f-efa422e9d3b0-kube-api-access-vtr7h\") pod \"01503096-6d72-49fc-9a8f-efa422e9d3b0\" (UID: \"01503096-6d72-49fc-9a8f-efa422e9d3b0\") " Nov 24 14:27:38 crc kubenswrapper[4790]: I1124 14:27:38.698152 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/01503096-6d72-49fc-9a8f-efa422e9d3b0-crc-storage\") pod \"01503096-6d72-49fc-9a8f-efa422e9d3b0\" (UID: \"01503096-6d72-49fc-9a8f-efa422e9d3b0\") " Nov 24 14:27:38 crc kubenswrapper[4790]: I1124 14:27:38.698224 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/01503096-6d72-49fc-9a8f-efa422e9d3b0-node-mnt\") pod \"01503096-6d72-49fc-9a8f-efa422e9d3b0\" (UID: \"01503096-6d72-49fc-9a8f-efa422e9d3b0\") " Nov 24 14:27:38 crc kubenswrapper[4790]: I1124 14:27:38.698451 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/01503096-6d72-49fc-9a8f-efa422e9d3b0-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "01503096-6d72-49fc-9a8f-efa422e9d3b0" (UID: "01503096-6d72-49fc-9a8f-efa422e9d3b0"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:27:38 crc kubenswrapper[4790]: I1124 14:27:38.699201 4790 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/01503096-6d72-49fc-9a8f-efa422e9d3b0-node-mnt\") on node \"crc\" DevicePath \"\"" Nov 24 14:27:38 crc kubenswrapper[4790]: I1124 14:27:38.707541 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01503096-6d72-49fc-9a8f-efa422e9d3b0-kube-api-access-vtr7h" (OuterVolumeSpecName: "kube-api-access-vtr7h") pod "01503096-6d72-49fc-9a8f-efa422e9d3b0" (UID: "01503096-6d72-49fc-9a8f-efa422e9d3b0"). InnerVolumeSpecName "kube-api-access-vtr7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:27:38 crc kubenswrapper[4790]: I1124 14:27:38.722790 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01503096-6d72-49fc-9a8f-efa422e9d3b0-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "01503096-6d72-49fc-9a8f-efa422e9d3b0" (UID: "01503096-6d72-49fc-9a8f-efa422e9d3b0"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:27:38 crc kubenswrapper[4790]: I1124 14:27:38.801664 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtr7h\" (UniqueName: \"kubernetes.io/projected/01503096-6d72-49fc-9a8f-efa422e9d3b0-kube-api-access-vtr7h\") on node \"crc\" DevicePath \"\"" Nov 24 14:27:38 crc kubenswrapper[4790]: I1124 14:27:38.801732 4790 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/01503096-6d72-49fc-9a8f-efa422e9d3b0-crc-storage\") on node \"crc\" DevicePath \"\"" Nov 24 14:27:39 crc kubenswrapper[4790]: I1124 14:27:39.148389 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-9bmbz" event={"ID":"01503096-6d72-49fc-9a8f-efa422e9d3b0","Type":"ContainerDied","Data":"d2d002c50d460057f018dffdfea6a26da12ba75ab6392bac5fba1429ccf518c5"} Nov 24 14:27:39 crc kubenswrapper[4790]: I1124 14:27:39.148428 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2d002c50d460057f018dffdfea6a26da12ba75ab6392bac5fba1429ccf518c5" Nov 24 14:27:39 crc kubenswrapper[4790]: I1124 14:27:39.148475 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-9bmbz" Nov 24 14:27:43 crc kubenswrapper[4790]: I1124 14:27:43.939643 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:27:43 crc kubenswrapper[4790]: I1124 14:27:43.940574 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:27:46 crc kubenswrapper[4790]: I1124 14:27:46.481806 4790 scope.go:117] "RemoveContainer" containerID="64bc3c80920485efa9acd6d9ada588bc80cbf57817a1a94698a53a2910fbf1d8" Nov 24 14:28:13 crc kubenswrapper[4790]: I1124 14:28:13.939102 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:28:13 crc kubenswrapper[4790]: I1124 14:28:13.939750 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:28:43 crc kubenswrapper[4790]: I1124 14:28:43.938534 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:28:43 crc kubenswrapper[4790]: I1124 14:28:43.939366 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:28:43 crc kubenswrapper[4790]: I1124 14:28:43.939443 4790 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 14:28:43 crc kubenswrapper[4790]: I1124 14:28:43.940422 4790 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"571bd0c27e6ad4756f7116ed8b5a40d06301f3d85436cea369594fdb322154a7"} pod="openshift-machine-config-operator/machine-config-daemon-xz49t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:28:43 crc kubenswrapper[4790]: I1124 14:28:43.940524 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" containerID="cri-o://571bd0c27e6ad4756f7116ed8b5a40d06301f3d85436cea369594fdb322154a7" gracePeriod=600 Nov 24 14:28:44 crc kubenswrapper[4790]: E1124 14:28:44.079608 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:28:44 crc kubenswrapper[4790]: I1124 14:28:44.749393 4790 generic.go:334] "Generic (PLEG): container finished" podID="0d73b133-48f0-455f-8f6a-742e633f631a" containerID="571bd0c27e6ad4756f7116ed8b5a40d06301f3d85436cea369594fdb322154a7" exitCode=0 Nov 24 14:28:44 crc kubenswrapper[4790]: I1124 14:28:44.749467 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerDied","Data":"571bd0c27e6ad4756f7116ed8b5a40d06301f3d85436cea369594fdb322154a7"} Nov 24 14:28:44 crc kubenswrapper[4790]: I1124 14:28:44.749510 4790 scope.go:117] "RemoveContainer" containerID="b25648b43123ad262f034d5f38126a0ec7bbda6c1db0e5dea770deb92fb3c707" Nov 24 14:28:44 crc kubenswrapper[4790]: I1124 14:28:44.750318 4790 scope.go:117] "RemoveContainer" containerID="571bd0c27e6ad4756f7116ed8b5a40d06301f3d85436cea369594fdb322154a7" Nov 24 14:28:44 crc kubenswrapper[4790]: E1124 14:28:44.750804 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:28:59 crc kubenswrapper[4790]: I1124 14:28:59.315425 4790 scope.go:117] "RemoveContainer" containerID="571bd0c27e6ad4756f7116ed8b5a40d06301f3d85436cea369594fdb322154a7" Nov 24 14:28:59 crc kubenswrapper[4790]: E1124 14:28:59.316572 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:29:11 crc kubenswrapper[4790]: I1124 14:29:11.315344 4790 scope.go:117] "RemoveContainer" containerID="571bd0c27e6ad4756f7116ed8b5a40d06301f3d85436cea369594fdb322154a7" Nov 24 14:29:11 crc kubenswrapper[4790]: E1124 14:29:11.316457 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:29:25 crc kubenswrapper[4790]: I1124 14:29:25.315083 4790 scope.go:117] "RemoveContainer" containerID="571bd0c27e6ad4756f7116ed8b5a40d06301f3d85436cea369594fdb322154a7" Nov 24 14:29:25 crc kubenswrapper[4790]: E1124 14:29:25.316119 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:29:40 crc kubenswrapper[4790]: I1124 14:29:40.314229 4790 scope.go:117] "RemoveContainer" containerID="571bd0c27e6ad4756f7116ed8b5a40d06301f3d85436cea369594fdb322154a7" Nov 24 14:29:40 crc kubenswrapper[4790]: E1124 14:29:40.314905 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:29:51 crc kubenswrapper[4790]: I1124 14:29:51.314869 4790 scope.go:117] "RemoveContainer" containerID="571bd0c27e6ad4756f7116ed8b5a40d06301f3d85436cea369594fdb322154a7" Nov 24 14:29:51 crc kubenswrapper[4790]: E1124 14:29:51.315793 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:30:00 crc kubenswrapper[4790]: I1124 14:30:00.160618 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399910-j9csk"] Nov 24 14:30:00 crc kubenswrapper[4790]: E1124 14:30:00.161855 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01503096-6d72-49fc-9a8f-efa422e9d3b0" containerName="storage" Nov 24 14:30:00 crc kubenswrapper[4790]: I1124 14:30:00.161878 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="01503096-6d72-49fc-9a8f-efa422e9d3b0" containerName="storage" Nov 24 14:30:00 crc kubenswrapper[4790]: I1124 14:30:00.162171 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="01503096-6d72-49fc-9a8f-efa422e9d3b0" containerName="storage" Nov 24 14:30:00 crc kubenswrapper[4790]: I1124 14:30:00.163055 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-j9csk" Nov 24 14:30:00 crc kubenswrapper[4790]: I1124 14:30:00.165690 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 14:30:00 crc kubenswrapper[4790]: I1124 14:30:00.165835 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 14:30:00 crc kubenswrapper[4790]: I1124 14:30:00.185029 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399910-j9csk"] Nov 24 14:30:00 crc kubenswrapper[4790]: I1124 14:30:00.277357 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qrhh\" (UniqueName: \"kubernetes.io/projected/39c253b2-56c2-46c1-bafd-587d40437a2d-kube-api-access-7qrhh\") pod \"collect-profiles-29399910-j9csk\" (UID: \"39c253b2-56c2-46c1-bafd-587d40437a2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-j9csk" Nov 24 14:30:00 crc kubenswrapper[4790]: I1124 14:30:00.277614 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/39c253b2-56c2-46c1-bafd-587d40437a2d-secret-volume\") pod \"collect-profiles-29399910-j9csk\" (UID: \"39c253b2-56c2-46c1-bafd-587d40437a2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-j9csk" Nov 24 14:30:00 crc kubenswrapper[4790]: I1124 14:30:00.277672 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/39c253b2-56c2-46c1-bafd-587d40437a2d-config-volume\") pod \"collect-profiles-29399910-j9csk\" (UID: \"39c253b2-56c2-46c1-bafd-587d40437a2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-j9csk" Nov 24 14:30:00 crc kubenswrapper[4790]: I1124 14:30:00.384213 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/39c253b2-56c2-46c1-bafd-587d40437a2d-secret-volume\") pod \"collect-profiles-29399910-j9csk\" (UID: \"39c253b2-56c2-46c1-bafd-587d40437a2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-j9csk" Nov 24 14:30:00 crc kubenswrapper[4790]: I1124 14:30:00.384338 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/39c253b2-56c2-46c1-bafd-587d40437a2d-config-volume\") pod \"collect-profiles-29399910-j9csk\" (UID: \"39c253b2-56c2-46c1-bafd-587d40437a2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-j9csk" Nov 24 14:30:00 crc kubenswrapper[4790]: I1124 14:30:00.384583 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qrhh\" (UniqueName: \"kubernetes.io/projected/39c253b2-56c2-46c1-bafd-587d40437a2d-kube-api-access-7qrhh\") pod \"collect-profiles-29399910-j9csk\" (UID: \"39c253b2-56c2-46c1-bafd-587d40437a2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-j9csk" Nov 24 14:30:00 crc kubenswrapper[4790]: I1124 14:30:00.392807 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/39c253b2-56c2-46c1-bafd-587d40437a2d-config-volume\") pod \"collect-profiles-29399910-j9csk\" (UID: \"39c253b2-56c2-46c1-bafd-587d40437a2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-j9csk" Nov 24 14:30:00 crc kubenswrapper[4790]: I1124 14:30:00.394926 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/39c253b2-56c2-46c1-bafd-587d40437a2d-secret-volume\") pod \"collect-profiles-29399910-j9csk\" (UID: \"39c253b2-56c2-46c1-bafd-587d40437a2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-j9csk" Nov 24 14:30:00 crc kubenswrapper[4790]: I1124 14:30:00.406734 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qrhh\" (UniqueName: \"kubernetes.io/projected/39c253b2-56c2-46c1-bafd-587d40437a2d-kube-api-access-7qrhh\") pod \"collect-profiles-29399910-j9csk\" (UID: \"39c253b2-56c2-46c1-bafd-587d40437a2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-j9csk" Nov 24 14:30:00 crc kubenswrapper[4790]: I1124 14:30:00.487990 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-j9csk" Nov 24 14:30:00 crc kubenswrapper[4790]: I1124 14:30:00.926385 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399910-j9csk"] Nov 24 14:30:01 crc kubenswrapper[4790]: I1124 14:30:01.417749 4790 generic.go:334] "Generic (PLEG): container finished" podID="39c253b2-56c2-46c1-bafd-587d40437a2d" containerID="6f24b1d494b226385df8a3047b512d4f54717c8d0cd839083dcb2d3da90e23eb" exitCode=0 Nov 24 14:30:01 crc kubenswrapper[4790]: I1124 14:30:01.418753 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-j9csk" event={"ID":"39c253b2-56c2-46c1-bafd-587d40437a2d","Type":"ContainerDied","Data":"6f24b1d494b226385df8a3047b512d4f54717c8d0cd839083dcb2d3da90e23eb"} Nov 24 14:30:01 crc kubenswrapper[4790]: I1124 14:30:01.418862 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-j9csk" event={"ID":"39c253b2-56c2-46c1-bafd-587d40437a2d","Type":"ContainerStarted","Data":"35bc9cfc245f36a4f9c21e9423f765342677926fd989341023177afbf3814413"} Nov 24 14:30:02 crc kubenswrapper[4790]: I1124 14:30:02.939020 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-j9csk" Nov 24 14:30:03 crc kubenswrapper[4790]: I1124 14:30:03.124325 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qrhh\" (UniqueName: \"kubernetes.io/projected/39c253b2-56c2-46c1-bafd-587d40437a2d-kube-api-access-7qrhh\") pod \"39c253b2-56c2-46c1-bafd-587d40437a2d\" (UID: \"39c253b2-56c2-46c1-bafd-587d40437a2d\") " Nov 24 14:30:03 crc kubenswrapper[4790]: I1124 14:30:03.124685 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/39c253b2-56c2-46c1-bafd-587d40437a2d-secret-volume\") pod \"39c253b2-56c2-46c1-bafd-587d40437a2d\" (UID: \"39c253b2-56c2-46c1-bafd-587d40437a2d\") " Nov 24 14:30:03 crc kubenswrapper[4790]: I1124 14:30:03.124858 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/39c253b2-56c2-46c1-bafd-587d40437a2d-config-volume\") pod \"39c253b2-56c2-46c1-bafd-587d40437a2d\" (UID: \"39c253b2-56c2-46c1-bafd-587d40437a2d\") " Nov 24 14:30:03 crc kubenswrapper[4790]: I1124 14:30:03.125615 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39c253b2-56c2-46c1-bafd-587d40437a2d-config-volume" (OuterVolumeSpecName: "config-volume") pod "39c253b2-56c2-46c1-bafd-587d40437a2d" (UID: "39c253b2-56c2-46c1-bafd-587d40437a2d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:30:03 crc kubenswrapper[4790]: I1124 14:30:03.129812 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39c253b2-56c2-46c1-bafd-587d40437a2d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "39c253b2-56c2-46c1-bafd-587d40437a2d" (UID: "39c253b2-56c2-46c1-bafd-587d40437a2d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:30:03 crc kubenswrapper[4790]: I1124 14:30:03.129895 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39c253b2-56c2-46c1-bafd-587d40437a2d-kube-api-access-7qrhh" (OuterVolumeSpecName: "kube-api-access-7qrhh") pod "39c253b2-56c2-46c1-bafd-587d40437a2d" (UID: "39c253b2-56c2-46c1-bafd-587d40437a2d"). InnerVolumeSpecName "kube-api-access-7qrhh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:30:03 crc kubenswrapper[4790]: I1124 14:30:03.225803 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qrhh\" (UniqueName: \"kubernetes.io/projected/39c253b2-56c2-46c1-bafd-587d40437a2d-kube-api-access-7qrhh\") on node \"crc\" DevicePath \"\"" Nov 24 14:30:03 crc kubenswrapper[4790]: I1124 14:30:03.226110 4790 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/39c253b2-56c2-46c1-bafd-587d40437a2d-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 14:30:03 crc kubenswrapper[4790]: I1124 14:30:03.226121 4790 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/39c253b2-56c2-46c1-bafd-587d40437a2d-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 14:30:03 crc kubenswrapper[4790]: I1124 14:30:03.314433 4790 scope.go:117] "RemoveContainer" containerID="571bd0c27e6ad4756f7116ed8b5a40d06301f3d85436cea369594fdb322154a7" Nov 24 14:30:03 crc kubenswrapper[4790]: E1124 14:30:03.314716 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:30:03 crc kubenswrapper[4790]: I1124 14:30:03.439131 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-j9csk" event={"ID":"39c253b2-56c2-46c1-bafd-587d40437a2d","Type":"ContainerDied","Data":"35bc9cfc245f36a4f9c21e9423f765342677926fd989341023177afbf3814413"} Nov 24 14:30:03 crc kubenswrapper[4790]: I1124 14:30:03.439185 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35bc9cfc245f36a4f9c21e9423f765342677926fd989341023177afbf3814413" Nov 24 14:30:03 crc kubenswrapper[4790]: I1124 14:30:03.439253 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-j9csk" Nov 24 14:30:04 crc kubenswrapper[4790]: I1124 14:30:04.025285 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399865-tdf2x"] Nov 24 14:30:04 crc kubenswrapper[4790]: I1124 14:30:04.032323 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399865-tdf2x"] Nov 24 14:30:04 crc kubenswrapper[4790]: I1124 14:30:04.325382 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b" path="/var/lib/kubelet/pods/12e47ad5-93c3-42e3-98ad-f7e2cb6bb66b/volumes" Nov 24 14:30:16 crc kubenswrapper[4790]: I1124 14:30:16.314165 4790 scope.go:117] "RemoveContainer" containerID="571bd0c27e6ad4756f7116ed8b5a40d06301f3d85436cea369594fdb322154a7" Nov 24 14:30:16 crc kubenswrapper[4790]: E1124 14:30:16.315050 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:30:31 crc kubenswrapper[4790]: I1124 14:30:31.314794 4790 scope.go:117] "RemoveContainer" containerID="571bd0c27e6ad4756f7116ed8b5a40d06301f3d85436cea369594fdb322154a7" Nov 24 14:30:31 crc kubenswrapper[4790]: E1124 14:30:31.316056 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.134910 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-gl96n"] Nov 24 14:30:45 crc kubenswrapper[4790]: E1124 14:30:45.135748 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39c253b2-56c2-46c1-bafd-587d40437a2d" containerName="collect-profiles" Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.135763 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="39c253b2-56c2-46c1-bafd-587d40437a2d" containerName="collect-profiles" Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.135924 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="39c253b2-56c2-46c1-bafd-587d40437a2d" containerName="collect-profiles" Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.136902 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-gl96n" Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.140584 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.140836 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.141056 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.141142 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-bwc5r" Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.151817 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.151859 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-gl96n"] Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.167951 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0954b3a7-0d87-4053-9c06-e700d51f3b94-config\") pod \"dnsmasq-dns-5d7b5456f5-gl96n\" (UID: \"0954b3a7-0d87-4053-9c06-e700d51f3b94\") " pod="openstack/dnsmasq-dns-5d7b5456f5-gl96n" Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.168051 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2w6v\" (UniqueName: \"kubernetes.io/projected/0954b3a7-0d87-4053-9c06-e700d51f3b94-kube-api-access-m2w6v\") pod \"dnsmasq-dns-5d7b5456f5-gl96n\" (UID: \"0954b3a7-0d87-4053-9c06-e700d51f3b94\") " pod="openstack/dnsmasq-dns-5d7b5456f5-gl96n" Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.168196 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0954b3a7-0d87-4053-9c06-e700d51f3b94-dns-svc\") pod \"dnsmasq-dns-5d7b5456f5-gl96n\" (UID: \"0954b3a7-0d87-4053-9c06-e700d51f3b94\") " pod="openstack/dnsmasq-dns-5d7b5456f5-gl96n" Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.269914 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0954b3a7-0d87-4053-9c06-e700d51f3b94-dns-svc\") pod \"dnsmasq-dns-5d7b5456f5-gl96n\" (UID: \"0954b3a7-0d87-4053-9c06-e700d51f3b94\") " pod="openstack/dnsmasq-dns-5d7b5456f5-gl96n" Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.270015 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0954b3a7-0d87-4053-9c06-e700d51f3b94-config\") pod \"dnsmasq-dns-5d7b5456f5-gl96n\" (UID: \"0954b3a7-0d87-4053-9c06-e700d51f3b94\") " pod="openstack/dnsmasq-dns-5d7b5456f5-gl96n" Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.270079 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2w6v\" (UniqueName: \"kubernetes.io/projected/0954b3a7-0d87-4053-9c06-e700d51f3b94-kube-api-access-m2w6v\") pod \"dnsmasq-dns-5d7b5456f5-gl96n\" (UID: \"0954b3a7-0d87-4053-9c06-e700d51f3b94\") " pod="openstack/dnsmasq-dns-5d7b5456f5-gl96n" Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.271557 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0954b3a7-0d87-4053-9c06-e700d51f3b94-config\") pod \"dnsmasq-dns-5d7b5456f5-gl96n\" (UID: \"0954b3a7-0d87-4053-9c06-e700d51f3b94\") " pod="openstack/dnsmasq-dns-5d7b5456f5-gl96n" Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.271558 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0954b3a7-0d87-4053-9c06-e700d51f3b94-dns-svc\") pod \"dnsmasq-dns-5d7b5456f5-gl96n\" (UID: \"0954b3a7-0d87-4053-9c06-e700d51f3b94\") " pod="openstack/dnsmasq-dns-5d7b5456f5-gl96n" Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.292992 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2w6v\" (UniqueName: \"kubernetes.io/projected/0954b3a7-0d87-4053-9c06-e700d51f3b94-kube-api-access-m2w6v\") pod \"dnsmasq-dns-5d7b5456f5-gl96n\" (UID: \"0954b3a7-0d87-4053-9c06-e700d51f3b94\") " pod="openstack/dnsmasq-dns-5d7b5456f5-gl96n" Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.314277 4790 scope.go:117] "RemoveContainer" containerID="571bd0c27e6ad4756f7116ed8b5a40d06301f3d85436cea369594fdb322154a7" Nov 24 14:30:45 crc kubenswrapper[4790]: E1124 14:30:45.314687 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.329634 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-54qbt"] Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.330829 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-54qbt" Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.340094 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-54qbt"] Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.371530 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50b76bef-0c9b-4f75-b902-adaa35fea421-config\") pod \"dnsmasq-dns-98ddfc8f-54qbt\" (UID: \"50b76bef-0c9b-4f75-b902-adaa35fea421\") " pod="openstack/dnsmasq-dns-98ddfc8f-54qbt" Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.371605 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/50b76bef-0c9b-4f75-b902-adaa35fea421-dns-svc\") pod \"dnsmasq-dns-98ddfc8f-54qbt\" (UID: \"50b76bef-0c9b-4f75-b902-adaa35fea421\") " pod="openstack/dnsmasq-dns-98ddfc8f-54qbt" Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.371651 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snl7g\" (UniqueName: \"kubernetes.io/projected/50b76bef-0c9b-4f75-b902-adaa35fea421-kube-api-access-snl7g\") pod \"dnsmasq-dns-98ddfc8f-54qbt\" (UID: \"50b76bef-0c9b-4f75-b902-adaa35fea421\") " pod="openstack/dnsmasq-dns-98ddfc8f-54qbt" Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.473101 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snl7g\" (UniqueName: \"kubernetes.io/projected/50b76bef-0c9b-4f75-b902-adaa35fea421-kube-api-access-snl7g\") pod \"dnsmasq-dns-98ddfc8f-54qbt\" (UID: \"50b76bef-0c9b-4f75-b902-adaa35fea421\") " pod="openstack/dnsmasq-dns-98ddfc8f-54qbt" Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.473220 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50b76bef-0c9b-4f75-b902-adaa35fea421-config\") pod \"dnsmasq-dns-98ddfc8f-54qbt\" (UID: \"50b76bef-0c9b-4f75-b902-adaa35fea421\") " pod="openstack/dnsmasq-dns-98ddfc8f-54qbt" Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.473252 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/50b76bef-0c9b-4f75-b902-adaa35fea421-dns-svc\") pod \"dnsmasq-dns-98ddfc8f-54qbt\" (UID: \"50b76bef-0c9b-4f75-b902-adaa35fea421\") " pod="openstack/dnsmasq-dns-98ddfc8f-54qbt" Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.474085 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/50b76bef-0c9b-4f75-b902-adaa35fea421-dns-svc\") pod \"dnsmasq-dns-98ddfc8f-54qbt\" (UID: \"50b76bef-0c9b-4f75-b902-adaa35fea421\") " pod="openstack/dnsmasq-dns-98ddfc8f-54qbt" Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.474417 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50b76bef-0c9b-4f75-b902-adaa35fea421-config\") pod \"dnsmasq-dns-98ddfc8f-54qbt\" (UID: \"50b76bef-0c9b-4f75-b902-adaa35fea421\") " pod="openstack/dnsmasq-dns-98ddfc8f-54qbt" Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.477715 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-gl96n" Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.492981 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snl7g\" (UniqueName: \"kubernetes.io/projected/50b76bef-0c9b-4f75-b902-adaa35fea421-kube-api-access-snl7g\") pod \"dnsmasq-dns-98ddfc8f-54qbt\" (UID: \"50b76bef-0c9b-4f75-b902-adaa35fea421\") " pod="openstack/dnsmasq-dns-98ddfc8f-54qbt" Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.654541 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-54qbt" Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.847519 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-54qbt"] Nov 24 14:30:45 crc kubenswrapper[4790]: W1124 14:30:45.849987 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod50b76bef_0c9b_4f75_b902_adaa35fea421.slice/crio-063ecd216bfa342df8f40e5b263b96a3ae2b4a6b646cd599f044c70f520d48ff WatchSource:0}: Error finding container 063ecd216bfa342df8f40e5b263b96a3ae2b4a6b646cd599f044c70f520d48ff: Status 404 returned error can't find the container with id 063ecd216bfa342df8f40e5b263b96a3ae2b4a6b646cd599f044c70f520d48ff Nov 24 14:30:45 crc kubenswrapper[4790]: I1124 14:30:45.920560 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-gl96n"] Nov 24 14:30:45 crc kubenswrapper[4790]: W1124 14:30:45.926586 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0954b3a7_0d87_4053_9c06_e700d51f3b94.slice/crio-89a65aea46f5664449145c6c4489777cdd7493c2dc50dd6c64cb7363f3d6d543 WatchSource:0}: Error finding container 89a65aea46f5664449145c6c4489777cdd7493c2dc50dd6c64cb7363f3d6d543: Status 404 returned error can't find the container with id 89a65aea46f5664449145c6c4489777cdd7493c2dc50dd6c64cb7363f3d6d543 Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.225743 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.227327 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.230189 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.230188 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.231145 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.233932 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.234030 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-nq6p2" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.239955 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.282825 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/aceb9aee-3ff6-4e3d-a406-2156b12d6265-server-conf\") pod \"rabbitmq-server-0\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " pod="openstack/rabbitmq-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.283162 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/aceb9aee-3ff6-4e3d-a406-2156b12d6265-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " pod="openstack/rabbitmq-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.283202 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-0e9b38f5-5501-4b3c-8d56-02c85887ccd6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0e9b38f5-5501-4b3c-8d56-02c85887ccd6\") pod \"rabbitmq-server-0\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " pod="openstack/rabbitmq-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.283221 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/aceb9aee-3ff6-4e3d-a406-2156b12d6265-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " pod="openstack/rabbitmq-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.283241 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/aceb9aee-3ff6-4e3d-a406-2156b12d6265-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " pod="openstack/rabbitmq-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.283307 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/aceb9aee-3ff6-4e3d-a406-2156b12d6265-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " pod="openstack/rabbitmq-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.283342 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/aceb9aee-3ff6-4e3d-a406-2156b12d6265-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " pod="openstack/rabbitmq-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.283534 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/aceb9aee-3ff6-4e3d-a406-2156b12d6265-pod-info\") pod \"rabbitmq-server-0\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " pod="openstack/rabbitmq-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.283557 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctfcz\" (UniqueName: \"kubernetes.io/projected/aceb9aee-3ff6-4e3d-a406-2156b12d6265-kube-api-access-ctfcz\") pod \"rabbitmq-server-0\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " pod="openstack/rabbitmq-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.384420 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/aceb9aee-3ff6-4e3d-a406-2156b12d6265-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " pod="openstack/rabbitmq-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.384465 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/aceb9aee-3ff6-4e3d-a406-2156b12d6265-pod-info\") pod \"rabbitmq-server-0\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " pod="openstack/rabbitmq-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.384483 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctfcz\" (UniqueName: \"kubernetes.io/projected/aceb9aee-3ff6-4e3d-a406-2156b12d6265-kube-api-access-ctfcz\") pod \"rabbitmq-server-0\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " pod="openstack/rabbitmq-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.384534 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/aceb9aee-3ff6-4e3d-a406-2156b12d6265-server-conf\") pod \"rabbitmq-server-0\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " pod="openstack/rabbitmq-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.384562 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/aceb9aee-3ff6-4e3d-a406-2156b12d6265-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " pod="openstack/rabbitmq-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.384596 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-0e9b38f5-5501-4b3c-8d56-02c85887ccd6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0e9b38f5-5501-4b3c-8d56-02c85887ccd6\") pod \"rabbitmq-server-0\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " pod="openstack/rabbitmq-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.384616 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/aceb9aee-3ff6-4e3d-a406-2156b12d6265-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " pod="openstack/rabbitmq-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.384637 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/aceb9aee-3ff6-4e3d-a406-2156b12d6265-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " pod="openstack/rabbitmq-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.384688 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/aceb9aee-3ff6-4e3d-a406-2156b12d6265-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " pod="openstack/rabbitmq-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.385437 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/aceb9aee-3ff6-4e3d-a406-2156b12d6265-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " pod="openstack/rabbitmq-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.386124 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/aceb9aee-3ff6-4e3d-a406-2156b12d6265-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " pod="openstack/rabbitmq-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.387280 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/aceb9aee-3ff6-4e3d-a406-2156b12d6265-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " pod="openstack/rabbitmq-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.387352 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/aceb9aee-3ff6-4e3d-a406-2156b12d6265-server-conf\") pod \"rabbitmq-server-0\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " pod="openstack/rabbitmq-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.390008 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/aceb9aee-3ff6-4e3d-a406-2156b12d6265-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " pod="openstack/rabbitmq-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.390354 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/aceb9aee-3ff6-4e3d-a406-2156b12d6265-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " pod="openstack/rabbitmq-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.390622 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/aceb9aee-3ff6-4e3d-a406-2156b12d6265-pod-info\") pod \"rabbitmq-server-0\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " pod="openstack/rabbitmq-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.391191 4790 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.391224 4790 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-0e9b38f5-5501-4b3c-8d56-02c85887ccd6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0e9b38f5-5501-4b3c-8d56-02c85887ccd6\") pod \"rabbitmq-server-0\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/65faf0adfc5f692cc530d451dc932ddb8871a251ab7abdecf7dab36ed8f4c11e/globalmount\"" pod="openstack/rabbitmq-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.415328 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctfcz\" (UniqueName: \"kubernetes.io/projected/aceb9aee-3ff6-4e3d-a406-2156b12d6265-kube-api-access-ctfcz\") pod \"rabbitmq-server-0\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " pod="openstack/rabbitmq-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.422835 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-0e9b38f5-5501-4b3c-8d56-02c85887ccd6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0e9b38f5-5501-4b3c-8d56-02c85887ccd6\") pod \"rabbitmq-server-0\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " pod="openstack/rabbitmq-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.468096 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.469733 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.472238 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.472249 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.472482 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-q8mlt" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.472631 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.472776 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.491898 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.552437 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.583642 4790 scope.go:117] "RemoveContainer" containerID="3f3ceba70968dd466f787b443c993444af51db55029a014cb89bd77d14331f7b" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.587235 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-0dc142ca-8752-419b-ae3a-93aa61837411\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0dc142ca-8752-419b-ae3a-93aa61837411\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.587275 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.587316 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.587337 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvwqs\" (UniqueName: \"kubernetes.io/projected/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-kube-api-access-qvwqs\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.587480 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.587580 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.587664 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.587722 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.587758 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.689215 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.689288 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.689324 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.689366 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.689443 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-0dc142ca-8752-419b-ae3a-93aa61837411\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0dc142ca-8752-419b-ae3a-93aa61837411\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.689494 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.689537 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.689573 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvwqs\" (UniqueName: \"kubernetes.io/projected/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-kube-api-access-qvwqs\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.689616 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.689983 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.690452 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.691087 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.691616 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.694080 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.694774 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.696773 4790 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.696821 4790 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-0dc142ca-8752-419b-ae3a-93aa61837411\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0dc142ca-8752-419b-ae3a-93aa61837411\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b46039fa1dd307f30c9f1f8971d3598f69f7cb989892e6b48b5c448c7c9ec3f0/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.697415 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.711990 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvwqs\" (UniqueName: \"kubernetes.io/projected/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-kube-api-access-qvwqs\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.740369 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-0dc142ca-8752-419b-ae3a-93aa61837411\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0dc142ca-8752-419b-ae3a-93aa61837411\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.788759 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.805256 4790 generic.go:334] "Generic (PLEG): container finished" podID="0954b3a7-0d87-4053-9c06-e700d51f3b94" containerID="a3841e38b4de196d36c201d0c2187ef2a0f7de8125fa03407d891e4b5f202b73" exitCode=0 Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.805362 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-gl96n" event={"ID":"0954b3a7-0d87-4053-9c06-e700d51f3b94","Type":"ContainerDied","Data":"a3841e38b4de196d36c201d0c2187ef2a0f7de8125fa03407d891e4b5f202b73"} Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.805391 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-gl96n" event={"ID":"0954b3a7-0d87-4053-9c06-e700d51f3b94","Type":"ContainerStarted","Data":"89a65aea46f5664449145c6c4489777cdd7493c2dc50dd6c64cb7363f3d6d543"} Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.807010 4790 generic.go:334] "Generic (PLEG): container finished" podID="50b76bef-0c9b-4f75-b902-adaa35fea421" containerID="e22026d8046d7f6bc2e5834577e923754455d12ad5d84eb624fb4204d5b59c22" exitCode=0 Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.807047 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-54qbt" event={"ID":"50b76bef-0c9b-4f75-b902-adaa35fea421","Type":"ContainerDied","Data":"e22026d8046d7f6bc2e5834577e923754455d12ad5d84eb624fb4204d5b59c22"} Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.807071 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-54qbt" event={"ID":"50b76bef-0c9b-4f75-b902-adaa35fea421","Type":"ContainerStarted","Data":"063ecd216bfa342df8f40e5b263b96a3ae2b4a6b646cd599f044c70f520d48ff"} Nov 24 14:30:46 crc kubenswrapper[4790]: I1124 14:30:46.986838 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 14:30:47 crc kubenswrapper[4790]: I1124 14:30:47.250528 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 14:30:47 crc kubenswrapper[4790]: W1124 14:30:47.254343 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode4d75dc1_2209_4c82_b562_9ad8ba78beb7.slice/crio-1e8f2e4d16e4c2d66d65fe9aeaca778ff69da1a3e9036fbabb83c29077584197 WatchSource:0}: Error finding container 1e8f2e4d16e4c2d66d65fe9aeaca778ff69da1a3e9036fbabb83c29077584197: Status 404 returned error can't find the container with id 1e8f2e4d16e4c2d66d65fe9aeaca778ff69da1a3e9036fbabb83c29077584197 Nov 24 14:30:47 crc kubenswrapper[4790]: I1124 14:30:47.759524 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 24 14:30:47 crc kubenswrapper[4790]: I1124 14:30:47.761118 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 14:30:47 crc kubenswrapper[4790]: I1124 14:30:47.762825 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-wwdcm" Nov 24 14:30:47 crc kubenswrapper[4790]: I1124 14:30:47.763014 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 24 14:30:47 crc kubenswrapper[4790]: I1124 14:30:47.763816 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 24 14:30:47 crc kubenswrapper[4790]: I1124 14:30:47.764064 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 24 14:30:47 crc kubenswrapper[4790]: I1124 14:30:47.773585 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 24 14:30:47 crc kubenswrapper[4790]: I1124 14:30:47.774141 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 14:30:47 crc kubenswrapper[4790]: I1124 14:30:47.813995 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e4d75dc1-2209-4c82-b562-9ad8ba78beb7","Type":"ContainerStarted","Data":"1e8f2e4d16e4c2d66d65fe9aeaca778ff69da1a3e9036fbabb83c29077584197"} Nov 24 14:30:47 crc kubenswrapper[4790]: I1124 14:30:47.816205 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-54qbt" event={"ID":"50b76bef-0c9b-4f75-b902-adaa35fea421","Type":"ContainerStarted","Data":"ee855f19e1b8ce5aecadd2150e60580a9a9d1536b0634249fea95b1ad77385bb"} Nov 24 14:30:47 crc kubenswrapper[4790]: I1124 14:30:47.816287 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-98ddfc8f-54qbt" Nov 24 14:30:47 crc kubenswrapper[4790]: I1124 14:30:47.817354 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"aceb9aee-3ff6-4e3d-a406-2156b12d6265","Type":"ContainerStarted","Data":"210984cd9937feaf77653007e54ad3cdedb5c62434b34e7d2fb61b5cbf93652a"} Nov 24 14:30:47 crc kubenswrapper[4790]: I1124 14:30:47.820135 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-gl96n" event={"ID":"0954b3a7-0d87-4053-9c06-e700d51f3b94","Type":"ContainerStarted","Data":"a5d4c518588c25b96c3a7e9b28cc5ac059289a2e7da3b2602638f127ca55c42e"} Nov 24 14:30:47 crc kubenswrapper[4790]: I1124 14:30:47.820373 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d7b5456f5-gl96n" Nov 24 14:30:47 crc kubenswrapper[4790]: I1124 14:30:47.833601 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-98ddfc8f-54qbt" podStartSLOduration=2.833578244 podStartE2EDuration="2.833578244s" podCreationTimestamp="2025-11-24 14:30:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:30:47.831612458 +0000 UTC m=+4696.211506120" watchObservedRunningTime="2025-11-24 14:30:47.833578244 +0000 UTC m=+4696.213471916" Nov 24 14:30:47 crc kubenswrapper[4790]: I1124 14:30:47.853751 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d7b5456f5-gl96n" podStartSLOduration=2.853732472 podStartE2EDuration="2.853732472s" podCreationTimestamp="2025-11-24 14:30:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:30:47.853689841 +0000 UTC m=+4696.233583503" watchObservedRunningTime="2025-11-24 14:30:47.853732472 +0000 UTC m=+4696.233626144" Nov 24 14:30:47 crc kubenswrapper[4790]: I1124 14:30:47.903992 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/95752c01-3e10-44b4-8af5-7b74622dbae3-kolla-config\") pod \"openstack-galera-0\" (UID: \"95752c01-3e10-44b4-8af5-7b74622dbae3\") " pod="openstack/openstack-galera-0" Nov 24 14:30:47 crc kubenswrapper[4790]: I1124 14:30:47.904061 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95752c01-3e10-44b4-8af5-7b74622dbae3-operator-scripts\") pod \"openstack-galera-0\" (UID: \"95752c01-3e10-44b4-8af5-7b74622dbae3\") " pod="openstack/openstack-galera-0" Nov 24 14:30:47 crc kubenswrapper[4790]: I1124 14:30:47.904084 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/95752c01-3e10-44b4-8af5-7b74622dbae3-config-data-generated\") pod \"openstack-galera-0\" (UID: \"95752c01-3e10-44b4-8af5-7b74622dbae3\") " pod="openstack/openstack-galera-0" Nov 24 14:30:47 crc kubenswrapper[4790]: I1124 14:30:47.904163 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-5caee7d1-e54b-4fa3-b0bc-8ed93645ac78\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5caee7d1-e54b-4fa3-b0bc-8ed93645ac78\") pod \"openstack-galera-0\" (UID: \"95752c01-3e10-44b4-8af5-7b74622dbae3\") " pod="openstack/openstack-galera-0" Nov 24 14:30:47 crc kubenswrapper[4790]: I1124 14:30:47.904371 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/95752c01-3e10-44b4-8af5-7b74622dbae3-config-data-default\") pod \"openstack-galera-0\" (UID: \"95752c01-3e10-44b4-8af5-7b74622dbae3\") " pod="openstack/openstack-galera-0" Nov 24 14:30:47 crc kubenswrapper[4790]: I1124 14:30:47.904497 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/95752c01-3e10-44b4-8af5-7b74622dbae3-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"95752c01-3e10-44b4-8af5-7b74622dbae3\") " pod="openstack/openstack-galera-0" Nov 24 14:30:47 crc kubenswrapper[4790]: I1124 14:30:47.904526 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95752c01-3e10-44b4-8af5-7b74622dbae3-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"95752c01-3e10-44b4-8af5-7b74622dbae3\") " pod="openstack/openstack-galera-0" Nov 24 14:30:47 crc kubenswrapper[4790]: I1124 14:30:47.904678 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snn49\" (UniqueName: \"kubernetes.io/projected/95752c01-3e10-44b4-8af5-7b74622dbae3-kube-api-access-snn49\") pod \"openstack-galera-0\" (UID: \"95752c01-3e10-44b4-8af5-7b74622dbae3\") " pod="openstack/openstack-galera-0" Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.006611 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snn49\" (UniqueName: \"kubernetes.io/projected/95752c01-3e10-44b4-8af5-7b74622dbae3-kube-api-access-snn49\") pod \"openstack-galera-0\" (UID: \"95752c01-3e10-44b4-8af5-7b74622dbae3\") " pod="openstack/openstack-galera-0" Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.006692 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/95752c01-3e10-44b4-8af5-7b74622dbae3-kolla-config\") pod \"openstack-galera-0\" (UID: \"95752c01-3e10-44b4-8af5-7b74622dbae3\") " pod="openstack/openstack-galera-0" Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.006747 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95752c01-3e10-44b4-8af5-7b74622dbae3-operator-scripts\") pod \"openstack-galera-0\" (UID: \"95752c01-3e10-44b4-8af5-7b74622dbae3\") " pod="openstack/openstack-galera-0" Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.006769 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/95752c01-3e10-44b4-8af5-7b74622dbae3-config-data-generated\") pod \"openstack-galera-0\" (UID: \"95752c01-3e10-44b4-8af5-7b74622dbae3\") " pod="openstack/openstack-galera-0" Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.006831 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-5caee7d1-e54b-4fa3-b0bc-8ed93645ac78\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5caee7d1-e54b-4fa3-b0bc-8ed93645ac78\") pod \"openstack-galera-0\" (UID: \"95752c01-3e10-44b4-8af5-7b74622dbae3\") " pod="openstack/openstack-galera-0" Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.006893 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/95752c01-3e10-44b4-8af5-7b74622dbae3-config-data-default\") pod \"openstack-galera-0\" (UID: \"95752c01-3e10-44b4-8af5-7b74622dbae3\") " pod="openstack/openstack-galera-0" Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.006945 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/95752c01-3e10-44b4-8af5-7b74622dbae3-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"95752c01-3e10-44b4-8af5-7b74622dbae3\") " pod="openstack/openstack-galera-0" Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.006969 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95752c01-3e10-44b4-8af5-7b74622dbae3-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"95752c01-3e10-44b4-8af5-7b74622dbae3\") " pod="openstack/openstack-galera-0" Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.007256 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/95752c01-3e10-44b4-8af5-7b74622dbae3-config-data-generated\") pod \"openstack-galera-0\" (UID: \"95752c01-3e10-44b4-8af5-7b74622dbae3\") " pod="openstack/openstack-galera-0" Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.007610 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/95752c01-3e10-44b4-8af5-7b74622dbae3-kolla-config\") pod \"openstack-galera-0\" (UID: \"95752c01-3e10-44b4-8af5-7b74622dbae3\") " pod="openstack/openstack-galera-0" Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.008038 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/95752c01-3e10-44b4-8af5-7b74622dbae3-config-data-default\") pod \"openstack-galera-0\" (UID: \"95752c01-3e10-44b4-8af5-7b74622dbae3\") " pod="openstack/openstack-galera-0" Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.008474 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95752c01-3e10-44b4-8af5-7b74622dbae3-operator-scripts\") pod \"openstack-galera-0\" (UID: \"95752c01-3e10-44b4-8af5-7b74622dbae3\") " pod="openstack/openstack-galera-0" Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.013589 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/95752c01-3e10-44b4-8af5-7b74622dbae3-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"95752c01-3e10-44b4-8af5-7b74622dbae3\") " pod="openstack/openstack-galera-0" Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.013641 4790 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.013685 4790 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-5caee7d1-e54b-4fa3-b0bc-8ed93645ac78\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5caee7d1-e54b-4fa3-b0bc-8ed93645ac78\") pod \"openstack-galera-0\" (UID: \"95752c01-3e10-44b4-8af5-7b74622dbae3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/286ca24998b61ee32070d19456c15a667634baf3e3d8e7049774417695f8fd14/globalmount\"" pod="openstack/openstack-galera-0" Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.020260 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95752c01-3e10-44b4-8af5-7b74622dbae3-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"95752c01-3e10-44b4-8af5-7b74622dbae3\") " pod="openstack/openstack-galera-0" Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.047954 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-5caee7d1-e54b-4fa3-b0bc-8ed93645ac78\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5caee7d1-e54b-4fa3-b0bc-8ed93645ac78\") pod \"openstack-galera-0\" (UID: \"95752c01-3e10-44b4-8af5-7b74622dbae3\") " pod="openstack/openstack-galera-0" Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.048995 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.050155 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.052561 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-hwbxc" Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.054039 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.063958 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.093850 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snn49\" (UniqueName: \"kubernetes.io/projected/95752c01-3e10-44b4-8af5-7b74622dbae3-kube-api-access-snn49\") pod \"openstack-galera-0\" (UID: \"95752c01-3e10-44b4-8af5-7b74622dbae3\") " pod="openstack/openstack-galera-0" Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.209864 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5f59a746-4906-45aa-b6ee-7feb652fa17a-config-data\") pod \"memcached-0\" (UID: \"5f59a746-4906-45aa-b6ee-7feb652fa17a\") " pod="openstack/memcached-0" Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.210223 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5f59a746-4906-45aa-b6ee-7feb652fa17a-kolla-config\") pod \"memcached-0\" (UID: \"5f59a746-4906-45aa-b6ee-7feb652fa17a\") " pod="openstack/memcached-0" Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.210247 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v87lc\" (UniqueName: \"kubernetes.io/projected/5f59a746-4906-45aa-b6ee-7feb652fa17a-kube-api-access-v87lc\") pod \"memcached-0\" (UID: \"5f59a746-4906-45aa-b6ee-7feb652fa17a\") " pod="openstack/memcached-0" Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.311268 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5f59a746-4906-45aa-b6ee-7feb652fa17a-config-data\") pod \"memcached-0\" (UID: \"5f59a746-4906-45aa-b6ee-7feb652fa17a\") " pod="openstack/memcached-0" Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.311327 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5f59a746-4906-45aa-b6ee-7feb652fa17a-kolla-config\") pod \"memcached-0\" (UID: \"5f59a746-4906-45aa-b6ee-7feb652fa17a\") " pod="openstack/memcached-0" Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.311358 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v87lc\" (UniqueName: \"kubernetes.io/projected/5f59a746-4906-45aa-b6ee-7feb652fa17a-kube-api-access-v87lc\") pod \"memcached-0\" (UID: \"5f59a746-4906-45aa-b6ee-7feb652fa17a\") " pod="openstack/memcached-0" Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.312277 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5f59a746-4906-45aa-b6ee-7feb652fa17a-config-data\") pod \"memcached-0\" (UID: \"5f59a746-4906-45aa-b6ee-7feb652fa17a\") " pod="openstack/memcached-0" Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.312297 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5f59a746-4906-45aa-b6ee-7feb652fa17a-kolla-config\") pod \"memcached-0\" (UID: \"5f59a746-4906-45aa-b6ee-7feb652fa17a\") " pod="openstack/memcached-0" Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.329055 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v87lc\" (UniqueName: \"kubernetes.io/projected/5f59a746-4906-45aa-b6ee-7feb652fa17a-kube-api-access-v87lc\") pod \"memcached-0\" (UID: \"5f59a746-4906-45aa-b6ee-7feb652fa17a\") " pod="openstack/memcached-0" Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.378246 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.532046 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.829232 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e4d75dc1-2209-4c82-b562-9ad8ba78beb7","Type":"ContainerStarted","Data":"eec7619ec057661e4f6c86d8f7e343e6ffc6cc4b1d97fe7cb9fd37282b4592ec"} Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.832133 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"aceb9aee-3ff6-4e3d-a406-2156b12d6265","Type":"ContainerStarted","Data":"b034c4d689b166a01dd04cb8c0670103f0f78292a162be11f936e20f6e05d278"} Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.923024 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 14:30:48 crc kubenswrapper[4790]: W1124 14:30:48.927007 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95752c01_3e10_44b4_8af5_7b74622dbae3.slice/crio-09cf5adffe4d97be7d60d0e4ca28ba24c2482c7f37d800f5f7846413af45662d WatchSource:0}: Error finding container 09cf5adffe4d97be7d60d0e4ca28ba24c2482c7f37d800f5f7846413af45662d: Status 404 returned error can't find the container with id 09cf5adffe4d97be7d60d0e4ca28ba24c2482c7f37d800f5f7846413af45662d Nov 24 14:30:48 crc kubenswrapper[4790]: I1124 14:30:48.988415 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 24 14:30:48 crc kubenswrapper[4790]: W1124 14:30:48.993221 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f59a746_4906_45aa_b6ee_7feb652fa17a.slice/crio-dcf6c027a617f39c7a8d594af840de0601eeceed69f1119753968e6abb15b729 WatchSource:0}: Error finding container dcf6c027a617f39c7a8d594af840de0601eeceed69f1119753968e6abb15b729: Status 404 returned error can't find the container with id dcf6c027a617f39c7a8d594af840de0601eeceed69f1119753968e6abb15b729 Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.199640 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.201170 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.203866 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.203872 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.204129 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-qpkpf" Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.204814 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.218418 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.331697 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09cadf79-68ab-4fc3-b391-4a7d8318a787-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"09cadf79-68ab-4fc3-b391-4a7d8318a787\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.331785 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b4438624-e9fa-4c56-bb20-320f16707924\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b4438624-e9fa-4c56-bb20-320f16707924\") pod \"openstack-cell1-galera-0\" (UID: \"09cadf79-68ab-4fc3-b391-4a7d8318a787\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.331821 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09cadf79-68ab-4fc3-b391-4a7d8318a787-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"09cadf79-68ab-4fc3-b391-4a7d8318a787\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.331854 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/09cadf79-68ab-4fc3-b391-4a7d8318a787-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"09cadf79-68ab-4fc3-b391-4a7d8318a787\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.331892 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/09cadf79-68ab-4fc3-b391-4a7d8318a787-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"09cadf79-68ab-4fc3-b391-4a7d8318a787\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.331939 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/09cadf79-68ab-4fc3-b391-4a7d8318a787-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"09cadf79-68ab-4fc3-b391-4a7d8318a787\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.331961 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4mvk\" (UniqueName: \"kubernetes.io/projected/09cadf79-68ab-4fc3-b391-4a7d8318a787-kube-api-access-v4mvk\") pod \"openstack-cell1-galera-0\" (UID: \"09cadf79-68ab-4fc3-b391-4a7d8318a787\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.331996 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/09cadf79-68ab-4fc3-b391-4a7d8318a787-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"09cadf79-68ab-4fc3-b391-4a7d8318a787\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.433317 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/09cadf79-68ab-4fc3-b391-4a7d8318a787-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"09cadf79-68ab-4fc3-b391-4a7d8318a787\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.433390 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09cadf79-68ab-4fc3-b391-4a7d8318a787-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"09cadf79-68ab-4fc3-b391-4a7d8318a787\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.433440 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b4438624-e9fa-4c56-bb20-320f16707924\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b4438624-e9fa-4c56-bb20-320f16707924\") pod \"openstack-cell1-galera-0\" (UID: \"09cadf79-68ab-4fc3-b391-4a7d8318a787\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.433460 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09cadf79-68ab-4fc3-b391-4a7d8318a787-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"09cadf79-68ab-4fc3-b391-4a7d8318a787\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.433483 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/09cadf79-68ab-4fc3-b391-4a7d8318a787-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"09cadf79-68ab-4fc3-b391-4a7d8318a787\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.433500 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/09cadf79-68ab-4fc3-b391-4a7d8318a787-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"09cadf79-68ab-4fc3-b391-4a7d8318a787\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.433536 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/09cadf79-68ab-4fc3-b391-4a7d8318a787-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"09cadf79-68ab-4fc3-b391-4a7d8318a787\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.433560 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4mvk\" (UniqueName: \"kubernetes.io/projected/09cadf79-68ab-4fc3-b391-4a7d8318a787-kube-api-access-v4mvk\") pod \"openstack-cell1-galera-0\" (UID: \"09cadf79-68ab-4fc3-b391-4a7d8318a787\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.434134 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/09cadf79-68ab-4fc3-b391-4a7d8318a787-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"09cadf79-68ab-4fc3-b391-4a7d8318a787\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.434493 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/09cadf79-68ab-4fc3-b391-4a7d8318a787-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"09cadf79-68ab-4fc3-b391-4a7d8318a787\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.434753 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/09cadf79-68ab-4fc3-b391-4a7d8318a787-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"09cadf79-68ab-4fc3-b391-4a7d8318a787\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.435119 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09cadf79-68ab-4fc3-b391-4a7d8318a787-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"09cadf79-68ab-4fc3-b391-4a7d8318a787\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.436268 4790 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.436305 4790 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b4438624-e9fa-4c56-bb20-320f16707924\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b4438624-e9fa-4c56-bb20-320f16707924\") pod \"openstack-cell1-galera-0\" (UID: \"09cadf79-68ab-4fc3-b391-4a7d8318a787\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/105725921aaaf41349303729376adc2b622f732683d38a8260d0dbfebe9192b2/globalmount\"" pod="openstack/openstack-cell1-galera-0" Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.438199 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/09cadf79-68ab-4fc3-b391-4a7d8318a787-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"09cadf79-68ab-4fc3-b391-4a7d8318a787\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.438433 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09cadf79-68ab-4fc3-b391-4a7d8318a787-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"09cadf79-68ab-4fc3-b391-4a7d8318a787\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.450691 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4mvk\" (UniqueName: \"kubernetes.io/projected/09cadf79-68ab-4fc3-b391-4a7d8318a787-kube-api-access-v4mvk\") pod \"openstack-cell1-galera-0\" (UID: \"09cadf79-68ab-4fc3-b391-4a7d8318a787\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.463898 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b4438624-e9fa-4c56-bb20-320f16707924\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b4438624-e9fa-4c56-bb20-320f16707924\") pod \"openstack-cell1-galera-0\" (UID: \"09cadf79-68ab-4fc3-b391-4a7d8318a787\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.529353 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.841980 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"5f59a746-4906-45aa-b6ee-7feb652fa17a","Type":"ContainerStarted","Data":"65775f46f5ab613a6e5dff733daf824d4a69b6959eb0af05bbe809a0c2bebf75"} Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.842261 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.842277 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"5f59a746-4906-45aa-b6ee-7feb652fa17a","Type":"ContainerStarted","Data":"dcf6c027a617f39c7a8d594af840de0601eeceed69f1119753968e6abb15b729"} Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.843661 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"95752c01-3e10-44b4-8af5-7b74622dbae3","Type":"ContainerStarted","Data":"0025cdd8d7b34e3c7e4e23f6ca121eb6450f23d392494121e0ba532ec322cd9c"} Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.843697 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"95752c01-3e10-44b4-8af5-7b74622dbae3","Type":"ContainerStarted","Data":"09cf5adffe4d97be7d60d0e4ca28ba24c2482c7f37d800f5f7846413af45662d"} Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.877288 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=1.877264738 podStartE2EDuration="1.877264738s" podCreationTimestamp="2025-11-24 14:30:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:30:49.856641146 +0000 UTC m=+4698.236534818" watchObservedRunningTime="2025-11-24 14:30:49.877264738 +0000 UTC m=+4698.257158400" Nov 24 14:30:49 crc kubenswrapper[4790]: I1124 14:30:49.955748 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 14:30:49 crc kubenswrapper[4790]: W1124 14:30:49.963087 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09cadf79_68ab_4fc3_b391_4a7d8318a787.slice/crio-2f6347401639372c133569e6b9ae73cb387eecc0a320abc28c02c8bb798eb39c WatchSource:0}: Error finding container 2f6347401639372c133569e6b9ae73cb387eecc0a320abc28c02c8bb798eb39c: Status 404 returned error can't find the container with id 2f6347401639372c133569e6b9ae73cb387eecc0a320abc28c02c8bb798eb39c Nov 24 14:30:50 crc kubenswrapper[4790]: I1124 14:30:50.853953 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"09cadf79-68ab-4fc3-b391-4a7d8318a787","Type":"ContainerStarted","Data":"549031b5b92dd82a5caa695fed568deb22110a93c82a1345e5d2bdf49484fe90"} Nov 24 14:30:50 crc kubenswrapper[4790]: I1124 14:30:50.854442 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"09cadf79-68ab-4fc3-b391-4a7d8318a787","Type":"ContainerStarted","Data":"2f6347401639372c133569e6b9ae73cb387eecc0a320abc28c02c8bb798eb39c"} Nov 24 14:30:55 crc kubenswrapper[4790]: I1124 14:30:55.480259 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5d7b5456f5-gl96n" Nov 24 14:30:55 crc kubenswrapper[4790]: I1124 14:30:55.657171 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-98ddfc8f-54qbt" Nov 24 14:30:55 crc kubenswrapper[4790]: I1124 14:30:55.697175 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-gl96n"] Nov 24 14:30:55 crc kubenswrapper[4790]: I1124 14:30:55.891828 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5d7b5456f5-gl96n" podUID="0954b3a7-0d87-4053-9c06-e700d51f3b94" containerName="dnsmasq-dns" containerID="cri-o://a5d4c518588c25b96c3a7e9b28cc5ac059289a2e7da3b2602638f127ca55c42e" gracePeriod=10 Nov 24 14:30:56 crc kubenswrapper[4790]: I1124 14:30:56.911692 4790 generic.go:334] "Generic (PLEG): container finished" podID="0954b3a7-0d87-4053-9c06-e700d51f3b94" containerID="a5d4c518588c25b96c3a7e9b28cc5ac059289a2e7da3b2602638f127ca55c42e" exitCode=0 Nov 24 14:30:56 crc kubenswrapper[4790]: I1124 14:30:56.911790 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-gl96n" event={"ID":"0954b3a7-0d87-4053-9c06-e700d51f3b94","Type":"ContainerDied","Data":"a5d4c518588c25b96c3a7e9b28cc5ac059289a2e7da3b2602638f127ca55c42e"} Nov 24 14:30:57 crc kubenswrapper[4790]: I1124 14:30:57.078968 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-gl96n" Nov 24 14:30:57 crc kubenswrapper[4790]: I1124 14:30:57.156666 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2w6v\" (UniqueName: \"kubernetes.io/projected/0954b3a7-0d87-4053-9c06-e700d51f3b94-kube-api-access-m2w6v\") pod \"0954b3a7-0d87-4053-9c06-e700d51f3b94\" (UID: \"0954b3a7-0d87-4053-9c06-e700d51f3b94\") " Nov 24 14:30:57 crc kubenswrapper[4790]: I1124 14:30:57.156734 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0954b3a7-0d87-4053-9c06-e700d51f3b94-config\") pod \"0954b3a7-0d87-4053-9c06-e700d51f3b94\" (UID: \"0954b3a7-0d87-4053-9c06-e700d51f3b94\") " Nov 24 14:30:57 crc kubenswrapper[4790]: I1124 14:30:57.156827 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0954b3a7-0d87-4053-9c06-e700d51f3b94-dns-svc\") pod \"0954b3a7-0d87-4053-9c06-e700d51f3b94\" (UID: \"0954b3a7-0d87-4053-9c06-e700d51f3b94\") " Nov 24 14:30:57 crc kubenswrapper[4790]: I1124 14:30:57.177293 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0954b3a7-0d87-4053-9c06-e700d51f3b94-kube-api-access-m2w6v" (OuterVolumeSpecName: "kube-api-access-m2w6v") pod "0954b3a7-0d87-4053-9c06-e700d51f3b94" (UID: "0954b3a7-0d87-4053-9c06-e700d51f3b94"). InnerVolumeSpecName "kube-api-access-m2w6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:30:57 crc kubenswrapper[4790]: I1124 14:30:57.193661 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0954b3a7-0d87-4053-9c06-e700d51f3b94-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0954b3a7-0d87-4053-9c06-e700d51f3b94" (UID: "0954b3a7-0d87-4053-9c06-e700d51f3b94"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:30:57 crc kubenswrapper[4790]: I1124 14:30:57.194206 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0954b3a7-0d87-4053-9c06-e700d51f3b94-config" (OuterVolumeSpecName: "config") pod "0954b3a7-0d87-4053-9c06-e700d51f3b94" (UID: "0954b3a7-0d87-4053-9c06-e700d51f3b94"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:30:57 crc kubenswrapper[4790]: I1124 14:30:57.258418 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2w6v\" (UniqueName: \"kubernetes.io/projected/0954b3a7-0d87-4053-9c06-e700d51f3b94-kube-api-access-m2w6v\") on node \"crc\" DevicePath \"\"" Nov 24 14:30:57 crc kubenswrapper[4790]: I1124 14:30:57.258456 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0954b3a7-0d87-4053-9c06-e700d51f3b94-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:30:57 crc kubenswrapper[4790]: I1124 14:30:57.258468 4790 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0954b3a7-0d87-4053-9c06-e700d51f3b94-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:30:57 crc kubenswrapper[4790]: I1124 14:30:57.926737 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-gl96n" event={"ID":"0954b3a7-0d87-4053-9c06-e700d51f3b94","Type":"ContainerDied","Data":"89a65aea46f5664449145c6c4489777cdd7493c2dc50dd6c64cb7363f3d6d543"} Nov 24 14:30:57 crc kubenswrapper[4790]: I1124 14:30:57.926846 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-gl96n" Nov 24 14:30:57 crc kubenswrapper[4790]: I1124 14:30:57.927401 4790 scope.go:117] "RemoveContainer" containerID="a5d4c518588c25b96c3a7e9b28cc5ac059289a2e7da3b2602638f127ca55c42e" Nov 24 14:30:57 crc kubenswrapper[4790]: I1124 14:30:57.957907 4790 scope.go:117] "RemoveContainer" containerID="a3841e38b4de196d36c201d0c2187ef2a0f7de8125fa03407d891e4b5f202b73" Nov 24 14:30:57 crc kubenswrapper[4790]: I1124 14:30:57.988838 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-gl96n"] Nov 24 14:30:57 crc kubenswrapper[4790]: I1124 14:30:57.999702 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-gl96n"] Nov 24 14:30:58 crc kubenswrapper[4790]: I1124 14:30:58.315394 4790 scope.go:117] "RemoveContainer" containerID="571bd0c27e6ad4756f7116ed8b5a40d06301f3d85436cea369594fdb322154a7" Nov 24 14:30:58 crc kubenswrapper[4790]: E1124 14:30:58.316168 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:30:58 crc kubenswrapper[4790]: I1124 14:30:58.329692 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0954b3a7-0d87-4053-9c06-e700d51f3b94" path="/var/lib/kubelet/pods/0954b3a7-0d87-4053-9c06-e700d51f3b94/volumes" Nov 24 14:30:58 crc kubenswrapper[4790]: I1124 14:30:58.534194 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 24 14:31:01 crc kubenswrapper[4790]: I1124 14:31:01.962479 4790 generic.go:334] "Generic (PLEG): container finished" podID="09cadf79-68ab-4fc3-b391-4a7d8318a787" containerID="549031b5b92dd82a5caa695fed568deb22110a93c82a1345e5d2bdf49484fe90" exitCode=0 Nov 24 14:31:01 crc kubenswrapper[4790]: I1124 14:31:01.962554 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"09cadf79-68ab-4fc3-b391-4a7d8318a787","Type":"ContainerDied","Data":"549031b5b92dd82a5caa695fed568deb22110a93c82a1345e5d2bdf49484fe90"} Nov 24 14:31:01 crc kubenswrapper[4790]: I1124 14:31:01.967101 4790 generic.go:334] "Generic (PLEG): container finished" podID="95752c01-3e10-44b4-8af5-7b74622dbae3" containerID="0025cdd8d7b34e3c7e4e23f6ca121eb6450f23d392494121e0ba532ec322cd9c" exitCode=0 Nov 24 14:31:01 crc kubenswrapper[4790]: I1124 14:31:01.967158 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"95752c01-3e10-44b4-8af5-7b74622dbae3","Type":"ContainerDied","Data":"0025cdd8d7b34e3c7e4e23f6ca121eb6450f23d392494121e0ba532ec322cd9c"} Nov 24 14:31:02 crc kubenswrapper[4790]: I1124 14:31:02.976622 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"95752c01-3e10-44b4-8af5-7b74622dbae3","Type":"ContainerStarted","Data":"38628cf3cef27197006bd6ee19de4dae3cd9b8e5ae6b158a5d71a412b2742d1c"} Nov 24 14:31:02 crc kubenswrapper[4790]: I1124 14:31:02.979158 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"09cadf79-68ab-4fc3-b391-4a7d8318a787","Type":"ContainerStarted","Data":"e9954d6d08c7b3affb646b0dbdf5946917ef772514db4b70741ad6e0f965f1be"} Nov 24 14:31:02 crc kubenswrapper[4790]: I1124 14:31:02.998856 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=16.998835763 podStartE2EDuration="16.998835763s" podCreationTimestamp="2025-11-24 14:30:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:31:02.99696606 +0000 UTC m=+4711.376859722" watchObservedRunningTime="2025-11-24 14:31:02.998835763 +0000 UTC m=+4711.378729425" Nov 24 14:31:03 crc kubenswrapper[4790]: I1124 14:31:03.018356 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=15.018335373 podStartE2EDuration="15.018335373s" podCreationTimestamp="2025-11-24 14:30:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:31:03.0150964 +0000 UTC m=+4711.394990062" watchObservedRunningTime="2025-11-24 14:31:03.018335373 +0000 UTC m=+4711.398229035" Nov 24 14:31:08 crc kubenswrapper[4790]: I1124 14:31:08.378445 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 24 14:31:08 crc kubenswrapper[4790]: I1124 14:31:08.379040 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 24 14:31:09 crc kubenswrapper[4790]: I1124 14:31:09.529872 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 24 14:31:09 crc kubenswrapper[4790]: I1124 14:31:09.530574 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 24 14:31:10 crc kubenswrapper[4790]: I1124 14:31:10.255034 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 24 14:31:10 crc kubenswrapper[4790]: I1124 14:31:10.585835 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 24 14:31:10 crc kubenswrapper[4790]: I1124 14:31:10.671902 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 24 14:31:11 crc kubenswrapper[4790]: I1124 14:31:11.121471 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 24 14:31:13 crc kubenswrapper[4790]: I1124 14:31:13.314507 4790 scope.go:117] "RemoveContainer" containerID="571bd0c27e6ad4756f7116ed8b5a40d06301f3d85436cea369594fdb322154a7" Nov 24 14:31:13 crc kubenswrapper[4790]: E1124 14:31:13.315592 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:31:21 crc kubenswrapper[4790]: I1124 14:31:21.145937 4790 generic.go:334] "Generic (PLEG): container finished" podID="e4d75dc1-2209-4c82-b562-9ad8ba78beb7" containerID="eec7619ec057661e4f6c86d8f7e343e6ffc6cc4b1d97fe7cb9fd37282b4592ec" exitCode=0 Nov 24 14:31:21 crc kubenswrapper[4790]: I1124 14:31:21.146082 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e4d75dc1-2209-4c82-b562-9ad8ba78beb7","Type":"ContainerDied","Data":"eec7619ec057661e4f6c86d8f7e343e6ffc6cc4b1d97fe7cb9fd37282b4592ec"} Nov 24 14:31:21 crc kubenswrapper[4790]: I1124 14:31:21.149345 4790 generic.go:334] "Generic (PLEG): container finished" podID="aceb9aee-3ff6-4e3d-a406-2156b12d6265" containerID="b034c4d689b166a01dd04cb8c0670103f0f78292a162be11f936e20f6e05d278" exitCode=0 Nov 24 14:31:21 crc kubenswrapper[4790]: I1124 14:31:21.149386 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"aceb9aee-3ff6-4e3d-a406-2156b12d6265","Type":"ContainerDied","Data":"b034c4d689b166a01dd04cb8c0670103f0f78292a162be11f936e20f6e05d278"} Nov 24 14:31:22 crc kubenswrapper[4790]: I1124 14:31:22.161188 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"aceb9aee-3ff6-4e3d-a406-2156b12d6265","Type":"ContainerStarted","Data":"0ec87d8d4a68e138da2ae0c5196b7726954ea36bc435477500ce77ef849abace"} Nov 24 14:31:22 crc kubenswrapper[4790]: I1124 14:31:22.164049 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 24 14:31:22 crc kubenswrapper[4790]: I1124 14:31:22.170391 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e4d75dc1-2209-4c82-b562-9ad8ba78beb7","Type":"ContainerStarted","Data":"14344e000f6145e53d02947e710228cb00c7ff5419e58822fc92e3d77bd4bee7"} Nov 24 14:31:22 crc kubenswrapper[4790]: I1124 14:31:22.170707 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:22 crc kubenswrapper[4790]: I1124 14:31:22.196568 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.196526644 podStartE2EDuration="37.196526644s" podCreationTimestamp="2025-11-24 14:30:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:31:22.189859182 +0000 UTC m=+4730.569752854" watchObservedRunningTime="2025-11-24 14:31:22.196526644 +0000 UTC m=+4730.576420306" Nov 24 14:31:22 crc kubenswrapper[4790]: I1124 14:31:22.225497 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.225477134 podStartE2EDuration="37.225477134s" podCreationTimestamp="2025-11-24 14:30:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:31:22.220339697 +0000 UTC m=+4730.600233359" watchObservedRunningTime="2025-11-24 14:31:22.225477134 +0000 UTC m=+4730.605370816" Nov 24 14:31:25 crc kubenswrapper[4790]: I1124 14:31:25.315398 4790 scope.go:117] "RemoveContainer" containerID="571bd0c27e6ad4756f7116ed8b5a40d06301f3d85436cea369594fdb322154a7" Nov 24 14:31:25 crc kubenswrapper[4790]: E1124 14:31:25.316130 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:31:36 crc kubenswrapper[4790]: I1124 14:31:36.315019 4790 scope.go:117] "RemoveContainer" containerID="571bd0c27e6ad4756f7116ed8b5a40d06301f3d85436cea369594fdb322154a7" Nov 24 14:31:36 crc kubenswrapper[4790]: E1124 14:31:36.316264 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:31:36 crc kubenswrapper[4790]: I1124 14:31:36.555576 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 24 14:31:36 crc kubenswrapper[4790]: I1124 14:31:36.792045 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:41 crc kubenswrapper[4790]: I1124 14:31:41.216105 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-wqrxr"] Nov 24 14:31:41 crc kubenswrapper[4790]: E1124 14:31:41.217598 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0954b3a7-0d87-4053-9c06-e700d51f3b94" containerName="init" Nov 24 14:31:41 crc kubenswrapper[4790]: I1124 14:31:41.217640 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="0954b3a7-0d87-4053-9c06-e700d51f3b94" containerName="init" Nov 24 14:31:41 crc kubenswrapper[4790]: E1124 14:31:41.217694 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0954b3a7-0d87-4053-9c06-e700d51f3b94" containerName="dnsmasq-dns" Nov 24 14:31:41 crc kubenswrapper[4790]: I1124 14:31:41.217713 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="0954b3a7-0d87-4053-9c06-e700d51f3b94" containerName="dnsmasq-dns" Nov 24 14:31:41 crc kubenswrapper[4790]: I1124 14:31:41.220171 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="0954b3a7-0d87-4053-9c06-e700d51f3b94" containerName="dnsmasq-dns" Nov 24 14:31:41 crc kubenswrapper[4790]: I1124 14:31:41.222181 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-wqrxr" Nov 24 14:31:41 crc kubenswrapper[4790]: I1124 14:31:41.231022 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-wqrxr"] Nov 24 14:31:41 crc kubenswrapper[4790]: I1124 14:31:41.367039 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecff746c-6546-41be-bee5-c07166f192dc-config\") pod \"dnsmasq-dns-5b7946d7b9-wqrxr\" (UID: \"ecff746c-6546-41be-bee5-c07166f192dc\") " pod="openstack/dnsmasq-dns-5b7946d7b9-wqrxr" Nov 24 14:31:41 crc kubenswrapper[4790]: I1124 14:31:41.367191 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xm4c\" (UniqueName: \"kubernetes.io/projected/ecff746c-6546-41be-bee5-c07166f192dc-kube-api-access-2xm4c\") pod \"dnsmasq-dns-5b7946d7b9-wqrxr\" (UID: \"ecff746c-6546-41be-bee5-c07166f192dc\") " pod="openstack/dnsmasq-dns-5b7946d7b9-wqrxr" Nov 24 14:31:41 crc kubenswrapper[4790]: I1124 14:31:41.367255 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ecff746c-6546-41be-bee5-c07166f192dc-dns-svc\") pod \"dnsmasq-dns-5b7946d7b9-wqrxr\" (UID: \"ecff746c-6546-41be-bee5-c07166f192dc\") " pod="openstack/dnsmasq-dns-5b7946d7b9-wqrxr" Nov 24 14:31:41 crc kubenswrapper[4790]: I1124 14:31:41.468228 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecff746c-6546-41be-bee5-c07166f192dc-config\") pod \"dnsmasq-dns-5b7946d7b9-wqrxr\" (UID: \"ecff746c-6546-41be-bee5-c07166f192dc\") " pod="openstack/dnsmasq-dns-5b7946d7b9-wqrxr" Nov 24 14:31:41 crc kubenswrapper[4790]: I1124 14:31:41.468312 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xm4c\" (UniqueName: \"kubernetes.io/projected/ecff746c-6546-41be-bee5-c07166f192dc-kube-api-access-2xm4c\") pod \"dnsmasq-dns-5b7946d7b9-wqrxr\" (UID: \"ecff746c-6546-41be-bee5-c07166f192dc\") " pod="openstack/dnsmasq-dns-5b7946d7b9-wqrxr" Nov 24 14:31:41 crc kubenswrapper[4790]: I1124 14:31:41.468352 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ecff746c-6546-41be-bee5-c07166f192dc-dns-svc\") pod \"dnsmasq-dns-5b7946d7b9-wqrxr\" (UID: \"ecff746c-6546-41be-bee5-c07166f192dc\") " pod="openstack/dnsmasq-dns-5b7946d7b9-wqrxr" Nov 24 14:31:41 crc kubenswrapper[4790]: I1124 14:31:41.469462 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecff746c-6546-41be-bee5-c07166f192dc-config\") pod \"dnsmasq-dns-5b7946d7b9-wqrxr\" (UID: \"ecff746c-6546-41be-bee5-c07166f192dc\") " pod="openstack/dnsmasq-dns-5b7946d7b9-wqrxr" Nov 24 14:31:41 crc kubenswrapper[4790]: I1124 14:31:41.469572 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ecff746c-6546-41be-bee5-c07166f192dc-dns-svc\") pod \"dnsmasq-dns-5b7946d7b9-wqrxr\" (UID: \"ecff746c-6546-41be-bee5-c07166f192dc\") " pod="openstack/dnsmasq-dns-5b7946d7b9-wqrxr" Nov 24 14:31:41 crc kubenswrapper[4790]: I1124 14:31:41.488661 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xm4c\" (UniqueName: \"kubernetes.io/projected/ecff746c-6546-41be-bee5-c07166f192dc-kube-api-access-2xm4c\") pod \"dnsmasq-dns-5b7946d7b9-wqrxr\" (UID: \"ecff746c-6546-41be-bee5-c07166f192dc\") " pod="openstack/dnsmasq-dns-5b7946d7b9-wqrxr" Nov 24 14:31:41 crc kubenswrapper[4790]: I1124 14:31:41.541213 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-wqrxr" Nov 24 14:31:41 crc kubenswrapper[4790]: I1124 14:31:41.815071 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 14:31:42 crc kubenswrapper[4790]: I1124 14:31:42.025566 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-wqrxr"] Nov 24 14:31:42 crc kubenswrapper[4790]: W1124 14:31:42.029995 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podecff746c_6546_41be_bee5_c07166f192dc.slice/crio-08741f67c2124c4017b63d958a30ea6ec29c9fad754febdac567a16e7806b4c8 WatchSource:0}: Error finding container 08741f67c2124c4017b63d958a30ea6ec29c9fad754febdac567a16e7806b4c8: Status 404 returned error can't find the container with id 08741f67c2124c4017b63d958a30ea6ec29c9fad754febdac567a16e7806b4c8 Nov 24 14:31:42 crc kubenswrapper[4790]: I1124 14:31:42.287349 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 14:31:42 crc kubenswrapper[4790]: I1124 14:31:42.353765 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-wqrxr" event={"ID":"ecff746c-6546-41be-bee5-c07166f192dc","Type":"ContainerStarted","Data":"1dcdc3f4c7a2479d6bd089ba9359c7e2bebafaa5f2620c830395377a0db42989"} Nov 24 14:31:42 crc kubenswrapper[4790]: I1124 14:31:42.353817 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-wqrxr" event={"ID":"ecff746c-6546-41be-bee5-c07166f192dc","Type":"ContainerStarted","Data":"08741f67c2124c4017b63d958a30ea6ec29c9fad754febdac567a16e7806b4c8"} Nov 24 14:31:43 crc kubenswrapper[4790]: I1124 14:31:43.361751 4790 generic.go:334] "Generic (PLEG): container finished" podID="ecff746c-6546-41be-bee5-c07166f192dc" containerID="1dcdc3f4c7a2479d6bd089ba9359c7e2bebafaa5f2620c830395377a0db42989" exitCode=0 Nov 24 14:31:43 crc kubenswrapper[4790]: I1124 14:31:43.361915 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-wqrxr" event={"ID":"ecff746c-6546-41be-bee5-c07166f192dc","Type":"ContainerDied","Data":"1dcdc3f4c7a2479d6bd089ba9359c7e2bebafaa5f2620c830395377a0db42989"} Nov 24 14:31:43 crc kubenswrapper[4790]: I1124 14:31:43.582355 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="aceb9aee-3ff6-4e3d-a406-2156b12d6265" containerName="rabbitmq" containerID="cri-o://0ec87d8d4a68e138da2ae0c5196b7726954ea36bc435477500ce77ef849abace" gracePeriod=604799 Nov 24 14:31:44 crc kubenswrapper[4790]: I1124 14:31:44.048929 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="e4d75dc1-2209-4c82-b562-9ad8ba78beb7" containerName="rabbitmq" containerID="cri-o://14344e000f6145e53d02947e710228cb00c7ff5419e58822fc92e3d77bd4bee7" gracePeriod=604799 Nov 24 14:31:44 crc kubenswrapper[4790]: I1124 14:31:44.375100 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-wqrxr" event={"ID":"ecff746c-6546-41be-bee5-c07166f192dc","Type":"ContainerStarted","Data":"fbc6287eaed7e81d9101554856b82bca2c140175830a008c338b860df97ff048"} Nov 24 14:31:44 crc kubenswrapper[4790]: I1124 14:31:44.375357 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b7946d7b9-wqrxr" Nov 24 14:31:44 crc kubenswrapper[4790]: I1124 14:31:44.399154 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b7946d7b9-wqrxr" podStartSLOduration=3.399130837 podStartE2EDuration="3.399130837s" podCreationTimestamp="2025-11-24 14:31:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:31:44.393450234 +0000 UTC m=+4752.773343906" watchObservedRunningTime="2025-11-24 14:31:44.399130837 +0000 UTC m=+4752.779024509" Nov 24 14:31:46 crc kubenswrapper[4790]: I1124 14:31:46.553495 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="aceb9aee-3ff6-4e3d-a406-2156b12d6265" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.238:5672: connect: connection refused" Nov 24 14:31:46 crc kubenswrapper[4790]: I1124 14:31:46.790816 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="e4d75dc1-2209-4c82-b562-9ad8ba78beb7" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.239:5672: connect: connection refused" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.054237 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.210148 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/aceb9aee-3ff6-4e3d-a406-2156b12d6265-plugins-conf\") pod \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.210326 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/aceb9aee-3ff6-4e3d-a406-2156b12d6265-rabbitmq-plugins\") pod \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.210419 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/aceb9aee-3ff6-4e3d-a406-2156b12d6265-rabbitmq-confd\") pod \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.210456 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/aceb9aee-3ff6-4e3d-a406-2156b12d6265-rabbitmq-erlang-cookie\") pod \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.210511 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/aceb9aee-3ff6-4e3d-a406-2156b12d6265-erlang-cookie-secret\") pod \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.210568 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/aceb9aee-3ff6-4e3d-a406-2156b12d6265-pod-info\") pod \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.210605 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctfcz\" (UniqueName: \"kubernetes.io/projected/aceb9aee-3ff6-4e3d-a406-2156b12d6265-kube-api-access-ctfcz\") pod \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.210684 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aceb9aee-3ff6-4e3d-a406-2156b12d6265-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "aceb9aee-3ff6-4e3d-a406-2156b12d6265" (UID: "aceb9aee-3ff6-4e3d-a406-2156b12d6265"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.210842 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0e9b38f5-5501-4b3c-8d56-02c85887ccd6\") pod \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.210945 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/aceb9aee-3ff6-4e3d-a406-2156b12d6265-server-conf\") pod \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\" (UID: \"aceb9aee-3ff6-4e3d-a406-2156b12d6265\") " Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.210964 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aceb9aee-3ff6-4e3d-a406-2156b12d6265-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "aceb9aee-3ff6-4e3d-a406-2156b12d6265" (UID: "aceb9aee-3ff6-4e3d-a406-2156b12d6265"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.211375 4790 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/aceb9aee-3ff6-4e3d-a406-2156b12d6265-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.211403 4790 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/aceb9aee-3ff6-4e3d-a406-2156b12d6265-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.211710 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aceb9aee-3ff6-4e3d-a406-2156b12d6265-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "aceb9aee-3ff6-4e3d-a406-2156b12d6265" (UID: "aceb9aee-3ff6-4e3d-a406-2156b12d6265"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.217199 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aceb9aee-3ff6-4e3d-a406-2156b12d6265-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "aceb9aee-3ff6-4e3d-a406-2156b12d6265" (UID: "aceb9aee-3ff6-4e3d-a406-2156b12d6265"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.217370 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aceb9aee-3ff6-4e3d-a406-2156b12d6265-kube-api-access-ctfcz" (OuterVolumeSpecName: "kube-api-access-ctfcz") pod "aceb9aee-3ff6-4e3d-a406-2156b12d6265" (UID: "aceb9aee-3ff6-4e3d-a406-2156b12d6265"). InnerVolumeSpecName "kube-api-access-ctfcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.218088 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/aceb9aee-3ff6-4e3d-a406-2156b12d6265-pod-info" (OuterVolumeSpecName: "pod-info") pod "aceb9aee-3ff6-4e3d-a406-2156b12d6265" (UID: "aceb9aee-3ff6-4e3d-a406-2156b12d6265"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.224358 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0e9b38f5-5501-4b3c-8d56-02c85887ccd6" (OuterVolumeSpecName: "persistence") pod "aceb9aee-3ff6-4e3d-a406-2156b12d6265" (UID: "aceb9aee-3ff6-4e3d-a406-2156b12d6265"). InnerVolumeSpecName "pvc-0e9b38f5-5501-4b3c-8d56-02c85887ccd6". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.234822 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aceb9aee-3ff6-4e3d-a406-2156b12d6265-server-conf" (OuterVolumeSpecName: "server-conf") pod "aceb9aee-3ff6-4e3d-a406-2156b12d6265" (UID: "aceb9aee-3ff6-4e3d-a406-2156b12d6265"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.295975 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aceb9aee-3ff6-4e3d-a406-2156b12d6265-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "aceb9aee-3ff6-4e3d-a406-2156b12d6265" (UID: "aceb9aee-3ff6-4e3d-a406-2156b12d6265"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.312553 4790 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-0e9b38f5-5501-4b3c-8d56-02c85887ccd6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0e9b38f5-5501-4b3c-8d56-02c85887ccd6\") on node \"crc\" " Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.312596 4790 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/aceb9aee-3ff6-4e3d-a406-2156b12d6265-server-conf\") on node \"crc\" DevicePath \"\"" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.312611 4790 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/aceb9aee-3ff6-4e3d-a406-2156b12d6265-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.312623 4790 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/aceb9aee-3ff6-4e3d-a406-2156b12d6265-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.312634 4790 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/aceb9aee-3ff6-4e3d-a406-2156b12d6265-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.312644 4790 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/aceb9aee-3ff6-4e3d-a406-2156b12d6265-pod-info\") on node \"crc\" DevicePath \"\"" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.312653 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctfcz\" (UniqueName: \"kubernetes.io/projected/aceb9aee-3ff6-4e3d-a406-2156b12d6265-kube-api-access-ctfcz\") on node \"crc\" DevicePath \"\"" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.335852 4790 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.336110 4790 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-0e9b38f5-5501-4b3c-8d56-02c85887ccd6" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0e9b38f5-5501-4b3c-8d56-02c85887ccd6") on node "crc" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.414986 4790 reconciler_common.go:293] "Volume detached for volume \"pvc-0e9b38f5-5501-4b3c-8d56-02c85887ccd6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0e9b38f5-5501-4b3c-8d56-02c85887ccd6\") on node \"crc\" DevicePath \"\"" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.428064 4790 generic.go:334] "Generic (PLEG): container finished" podID="aceb9aee-3ff6-4e3d-a406-2156b12d6265" containerID="0ec87d8d4a68e138da2ae0c5196b7726954ea36bc435477500ce77ef849abace" exitCode=0 Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.428139 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.428157 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"aceb9aee-3ff6-4e3d-a406-2156b12d6265","Type":"ContainerDied","Data":"0ec87d8d4a68e138da2ae0c5196b7726954ea36bc435477500ce77ef849abace"} Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.428621 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"aceb9aee-3ff6-4e3d-a406-2156b12d6265","Type":"ContainerDied","Data":"210984cd9937feaf77653007e54ad3cdedb5c62434b34e7d2fb61b5cbf93652a"} Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.428641 4790 scope.go:117] "RemoveContainer" containerID="0ec87d8d4a68e138da2ae0c5196b7726954ea36bc435477500ce77ef849abace" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.433439 4790 generic.go:334] "Generic (PLEG): container finished" podID="e4d75dc1-2209-4c82-b562-9ad8ba78beb7" containerID="14344e000f6145e53d02947e710228cb00c7ff5419e58822fc92e3d77bd4bee7" exitCode=0 Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.433464 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e4d75dc1-2209-4c82-b562-9ad8ba78beb7","Type":"ContainerDied","Data":"14344e000f6145e53d02947e710228cb00c7ff5419e58822fc92e3d77bd4bee7"} Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.456368 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.465538 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.465860 4790 scope.go:117] "RemoveContainer" containerID="b034c4d689b166a01dd04cb8c0670103f0f78292a162be11f936e20f6e05d278" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.478091 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 14:31:50 crc kubenswrapper[4790]: E1124 14:31:50.478425 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aceb9aee-3ff6-4e3d-a406-2156b12d6265" containerName="rabbitmq" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.478437 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="aceb9aee-3ff6-4e3d-a406-2156b12d6265" containerName="rabbitmq" Nov 24 14:31:50 crc kubenswrapper[4790]: E1124 14:31:50.478470 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aceb9aee-3ff6-4e3d-a406-2156b12d6265" containerName="setup-container" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.478476 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="aceb9aee-3ff6-4e3d-a406-2156b12d6265" containerName="setup-container" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.478629 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="aceb9aee-3ff6-4e3d-a406-2156b12d6265" containerName="rabbitmq" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.480733 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.482518 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-nq6p2" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.482772 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.482793 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.485142 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.486609 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.486939 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.506495 4790 scope.go:117] "RemoveContainer" containerID="0ec87d8d4a68e138da2ae0c5196b7726954ea36bc435477500ce77ef849abace" Nov 24 14:31:50 crc kubenswrapper[4790]: E1124 14:31:50.508129 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ec87d8d4a68e138da2ae0c5196b7726954ea36bc435477500ce77ef849abace\": container with ID starting with 0ec87d8d4a68e138da2ae0c5196b7726954ea36bc435477500ce77ef849abace not found: ID does not exist" containerID="0ec87d8d4a68e138da2ae0c5196b7726954ea36bc435477500ce77ef849abace" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.508173 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ec87d8d4a68e138da2ae0c5196b7726954ea36bc435477500ce77ef849abace"} err="failed to get container status \"0ec87d8d4a68e138da2ae0c5196b7726954ea36bc435477500ce77ef849abace\": rpc error: code = NotFound desc = could not find container \"0ec87d8d4a68e138da2ae0c5196b7726954ea36bc435477500ce77ef849abace\": container with ID starting with 0ec87d8d4a68e138da2ae0c5196b7726954ea36bc435477500ce77ef849abace not found: ID does not exist" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.508198 4790 scope.go:117] "RemoveContainer" containerID="b034c4d689b166a01dd04cb8c0670103f0f78292a162be11f936e20f6e05d278" Nov 24 14:31:50 crc kubenswrapper[4790]: E1124 14:31:50.508442 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b034c4d689b166a01dd04cb8c0670103f0f78292a162be11f936e20f6e05d278\": container with ID starting with b034c4d689b166a01dd04cb8c0670103f0f78292a162be11f936e20f6e05d278 not found: ID does not exist" containerID="b034c4d689b166a01dd04cb8c0670103f0f78292a162be11f936e20f6e05d278" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.508466 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b034c4d689b166a01dd04cb8c0670103f0f78292a162be11f936e20f6e05d278"} err="failed to get container status \"b034c4d689b166a01dd04cb8c0670103f0f78292a162be11f936e20f6e05d278\": rpc error: code = NotFound desc = could not find container \"b034c4d689b166a01dd04cb8c0670103f0f78292a162be11f936e20f6e05d278\": container with ID starting with b034c4d689b166a01dd04cb8c0670103f0f78292a162be11f936e20f6e05d278 not found: ID does not exist" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.540786 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.617859 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c1043d54-fc1a-4758-abaa-c53a371cebb8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c1043d54-fc1a-4758-abaa-c53a371cebb8\") " pod="openstack/rabbitmq-server-0" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.617945 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-0e9b38f5-5501-4b3c-8d56-02c85887ccd6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0e9b38f5-5501-4b3c-8d56-02c85887ccd6\") pod \"rabbitmq-server-0\" (UID: \"c1043d54-fc1a-4758-abaa-c53a371cebb8\") " pod="openstack/rabbitmq-server-0" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.618097 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c1043d54-fc1a-4758-abaa-c53a371cebb8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c1043d54-fc1a-4758-abaa-c53a371cebb8\") " pod="openstack/rabbitmq-server-0" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.618157 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c1043d54-fc1a-4758-abaa-c53a371cebb8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c1043d54-fc1a-4758-abaa-c53a371cebb8\") " pod="openstack/rabbitmq-server-0" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.618282 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c1043d54-fc1a-4758-abaa-c53a371cebb8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c1043d54-fc1a-4758-abaa-c53a371cebb8\") " pod="openstack/rabbitmq-server-0" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.618348 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c1043d54-fc1a-4758-abaa-c53a371cebb8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c1043d54-fc1a-4758-abaa-c53a371cebb8\") " pod="openstack/rabbitmq-server-0" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.618376 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c1043d54-fc1a-4758-abaa-c53a371cebb8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c1043d54-fc1a-4758-abaa-c53a371cebb8\") " pod="openstack/rabbitmq-server-0" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.618425 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbkjn\" (UniqueName: \"kubernetes.io/projected/c1043d54-fc1a-4758-abaa-c53a371cebb8-kube-api-access-xbkjn\") pod \"rabbitmq-server-0\" (UID: \"c1043d54-fc1a-4758-abaa-c53a371cebb8\") " pod="openstack/rabbitmq-server-0" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.618520 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c1043d54-fc1a-4758-abaa-c53a371cebb8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c1043d54-fc1a-4758-abaa-c53a371cebb8\") " pod="openstack/rabbitmq-server-0" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.719304 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-erlang-cookie-secret\") pod \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.719355 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-plugins-conf\") pod \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.719399 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-rabbitmq-erlang-cookie\") pod \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.719430 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-rabbitmq-plugins\") pod \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.719586 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0dc142ca-8752-419b-ae3a-93aa61837411\") pod \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.719615 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-pod-info\") pod \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.719691 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvwqs\" (UniqueName: \"kubernetes.io/projected/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-kube-api-access-qvwqs\") pod \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.719725 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-rabbitmq-confd\") pod \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.719801 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-server-conf\") pod \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\" (UID: \"e4d75dc1-2209-4c82-b562-9ad8ba78beb7\") " Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.720032 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c1043d54-fc1a-4758-abaa-c53a371cebb8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c1043d54-fc1a-4758-abaa-c53a371cebb8\") " pod="openstack/rabbitmq-server-0" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.720076 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c1043d54-fc1a-4758-abaa-c53a371cebb8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c1043d54-fc1a-4758-abaa-c53a371cebb8\") " pod="openstack/rabbitmq-server-0" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.720100 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c1043d54-fc1a-4758-abaa-c53a371cebb8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c1043d54-fc1a-4758-abaa-c53a371cebb8\") " pod="openstack/rabbitmq-server-0" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.720123 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c1043d54-fc1a-4758-abaa-c53a371cebb8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c1043d54-fc1a-4758-abaa-c53a371cebb8\") " pod="openstack/rabbitmq-server-0" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.720145 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c1043d54-fc1a-4758-abaa-c53a371cebb8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c1043d54-fc1a-4758-abaa-c53a371cebb8\") " pod="openstack/rabbitmq-server-0" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.720180 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbkjn\" (UniqueName: \"kubernetes.io/projected/c1043d54-fc1a-4758-abaa-c53a371cebb8-kube-api-access-xbkjn\") pod \"rabbitmq-server-0\" (UID: \"c1043d54-fc1a-4758-abaa-c53a371cebb8\") " pod="openstack/rabbitmq-server-0" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.720400 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c1043d54-fc1a-4758-abaa-c53a371cebb8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c1043d54-fc1a-4758-abaa-c53a371cebb8\") " pod="openstack/rabbitmq-server-0" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.720461 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c1043d54-fc1a-4758-abaa-c53a371cebb8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c1043d54-fc1a-4758-abaa-c53a371cebb8\") " pod="openstack/rabbitmq-server-0" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.720534 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-0e9b38f5-5501-4b3c-8d56-02c85887ccd6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0e9b38f5-5501-4b3c-8d56-02c85887ccd6\") pod \"rabbitmq-server-0\" (UID: \"c1043d54-fc1a-4758-abaa-c53a371cebb8\") " pod="openstack/rabbitmq-server-0" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.721821 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c1043d54-fc1a-4758-abaa-c53a371cebb8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c1043d54-fc1a-4758-abaa-c53a371cebb8\") " pod="openstack/rabbitmq-server-0" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.722006 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "e4d75dc1-2209-4c82-b562-9ad8ba78beb7" (UID: "e4d75dc1-2209-4c82-b562-9ad8ba78beb7"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.722261 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c1043d54-fc1a-4758-abaa-c53a371cebb8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c1043d54-fc1a-4758-abaa-c53a371cebb8\") " pod="openstack/rabbitmq-server-0" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.722345 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "e4d75dc1-2209-4c82-b562-9ad8ba78beb7" (UID: "e4d75dc1-2209-4c82-b562-9ad8ba78beb7"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.722433 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c1043d54-fc1a-4758-abaa-c53a371cebb8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c1043d54-fc1a-4758-abaa-c53a371cebb8\") " pod="openstack/rabbitmq-server-0" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.722575 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "e4d75dc1-2209-4c82-b562-9ad8ba78beb7" (UID: "e4d75dc1-2209-4c82-b562-9ad8ba78beb7"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.726592 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-kube-api-access-qvwqs" (OuterVolumeSpecName: "kube-api-access-qvwqs") pod "e4d75dc1-2209-4c82-b562-9ad8ba78beb7" (UID: "e4d75dc1-2209-4c82-b562-9ad8ba78beb7"). InnerVolumeSpecName "kube-api-access-qvwqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.726744 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c1043d54-fc1a-4758-abaa-c53a371cebb8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c1043d54-fc1a-4758-abaa-c53a371cebb8\") " pod="openstack/rabbitmq-server-0" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.728920 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c1043d54-fc1a-4758-abaa-c53a371cebb8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c1043d54-fc1a-4758-abaa-c53a371cebb8\") " pod="openstack/rabbitmq-server-0" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.732745 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-pod-info" (OuterVolumeSpecName: "pod-info") pod "e4d75dc1-2209-4c82-b562-9ad8ba78beb7" (UID: "e4d75dc1-2209-4c82-b562-9ad8ba78beb7"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.733136 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "e4d75dc1-2209-4c82-b562-9ad8ba78beb7" (UID: "e4d75dc1-2209-4c82-b562-9ad8ba78beb7"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.733712 4790 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.733746 4790 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-0e9b38f5-5501-4b3c-8d56-02c85887ccd6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0e9b38f5-5501-4b3c-8d56-02c85887ccd6\") pod \"rabbitmq-server-0\" (UID: \"c1043d54-fc1a-4758-abaa-c53a371cebb8\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/65faf0adfc5f692cc530d451dc932ddb8871a251ab7abdecf7dab36ed8f4c11e/globalmount\"" pod="openstack/rabbitmq-server-0" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.734340 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c1043d54-fc1a-4758-abaa-c53a371cebb8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c1043d54-fc1a-4758-abaa-c53a371cebb8\") " pod="openstack/rabbitmq-server-0" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.738398 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0dc142ca-8752-419b-ae3a-93aa61837411" (OuterVolumeSpecName: "persistence") pod "e4d75dc1-2209-4c82-b562-9ad8ba78beb7" (UID: "e4d75dc1-2209-4c82-b562-9ad8ba78beb7"). InnerVolumeSpecName "pvc-0dc142ca-8752-419b-ae3a-93aa61837411". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.738919 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c1043d54-fc1a-4758-abaa-c53a371cebb8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c1043d54-fc1a-4758-abaa-c53a371cebb8\") " pod="openstack/rabbitmq-server-0" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.742443 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbkjn\" (UniqueName: \"kubernetes.io/projected/c1043d54-fc1a-4758-abaa-c53a371cebb8-kube-api-access-xbkjn\") pod \"rabbitmq-server-0\" (UID: \"c1043d54-fc1a-4758-abaa-c53a371cebb8\") " pod="openstack/rabbitmq-server-0" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.748765 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-server-conf" (OuterVolumeSpecName: "server-conf") pod "e4d75dc1-2209-4c82-b562-9ad8ba78beb7" (UID: "e4d75dc1-2209-4c82-b562-9ad8ba78beb7"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.759225 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-0e9b38f5-5501-4b3c-8d56-02c85887ccd6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0e9b38f5-5501-4b3c-8d56-02c85887ccd6\") pod \"rabbitmq-server-0\" (UID: \"c1043d54-fc1a-4758-abaa-c53a371cebb8\") " pod="openstack/rabbitmq-server-0" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.806182 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "e4d75dc1-2209-4c82-b562-9ad8ba78beb7" (UID: "e4d75dc1-2209-4c82-b562-9ad8ba78beb7"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.811298 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.821987 4790 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-0dc142ca-8752-419b-ae3a-93aa61837411\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0dc142ca-8752-419b-ae3a-93aa61837411\") on node \"crc\" " Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.822032 4790 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-pod-info\") on node \"crc\" DevicePath \"\"" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.822048 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvwqs\" (UniqueName: \"kubernetes.io/projected/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-kube-api-access-qvwqs\") on node \"crc\" DevicePath \"\"" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.822061 4790 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.822073 4790 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-server-conf\") on node \"crc\" DevicePath \"\"" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.822083 4790 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.822093 4790 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.822104 4790 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.822143 4790 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e4d75dc1-2209-4c82-b562-9ad8ba78beb7-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.847324 4790 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.847476 4790 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-0dc142ca-8752-419b-ae3a-93aa61837411" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0dc142ca-8752-419b-ae3a-93aa61837411") on node "crc" Nov 24 14:31:50 crc kubenswrapper[4790]: I1124 14:31:50.923765 4790 reconciler_common.go:293] "Volume detached for volume \"pvc-0dc142ca-8752-419b-ae3a-93aa61837411\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0dc142ca-8752-419b-ae3a-93aa61837411\") on node \"crc\" DevicePath \"\"" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.232507 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.314581 4790 scope.go:117] "RemoveContainer" containerID="571bd0c27e6ad4756f7116ed8b5a40d06301f3d85436cea369594fdb322154a7" Nov 24 14:31:51 crc kubenswrapper[4790]: E1124 14:31:51.314849 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.443336 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c1043d54-fc1a-4758-abaa-c53a371cebb8","Type":"ContainerStarted","Data":"d0a07fde67691713073b94e4eca82478e5415ba551dff2c450829584c5ec52bf"} Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.445836 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.445842 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e4d75dc1-2209-4c82-b562-9ad8ba78beb7","Type":"ContainerDied","Data":"1e8f2e4d16e4c2d66d65fe9aeaca778ff69da1a3e9036fbabb83c29077584197"} Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.445951 4790 scope.go:117] "RemoveContainer" containerID="14344e000f6145e53d02947e710228cb00c7ff5419e58822fc92e3d77bd4bee7" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.471725 4790 scope.go:117] "RemoveContainer" containerID="eec7619ec057661e4f6c86d8f7e343e6ffc6cc4b1d97fe7cb9fd37282b4592ec" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.489964 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.498278 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.513762 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 14:31:51 crc kubenswrapper[4790]: E1124 14:31:51.514223 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4d75dc1-2209-4c82-b562-9ad8ba78beb7" containerName="rabbitmq" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.514255 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4d75dc1-2209-4c82-b562-9ad8ba78beb7" containerName="rabbitmq" Nov 24 14:31:51 crc kubenswrapper[4790]: E1124 14:31:51.514281 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4d75dc1-2209-4c82-b562-9ad8ba78beb7" containerName="setup-container" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.514293 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4d75dc1-2209-4c82-b562-9ad8ba78beb7" containerName="setup-container" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.514567 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4d75dc1-2209-4c82-b562-9ad8ba78beb7" containerName="rabbitmq" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.517734 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.520964 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.521930 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.523112 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-q8mlt" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.523395 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.523660 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.543778 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.546017 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b7946d7b9-wqrxr" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.620490 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-54qbt"] Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.620892 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-98ddfc8f-54qbt" podUID="50b76bef-0c9b-4f75-b902-adaa35fea421" containerName="dnsmasq-dns" containerID="cri-o://ee855f19e1b8ce5aecadd2150e60580a9a9d1536b0634249fea95b1ad77385bb" gracePeriod=10 Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.636541 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hkjl\" (UniqueName: \"kubernetes.io/projected/260598aa-c7fe-453b-8f55-c40dd2037e8d-kube-api-access-2hkjl\") pod \"rabbitmq-cell1-server-0\" (UID: \"260598aa-c7fe-453b-8f55-c40dd2037e8d\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.636615 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/260598aa-c7fe-453b-8f55-c40dd2037e8d-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"260598aa-c7fe-453b-8f55-c40dd2037e8d\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.636641 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/260598aa-c7fe-453b-8f55-c40dd2037e8d-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"260598aa-c7fe-453b-8f55-c40dd2037e8d\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.636699 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-0dc142ca-8752-419b-ae3a-93aa61837411\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0dc142ca-8752-419b-ae3a-93aa61837411\") pod \"rabbitmq-cell1-server-0\" (UID: \"260598aa-c7fe-453b-8f55-c40dd2037e8d\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.636747 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/260598aa-c7fe-453b-8f55-c40dd2037e8d-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"260598aa-c7fe-453b-8f55-c40dd2037e8d\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.636802 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/260598aa-c7fe-453b-8f55-c40dd2037e8d-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"260598aa-c7fe-453b-8f55-c40dd2037e8d\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.636823 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/260598aa-c7fe-453b-8f55-c40dd2037e8d-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"260598aa-c7fe-453b-8f55-c40dd2037e8d\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.637013 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/260598aa-c7fe-453b-8f55-c40dd2037e8d-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"260598aa-c7fe-453b-8f55-c40dd2037e8d\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.637211 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/260598aa-c7fe-453b-8f55-c40dd2037e8d-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"260598aa-c7fe-453b-8f55-c40dd2037e8d\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.738207 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/260598aa-c7fe-453b-8f55-c40dd2037e8d-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"260598aa-c7fe-453b-8f55-c40dd2037e8d\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.738656 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/260598aa-c7fe-453b-8f55-c40dd2037e8d-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"260598aa-c7fe-453b-8f55-c40dd2037e8d\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.738690 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hkjl\" (UniqueName: \"kubernetes.io/projected/260598aa-c7fe-453b-8f55-c40dd2037e8d-kube-api-access-2hkjl\") pod \"rabbitmq-cell1-server-0\" (UID: \"260598aa-c7fe-453b-8f55-c40dd2037e8d\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.738714 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/260598aa-c7fe-453b-8f55-c40dd2037e8d-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"260598aa-c7fe-453b-8f55-c40dd2037e8d\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.738736 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/260598aa-c7fe-453b-8f55-c40dd2037e8d-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"260598aa-c7fe-453b-8f55-c40dd2037e8d\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.738774 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-0dc142ca-8752-419b-ae3a-93aa61837411\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0dc142ca-8752-419b-ae3a-93aa61837411\") pod \"rabbitmq-cell1-server-0\" (UID: \"260598aa-c7fe-453b-8f55-c40dd2037e8d\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.738824 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/260598aa-c7fe-453b-8f55-c40dd2037e8d-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"260598aa-c7fe-453b-8f55-c40dd2037e8d\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.738859 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/260598aa-c7fe-453b-8f55-c40dd2037e8d-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"260598aa-c7fe-453b-8f55-c40dd2037e8d\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.738899 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/260598aa-c7fe-453b-8f55-c40dd2037e8d-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"260598aa-c7fe-453b-8f55-c40dd2037e8d\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.739181 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/260598aa-c7fe-453b-8f55-c40dd2037e8d-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"260598aa-c7fe-453b-8f55-c40dd2037e8d\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.739816 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/260598aa-c7fe-453b-8f55-c40dd2037e8d-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"260598aa-c7fe-453b-8f55-c40dd2037e8d\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.740191 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/260598aa-c7fe-453b-8f55-c40dd2037e8d-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"260598aa-c7fe-453b-8f55-c40dd2037e8d\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.740950 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/260598aa-c7fe-453b-8f55-c40dd2037e8d-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"260598aa-c7fe-453b-8f55-c40dd2037e8d\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.742465 4790 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.742580 4790 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-0dc142ca-8752-419b-ae3a-93aa61837411\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0dc142ca-8752-419b-ae3a-93aa61837411\") pod \"rabbitmq-cell1-server-0\" (UID: \"260598aa-c7fe-453b-8f55-c40dd2037e8d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b46039fa1dd307f30c9f1f8971d3598f69f7cb989892e6b48b5c448c7c9ec3f0/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.745844 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/260598aa-c7fe-453b-8f55-c40dd2037e8d-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"260598aa-c7fe-453b-8f55-c40dd2037e8d\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.746429 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/260598aa-c7fe-453b-8f55-c40dd2037e8d-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"260598aa-c7fe-453b-8f55-c40dd2037e8d\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.751075 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/260598aa-c7fe-453b-8f55-c40dd2037e8d-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"260598aa-c7fe-453b-8f55-c40dd2037e8d\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.757376 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hkjl\" (UniqueName: \"kubernetes.io/projected/260598aa-c7fe-453b-8f55-c40dd2037e8d-kube-api-access-2hkjl\") pod \"rabbitmq-cell1-server-0\" (UID: \"260598aa-c7fe-453b-8f55-c40dd2037e8d\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.780696 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-0dc142ca-8752-419b-ae3a-93aa61837411\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0dc142ca-8752-419b-ae3a-93aa61837411\") pod \"rabbitmq-cell1-server-0\" (UID: \"260598aa-c7fe-453b-8f55-c40dd2037e8d\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.856664 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:31:51 crc kubenswrapper[4790]: I1124 14:31:51.979161 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-54qbt" Nov 24 14:31:52 crc kubenswrapper[4790]: I1124 14:31:52.145586 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/50b76bef-0c9b-4f75-b902-adaa35fea421-dns-svc\") pod \"50b76bef-0c9b-4f75-b902-adaa35fea421\" (UID: \"50b76bef-0c9b-4f75-b902-adaa35fea421\") " Nov 24 14:31:52 crc kubenswrapper[4790]: I1124 14:31:52.145985 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50b76bef-0c9b-4f75-b902-adaa35fea421-config\") pod \"50b76bef-0c9b-4f75-b902-adaa35fea421\" (UID: \"50b76bef-0c9b-4f75-b902-adaa35fea421\") " Nov 24 14:31:52 crc kubenswrapper[4790]: I1124 14:31:52.146037 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snl7g\" (UniqueName: \"kubernetes.io/projected/50b76bef-0c9b-4f75-b902-adaa35fea421-kube-api-access-snl7g\") pod \"50b76bef-0c9b-4f75-b902-adaa35fea421\" (UID: \"50b76bef-0c9b-4f75-b902-adaa35fea421\") " Nov 24 14:31:52 crc kubenswrapper[4790]: I1124 14:31:52.151114 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50b76bef-0c9b-4f75-b902-adaa35fea421-kube-api-access-snl7g" (OuterVolumeSpecName: "kube-api-access-snl7g") pod "50b76bef-0c9b-4f75-b902-adaa35fea421" (UID: "50b76bef-0c9b-4f75-b902-adaa35fea421"). InnerVolumeSpecName "kube-api-access-snl7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:31:52 crc kubenswrapper[4790]: I1124 14:31:52.207655 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50b76bef-0c9b-4f75-b902-adaa35fea421-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "50b76bef-0c9b-4f75-b902-adaa35fea421" (UID: "50b76bef-0c9b-4f75-b902-adaa35fea421"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:31:52 crc kubenswrapper[4790]: I1124 14:31:52.247715 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snl7g\" (UniqueName: \"kubernetes.io/projected/50b76bef-0c9b-4f75-b902-adaa35fea421-kube-api-access-snl7g\") on node \"crc\" DevicePath \"\"" Nov 24 14:31:52 crc kubenswrapper[4790]: I1124 14:31:52.247752 4790 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/50b76bef-0c9b-4f75-b902-adaa35fea421-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:31:52 crc kubenswrapper[4790]: I1124 14:31:52.291668 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50b76bef-0c9b-4f75-b902-adaa35fea421-config" (OuterVolumeSpecName: "config") pod "50b76bef-0c9b-4f75-b902-adaa35fea421" (UID: "50b76bef-0c9b-4f75-b902-adaa35fea421"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:31:52 crc kubenswrapper[4790]: I1124 14:31:52.300783 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 14:31:52 crc kubenswrapper[4790]: I1124 14:31:52.325729 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aceb9aee-3ff6-4e3d-a406-2156b12d6265" path="/var/lib/kubelet/pods/aceb9aee-3ff6-4e3d-a406-2156b12d6265/volumes" Nov 24 14:31:52 crc kubenswrapper[4790]: I1124 14:31:52.326697 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4d75dc1-2209-4c82-b562-9ad8ba78beb7" path="/var/lib/kubelet/pods/e4d75dc1-2209-4c82-b562-9ad8ba78beb7/volumes" Nov 24 14:31:52 crc kubenswrapper[4790]: I1124 14:31:52.349043 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50b76bef-0c9b-4f75-b902-adaa35fea421-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:31:52 crc kubenswrapper[4790]: W1124 14:31:52.395222 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod260598aa_c7fe_453b_8f55_c40dd2037e8d.slice/crio-9ed06b1266a7dd59d2ac4755c9b26118ea62042787dc4aba4c2e4834913eceac WatchSource:0}: Error finding container 9ed06b1266a7dd59d2ac4755c9b26118ea62042787dc4aba4c2e4834913eceac: Status 404 returned error can't find the container with id 9ed06b1266a7dd59d2ac4755c9b26118ea62042787dc4aba4c2e4834913eceac Nov 24 14:31:52 crc kubenswrapper[4790]: I1124 14:31:52.459969 4790 generic.go:334] "Generic (PLEG): container finished" podID="50b76bef-0c9b-4f75-b902-adaa35fea421" containerID="ee855f19e1b8ce5aecadd2150e60580a9a9d1536b0634249fea95b1ad77385bb" exitCode=0 Nov 24 14:31:52 crc kubenswrapper[4790]: I1124 14:31:52.460032 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-54qbt" Nov 24 14:31:52 crc kubenswrapper[4790]: I1124 14:31:52.460072 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-54qbt" event={"ID":"50b76bef-0c9b-4f75-b902-adaa35fea421","Type":"ContainerDied","Data":"ee855f19e1b8ce5aecadd2150e60580a9a9d1536b0634249fea95b1ad77385bb"} Nov 24 14:31:52 crc kubenswrapper[4790]: I1124 14:31:52.460112 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-54qbt" event={"ID":"50b76bef-0c9b-4f75-b902-adaa35fea421","Type":"ContainerDied","Data":"063ecd216bfa342df8f40e5b263b96a3ae2b4a6b646cd599f044c70f520d48ff"} Nov 24 14:31:52 crc kubenswrapper[4790]: I1124 14:31:52.460138 4790 scope.go:117] "RemoveContainer" containerID="ee855f19e1b8ce5aecadd2150e60580a9a9d1536b0634249fea95b1ad77385bb" Nov 24 14:31:52 crc kubenswrapper[4790]: I1124 14:31:52.463672 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"260598aa-c7fe-453b-8f55-c40dd2037e8d","Type":"ContainerStarted","Data":"9ed06b1266a7dd59d2ac4755c9b26118ea62042787dc4aba4c2e4834913eceac"} Nov 24 14:31:52 crc kubenswrapper[4790]: I1124 14:31:52.486730 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-54qbt"] Nov 24 14:31:52 crc kubenswrapper[4790]: I1124 14:31:52.492680 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-54qbt"] Nov 24 14:31:52 crc kubenswrapper[4790]: I1124 14:31:52.494047 4790 scope.go:117] "RemoveContainer" containerID="e22026d8046d7f6bc2e5834577e923754455d12ad5d84eb624fb4204d5b59c22" Nov 24 14:31:52 crc kubenswrapper[4790]: I1124 14:31:52.513349 4790 scope.go:117] "RemoveContainer" containerID="ee855f19e1b8ce5aecadd2150e60580a9a9d1536b0634249fea95b1ad77385bb" Nov 24 14:31:52 crc kubenswrapper[4790]: E1124 14:31:52.513785 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee855f19e1b8ce5aecadd2150e60580a9a9d1536b0634249fea95b1ad77385bb\": container with ID starting with ee855f19e1b8ce5aecadd2150e60580a9a9d1536b0634249fea95b1ad77385bb not found: ID does not exist" containerID="ee855f19e1b8ce5aecadd2150e60580a9a9d1536b0634249fea95b1ad77385bb" Nov 24 14:31:52 crc kubenswrapper[4790]: I1124 14:31:52.513876 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee855f19e1b8ce5aecadd2150e60580a9a9d1536b0634249fea95b1ad77385bb"} err="failed to get container status \"ee855f19e1b8ce5aecadd2150e60580a9a9d1536b0634249fea95b1ad77385bb\": rpc error: code = NotFound desc = could not find container \"ee855f19e1b8ce5aecadd2150e60580a9a9d1536b0634249fea95b1ad77385bb\": container with ID starting with ee855f19e1b8ce5aecadd2150e60580a9a9d1536b0634249fea95b1ad77385bb not found: ID does not exist" Nov 24 14:31:52 crc kubenswrapper[4790]: I1124 14:31:52.514001 4790 scope.go:117] "RemoveContainer" containerID="e22026d8046d7f6bc2e5834577e923754455d12ad5d84eb624fb4204d5b59c22" Nov 24 14:31:52 crc kubenswrapper[4790]: E1124 14:31:52.514348 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e22026d8046d7f6bc2e5834577e923754455d12ad5d84eb624fb4204d5b59c22\": container with ID starting with e22026d8046d7f6bc2e5834577e923754455d12ad5d84eb624fb4204d5b59c22 not found: ID does not exist" containerID="e22026d8046d7f6bc2e5834577e923754455d12ad5d84eb624fb4204d5b59c22" Nov 24 14:31:52 crc kubenswrapper[4790]: I1124 14:31:52.514390 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e22026d8046d7f6bc2e5834577e923754455d12ad5d84eb624fb4204d5b59c22"} err="failed to get container status \"e22026d8046d7f6bc2e5834577e923754455d12ad5d84eb624fb4204d5b59c22\": rpc error: code = NotFound desc = could not find container \"e22026d8046d7f6bc2e5834577e923754455d12ad5d84eb624fb4204d5b59c22\": container with ID starting with e22026d8046d7f6bc2e5834577e923754455d12ad5d84eb624fb4204d5b59c22 not found: ID does not exist" Nov 24 14:31:53 crc kubenswrapper[4790]: I1124 14:31:53.480307 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c1043d54-fc1a-4758-abaa-c53a371cebb8","Type":"ContainerStarted","Data":"94abdc18fc80af2484b962a9037a6d679ca4c784aa9ec536f4fdbc5aa3d8cc62"} Nov 24 14:31:54 crc kubenswrapper[4790]: I1124 14:31:54.327038 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50b76bef-0c9b-4f75-b902-adaa35fea421" path="/var/lib/kubelet/pods/50b76bef-0c9b-4f75-b902-adaa35fea421/volumes" Nov 24 14:31:54 crc kubenswrapper[4790]: I1124 14:31:54.490358 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"260598aa-c7fe-453b-8f55-c40dd2037e8d","Type":"ContainerStarted","Data":"e130154f058ec3e7968692e7ccdee4d18bb63f2c768752ff9c116a3c77b2ecf9"} Nov 24 14:32:03 crc kubenswrapper[4790]: I1124 14:32:03.316094 4790 scope.go:117] "RemoveContainer" containerID="571bd0c27e6ad4756f7116ed8b5a40d06301f3d85436cea369594fdb322154a7" Nov 24 14:32:03 crc kubenswrapper[4790]: E1124 14:32:03.317168 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:32:13 crc kubenswrapper[4790]: I1124 14:32:13.773431 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-67gl4"] Nov 24 14:32:13 crc kubenswrapper[4790]: E1124 14:32:13.774769 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50b76bef-0c9b-4f75-b902-adaa35fea421" containerName="dnsmasq-dns" Nov 24 14:32:13 crc kubenswrapper[4790]: I1124 14:32:13.774789 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="50b76bef-0c9b-4f75-b902-adaa35fea421" containerName="dnsmasq-dns" Nov 24 14:32:13 crc kubenswrapper[4790]: E1124 14:32:13.774801 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50b76bef-0c9b-4f75-b902-adaa35fea421" containerName="init" Nov 24 14:32:13 crc kubenswrapper[4790]: I1124 14:32:13.774812 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="50b76bef-0c9b-4f75-b902-adaa35fea421" containerName="init" Nov 24 14:32:13 crc kubenswrapper[4790]: I1124 14:32:13.775059 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="50b76bef-0c9b-4f75-b902-adaa35fea421" containerName="dnsmasq-dns" Nov 24 14:32:13 crc kubenswrapper[4790]: I1124 14:32:13.776515 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-67gl4" Nov 24 14:32:13 crc kubenswrapper[4790]: I1124 14:32:13.788225 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-67gl4"] Nov 24 14:32:13 crc kubenswrapper[4790]: I1124 14:32:13.926464 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86-catalog-content\") pod \"certified-operators-67gl4\" (UID: \"cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86\") " pod="openshift-marketplace/certified-operators-67gl4" Nov 24 14:32:13 crc kubenswrapper[4790]: I1124 14:32:13.926698 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6vtn\" (UniqueName: \"kubernetes.io/projected/cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86-kube-api-access-x6vtn\") pod \"certified-operators-67gl4\" (UID: \"cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86\") " pod="openshift-marketplace/certified-operators-67gl4" Nov 24 14:32:13 crc kubenswrapper[4790]: I1124 14:32:13.926745 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86-utilities\") pod \"certified-operators-67gl4\" (UID: \"cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86\") " pod="openshift-marketplace/certified-operators-67gl4" Nov 24 14:32:14 crc kubenswrapper[4790]: I1124 14:32:14.028156 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86-catalog-content\") pod \"certified-operators-67gl4\" (UID: \"cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86\") " pod="openshift-marketplace/certified-operators-67gl4" Nov 24 14:32:14 crc kubenswrapper[4790]: I1124 14:32:14.028267 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6vtn\" (UniqueName: \"kubernetes.io/projected/cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86-kube-api-access-x6vtn\") pod \"certified-operators-67gl4\" (UID: \"cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86\") " pod="openshift-marketplace/certified-operators-67gl4" Nov 24 14:32:14 crc kubenswrapper[4790]: I1124 14:32:14.028296 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86-utilities\") pod \"certified-operators-67gl4\" (UID: \"cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86\") " pod="openshift-marketplace/certified-operators-67gl4" Nov 24 14:32:14 crc kubenswrapper[4790]: I1124 14:32:14.028693 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86-catalog-content\") pod \"certified-operators-67gl4\" (UID: \"cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86\") " pod="openshift-marketplace/certified-operators-67gl4" Nov 24 14:32:14 crc kubenswrapper[4790]: I1124 14:32:14.028811 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86-utilities\") pod \"certified-operators-67gl4\" (UID: \"cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86\") " pod="openshift-marketplace/certified-operators-67gl4" Nov 24 14:32:14 crc kubenswrapper[4790]: I1124 14:32:14.046301 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6vtn\" (UniqueName: \"kubernetes.io/projected/cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86-kube-api-access-x6vtn\") pod \"certified-operators-67gl4\" (UID: \"cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86\") " pod="openshift-marketplace/certified-operators-67gl4" Nov 24 14:32:14 crc kubenswrapper[4790]: I1124 14:32:14.095207 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-67gl4" Nov 24 14:32:14 crc kubenswrapper[4790]: I1124 14:32:14.575341 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-67gl4"] Nov 24 14:32:14 crc kubenswrapper[4790]: I1124 14:32:14.674751 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-67gl4" event={"ID":"cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86","Type":"ContainerStarted","Data":"7df287ba39c253793a7839b92432deb75cb8830a50e9a36c1d371ce7b967fbba"} Nov 24 14:32:15 crc kubenswrapper[4790]: I1124 14:32:15.684358 4790 generic.go:334] "Generic (PLEG): container finished" podID="cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86" containerID="124fc9bc46bdb40674ea3aa4e645886ae37828488581dbcad54594dacd8992ee" exitCode=0 Nov 24 14:32:15 crc kubenswrapper[4790]: I1124 14:32:15.684469 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-67gl4" event={"ID":"cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86","Type":"ContainerDied","Data":"124fc9bc46bdb40674ea3aa4e645886ae37828488581dbcad54594dacd8992ee"} Nov 24 14:32:15 crc kubenswrapper[4790]: I1124 14:32:15.688122 4790 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 14:32:16 crc kubenswrapper[4790]: I1124 14:32:16.314385 4790 scope.go:117] "RemoveContainer" containerID="571bd0c27e6ad4756f7116ed8b5a40d06301f3d85436cea369594fdb322154a7" Nov 24 14:32:16 crc kubenswrapper[4790]: E1124 14:32:16.314649 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:32:16 crc kubenswrapper[4790]: I1124 14:32:16.694781 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-67gl4" event={"ID":"cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86","Type":"ContainerStarted","Data":"b6ba943229ee26a1afe9edacb9f2c95f991ee96cbbca186aaaec6198fb1f3cbe"} Nov 24 14:32:17 crc kubenswrapper[4790]: I1124 14:32:17.702654 4790 generic.go:334] "Generic (PLEG): container finished" podID="cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86" containerID="b6ba943229ee26a1afe9edacb9f2c95f991ee96cbbca186aaaec6198fb1f3cbe" exitCode=0 Nov 24 14:32:17 crc kubenswrapper[4790]: I1124 14:32:17.702731 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-67gl4" event={"ID":"cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86","Type":"ContainerDied","Data":"b6ba943229ee26a1afe9edacb9f2c95f991ee96cbbca186aaaec6198fb1f3cbe"} Nov 24 14:32:18 crc kubenswrapper[4790]: I1124 14:32:18.713353 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-67gl4" event={"ID":"cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86","Type":"ContainerStarted","Data":"478c803a8567e3c904d954f3475baa6bde42ebbcdd6e6ded53d0b0ccaa3cbce5"} Nov 24 14:32:18 crc kubenswrapper[4790]: I1124 14:32:18.730486 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-67gl4" podStartSLOduration=3.297038125 podStartE2EDuration="5.730467196s" podCreationTimestamp="2025-11-24 14:32:13 +0000 UTC" firstStartedPulling="2025-11-24 14:32:15.68780362 +0000 UTC m=+4784.067697302" lastFinishedPulling="2025-11-24 14:32:18.121232711 +0000 UTC m=+4786.501126373" observedRunningTime="2025-11-24 14:32:18.728833459 +0000 UTC m=+4787.108727191" watchObservedRunningTime="2025-11-24 14:32:18.730467196 +0000 UTC m=+4787.110360858" Nov 24 14:32:24 crc kubenswrapper[4790]: I1124 14:32:24.095947 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-67gl4" Nov 24 14:32:24 crc kubenswrapper[4790]: I1124 14:32:24.096752 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-67gl4" Nov 24 14:32:24 crc kubenswrapper[4790]: I1124 14:32:24.168140 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-67gl4" Nov 24 14:32:24 crc kubenswrapper[4790]: I1124 14:32:24.804496 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-67gl4" Nov 24 14:32:24 crc kubenswrapper[4790]: I1124 14:32:24.851067 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-67gl4"] Nov 24 14:32:25 crc kubenswrapper[4790]: I1124 14:32:25.790333 4790 generic.go:334] "Generic (PLEG): container finished" podID="c1043d54-fc1a-4758-abaa-c53a371cebb8" containerID="94abdc18fc80af2484b962a9037a6d679ca4c784aa9ec536f4fdbc5aa3d8cc62" exitCode=0 Nov 24 14:32:25 crc kubenswrapper[4790]: I1124 14:32:25.790466 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c1043d54-fc1a-4758-abaa-c53a371cebb8","Type":"ContainerDied","Data":"94abdc18fc80af2484b962a9037a6d679ca4c784aa9ec536f4fdbc5aa3d8cc62"} Nov 24 14:32:26 crc kubenswrapper[4790]: I1124 14:32:26.801225 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c1043d54-fc1a-4758-abaa-c53a371cebb8","Type":"ContainerStarted","Data":"9f99516547b94939c50f5e15faecc356587859fc2e7747cce26a8a54422be4c3"} Nov 24 14:32:26 crc kubenswrapper[4790]: I1124 14:32:26.801374 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-67gl4" podUID="cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86" containerName="registry-server" containerID="cri-o://478c803a8567e3c904d954f3475baa6bde42ebbcdd6e6ded53d0b0ccaa3cbce5" gracePeriod=2 Nov 24 14:32:26 crc kubenswrapper[4790]: I1124 14:32:26.801815 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 24 14:32:26 crc kubenswrapper[4790]: I1124 14:32:26.850359 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.85034283 podStartE2EDuration="36.85034283s" podCreationTimestamp="2025-11-24 14:31:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:32:26.845053438 +0000 UTC m=+4795.224947120" watchObservedRunningTime="2025-11-24 14:32:26.85034283 +0000 UTC m=+4795.230236492" Nov 24 14:32:27 crc kubenswrapper[4790]: I1124 14:32:27.202769 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-67gl4" Nov 24 14:32:27 crc kubenswrapper[4790]: I1124 14:32:27.314710 4790 scope.go:117] "RemoveContainer" containerID="571bd0c27e6ad4756f7116ed8b5a40d06301f3d85436cea369594fdb322154a7" Nov 24 14:32:27 crc kubenswrapper[4790]: E1124 14:32:27.315094 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:32:27 crc kubenswrapper[4790]: I1124 14:32:27.348265 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6vtn\" (UniqueName: \"kubernetes.io/projected/cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86-kube-api-access-x6vtn\") pod \"cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86\" (UID: \"cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86\") " Nov 24 14:32:27 crc kubenswrapper[4790]: I1124 14:32:27.348471 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86-catalog-content\") pod \"cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86\" (UID: \"cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86\") " Nov 24 14:32:27 crc kubenswrapper[4790]: I1124 14:32:27.348539 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86-utilities\") pod \"cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86\" (UID: \"cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86\") " Nov 24 14:32:27 crc kubenswrapper[4790]: I1124 14:32:27.349644 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86-utilities" (OuterVolumeSpecName: "utilities") pod "cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86" (UID: "cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:32:27 crc kubenswrapper[4790]: I1124 14:32:27.354300 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86-kube-api-access-x6vtn" (OuterVolumeSpecName: "kube-api-access-x6vtn") pod "cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86" (UID: "cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86"). InnerVolumeSpecName "kube-api-access-x6vtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:32:27 crc kubenswrapper[4790]: I1124 14:32:27.450021 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6vtn\" (UniqueName: \"kubernetes.io/projected/cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86-kube-api-access-x6vtn\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:27 crc kubenswrapper[4790]: I1124 14:32:27.450066 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:27 crc kubenswrapper[4790]: I1124 14:32:27.557979 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86" (UID: "cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:32:27 crc kubenswrapper[4790]: I1124 14:32:27.652742 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:27 crc kubenswrapper[4790]: I1124 14:32:27.810831 4790 generic.go:334] "Generic (PLEG): container finished" podID="260598aa-c7fe-453b-8f55-c40dd2037e8d" containerID="e130154f058ec3e7968692e7ccdee4d18bb63f2c768752ff9c116a3c77b2ecf9" exitCode=0 Nov 24 14:32:27 crc kubenswrapper[4790]: I1124 14:32:27.811001 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"260598aa-c7fe-453b-8f55-c40dd2037e8d","Type":"ContainerDied","Data":"e130154f058ec3e7968692e7ccdee4d18bb63f2c768752ff9c116a3c77b2ecf9"} Nov 24 14:32:27 crc kubenswrapper[4790]: I1124 14:32:27.813249 4790 generic.go:334] "Generic (PLEG): container finished" podID="cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86" containerID="478c803a8567e3c904d954f3475baa6bde42ebbcdd6e6ded53d0b0ccaa3cbce5" exitCode=0 Nov 24 14:32:27 crc kubenswrapper[4790]: I1124 14:32:27.813392 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-67gl4" Nov 24 14:32:27 crc kubenswrapper[4790]: I1124 14:32:27.813389 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-67gl4" event={"ID":"cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86","Type":"ContainerDied","Data":"478c803a8567e3c904d954f3475baa6bde42ebbcdd6e6ded53d0b0ccaa3cbce5"} Nov 24 14:32:27 crc kubenswrapper[4790]: I1124 14:32:27.813649 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-67gl4" event={"ID":"cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86","Type":"ContainerDied","Data":"7df287ba39c253793a7839b92432deb75cb8830a50e9a36c1d371ce7b967fbba"} Nov 24 14:32:27 crc kubenswrapper[4790]: I1124 14:32:27.813695 4790 scope.go:117] "RemoveContainer" containerID="478c803a8567e3c904d954f3475baa6bde42ebbcdd6e6ded53d0b0ccaa3cbce5" Nov 24 14:32:27 crc kubenswrapper[4790]: I1124 14:32:27.850529 4790 scope.go:117] "RemoveContainer" containerID="b6ba943229ee26a1afe9edacb9f2c95f991ee96cbbca186aaaec6198fb1f3cbe" Nov 24 14:32:27 crc kubenswrapper[4790]: I1124 14:32:27.892196 4790 scope.go:117] "RemoveContainer" containerID="124fc9bc46bdb40674ea3aa4e645886ae37828488581dbcad54594dacd8992ee" Nov 24 14:32:27 crc kubenswrapper[4790]: I1124 14:32:27.902785 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-67gl4"] Nov 24 14:32:27 crc kubenswrapper[4790]: I1124 14:32:27.911831 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-67gl4"] Nov 24 14:32:27 crc kubenswrapper[4790]: I1124 14:32:27.953987 4790 scope.go:117] "RemoveContainer" containerID="478c803a8567e3c904d954f3475baa6bde42ebbcdd6e6ded53d0b0ccaa3cbce5" Nov 24 14:32:27 crc kubenswrapper[4790]: E1124 14:32:27.954528 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"478c803a8567e3c904d954f3475baa6bde42ebbcdd6e6ded53d0b0ccaa3cbce5\": container with ID starting with 478c803a8567e3c904d954f3475baa6bde42ebbcdd6e6ded53d0b0ccaa3cbce5 not found: ID does not exist" containerID="478c803a8567e3c904d954f3475baa6bde42ebbcdd6e6ded53d0b0ccaa3cbce5" Nov 24 14:32:27 crc kubenswrapper[4790]: I1124 14:32:27.954604 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"478c803a8567e3c904d954f3475baa6bde42ebbcdd6e6ded53d0b0ccaa3cbce5"} err="failed to get container status \"478c803a8567e3c904d954f3475baa6bde42ebbcdd6e6ded53d0b0ccaa3cbce5\": rpc error: code = NotFound desc = could not find container \"478c803a8567e3c904d954f3475baa6bde42ebbcdd6e6ded53d0b0ccaa3cbce5\": container with ID starting with 478c803a8567e3c904d954f3475baa6bde42ebbcdd6e6ded53d0b0ccaa3cbce5 not found: ID does not exist" Nov 24 14:32:27 crc kubenswrapper[4790]: I1124 14:32:27.954655 4790 scope.go:117] "RemoveContainer" containerID="b6ba943229ee26a1afe9edacb9f2c95f991ee96cbbca186aaaec6198fb1f3cbe" Nov 24 14:32:27 crc kubenswrapper[4790]: E1124 14:32:27.955170 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6ba943229ee26a1afe9edacb9f2c95f991ee96cbbca186aaaec6198fb1f3cbe\": container with ID starting with b6ba943229ee26a1afe9edacb9f2c95f991ee96cbbca186aaaec6198fb1f3cbe not found: ID does not exist" containerID="b6ba943229ee26a1afe9edacb9f2c95f991ee96cbbca186aaaec6198fb1f3cbe" Nov 24 14:32:27 crc kubenswrapper[4790]: I1124 14:32:27.955229 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6ba943229ee26a1afe9edacb9f2c95f991ee96cbbca186aaaec6198fb1f3cbe"} err="failed to get container status \"b6ba943229ee26a1afe9edacb9f2c95f991ee96cbbca186aaaec6198fb1f3cbe\": rpc error: code = NotFound desc = could not find container \"b6ba943229ee26a1afe9edacb9f2c95f991ee96cbbca186aaaec6198fb1f3cbe\": container with ID starting with b6ba943229ee26a1afe9edacb9f2c95f991ee96cbbca186aaaec6198fb1f3cbe not found: ID does not exist" Nov 24 14:32:27 crc kubenswrapper[4790]: I1124 14:32:27.955271 4790 scope.go:117] "RemoveContainer" containerID="124fc9bc46bdb40674ea3aa4e645886ae37828488581dbcad54594dacd8992ee" Nov 24 14:32:27 crc kubenswrapper[4790]: E1124 14:32:27.955629 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"124fc9bc46bdb40674ea3aa4e645886ae37828488581dbcad54594dacd8992ee\": container with ID starting with 124fc9bc46bdb40674ea3aa4e645886ae37828488581dbcad54594dacd8992ee not found: ID does not exist" containerID="124fc9bc46bdb40674ea3aa4e645886ae37828488581dbcad54594dacd8992ee" Nov 24 14:32:27 crc kubenswrapper[4790]: I1124 14:32:27.955867 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"124fc9bc46bdb40674ea3aa4e645886ae37828488581dbcad54594dacd8992ee"} err="failed to get container status \"124fc9bc46bdb40674ea3aa4e645886ae37828488581dbcad54594dacd8992ee\": rpc error: code = NotFound desc = could not find container \"124fc9bc46bdb40674ea3aa4e645886ae37828488581dbcad54594dacd8992ee\": container with ID starting with 124fc9bc46bdb40674ea3aa4e645886ae37828488581dbcad54594dacd8992ee not found: ID does not exist" Nov 24 14:32:28 crc kubenswrapper[4790]: I1124 14:32:28.327147 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86" path="/var/lib/kubelet/pods/cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86/volumes" Nov 24 14:32:28 crc kubenswrapper[4790]: I1124 14:32:28.824099 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"260598aa-c7fe-453b-8f55-c40dd2037e8d","Type":"ContainerStarted","Data":"95c9cd0701154aef9af91e006d228364df856490f52e9b93f4a9275557c65228"} Nov 24 14:32:28 crc kubenswrapper[4790]: I1124 14:32:28.824345 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:32:28 crc kubenswrapper[4790]: I1124 14:32:28.847056 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.847035267 podStartE2EDuration="37.847035267s" podCreationTimestamp="2025-11-24 14:31:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:32:28.843296499 +0000 UTC m=+4797.223190151" watchObservedRunningTime="2025-11-24 14:32:28.847035267 +0000 UTC m=+4797.226928929" Nov 24 14:32:40 crc kubenswrapper[4790]: I1124 14:32:40.315570 4790 scope.go:117] "RemoveContainer" containerID="571bd0c27e6ad4756f7116ed8b5a40d06301f3d85436cea369594fdb322154a7" Nov 24 14:32:40 crc kubenswrapper[4790]: E1124 14:32:40.316609 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:32:40 crc kubenswrapper[4790]: I1124 14:32:40.816390 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 24 14:32:41 crc kubenswrapper[4790]: I1124 14:32:41.862144 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:32:52 crc kubenswrapper[4790]: I1124 14:32:52.957052 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1-default"] Nov 24 14:32:52 crc kubenswrapper[4790]: E1124 14:32:52.957993 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86" containerName="registry-server" Nov 24 14:32:52 crc kubenswrapper[4790]: I1124 14:32:52.958011 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86" containerName="registry-server" Nov 24 14:32:52 crc kubenswrapper[4790]: E1124 14:32:52.958055 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86" containerName="extract-utilities" Nov 24 14:32:52 crc kubenswrapper[4790]: I1124 14:32:52.958064 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86" containerName="extract-utilities" Nov 24 14:32:52 crc kubenswrapper[4790]: E1124 14:32:52.958079 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86" containerName="extract-content" Nov 24 14:32:52 crc kubenswrapper[4790]: I1124 14:32:52.958086 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86" containerName="extract-content" Nov 24 14:32:52 crc kubenswrapper[4790]: I1124 14:32:52.958283 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd87fabe-fafe-4dfc-ae0e-72c1ef91ba86" containerName="registry-server" Nov 24 14:32:52 crc kubenswrapper[4790]: I1124 14:32:52.958919 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Nov 24 14:32:52 crc kubenswrapper[4790]: I1124 14:32:52.960955 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-fx2g4" Nov 24 14:32:52 crc kubenswrapper[4790]: I1124 14:32:52.963287 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Nov 24 14:32:53 crc kubenswrapper[4790]: I1124 14:32:53.062120 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsz25\" (UniqueName: \"kubernetes.io/projected/0584256d-35c9-4cf1-953b-78d1fc050fc6-kube-api-access-vsz25\") pod \"mariadb-client-1-default\" (UID: \"0584256d-35c9-4cf1-953b-78d1fc050fc6\") " pod="openstack/mariadb-client-1-default" Nov 24 14:32:53 crc kubenswrapper[4790]: I1124 14:32:53.163763 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsz25\" (UniqueName: \"kubernetes.io/projected/0584256d-35c9-4cf1-953b-78d1fc050fc6-kube-api-access-vsz25\") pod \"mariadb-client-1-default\" (UID: \"0584256d-35c9-4cf1-953b-78d1fc050fc6\") " pod="openstack/mariadb-client-1-default" Nov 24 14:32:53 crc kubenswrapper[4790]: I1124 14:32:53.180357 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsz25\" (UniqueName: \"kubernetes.io/projected/0584256d-35c9-4cf1-953b-78d1fc050fc6-kube-api-access-vsz25\") pod \"mariadb-client-1-default\" (UID: \"0584256d-35c9-4cf1-953b-78d1fc050fc6\") " pod="openstack/mariadb-client-1-default" Nov 24 14:32:53 crc kubenswrapper[4790]: I1124 14:32:53.277094 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Nov 24 14:32:53 crc kubenswrapper[4790]: I1124 14:32:53.816486 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Nov 24 14:32:53 crc kubenswrapper[4790]: W1124 14:32:53.829321 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0584256d_35c9_4cf1_953b_78d1fc050fc6.slice/crio-d426a01fdb827a55ccdf6ffdfe476ea34870d952bd9af8564f53d35c5021109a WatchSource:0}: Error finding container d426a01fdb827a55ccdf6ffdfe476ea34870d952bd9af8564f53d35c5021109a: Status 404 returned error can't find the container with id d426a01fdb827a55ccdf6ffdfe476ea34870d952bd9af8564f53d35c5021109a Nov 24 14:32:54 crc kubenswrapper[4790]: I1124 14:32:54.047818 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"0584256d-35c9-4cf1-953b-78d1fc050fc6","Type":"ContainerStarted","Data":"d426a01fdb827a55ccdf6ffdfe476ea34870d952bd9af8564f53d35c5021109a"} Nov 24 14:32:55 crc kubenswrapper[4790]: I1124 14:32:55.058501 4790 generic.go:334] "Generic (PLEG): container finished" podID="0584256d-35c9-4cf1-953b-78d1fc050fc6" containerID="ca0870a13339f5a612e870cbdb9d7e0b9dfe2acfb69a8e31c89d5756866515c3" exitCode=0 Nov 24 14:32:55 crc kubenswrapper[4790]: I1124 14:32:55.059472 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"0584256d-35c9-4cf1-953b-78d1fc050fc6","Type":"ContainerDied","Data":"ca0870a13339f5a612e870cbdb9d7e0b9dfe2acfb69a8e31c89d5756866515c3"} Nov 24 14:32:55 crc kubenswrapper[4790]: I1124 14:32:55.314739 4790 scope.go:117] "RemoveContainer" containerID="571bd0c27e6ad4756f7116ed8b5a40d06301f3d85436cea369594fdb322154a7" Nov 24 14:32:55 crc kubenswrapper[4790]: E1124 14:32:55.315303 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:32:56 crc kubenswrapper[4790]: I1124 14:32:56.495515 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Nov 24 14:32:56 crc kubenswrapper[4790]: I1124 14:32:56.523232 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsz25\" (UniqueName: \"kubernetes.io/projected/0584256d-35c9-4cf1-953b-78d1fc050fc6-kube-api-access-vsz25\") pod \"0584256d-35c9-4cf1-953b-78d1fc050fc6\" (UID: \"0584256d-35c9-4cf1-953b-78d1fc050fc6\") " Nov 24 14:32:56 crc kubenswrapper[4790]: I1124 14:32:56.529594 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0584256d-35c9-4cf1-953b-78d1fc050fc6-kube-api-access-vsz25" (OuterVolumeSpecName: "kube-api-access-vsz25") pod "0584256d-35c9-4cf1-953b-78d1fc050fc6" (UID: "0584256d-35c9-4cf1-953b-78d1fc050fc6"). InnerVolumeSpecName "kube-api-access-vsz25". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:32:56 crc kubenswrapper[4790]: I1124 14:32:56.533997 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1-default_0584256d-35c9-4cf1-953b-78d1fc050fc6/mariadb-client-1-default/0.log" Nov 24 14:32:56 crc kubenswrapper[4790]: I1124 14:32:56.563958 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1-default"] Nov 24 14:32:56 crc kubenswrapper[4790]: I1124 14:32:56.571791 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1-default"] Nov 24 14:32:56 crc kubenswrapper[4790]: I1124 14:32:56.625453 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsz25\" (UniqueName: \"kubernetes.io/projected/0584256d-35c9-4cf1-953b-78d1fc050fc6-kube-api-access-vsz25\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:56 crc kubenswrapper[4790]: I1124 14:32:56.980595 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2-default"] Nov 24 14:32:56 crc kubenswrapper[4790]: E1124 14:32:56.981196 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0584256d-35c9-4cf1-953b-78d1fc050fc6" containerName="mariadb-client-1-default" Nov 24 14:32:56 crc kubenswrapper[4790]: I1124 14:32:56.981231 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="0584256d-35c9-4cf1-953b-78d1fc050fc6" containerName="mariadb-client-1-default" Nov 24 14:32:56 crc kubenswrapper[4790]: I1124 14:32:56.981536 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="0584256d-35c9-4cf1-953b-78d1fc050fc6" containerName="mariadb-client-1-default" Nov 24 14:32:56 crc kubenswrapper[4790]: I1124 14:32:56.982448 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Nov 24 14:32:56 crc kubenswrapper[4790]: I1124 14:32:56.992691 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Nov 24 14:32:57 crc kubenswrapper[4790]: I1124 14:32:57.029956 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhtz7\" (UniqueName: \"kubernetes.io/projected/e5d5fa4a-219e-47bd-b7a6-837c6a2c7170-kube-api-access-nhtz7\") pod \"mariadb-client-2-default\" (UID: \"e5d5fa4a-219e-47bd-b7a6-837c6a2c7170\") " pod="openstack/mariadb-client-2-default" Nov 24 14:32:57 crc kubenswrapper[4790]: I1124 14:32:57.079953 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d426a01fdb827a55ccdf6ffdfe476ea34870d952bd9af8564f53d35c5021109a" Nov 24 14:32:57 crc kubenswrapper[4790]: I1124 14:32:57.080052 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Nov 24 14:32:57 crc kubenswrapper[4790]: I1124 14:32:57.132011 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhtz7\" (UniqueName: \"kubernetes.io/projected/e5d5fa4a-219e-47bd-b7a6-837c6a2c7170-kube-api-access-nhtz7\") pod \"mariadb-client-2-default\" (UID: \"e5d5fa4a-219e-47bd-b7a6-837c6a2c7170\") " pod="openstack/mariadb-client-2-default" Nov 24 14:32:57 crc kubenswrapper[4790]: I1124 14:32:57.146690 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhtz7\" (UniqueName: \"kubernetes.io/projected/e5d5fa4a-219e-47bd-b7a6-837c6a2c7170-kube-api-access-nhtz7\") pod \"mariadb-client-2-default\" (UID: \"e5d5fa4a-219e-47bd-b7a6-837c6a2c7170\") " pod="openstack/mariadb-client-2-default" Nov 24 14:32:57 crc kubenswrapper[4790]: I1124 14:32:57.310052 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Nov 24 14:32:57 crc kubenswrapper[4790]: I1124 14:32:57.911170 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Nov 24 14:32:57 crc kubenswrapper[4790]: W1124 14:32:57.918460 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5d5fa4a_219e_47bd_b7a6_837c6a2c7170.slice/crio-52abf9590f6d6b9e420362bd04f7e4374477b1a089282bfa1df2997828e7f978 WatchSource:0}: Error finding container 52abf9590f6d6b9e420362bd04f7e4374477b1a089282bfa1df2997828e7f978: Status 404 returned error can't find the container with id 52abf9590f6d6b9e420362bd04f7e4374477b1a089282bfa1df2997828e7f978 Nov 24 14:32:58 crc kubenswrapper[4790]: I1124 14:32:58.092763 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"e5d5fa4a-219e-47bd-b7a6-837c6a2c7170","Type":"ContainerStarted","Data":"52abf9590f6d6b9e420362bd04f7e4374477b1a089282bfa1df2997828e7f978"} Nov 24 14:32:58 crc kubenswrapper[4790]: I1124 14:32:58.336677 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0584256d-35c9-4cf1-953b-78d1fc050fc6" path="/var/lib/kubelet/pods/0584256d-35c9-4cf1-953b-78d1fc050fc6/volumes" Nov 24 14:32:59 crc kubenswrapper[4790]: I1124 14:32:59.102391 4790 generic.go:334] "Generic (PLEG): container finished" podID="e5d5fa4a-219e-47bd-b7a6-837c6a2c7170" containerID="37e005f44c60754f8dcfc8dda04f4ecba44c5a6391db55078ba6521473c59745" exitCode=1 Nov 24 14:32:59 crc kubenswrapper[4790]: I1124 14:32:59.102438 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"e5d5fa4a-219e-47bd-b7a6-837c6a2c7170","Type":"ContainerDied","Data":"37e005f44c60754f8dcfc8dda04f4ecba44c5a6391db55078ba6521473c59745"} Nov 24 14:33:00 crc kubenswrapper[4790]: I1124 14:33:00.491493 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Nov 24 14:33:00 crc kubenswrapper[4790]: I1124 14:33:00.533436 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2-default_e5d5fa4a-219e-47bd-b7a6-837c6a2c7170/mariadb-client-2-default/0.log" Nov 24 14:33:00 crc kubenswrapper[4790]: I1124 14:33:00.562634 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2-default"] Nov 24 14:33:00 crc kubenswrapper[4790]: I1124 14:33:00.567484 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2-default"] Nov 24 14:33:00 crc kubenswrapper[4790]: I1124 14:33:00.584953 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhtz7\" (UniqueName: \"kubernetes.io/projected/e5d5fa4a-219e-47bd-b7a6-837c6a2c7170-kube-api-access-nhtz7\") pod \"e5d5fa4a-219e-47bd-b7a6-837c6a2c7170\" (UID: \"e5d5fa4a-219e-47bd-b7a6-837c6a2c7170\") " Nov 24 14:33:00 crc kubenswrapper[4790]: I1124 14:33:00.591834 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5d5fa4a-219e-47bd-b7a6-837c6a2c7170-kube-api-access-nhtz7" (OuterVolumeSpecName: "kube-api-access-nhtz7") pod "e5d5fa4a-219e-47bd-b7a6-837c6a2c7170" (UID: "e5d5fa4a-219e-47bd-b7a6-837c6a2c7170"). InnerVolumeSpecName "kube-api-access-nhtz7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:33:00 crc kubenswrapper[4790]: I1124 14:33:00.687460 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhtz7\" (UniqueName: \"kubernetes.io/projected/e5d5fa4a-219e-47bd-b7a6-837c6a2c7170-kube-api-access-nhtz7\") on node \"crc\" DevicePath \"\"" Nov 24 14:33:00 crc kubenswrapper[4790]: I1124 14:33:00.943653 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1"] Nov 24 14:33:00 crc kubenswrapper[4790]: E1124 14:33:00.943998 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5d5fa4a-219e-47bd-b7a6-837c6a2c7170" containerName="mariadb-client-2-default" Nov 24 14:33:00 crc kubenswrapper[4790]: I1124 14:33:00.944019 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5d5fa4a-219e-47bd-b7a6-837c6a2c7170" containerName="mariadb-client-2-default" Nov 24 14:33:00 crc kubenswrapper[4790]: I1124 14:33:00.944191 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5d5fa4a-219e-47bd-b7a6-837c6a2c7170" containerName="mariadb-client-2-default" Nov 24 14:33:00 crc kubenswrapper[4790]: I1124 14:33:00.944773 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Nov 24 14:33:00 crc kubenswrapper[4790]: I1124 14:33:00.951230 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Nov 24 14:33:00 crc kubenswrapper[4790]: I1124 14:33:00.991916 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lff7n\" (UniqueName: \"kubernetes.io/projected/461dc8e2-1798-4dce-95a6-f11adbe25c6a-kube-api-access-lff7n\") pod \"mariadb-client-1\" (UID: \"461dc8e2-1798-4dce-95a6-f11adbe25c6a\") " pod="openstack/mariadb-client-1" Nov 24 14:33:01 crc kubenswrapper[4790]: I1124 14:33:01.093089 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lff7n\" (UniqueName: \"kubernetes.io/projected/461dc8e2-1798-4dce-95a6-f11adbe25c6a-kube-api-access-lff7n\") pod \"mariadb-client-1\" (UID: \"461dc8e2-1798-4dce-95a6-f11adbe25c6a\") " pod="openstack/mariadb-client-1" Nov 24 14:33:01 crc kubenswrapper[4790]: I1124 14:33:01.109056 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lff7n\" (UniqueName: \"kubernetes.io/projected/461dc8e2-1798-4dce-95a6-f11adbe25c6a-kube-api-access-lff7n\") pod \"mariadb-client-1\" (UID: \"461dc8e2-1798-4dce-95a6-f11adbe25c6a\") " pod="openstack/mariadb-client-1" Nov 24 14:33:01 crc kubenswrapper[4790]: I1124 14:33:01.120793 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52abf9590f6d6b9e420362bd04f7e4374477b1a089282bfa1df2997828e7f978" Nov 24 14:33:01 crc kubenswrapper[4790]: I1124 14:33:01.120838 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Nov 24 14:33:01 crc kubenswrapper[4790]: I1124 14:33:01.259844 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Nov 24 14:33:01 crc kubenswrapper[4790]: I1124 14:33:01.726767 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Nov 24 14:33:02 crc kubenswrapper[4790]: I1124 14:33:02.128593 4790 generic.go:334] "Generic (PLEG): container finished" podID="461dc8e2-1798-4dce-95a6-f11adbe25c6a" containerID="37e69b339f97ae0520e33f147dfb96088843ee20fb88a78f589302602baa82d1" exitCode=0 Nov 24 14:33:02 crc kubenswrapper[4790]: I1124 14:33:02.128645 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"461dc8e2-1798-4dce-95a6-f11adbe25c6a","Type":"ContainerDied","Data":"37e69b339f97ae0520e33f147dfb96088843ee20fb88a78f589302602baa82d1"} Nov 24 14:33:02 crc kubenswrapper[4790]: I1124 14:33:02.128924 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"461dc8e2-1798-4dce-95a6-f11adbe25c6a","Type":"ContainerStarted","Data":"45c45601060f99f0538b3c2fb92a1aa96be8e50774a731c75b256f3b28e437d0"} Nov 24 14:33:02 crc kubenswrapper[4790]: I1124 14:33:02.330672 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5d5fa4a-219e-47bd-b7a6-837c6a2c7170" path="/var/lib/kubelet/pods/e5d5fa4a-219e-47bd-b7a6-837c6a2c7170/volumes" Nov 24 14:33:04 crc kubenswrapper[4790]: I1124 14:33:03.462454 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Nov 24 14:33:04 crc kubenswrapper[4790]: I1124 14:33:03.481296 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1_461dc8e2-1798-4dce-95a6-f11adbe25c6a/mariadb-client-1/0.log" Nov 24 14:33:04 crc kubenswrapper[4790]: I1124 14:33:03.506076 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1"] Nov 24 14:33:04 crc kubenswrapper[4790]: I1124 14:33:03.512519 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1"] Nov 24 14:33:04 crc kubenswrapper[4790]: I1124 14:33:03.533365 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lff7n\" (UniqueName: \"kubernetes.io/projected/461dc8e2-1798-4dce-95a6-f11adbe25c6a-kube-api-access-lff7n\") pod \"461dc8e2-1798-4dce-95a6-f11adbe25c6a\" (UID: \"461dc8e2-1798-4dce-95a6-f11adbe25c6a\") " Nov 24 14:33:04 crc kubenswrapper[4790]: I1124 14:33:03.539146 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/461dc8e2-1798-4dce-95a6-f11adbe25c6a-kube-api-access-lff7n" (OuterVolumeSpecName: "kube-api-access-lff7n") pod "461dc8e2-1798-4dce-95a6-f11adbe25c6a" (UID: "461dc8e2-1798-4dce-95a6-f11adbe25c6a"). InnerVolumeSpecName "kube-api-access-lff7n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:33:04 crc kubenswrapper[4790]: I1124 14:33:03.635467 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lff7n\" (UniqueName: \"kubernetes.io/projected/461dc8e2-1798-4dce-95a6-f11adbe25c6a-kube-api-access-lff7n\") on node \"crc\" DevicePath \"\"" Nov 24 14:33:04 crc kubenswrapper[4790]: I1124 14:33:03.891056 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-4-default"] Nov 24 14:33:04 crc kubenswrapper[4790]: E1124 14:33:03.891601 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="461dc8e2-1798-4dce-95a6-f11adbe25c6a" containerName="mariadb-client-1" Nov 24 14:33:04 crc kubenswrapper[4790]: I1124 14:33:03.891621 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="461dc8e2-1798-4dce-95a6-f11adbe25c6a" containerName="mariadb-client-1" Nov 24 14:33:04 crc kubenswrapper[4790]: I1124 14:33:03.891862 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="461dc8e2-1798-4dce-95a6-f11adbe25c6a" containerName="mariadb-client-1" Nov 24 14:33:04 crc kubenswrapper[4790]: I1124 14:33:03.892680 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Nov 24 14:33:04 crc kubenswrapper[4790]: I1124 14:33:03.900242 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Nov 24 14:33:04 crc kubenswrapper[4790]: I1124 14:33:03.940519 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlfqt\" (UniqueName: \"kubernetes.io/projected/e8f0a5de-e6bf-49e7-8a3c-f67ba9c09ada-kube-api-access-jlfqt\") pod \"mariadb-client-4-default\" (UID: \"e8f0a5de-e6bf-49e7-8a3c-f67ba9c09ada\") " pod="openstack/mariadb-client-4-default" Nov 24 14:33:04 crc kubenswrapper[4790]: I1124 14:33:04.041929 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlfqt\" (UniqueName: \"kubernetes.io/projected/e8f0a5de-e6bf-49e7-8a3c-f67ba9c09ada-kube-api-access-jlfqt\") pod \"mariadb-client-4-default\" (UID: \"e8f0a5de-e6bf-49e7-8a3c-f67ba9c09ada\") " pod="openstack/mariadb-client-4-default" Nov 24 14:33:04 crc kubenswrapper[4790]: I1124 14:33:04.059910 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlfqt\" (UniqueName: \"kubernetes.io/projected/e8f0a5de-e6bf-49e7-8a3c-f67ba9c09ada-kube-api-access-jlfqt\") pod \"mariadb-client-4-default\" (UID: \"e8f0a5de-e6bf-49e7-8a3c-f67ba9c09ada\") " pod="openstack/mariadb-client-4-default" Nov 24 14:33:04 crc kubenswrapper[4790]: I1124 14:33:04.143695 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45c45601060f99f0538b3c2fb92a1aa96be8e50774a731c75b256f3b28e437d0" Nov 24 14:33:04 crc kubenswrapper[4790]: I1124 14:33:04.143770 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Nov 24 14:33:04 crc kubenswrapper[4790]: I1124 14:33:04.212831 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Nov 24 14:33:04 crc kubenswrapper[4790]: I1124 14:33:04.323968 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="461dc8e2-1798-4dce-95a6-f11adbe25c6a" path="/var/lib/kubelet/pods/461dc8e2-1798-4dce-95a6-f11adbe25c6a/volumes" Nov 24 14:33:04 crc kubenswrapper[4790]: I1124 14:33:04.737935 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Nov 24 14:33:04 crc kubenswrapper[4790]: W1124 14:33:04.747090 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8f0a5de_e6bf_49e7_8a3c_f67ba9c09ada.slice/crio-c351cb292b0c4ceb1b63fb2f46ffc1837ce585b89944c0fd45e113210eddf3aa WatchSource:0}: Error finding container c351cb292b0c4ceb1b63fb2f46ffc1837ce585b89944c0fd45e113210eddf3aa: Status 404 returned error can't find the container with id c351cb292b0c4ceb1b63fb2f46ffc1837ce585b89944c0fd45e113210eddf3aa Nov 24 14:33:05 crc kubenswrapper[4790]: I1124 14:33:05.156474 4790 generic.go:334] "Generic (PLEG): container finished" podID="e8f0a5de-e6bf-49e7-8a3c-f67ba9c09ada" containerID="798103c1fdecc46c8858e184adbd64a086eafc76c05f970f4526cde26b4c9719" exitCode=0 Nov 24 14:33:05 crc kubenswrapper[4790]: I1124 14:33:05.156529 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"e8f0a5de-e6bf-49e7-8a3c-f67ba9c09ada","Type":"ContainerDied","Data":"798103c1fdecc46c8858e184adbd64a086eafc76c05f970f4526cde26b4c9719"} Nov 24 14:33:05 crc kubenswrapper[4790]: I1124 14:33:05.156568 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"e8f0a5de-e6bf-49e7-8a3c-f67ba9c09ada","Type":"ContainerStarted","Data":"c351cb292b0c4ceb1b63fb2f46ffc1837ce585b89944c0fd45e113210eddf3aa"} Nov 24 14:33:06 crc kubenswrapper[4790]: I1124 14:33:06.510990 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Nov 24 14:33:06 crc kubenswrapper[4790]: I1124 14:33:06.532189 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-4-default_e8f0a5de-e6bf-49e7-8a3c-f67ba9c09ada/mariadb-client-4-default/0.log" Nov 24 14:33:06 crc kubenswrapper[4790]: I1124 14:33:06.555758 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-4-default"] Nov 24 14:33:06 crc kubenswrapper[4790]: I1124 14:33:06.560634 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-4-default"] Nov 24 14:33:06 crc kubenswrapper[4790]: I1124 14:33:06.576607 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlfqt\" (UniqueName: \"kubernetes.io/projected/e8f0a5de-e6bf-49e7-8a3c-f67ba9c09ada-kube-api-access-jlfqt\") pod \"e8f0a5de-e6bf-49e7-8a3c-f67ba9c09ada\" (UID: \"e8f0a5de-e6bf-49e7-8a3c-f67ba9c09ada\") " Nov 24 14:33:06 crc kubenswrapper[4790]: I1124 14:33:06.589181 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8f0a5de-e6bf-49e7-8a3c-f67ba9c09ada-kube-api-access-jlfqt" (OuterVolumeSpecName: "kube-api-access-jlfqt") pod "e8f0a5de-e6bf-49e7-8a3c-f67ba9c09ada" (UID: "e8f0a5de-e6bf-49e7-8a3c-f67ba9c09ada"). InnerVolumeSpecName "kube-api-access-jlfqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:33:06 crc kubenswrapper[4790]: I1124 14:33:06.678624 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlfqt\" (UniqueName: \"kubernetes.io/projected/e8f0a5de-e6bf-49e7-8a3c-f67ba9c09ada-kube-api-access-jlfqt\") on node \"crc\" DevicePath \"\"" Nov 24 14:33:07 crc kubenswrapper[4790]: I1124 14:33:07.175503 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c351cb292b0c4ceb1b63fb2f46ffc1837ce585b89944c0fd45e113210eddf3aa" Nov 24 14:33:07 crc kubenswrapper[4790]: I1124 14:33:07.175568 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Nov 24 14:33:08 crc kubenswrapper[4790]: I1124 14:33:08.314130 4790 scope.go:117] "RemoveContainer" containerID="571bd0c27e6ad4756f7116ed8b5a40d06301f3d85436cea369594fdb322154a7" Nov 24 14:33:08 crc kubenswrapper[4790]: E1124 14:33:08.314647 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:33:08 crc kubenswrapper[4790]: I1124 14:33:08.330224 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8f0a5de-e6bf-49e7-8a3c-f67ba9c09ada" path="/var/lib/kubelet/pods/e8f0a5de-e6bf-49e7-8a3c-f67ba9c09ada/volumes" Nov 24 14:33:10 crc kubenswrapper[4790]: I1124 14:33:10.998452 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-5-default"] Nov 24 14:33:10 crc kubenswrapper[4790]: E1124 14:33:10.999408 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8f0a5de-e6bf-49e7-8a3c-f67ba9c09ada" containerName="mariadb-client-4-default" Nov 24 14:33:10 crc kubenswrapper[4790]: I1124 14:33:10.999433 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8f0a5de-e6bf-49e7-8a3c-f67ba9c09ada" containerName="mariadb-client-4-default" Nov 24 14:33:10 crc kubenswrapper[4790]: I1124 14:33:10.999723 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8f0a5de-e6bf-49e7-8a3c-f67ba9c09ada" containerName="mariadb-client-4-default" Nov 24 14:33:11 crc kubenswrapper[4790]: I1124 14:33:11.000561 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Nov 24 14:33:11 crc kubenswrapper[4790]: I1124 14:33:11.003571 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-fx2g4" Nov 24 14:33:11 crc kubenswrapper[4790]: I1124 14:33:11.009138 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Nov 24 14:33:11 crc kubenswrapper[4790]: I1124 14:33:11.049401 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdsd8\" (UniqueName: \"kubernetes.io/projected/a73df7fd-f387-4e45-acd2-8077d73a48aa-kube-api-access-zdsd8\") pod \"mariadb-client-5-default\" (UID: \"a73df7fd-f387-4e45-acd2-8077d73a48aa\") " pod="openstack/mariadb-client-5-default" Nov 24 14:33:11 crc kubenswrapper[4790]: I1124 14:33:11.151071 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdsd8\" (UniqueName: \"kubernetes.io/projected/a73df7fd-f387-4e45-acd2-8077d73a48aa-kube-api-access-zdsd8\") pod \"mariadb-client-5-default\" (UID: \"a73df7fd-f387-4e45-acd2-8077d73a48aa\") " pod="openstack/mariadb-client-5-default" Nov 24 14:33:11 crc kubenswrapper[4790]: I1124 14:33:11.179231 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdsd8\" (UniqueName: \"kubernetes.io/projected/a73df7fd-f387-4e45-acd2-8077d73a48aa-kube-api-access-zdsd8\") pod \"mariadb-client-5-default\" (UID: \"a73df7fd-f387-4e45-acd2-8077d73a48aa\") " pod="openstack/mariadb-client-5-default" Nov 24 14:33:11 crc kubenswrapper[4790]: I1124 14:33:11.326014 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Nov 24 14:33:11 crc kubenswrapper[4790]: I1124 14:33:11.935661 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Nov 24 14:33:12 crc kubenswrapper[4790]: I1124 14:33:12.230827 4790 generic.go:334] "Generic (PLEG): container finished" podID="a73df7fd-f387-4e45-acd2-8077d73a48aa" containerID="01b7eddcd71f1f402684449f51f27b0a55c1785a6727bc6d135cff897cba5157" exitCode=0 Nov 24 14:33:12 crc kubenswrapper[4790]: I1124 14:33:12.231001 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"a73df7fd-f387-4e45-acd2-8077d73a48aa","Type":"ContainerDied","Data":"01b7eddcd71f1f402684449f51f27b0a55c1785a6727bc6d135cff897cba5157"} Nov 24 14:33:12 crc kubenswrapper[4790]: I1124 14:33:12.231073 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"a73df7fd-f387-4e45-acd2-8077d73a48aa","Type":"ContainerStarted","Data":"da6fd9c408fbfb1e9953b8b2c10ecbe494693463662658e808967694d527397f"} Nov 24 14:33:13 crc kubenswrapper[4790]: I1124 14:33:13.543996 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-98wr6"] Nov 24 14:33:13 crc kubenswrapper[4790]: I1124 14:33:13.547016 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-98wr6" Nov 24 14:33:13 crc kubenswrapper[4790]: I1124 14:33:13.564766 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-98wr6"] Nov 24 14:33:13 crc kubenswrapper[4790]: I1124 14:33:13.600442 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Nov 24 14:33:13 crc kubenswrapper[4790]: I1124 14:33:13.626138 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-5-default_a73df7fd-f387-4e45-acd2-8077d73a48aa/mariadb-client-5-default/0.log" Nov 24 14:33:13 crc kubenswrapper[4790]: I1124 14:33:13.656954 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-5-default"] Nov 24 14:33:13 crc kubenswrapper[4790]: I1124 14:33:13.663943 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-5-default"] Nov 24 14:33:13 crc kubenswrapper[4790]: I1124 14:33:13.697289 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/274df355-8bfc-4461-af5a-f08b068c6805-catalog-content\") pod \"redhat-marketplace-98wr6\" (UID: \"274df355-8bfc-4461-af5a-f08b068c6805\") " pod="openshift-marketplace/redhat-marketplace-98wr6" Nov 24 14:33:13 crc kubenswrapper[4790]: I1124 14:33:13.697360 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6z4gj\" (UniqueName: \"kubernetes.io/projected/274df355-8bfc-4461-af5a-f08b068c6805-kube-api-access-6z4gj\") pod \"redhat-marketplace-98wr6\" (UID: \"274df355-8bfc-4461-af5a-f08b068c6805\") " pod="openshift-marketplace/redhat-marketplace-98wr6" Nov 24 14:33:13 crc kubenswrapper[4790]: I1124 14:33:13.697434 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/274df355-8bfc-4461-af5a-f08b068c6805-utilities\") pod \"redhat-marketplace-98wr6\" (UID: \"274df355-8bfc-4461-af5a-f08b068c6805\") " pod="openshift-marketplace/redhat-marketplace-98wr6" Nov 24 14:33:13 crc kubenswrapper[4790]: I1124 14:33:13.772922 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-6-default"] Nov 24 14:33:13 crc kubenswrapper[4790]: E1124 14:33:13.773570 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a73df7fd-f387-4e45-acd2-8077d73a48aa" containerName="mariadb-client-5-default" Nov 24 14:33:13 crc kubenswrapper[4790]: I1124 14:33:13.773586 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="a73df7fd-f387-4e45-acd2-8077d73a48aa" containerName="mariadb-client-5-default" Nov 24 14:33:13 crc kubenswrapper[4790]: I1124 14:33:13.773721 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="a73df7fd-f387-4e45-acd2-8077d73a48aa" containerName="mariadb-client-5-default" Nov 24 14:33:13 crc kubenswrapper[4790]: I1124 14:33:13.774312 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Nov 24 14:33:13 crc kubenswrapper[4790]: I1124 14:33:13.784791 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Nov 24 14:33:13 crc kubenswrapper[4790]: I1124 14:33:13.799074 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdsd8\" (UniqueName: \"kubernetes.io/projected/a73df7fd-f387-4e45-acd2-8077d73a48aa-kube-api-access-zdsd8\") pod \"a73df7fd-f387-4e45-acd2-8077d73a48aa\" (UID: \"a73df7fd-f387-4e45-acd2-8077d73a48aa\") " Nov 24 14:33:13 crc kubenswrapper[4790]: I1124 14:33:13.799328 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/274df355-8bfc-4461-af5a-f08b068c6805-catalog-content\") pod \"redhat-marketplace-98wr6\" (UID: \"274df355-8bfc-4461-af5a-f08b068c6805\") " pod="openshift-marketplace/redhat-marketplace-98wr6" Nov 24 14:33:13 crc kubenswrapper[4790]: I1124 14:33:13.799374 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6z4gj\" (UniqueName: \"kubernetes.io/projected/274df355-8bfc-4461-af5a-f08b068c6805-kube-api-access-6z4gj\") pod \"redhat-marketplace-98wr6\" (UID: \"274df355-8bfc-4461-af5a-f08b068c6805\") " pod="openshift-marketplace/redhat-marketplace-98wr6" Nov 24 14:33:13 crc kubenswrapper[4790]: I1124 14:33:13.799407 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/274df355-8bfc-4461-af5a-f08b068c6805-utilities\") pod \"redhat-marketplace-98wr6\" (UID: \"274df355-8bfc-4461-af5a-f08b068c6805\") " pod="openshift-marketplace/redhat-marketplace-98wr6" Nov 24 14:33:13 crc kubenswrapper[4790]: I1124 14:33:13.799449 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vmpn\" (UniqueName: \"kubernetes.io/projected/f878eedc-758a-472a-8803-4d444bf7bf0d-kube-api-access-9vmpn\") pod \"mariadb-client-6-default\" (UID: \"f878eedc-758a-472a-8803-4d444bf7bf0d\") " pod="openstack/mariadb-client-6-default" Nov 24 14:33:13 crc kubenswrapper[4790]: I1124 14:33:13.800005 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/274df355-8bfc-4461-af5a-f08b068c6805-catalog-content\") pod \"redhat-marketplace-98wr6\" (UID: \"274df355-8bfc-4461-af5a-f08b068c6805\") " pod="openshift-marketplace/redhat-marketplace-98wr6" Nov 24 14:33:13 crc kubenswrapper[4790]: I1124 14:33:13.800634 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/274df355-8bfc-4461-af5a-f08b068c6805-utilities\") pod \"redhat-marketplace-98wr6\" (UID: \"274df355-8bfc-4461-af5a-f08b068c6805\") " pod="openshift-marketplace/redhat-marketplace-98wr6" Nov 24 14:33:13 crc kubenswrapper[4790]: I1124 14:33:13.810061 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a73df7fd-f387-4e45-acd2-8077d73a48aa-kube-api-access-zdsd8" (OuterVolumeSpecName: "kube-api-access-zdsd8") pod "a73df7fd-f387-4e45-acd2-8077d73a48aa" (UID: "a73df7fd-f387-4e45-acd2-8077d73a48aa"). InnerVolumeSpecName "kube-api-access-zdsd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:33:13 crc kubenswrapper[4790]: I1124 14:33:13.822283 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6z4gj\" (UniqueName: \"kubernetes.io/projected/274df355-8bfc-4461-af5a-f08b068c6805-kube-api-access-6z4gj\") pod \"redhat-marketplace-98wr6\" (UID: \"274df355-8bfc-4461-af5a-f08b068c6805\") " pod="openshift-marketplace/redhat-marketplace-98wr6" Nov 24 14:33:13 crc kubenswrapper[4790]: I1124 14:33:13.900652 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vmpn\" (UniqueName: \"kubernetes.io/projected/f878eedc-758a-472a-8803-4d444bf7bf0d-kube-api-access-9vmpn\") pod \"mariadb-client-6-default\" (UID: \"f878eedc-758a-472a-8803-4d444bf7bf0d\") " pod="openstack/mariadb-client-6-default" Nov 24 14:33:13 crc kubenswrapper[4790]: I1124 14:33:13.900744 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdsd8\" (UniqueName: \"kubernetes.io/projected/a73df7fd-f387-4e45-acd2-8077d73a48aa-kube-api-access-zdsd8\") on node \"crc\" DevicePath \"\"" Nov 24 14:33:13 crc kubenswrapper[4790]: I1124 14:33:13.918068 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vmpn\" (UniqueName: \"kubernetes.io/projected/f878eedc-758a-472a-8803-4d444bf7bf0d-kube-api-access-9vmpn\") pod \"mariadb-client-6-default\" (UID: \"f878eedc-758a-472a-8803-4d444bf7bf0d\") " pod="openstack/mariadb-client-6-default" Nov 24 14:33:13 crc kubenswrapper[4790]: I1124 14:33:13.918481 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-98wr6" Nov 24 14:33:14 crc kubenswrapper[4790]: I1124 14:33:14.089052 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Nov 24 14:33:14 crc kubenswrapper[4790]: I1124 14:33:14.247892 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da6fd9c408fbfb1e9953b8b2c10ecbe494693463662658e808967694d527397f" Nov 24 14:33:14 crc kubenswrapper[4790]: I1124 14:33:14.247941 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Nov 24 14:33:14 crc kubenswrapper[4790]: I1124 14:33:14.332717 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a73df7fd-f387-4e45-acd2-8077d73a48aa" path="/var/lib/kubelet/pods/a73df7fd-f387-4e45-acd2-8077d73a48aa/volumes" Nov 24 14:33:14 crc kubenswrapper[4790]: I1124 14:33:14.358301 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-98wr6"] Nov 24 14:33:14 crc kubenswrapper[4790]: W1124 14:33:14.366763 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod274df355_8bfc_4461_af5a_f08b068c6805.slice/crio-77a869d46eff589ced41be0e6a0b2029fb580c761311dcd25fa59d55b8e767de WatchSource:0}: Error finding container 77a869d46eff589ced41be0e6a0b2029fb580c761311dcd25fa59d55b8e767de: Status 404 returned error can't find the container with id 77a869d46eff589ced41be0e6a0b2029fb580c761311dcd25fa59d55b8e767de Nov 24 14:33:14 crc kubenswrapper[4790]: I1124 14:33:14.573382 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Nov 24 14:33:14 crc kubenswrapper[4790]: W1124 14:33:14.605225 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf878eedc_758a_472a_8803_4d444bf7bf0d.slice/crio-bbe99b3cf0f95181ac5f0f020b78b500b664d3836eb6d488de4c13ee89eedd5b WatchSource:0}: Error finding container bbe99b3cf0f95181ac5f0f020b78b500b664d3836eb6d488de4c13ee89eedd5b: Status 404 returned error can't find the container with id bbe99b3cf0f95181ac5f0f020b78b500b664d3836eb6d488de4c13ee89eedd5b Nov 24 14:33:15 crc kubenswrapper[4790]: I1124 14:33:15.260468 4790 generic.go:334] "Generic (PLEG): container finished" podID="274df355-8bfc-4461-af5a-f08b068c6805" containerID="45d031faa1429083def26e782c4999da0e6268650b3a941d7b52fd14bef2f675" exitCode=0 Nov 24 14:33:15 crc kubenswrapper[4790]: I1124 14:33:15.260595 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-98wr6" event={"ID":"274df355-8bfc-4461-af5a-f08b068c6805","Type":"ContainerDied","Data":"45d031faa1429083def26e782c4999da0e6268650b3a941d7b52fd14bef2f675"} Nov 24 14:33:15 crc kubenswrapper[4790]: I1124 14:33:15.260976 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-98wr6" event={"ID":"274df355-8bfc-4461-af5a-f08b068c6805","Type":"ContainerStarted","Data":"77a869d46eff589ced41be0e6a0b2029fb580c761311dcd25fa59d55b8e767de"} Nov 24 14:33:15 crc kubenswrapper[4790]: I1124 14:33:15.262382 4790 generic.go:334] "Generic (PLEG): container finished" podID="f878eedc-758a-472a-8803-4d444bf7bf0d" containerID="c091dbbfabd40597bf770f01f653a630e1c315e029f65d1cf4747dde8a4a16a8" exitCode=1 Nov 24 14:33:15 crc kubenswrapper[4790]: I1124 14:33:15.262420 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"f878eedc-758a-472a-8803-4d444bf7bf0d","Type":"ContainerDied","Data":"c091dbbfabd40597bf770f01f653a630e1c315e029f65d1cf4747dde8a4a16a8"} Nov 24 14:33:15 crc kubenswrapper[4790]: I1124 14:33:15.262444 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"f878eedc-758a-472a-8803-4d444bf7bf0d","Type":"ContainerStarted","Data":"bbe99b3cf0f95181ac5f0f020b78b500b664d3836eb6d488de4c13ee89eedd5b"} Nov 24 14:33:16 crc kubenswrapper[4790]: I1124 14:33:16.692871 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Nov 24 14:33:16 crc kubenswrapper[4790]: I1124 14:33:16.712677 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-6-default_f878eedc-758a-472a-8803-4d444bf7bf0d/mariadb-client-6-default/0.log" Nov 24 14:33:16 crc kubenswrapper[4790]: I1124 14:33:16.738584 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-6-default"] Nov 24 14:33:16 crc kubenswrapper[4790]: I1124 14:33:16.743403 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-6-default"] Nov 24 14:33:16 crc kubenswrapper[4790]: I1124 14:33:16.848541 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vmpn\" (UniqueName: \"kubernetes.io/projected/f878eedc-758a-472a-8803-4d444bf7bf0d-kube-api-access-9vmpn\") pod \"f878eedc-758a-472a-8803-4d444bf7bf0d\" (UID: \"f878eedc-758a-472a-8803-4d444bf7bf0d\") " Nov 24 14:33:16 crc kubenswrapper[4790]: I1124 14:33:16.855252 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f878eedc-758a-472a-8803-4d444bf7bf0d-kube-api-access-9vmpn" (OuterVolumeSpecName: "kube-api-access-9vmpn") pod "f878eedc-758a-472a-8803-4d444bf7bf0d" (UID: "f878eedc-758a-472a-8803-4d444bf7bf0d"). InnerVolumeSpecName "kube-api-access-9vmpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:33:16 crc kubenswrapper[4790]: I1124 14:33:16.873196 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-7-default"] Nov 24 14:33:16 crc kubenswrapper[4790]: E1124 14:33:16.873614 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f878eedc-758a-472a-8803-4d444bf7bf0d" containerName="mariadb-client-6-default" Nov 24 14:33:16 crc kubenswrapper[4790]: I1124 14:33:16.873631 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="f878eedc-758a-472a-8803-4d444bf7bf0d" containerName="mariadb-client-6-default" Nov 24 14:33:16 crc kubenswrapper[4790]: I1124 14:33:16.873841 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="f878eedc-758a-472a-8803-4d444bf7bf0d" containerName="mariadb-client-6-default" Nov 24 14:33:16 crc kubenswrapper[4790]: I1124 14:33:16.874618 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Nov 24 14:33:16 crc kubenswrapper[4790]: I1124 14:33:16.877165 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Nov 24 14:33:16 crc kubenswrapper[4790]: I1124 14:33:16.950504 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vmpn\" (UniqueName: \"kubernetes.io/projected/f878eedc-758a-472a-8803-4d444bf7bf0d-kube-api-access-9vmpn\") on node \"crc\" DevicePath \"\"" Nov 24 14:33:17 crc kubenswrapper[4790]: I1124 14:33:17.052253 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2v42\" (UniqueName: \"kubernetes.io/projected/c24cdf84-e914-4faa-9fa7-e2149714164e-kube-api-access-w2v42\") pod \"mariadb-client-7-default\" (UID: \"c24cdf84-e914-4faa-9fa7-e2149714164e\") " pod="openstack/mariadb-client-7-default" Nov 24 14:33:17 crc kubenswrapper[4790]: I1124 14:33:17.153712 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2v42\" (UniqueName: \"kubernetes.io/projected/c24cdf84-e914-4faa-9fa7-e2149714164e-kube-api-access-w2v42\") pod \"mariadb-client-7-default\" (UID: \"c24cdf84-e914-4faa-9fa7-e2149714164e\") " pod="openstack/mariadb-client-7-default" Nov 24 14:33:17 crc kubenswrapper[4790]: I1124 14:33:17.175509 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2v42\" (UniqueName: \"kubernetes.io/projected/c24cdf84-e914-4faa-9fa7-e2149714164e-kube-api-access-w2v42\") pod \"mariadb-client-7-default\" (UID: \"c24cdf84-e914-4faa-9fa7-e2149714164e\") " pod="openstack/mariadb-client-7-default" Nov 24 14:33:17 crc kubenswrapper[4790]: I1124 14:33:17.223002 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Nov 24 14:33:17 crc kubenswrapper[4790]: I1124 14:33:17.293858 4790 generic.go:334] "Generic (PLEG): container finished" podID="274df355-8bfc-4461-af5a-f08b068c6805" containerID="574b000117c2ac47a7d6976a948166e34f01074e0445b646354eba12393da051" exitCode=0 Nov 24 14:33:17 crc kubenswrapper[4790]: I1124 14:33:17.293936 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-98wr6" event={"ID":"274df355-8bfc-4461-af5a-f08b068c6805","Type":"ContainerDied","Data":"574b000117c2ac47a7d6976a948166e34f01074e0445b646354eba12393da051"} Nov 24 14:33:17 crc kubenswrapper[4790]: I1124 14:33:17.302732 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bbe99b3cf0f95181ac5f0f020b78b500b664d3836eb6d488de4c13ee89eedd5b" Nov 24 14:33:17 crc kubenswrapper[4790]: I1124 14:33:17.302809 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Nov 24 14:33:17 crc kubenswrapper[4790]: I1124 14:33:17.549104 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Nov 24 14:33:17 crc kubenswrapper[4790]: W1124 14:33:17.557999 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc24cdf84_e914_4faa_9fa7_e2149714164e.slice/crio-3b0c8785e909aea0a9b65e355cecc8a2e017e546a41485e311a233c50e3d8bf1 WatchSource:0}: Error finding container 3b0c8785e909aea0a9b65e355cecc8a2e017e546a41485e311a233c50e3d8bf1: Status 404 returned error can't find the container with id 3b0c8785e909aea0a9b65e355cecc8a2e017e546a41485e311a233c50e3d8bf1 Nov 24 14:33:18 crc kubenswrapper[4790]: I1124 14:33:18.323287 4790 generic.go:334] "Generic (PLEG): container finished" podID="c24cdf84-e914-4faa-9fa7-e2149714164e" containerID="76927d6aa40419a38e64fbb9c14a26948319fe250a4986431acff969b9bd20d0" exitCode=0 Nov 24 14:33:18 crc kubenswrapper[4790]: I1124 14:33:18.327088 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f878eedc-758a-472a-8803-4d444bf7bf0d" path="/var/lib/kubelet/pods/f878eedc-758a-472a-8803-4d444bf7bf0d/volumes" Nov 24 14:33:18 crc kubenswrapper[4790]: I1124 14:33:18.327563 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-98wr6" event={"ID":"274df355-8bfc-4461-af5a-f08b068c6805","Type":"ContainerStarted","Data":"ae2205dce5742e5a7ca1ccbde0ad16bb3ead6e4fb0516b123c0b383d65f0f202"} Nov 24 14:33:18 crc kubenswrapper[4790]: I1124 14:33:18.327638 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"c24cdf84-e914-4faa-9fa7-e2149714164e","Type":"ContainerDied","Data":"76927d6aa40419a38e64fbb9c14a26948319fe250a4986431acff969b9bd20d0"} Nov 24 14:33:18 crc kubenswrapper[4790]: I1124 14:33:18.327652 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"c24cdf84-e914-4faa-9fa7-e2149714164e","Type":"ContainerStarted","Data":"3b0c8785e909aea0a9b65e355cecc8a2e017e546a41485e311a233c50e3d8bf1"} Nov 24 14:33:18 crc kubenswrapper[4790]: I1124 14:33:18.352514 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-98wr6" podStartSLOduration=2.667195497 podStartE2EDuration="5.352495706s" podCreationTimestamp="2025-11-24 14:33:13 +0000 UTC" firstStartedPulling="2025-11-24 14:33:15.263069508 +0000 UTC m=+4843.642963190" lastFinishedPulling="2025-11-24 14:33:17.948369687 +0000 UTC m=+4846.328263399" observedRunningTime="2025-11-24 14:33:18.348302456 +0000 UTC m=+4846.728196138" watchObservedRunningTime="2025-11-24 14:33:18.352495706 +0000 UTC m=+4846.732389368" Nov 24 14:33:19 crc kubenswrapper[4790]: I1124 14:33:19.739326 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Nov 24 14:33:19 crc kubenswrapper[4790]: I1124 14:33:19.757113 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-7-default_c24cdf84-e914-4faa-9fa7-e2149714164e/mariadb-client-7-default/0.log" Nov 24 14:33:19 crc kubenswrapper[4790]: I1124 14:33:19.780786 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-7-default"] Nov 24 14:33:19 crc kubenswrapper[4790]: I1124 14:33:19.786334 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-7-default"] Nov 24 14:33:19 crc kubenswrapper[4790]: I1124 14:33:19.796533 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2v42\" (UniqueName: \"kubernetes.io/projected/c24cdf84-e914-4faa-9fa7-e2149714164e-kube-api-access-w2v42\") pod \"c24cdf84-e914-4faa-9fa7-e2149714164e\" (UID: \"c24cdf84-e914-4faa-9fa7-e2149714164e\") " Nov 24 14:33:19 crc kubenswrapper[4790]: I1124 14:33:19.804794 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c24cdf84-e914-4faa-9fa7-e2149714164e-kube-api-access-w2v42" (OuterVolumeSpecName: "kube-api-access-w2v42") pod "c24cdf84-e914-4faa-9fa7-e2149714164e" (UID: "c24cdf84-e914-4faa-9fa7-e2149714164e"). InnerVolumeSpecName "kube-api-access-w2v42". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:33:19 crc kubenswrapper[4790]: I1124 14:33:19.898173 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2v42\" (UniqueName: \"kubernetes.io/projected/c24cdf84-e914-4faa-9fa7-e2149714164e-kube-api-access-w2v42\") on node \"crc\" DevicePath \"\"" Nov 24 14:33:19 crc kubenswrapper[4790]: I1124 14:33:19.905188 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2"] Nov 24 14:33:19 crc kubenswrapper[4790]: E1124 14:33:19.907221 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c24cdf84-e914-4faa-9fa7-e2149714164e" containerName="mariadb-client-7-default" Nov 24 14:33:19 crc kubenswrapper[4790]: I1124 14:33:19.907268 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="c24cdf84-e914-4faa-9fa7-e2149714164e" containerName="mariadb-client-7-default" Nov 24 14:33:19 crc kubenswrapper[4790]: I1124 14:33:19.908941 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="c24cdf84-e914-4faa-9fa7-e2149714164e" containerName="mariadb-client-7-default" Nov 24 14:33:19 crc kubenswrapper[4790]: I1124 14:33:19.910205 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Nov 24 14:33:19 crc kubenswrapper[4790]: I1124 14:33:19.931338 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Nov 24 14:33:20 crc kubenswrapper[4790]: I1124 14:33:20.000551 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qq5t8\" (UniqueName: \"kubernetes.io/projected/ba21e196-fd26-409e-93d9-83b01694b9e7-kube-api-access-qq5t8\") pod \"mariadb-client-2\" (UID: \"ba21e196-fd26-409e-93d9-83b01694b9e7\") " pod="openstack/mariadb-client-2" Nov 24 14:33:20 crc kubenswrapper[4790]: I1124 14:33:20.101996 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qq5t8\" (UniqueName: \"kubernetes.io/projected/ba21e196-fd26-409e-93d9-83b01694b9e7-kube-api-access-qq5t8\") pod \"mariadb-client-2\" (UID: \"ba21e196-fd26-409e-93d9-83b01694b9e7\") " pod="openstack/mariadb-client-2" Nov 24 14:33:20 crc kubenswrapper[4790]: I1124 14:33:20.129996 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qq5t8\" (UniqueName: \"kubernetes.io/projected/ba21e196-fd26-409e-93d9-83b01694b9e7-kube-api-access-qq5t8\") pod \"mariadb-client-2\" (UID: \"ba21e196-fd26-409e-93d9-83b01694b9e7\") " pod="openstack/mariadb-client-2" Nov 24 14:33:20 crc kubenswrapper[4790]: I1124 14:33:20.241136 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Nov 24 14:33:20 crc kubenswrapper[4790]: I1124 14:33:20.325631 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c24cdf84-e914-4faa-9fa7-e2149714164e" path="/var/lib/kubelet/pods/c24cdf84-e914-4faa-9fa7-e2149714164e/volumes" Nov 24 14:33:20 crc kubenswrapper[4790]: I1124 14:33:20.347735 4790 scope.go:117] "RemoveContainer" containerID="76927d6aa40419a38e64fbb9c14a26948319fe250a4986431acff969b9bd20d0" Nov 24 14:33:20 crc kubenswrapper[4790]: I1124 14:33:20.347784 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Nov 24 14:33:20 crc kubenswrapper[4790]: I1124 14:33:20.563011 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Nov 24 14:33:20 crc kubenswrapper[4790]: W1124 14:33:20.568318 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podba21e196_fd26_409e_93d9_83b01694b9e7.slice/crio-c47acae291949f386da1c897d6a433dec39510499bfea6a23b6437eaf9e246cb WatchSource:0}: Error finding container c47acae291949f386da1c897d6a433dec39510499bfea6a23b6437eaf9e246cb: Status 404 returned error can't find the container with id c47acae291949f386da1c897d6a433dec39510499bfea6a23b6437eaf9e246cb Nov 24 14:33:21 crc kubenswrapper[4790]: I1124 14:33:21.315230 4790 scope.go:117] "RemoveContainer" containerID="571bd0c27e6ad4756f7116ed8b5a40d06301f3d85436cea369594fdb322154a7" Nov 24 14:33:21 crc kubenswrapper[4790]: E1124 14:33:21.316129 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:33:21 crc kubenswrapper[4790]: I1124 14:33:21.362603 4790 generic.go:334] "Generic (PLEG): container finished" podID="ba21e196-fd26-409e-93d9-83b01694b9e7" containerID="aa3082ba93f1d59782e9be84026e161036aff7db70b05cb7ff37f9566a5819c3" exitCode=0 Nov 24 14:33:21 crc kubenswrapper[4790]: I1124 14:33:21.362671 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"ba21e196-fd26-409e-93d9-83b01694b9e7","Type":"ContainerDied","Data":"aa3082ba93f1d59782e9be84026e161036aff7db70b05cb7ff37f9566a5819c3"} Nov 24 14:33:21 crc kubenswrapper[4790]: I1124 14:33:21.362707 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"ba21e196-fd26-409e-93d9-83b01694b9e7","Type":"ContainerStarted","Data":"c47acae291949f386da1c897d6a433dec39510499bfea6a23b6437eaf9e246cb"} Nov 24 14:33:22 crc kubenswrapper[4790]: I1124 14:33:22.841344 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Nov 24 14:33:22 crc kubenswrapper[4790]: I1124 14:33:22.865798 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2_ba21e196-fd26-409e-93d9-83b01694b9e7/mariadb-client-2/0.log" Nov 24 14:33:22 crc kubenswrapper[4790]: I1124 14:33:22.893340 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2"] Nov 24 14:33:22 crc kubenswrapper[4790]: I1124 14:33:22.898846 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2"] Nov 24 14:33:22 crc kubenswrapper[4790]: I1124 14:33:22.947779 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qq5t8\" (UniqueName: \"kubernetes.io/projected/ba21e196-fd26-409e-93d9-83b01694b9e7-kube-api-access-qq5t8\") pod \"ba21e196-fd26-409e-93d9-83b01694b9e7\" (UID: \"ba21e196-fd26-409e-93d9-83b01694b9e7\") " Nov 24 14:33:22 crc kubenswrapper[4790]: I1124 14:33:22.955384 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba21e196-fd26-409e-93d9-83b01694b9e7-kube-api-access-qq5t8" (OuterVolumeSpecName: "kube-api-access-qq5t8") pod "ba21e196-fd26-409e-93d9-83b01694b9e7" (UID: "ba21e196-fd26-409e-93d9-83b01694b9e7"). InnerVolumeSpecName "kube-api-access-qq5t8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:33:23 crc kubenswrapper[4790]: I1124 14:33:23.051566 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qq5t8\" (UniqueName: \"kubernetes.io/projected/ba21e196-fd26-409e-93d9-83b01694b9e7-kube-api-access-qq5t8\") on node \"crc\" DevicePath \"\"" Nov 24 14:33:23 crc kubenswrapper[4790]: I1124 14:33:23.382072 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c47acae291949f386da1c897d6a433dec39510499bfea6a23b6437eaf9e246cb" Nov 24 14:33:23 crc kubenswrapper[4790]: I1124 14:33:23.382172 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Nov 24 14:33:23 crc kubenswrapper[4790]: I1124 14:33:23.919248 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-98wr6" Nov 24 14:33:23 crc kubenswrapper[4790]: I1124 14:33:23.919326 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-98wr6" Nov 24 14:33:23 crc kubenswrapper[4790]: I1124 14:33:23.992513 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-98wr6" Nov 24 14:33:24 crc kubenswrapper[4790]: I1124 14:33:24.331695 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba21e196-fd26-409e-93d9-83b01694b9e7" path="/var/lib/kubelet/pods/ba21e196-fd26-409e-93d9-83b01694b9e7/volumes" Nov 24 14:33:24 crc kubenswrapper[4790]: I1124 14:33:24.449532 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-98wr6" Nov 24 14:33:24 crc kubenswrapper[4790]: I1124 14:33:24.505732 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-98wr6"] Nov 24 14:33:26 crc kubenswrapper[4790]: I1124 14:33:26.407795 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-98wr6" podUID="274df355-8bfc-4461-af5a-f08b068c6805" containerName="registry-server" containerID="cri-o://ae2205dce5742e5a7ca1ccbde0ad16bb3ead6e4fb0516b123c0b383d65f0f202" gracePeriod=2 Nov 24 14:33:26 crc kubenswrapper[4790]: I1124 14:33:26.849046 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-98wr6" Nov 24 14:33:26 crc kubenswrapper[4790]: I1124 14:33:26.919873 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6z4gj\" (UniqueName: \"kubernetes.io/projected/274df355-8bfc-4461-af5a-f08b068c6805-kube-api-access-6z4gj\") pod \"274df355-8bfc-4461-af5a-f08b068c6805\" (UID: \"274df355-8bfc-4461-af5a-f08b068c6805\") " Nov 24 14:33:26 crc kubenswrapper[4790]: I1124 14:33:26.920344 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/274df355-8bfc-4461-af5a-f08b068c6805-utilities\") pod \"274df355-8bfc-4461-af5a-f08b068c6805\" (UID: \"274df355-8bfc-4461-af5a-f08b068c6805\") " Nov 24 14:33:26 crc kubenswrapper[4790]: I1124 14:33:26.920376 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/274df355-8bfc-4461-af5a-f08b068c6805-catalog-content\") pod \"274df355-8bfc-4461-af5a-f08b068c6805\" (UID: \"274df355-8bfc-4461-af5a-f08b068c6805\") " Nov 24 14:33:26 crc kubenswrapper[4790]: I1124 14:33:26.921946 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/274df355-8bfc-4461-af5a-f08b068c6805-utilities" (OuterVolumeSpecName: "utilities") pod "274df355-8bfc-4461-af5a-f08b068c6805" (UID: "274df355-8bfc-4461-af5a-f08b068c6805"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:33:26 crc kubenswrapper[4790]: I1124 14:33:26.924923 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/274df355-8bfc-4461-af5a-f08b068c6805-kube-api-access-6z4gj" (OuterVolumeSpecName: "kube-api-access-6z4gj") pod "274df355-8bfc-4461-af5a-f08b068c6805" (UID: "274df355-8bfc-4461-af5a-f08b068c6805"). InnerVolumeSpecName "kube-api-access-6z4gj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:33:26 crc kubenswrapper[4790]: I1124 14:33:26.951713 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/274df355-8bfc-4461-af5a-f08b068c6805-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "274df355-8bfc-4461-af5a-f08b068c6805" (UID: "274df355-8bfc-4461-af5a-f08b068c6805"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:33:27 crc kubenswrapper[4790]: I1124 14:33:27.022594 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6z4gj\" (UniqueName: \"kubernetes.io/projected/274df355-8bfc-4461-af5a-f08b068c6805-kube-api-access-6z4gj\") on node \"crc\" DevicePath \"\"" Nov 24 14:33:27 crc kubenswrapper[4790]: I1124 14:33:27.022641 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/274df355-8bfc-4461-af5a-f08b068c6805-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:33:27 crc kubenswrapper[4790]: I1124 14:33:27.022658 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/274df355-8bfc-4461-af5a-f08b068c6805-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:33:27 crc kubenswrapper[4790]: I1124 14:33:27.421064 4790 generic.go:334] "Generic (PLEG): container finished" podID="274df355-8bfc-4461-af5a-f08b068c6805" containerID="ae2205dce5742e5a7ca1ccbde0ad16bb3ead6e4fb0516b123c0b383d65f0f202" exitCode=0 Nov 24 14:33:27 crc kubenswrapper[4790]: I1124 14:33:27.421131 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-98wr6" event={"ID":"274df355-8bfc-4461-af5a-f08b068c6805","Type":"ContainerDied","Data":"ae2205dce5742e5a7ca1ccbde0ad16bb3ead6e4fb0516b123c0b383d65f0f202"} Nov 24 14:33:27 crc kubenswrapper[4790]: I1124 14:33:27.421146 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-98wr6" Nov 24 14:33:27 crc kubenswrapper[4790]: I1124 14:33:27.421179 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-98wr6" event={"ID":"274df355-8bfc-4461-af5a-f08b068c6805","Type":"ContainerDied","Data":"77a869d46eff589ced41be0e6a0b2029fb580c761311dcd25fa59d55b8e767de"} Nov 24 14:33:27 crc kubenswrapper[4790]: I1124 14:33:27.421225 4790 scope.go:117] "RemoveContainer" containerID="ae2205dce5742e5a7ca1ccbde0ad16bb3ead6e4fb0516b123c0b383d65f0f202" Nov 24 14:33:27 crc kubenswrapper[4790]: I1124 14:33:27.442299 4790 scope.go:117] "RemoveContainer" containerID="574b000117c2ac47a7d6976a948166e34f01074e0445b646354eba12393da051" Nov 24 14:33:27 crc kubenswrapper[4790]: I1124 14:33:27.457838 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-98wr6"] Nov 24 14:33:27 crc kubenswrapper[4790]: I1124 14:33:27.465787 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-98wr6"] Nov 24 14:33:27 crc kubenswrapper[4790]: I1124 14:33:27.488169 4790 scope.go:117] "RemoveContainer" containerID="45d031faa1429083def26e782c4999da0e6268650b3a941d7b52fd14bef2f675" Nov 24 14:33:27 crc kubenswrapper[4790]: I1124 14:33:27.516854 4790 scope.go:117] "RemoveContainer" containerID="ae2205dce5742e5a7ca1ccbde0ad16bb3ead6e4fb0516b123c0b383d65f0f202" Nov 24 14:33:27 crc kubenswrapper[4790]: E1124 14:33:27.517642 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae2205dce5742e5a7ca1ccbde0ad16bb3ead6e4fb0516b123c0b383d65f0f202\": container with ID starting with ae2205dce5742e5a7ca1ccbde0ad16bb3ead6e4fb0516b123c0b383d65f0f202 not found: ID does not exist" containerID="ae2205dce5742e5a7ca1ccbde0ad16bb3ead6e4fb0516b123c0b383d65f0f202" Nov 24 14:33:27 crc kubenswrapper[4790]: I1124 14:33:27.517671 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae2205dce5742e5a7ca1ccbde0ad16bb3ead6e4fb0516b123c0b383d65f0f202"} err="failed to get container status \"ae2205dce5742e5a7ca1ccbde0ad16bb3ead6e4fb0516b123c0b383d65f0f202\": rpc error: code = NotFound desc = could not find container \"ae2205dce5742e5a7ca1ccbde0ad16bb3ead6e4fb0516b123c0b383d65f0f202\": container with ID starting with ae2205dce5742e5a7ca1ccbde0ad16bb3ead6e4fb0516b123c0b383d65f0f202 not found: ID does not exist" Nov 24 14:33:27 crc kubenswrapper[4790]: I1124 14:33:27.517691 4790 scope.go:117] "RemoveContainer" containerID="574b000117c2ac47a7d6976a948166e34f01074e0445b646354eba12393da051" Nov 24 14:33:27 crc kubenswrapper[4790]: E1124 14:33:27.519351 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"574b000117c2ac47a7d6976a948166e34f01074e0445b646354eba12393da051\": container with ID starting with 574b000117c2ac47a7d6976a948166e34f01074e0445b646354eba12393da051 not found: ID does not exist" containerID="574b000117c2ac47a7d6976a948166e34f01074e0445b646354eba12393da051" Nov 24 14:33:27 crc kubenswrapper[4790]: I1124 14:33:27.519371 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"574b000117c2ac47a7d6976a948166e34f01074e0445b646354eba12393da051"} err="failed to get container status \"574b000117c2ac47a7d6976a948166e34f01074e0445b646354eba12393da051\": rpc error: code = NotFound desc = could not find container \"574b000117c2ac47a7d6976a948166e34f01074e0445b646354eba12393da051\": container with ID starting with 574b000117c2ac47a7d6976a948166e34f01074e0445b646354eba12393da051 not found: ID does not exist" Nov 24 14:33:27 crc kubenswrapper[4790]: I1124 14:33:27.519384 4790 scope.go:117] "RemoveContainer" containerID="45d031faa1429083def26e782c4999da0e6268650b3a941d7b52fd14bef2f675" Nov 24 14:33:27 crc kubenswrapper[4790]: E1124 14:33:27.520365 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45d031faa1429083def26e782c4999da0e6268650b3a941d7b52fd14bef2f675\": container with ID starting with 45d031faa1429083def26e782c4999da0e6268650b3a941d7b52fd14bef2f675 not found: ID does not exist" containerID="45d031faa1429083def26e782c4999da0e6268650b3a941d7b52fd14bef2f675" Nov 24 14:33:27 crc kubenswrapper[4790]: I1124 14:33:27.520450 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45d031faa1429083def26e782c4999da0e6268650b3a941d7b52fd14bef2f675"} err="failed to get container status \"45d031faa1429083def26e782c4999da0e6268650b3a941d7b52fd14bef2f675\": rpc error: code = NotFound desc = could not find container \"45d031faa1429083def26e782c4999da0e6268650b3a941d7b52fd14bef2f675\": container with ID starting with 45d031faa1429083def26e782c4999da0e6268650b3a941d7b52fd14bef2f675 not found: ID does not exist" Nov 24 14:33:28 crc kubenswrapper[4790]: I1124 14:33:28.325327 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="274df355-8bfc-4461-af5a-f08b068c6805" path="/var/lib/kubelet/pods/274df355-8bfc-4461-af5a-f08b068c6805/volumes" Nov 24 14:33:29 crc kubenswrapper[4790]: I1124 14:33:29.660393 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mxvl9"] Nov 24 14:33:29 crc kubenswrapper[4790]: E1124 14:33:29.661093 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="274df355-8bfc-4461-af5a-f08b068c6805" containerName="extract-utilities" Nov 24 14:33:29 crc kubenswrapper[4790]: I1124 14:33:29.661114 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="274df355-8bfc-4461-af5a-f08b068c6805" containerName="extract-utilities" Nov 24 14:33:29 crc kubenswrapper[4790]: E1124 14:33:29.661135 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba21e196-fd26-409e-93d9-83b01694b9e7" containerName="mariadb-client-2" Nov 24 14:33:29 crc kubenswrapper[4790]: I1124 14:33:29.661143 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba21e196-fd26-409e-93d9-83b01694b9e7" containerName="mariadb-client-2" Nov 24 14:33:29 crc kubenswrapper[4790]: E1124 14:33:29.661166 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="274df355-8bfc-4461-af5a-f08b068c6805" containerName="extract-content" Nov 24 14:33:29 crc kubenswrapper[4790]: I1124 14:33:29.661174 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="274df355-8bfc-4461-af5a-f08b068c6805" containerName="extract-content" Nov 24 14:33:29 crc kubenswrapper[4790]: E1124 14:33:29.661204 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="274df355-8bfc-4461-af5a-f08b068c6805" containerName="registry-server" Nov 24 14:33:29 crc kubenswrapper[4790]: I1124 14:33:29.661211 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="274df355-8bfc-4461-af5a-f08b068c6805" containerName="registry-server" Nov 24 14:33:29 crc kubenswrapper[4790]: I1124 14:33:29.661397 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba21e196-fd26-409e-93d9-83b01694b9e7" containerName="mariadb-client-2" Nov 24 14:33:29 crc kubenswrapper[4790]: I1124 14:33:29.661415 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="274df355-8bfc-4461-af5a-f08b068c6805" containerName="registry-server" Nov 24 14:33:29 crc kubenswrapper[4790]: I1124 14:33:29.662766 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mxvl9" Nov 24 14:33:29 crc kubenswrapper[4790]: I1124 14:33:29.680495 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mxvl9"] Nov 24 14:33:29 crc kubenswrapper[4790]: I1124 14:33:29.767229 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34330f45-b4e4-44a1-ae2c-58982b2f0ae7-catalog-content\") pod \"redhat-operators-mxvl9\" (UID: \"34330f45-b4e4-44a1-ae2c-58982b2f0ae7\") " pod="openshift-marketplace/redhat-operators-mxvl9" Nov 24 14:33:29 crc kubenswrapper[4790]: I1124 14:33:29.767355 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4q8rp\" (UniqueName: \"kubernetes.io/projected/34330f45-b4e4-44a1-ae2c-58982b2f0ae7-kube-api-access-4q8rp\") pod \"redhat-operators-mxvl9\" (UID: \"34330f45-b4e4-44a1-ae2c-58982b2f0ae7\") " pod="openshift-marketplace/redhat-operators-mxvl9" Nov 24 14:33:29 crc kubenswrapper[4790]: I1124 14:33:29.767394 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34330f45-b4e4-44a1-ae2c-58982b2f0ae7-utilities\") pod \"redhat-operators-mxvl9\" (UID: \"34330f45-b4e4-44a1-ae2c-58982b2f0ae7\") " pod="openshift-marketplace/redhat-operators-mxvl9" Nov 24 14:33:29 crc kubenswrapper[4790]: I1124 14:33:29.868816 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34330f45-b4e4-44a1-ae2c-58982b2f0ae7-catalog-content\") pod \"redhat-operators-mxvl9\" (UID: \"34330f45-b4e4-44a1-ae2c-58982b2f0ae7\") " pod="openshift-marketplace/redhat-operators-mxvl9" Nov 24 14:33:29 crc kubenswrapper[4790]: I1124 14:33:29.868921 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4q8rp\" (UniqueName: \"kubernetes.io/projected/34330f45-b4e4-44a1-ae2c-58982b2f0ae7-kube-api-access-4q8rp\") pod \"redhat-operators-mxvl9\" (UID: \"34330f45-b4e4-44a1-ae2c-58982b2f0ae7\") " pod="openshift-marketplace/redhat-operators-mxvl9" Nov 24 14:33:29 crc kubenswrapper[4790]: I1124 14:33:29.868951 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34330f45-b4e4-44a1-ae2c-58982b2f0ae7-utilities\") pod \"redhat-operators-mxvl9\" (UID: \"34330f45-b4e4-44a1-ae2c-58982b2f0ae7\") " pod="openshift-marketplace/redhat-operators-mxvl9" Nov 24 14:33:29 crc kubenswrapper[4790]: I1124 14:33:29.869393 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34330f45-b4e4-44a1-ae2c-58982b2f0ae7-utilities\") pod \"redhat-operators-mxvl9\" (UID: \"34330f45-b4e4-44a1-ae2c-58982b2f0ae7\") " pod="openshift-marketplace/redhat-operators-mxvl9" Nov 24 14:33:29 crc kubenswrapper[4790]: I1124 14:33:29.869589 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34330f45-b4e4-44a1-ae2c-58982b2f0ae7-catalog-content\") pod \"redhat-operators-mxvl9\" (UID: \"34330f45-b4e4-44a1-ae2c-58982b2f0ae7\") " pod="openshift-marketplace/redhat-operators-mxvl9" Nov 24 14:33:29 crc kubenswrapper[4790]: I1124 14:33:29.890161 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4q8rp\" (UniqueName: \"kubernetes.io/projected/34330f45-b4e4-44a1-ae2c-58982b2f0ae7-kube-api-access-4q8rp\") pod \"redhat-operators-mxvl9\" (UID: \"34330f45-b4e4-44a1-ae2c-58982b2f0ae7\") " pod="openshift-marketplace/redhat-operators-mxvl9" Nov 24 14:33:29 crc kubenswrapper[4790]: I1124 14:33:29.994561 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mxvl9" Nov 24 14:33:30 crc kubenswrapper[4790]: I1124 14:33:30.414615 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mxvl9"] Nov 24 14:33:30 crc kubenswrapper[4790]: I1124 14:33:30.452114 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mxvl9" event={"ID":"34330f45-b4e4-44a1-ae2c-58982b2f0ae7","Type":"ContainerStarted","Data":"c95e60192c34d56c963567f3c49b8495811dba56ddad8e76d48f90e5447c0a9c"} Nov 24 14:33:31 crc kubenswrapper[4790]: I1124 14:33:31.461481 4790 generic.go:334] "Generic (PLEG): container finished" podID="34330f45-b4e4-44a1-ae2c-58982b2f0ae7" containerID="ffbf2bd0b5c199ec2c9fb87e3563dfd179ba42984badcf22ad62ba0bc352c077" exitCode=0 Nov 24 14:33:31 crc kubenswrapper[4790]: I1124 14:33:31.461553 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mxvl9" event={"ID":"34330f45-b4e4-44a1-ae2c-58982b2f0ae7","Type":"ContainerDied","Data":"ffbf2bd0b5c199ec2c9fb87e3563dfd179ba42984badcf22ad62ba0bc352c077"} Nov 24 14:33:32 crc kubenswrapper[4790]: I1124 14:33:32.318515 4790 scope.go:117] "RemoveContainer" containerID="571bd0c27e6ad4756f7116ed8b5a40d06301f3d85436cea369594fdb322154a7" Nov 24 14:33:32 crc kubenswrapper[4790]: E1124 14:33:32.319012 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:33:34 crc kubenswrapper[4790]: I1124 14:33:34.481859 4790 generic.go:334] "Generic (PLEG): container finished" podID="34330f45-b4e4-44a1-ae2c-58982b2f0ae7" containerID="222e6323eefc74b29d9da853302e512f1fcf8f3d3cda65563dbc350271f4da6e" exitCode=0 Nov 24 14:33:34 crc kubenswrapper[4790]: I1124 14:33:34.481958 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mxvl9" event={"ID":"34330f45-b4e4-44a1-ae2c-58982b2f0ae7","Type":"ContainerDied","Data":"222e6323eefc74b29d9da853302e512f1fcf8f3d3cda65563dbc350271f4da6e"} Nov 24 14:33:35 crc kubenswrapper[4790]: I1124 14:33:35.492703 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mxvl9" event={"ID":"34330f45-b4e4-44a1-ae2c-58982b2f0ae7","Type":"ContainerStarted","Data":"76fbebab97ec412a77bf261df414349214df5580a5c2f0058104ff1a38e0631e"} Nov 24 14:33:35 crc kubenswrapper[4790]: I1124 14:33:35.515321 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mxvl9" podStartSLOduration=3.0667945 podStartE2EDuration="6.515298814s" podCreationTimestamp="2025-11-24 14:33:29 +0000 UTC" firstStartedPulling="2025-11-24 14:33:31.462806666 +0000 UTC m=+4859.842700328" lastFinishedPulling="2025-11-24 14:33:34.91131098 +0000 UTC m=+4863.291204642" observedRunningTime="2025-11-24 14:33:35.510041844 +0000 UTC m=+4863.889935516" watchObservedRunningTime="2025-11-24 14:33:35.515298814 +0000 UTC m=+4863.895192486" Nov 24 14:33:39 crc kubenswrapper[4790]: I1124 14:33:39.995207 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mxvl9" Nov 24 14:33:39 crc kubenswrapper[4790]: I1124 14:33:39.996038 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mxvl9" Nov 24 14:33:41 crc kubenswrapper[4790]: I1124 14:33:41.056142 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mxvl9" podUID="34330f45-b4e4-44a1-ae2c-58982b2f0ae7" containerName="registry-server" probeResult="failure" output=< Nov 24 14:33:41 crc kubenswrapper[4790]: timeout: failed to connect service ":50051" within 1s Nov 24 14:33:41 crc kubenswrapper[4790]: > Nov 24 14:33:45 crc kubenswrapper[4790]: I1124 14:33:45.314611 4790 scope.go:117] "RemoveContainer" containerID="571bd0c27e6ad4756f7116ed8b5a40d06301f3d85436cea369594fdb322154a7" Nov 24 14:33:46 crc kubenswrapper[4790]: I1124 14:33:46.587834 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"1551c55c728005d2d8114bb5394315ba445d97c4e7ab35c743e2f23da5188013"} Nov 24 14:33:46 crc kubenswrapper[4790]: I1124 14:33:46.762355 4790 scope.go:117] "RemoveContainer" containerID="327852f78d1c40248ad2a4e37cb155535377c520bc90cb223080cf976bf095fa" Nov 24 14:33:50 crc kubenswrapper[4790]: I1124 14:33:50.046699 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mxvl9" Nov 24 14:33:50 crc kubenswrapper[4790]: I1124 14:33:50.096089 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mxvl9" Nov 24 14:33:50 crc kubenswrapper[4790]: I1124 14:33:50.285117 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mxvl9"] Nov 24 14:33:51 crc kubenswrapper[4790]: I1124 14:33:51.631634 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mxvl9" podUID="34330f45-b4e4-44a1-ae2c-58982b2f0ae7" containerName="registry-server" containerID="cri-o://76fbebab97ec412a77bf261df414349214df5580a5c2f0058104ff1a38e0631e" gracePeriod=2 Nov 24 14:33:52 crc kubenswrapper[4790]: I1124 14:33:52.071317 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mxvl9" Nov 24 14:33:52 crc kubenswrapper[4790]: I1124 14:33:52.115333 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34330f45-b4e4-44a1-ae2c-58982b2f0ae7-catalog-content\") pod \"34330f45-b4e4-44a1-ae2c-58982b2f0ae7\" (UID: \"34330f45-b4e4-44a1-ae2c-58982b2f0ae7\") " Nov 24 14:33:52 crc kubenswrapper[4790]: I1124 14:33:52.115395 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34330f45-b4e4-44a1-ae2c-58982b2f0ae7-utilities\") pod \"34330f45-b4e4-44a1-ae2c-58982b2f0ae7\" (UID: \"34330f45-b4e4-44a1-ae2c-58982b2f0ae7\") " Nov 24 14:33:52 crc kubenswrapper[4790]: I1124 14:33:52.115486 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4q8rp\" (UniqueName: \"kubernetes.io/projected/34330f45-b4e4-44a1-ae2c-58982b2f0ae7-kube-api-access-4q8rp\") pod \"34330f45-b4e4-44a1-ae2c-58982b2f0ae7\" (UID: \"34330f45-b4e4-44a1-ae2c-58982b2f0ae7\") " Nov 24 14:33:52 crc kubenswrapper[4790]: I1124 14:33:52.117612 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34330f45-b4e4-44a1-ae2c-58982b2f0ae7-utilities" (OuterVolumeSpecName: "utilities") pod "34330f45-b4e4-44a1-ae2c-58982b2f0ae7" (UID: "34330f45-b4e4-44a1-ae2c-58982b2f0ae7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:33:52 crc kubenswrapper[4790]: I1124 14:33:52.124498 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34330f45-b4e4-44a1-ae2c-58982b2f0ae7-kube-api-access-4q8rp" (OuterVolumeSpecName: "kube-api-access-4q8rp") pod "34330f45-b4e4-44a1-ae2c-58982b2f0ae7" (UID: "34330f45-b4e4-44a1-ae2c-58982b2f0ae7"). InnerVolumeSpecName "kube-api-access-4q8rp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:33:52 crc kubenswrapper[4790]: I1124 14:33:52.210602 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34330f45-b4e4-44a1-ae2c-58982b2f0ae7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "34330f45-b4e4-44a1-ae2c-58982b2f0ae7" (UID: "34330f45-b4e4-44a1-ae2c-58982b2f0ae7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:33:52 crc kubenswrapper[4790]: I1124 14:33:52.216502 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4q8rp\" (UniqueName: \"kubernetes.io/projected/34330f45-b4e4-44a1-ae2c-58982b2f0ae7-kube-api-access-4q8rp\") on node \"crc\" DevicePath \"\"" Nov 24 14:33:52 crc kubenswrapper[4790]: I1124 14:33:52.216553 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34330f45-b4e4-44a1-ae2c-58982b2f0ae7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:33:52 crc kubenswrapper[4790]: I1124 14:33:52.216562 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34330f45-b4e4-44a1-ae2c-58982b2f0ae7-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:33:52 crc kubenswrapper[4790]: I1124 14:33:52.647443 4790 generic.go:334] "Generic (PLEG): container finished" podID="34330f45-b4e4-44a1-ae2c-58982b2f0ae7" containerID="76fbebab97ec412a77bf261df414349214df5580a5c2f0058104ff1a38e0631e" exitCode=0 Nov 24 14:33:52 crc kubenswrapper[4790]: I1124 14:33:52.647513 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mxvl9" event={"ID":"34330f45-b4e4-44a1-ae2c-58982b2f0ae7","Type":"ContainerDied","Data":"76fbebab97ec412a77bf261df414349214df5580a5c2f0058104ff1a38e0631e"} Nov 24 14:33:52 crc kubenswrapper[4790]: I1124 14:33:52.647555 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mxvl9" event={"ID":"34330f45-b4e4-44a1-ae2c-58982b2f0ae7","Type":"ContainerDied","Data":"c95e60192c34d56c963567f3c49b8495811dba56ddad8e76d48f90e5447c0a9c"} Nov 24 14:33:52 crc kubenswrapper[4790]: I1124 14:33:52.647566 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mxvl9" Nov 24 14:33:52 crc kubenswrapper[4790]: I1124 14:33:52.647578 4790 scope.go:117] "RemoveContainer" containerID="76fbebab97ec412a77bf261df414349214df5580a5c2f0058104ff1a38e0631e" Nov 24 14:33:52 crc kubenswrapper[4790]: I1124 14:33:52.675223 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mxvl9"] Nov 24 14:33:52 crc kubenswrapper[4790]: I1124 14:33:52.680473 4790 scope.go:117] "RemoveContainer" containerID="222e6323eefc74b29d9da853302e512f1fcf8f3d3cda65563dbc350271f4da6e" Nov 24 14:33:52 crc kubenswrapper[4790]: I1124 14:33:52.683128 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mxvl9"] Nov 24 14:33:52 crc kubenswrapper[4790]: I1124 14:33:52.698605 4790 scope.go:117] "RemoveContainer" containerID="ffbf2bd0b5c199ec2c9fb87e3563dfd179ba42984badcf22ad62ba0bc352c077" Nov 24 14:33:52 crc kubenswrapper[4790]: I1124 14:33:52.748073 4790 scope.go:117] "RemoveContainer" containerID="76fbebab97ec412a77bf261df414349214df5580a5c2f0058104ff1a38e0631e" Nov 24 14:33:52 crc kubenswrapper[4790]: E1124 14:33:52.748608 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76fbebab97ec412a77bf261df414349214df5580a5c2f0058104ff1a38e0631e\": container with ID starting with 76fbebab97ec412a77bf261df414349214df5580a5c2f0058104ff1a38e0631e not found: ID does not exist" containerID="76fbebab97ec412a77bf261df414349214df5580a5c2f0058104ff1a38e0631e" Nov 24 14:33:52 crc kubenswrapper[4790]: I1124 14:33:52.748662 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76fbebab97ec412a77bf261df414349214df5580a5c2f0058104ff1a38e0631e"} err="failed to get container status \"76fbebab97ec412a77bf261df414349214df5580a5c2f0058104ff1a38e0631e\": rpc error: code = NotFound desc = could not find container \"76fbebab97ec412a77bf261df414349214df5580a5c2f0058104ff1a38e0631e\": container with ID starting with 76fbebab97ec412a77bf261df414349214df5580a5c2f0058104ff1a38e0631e not found: ID does not exist" Nov 24 14:33:52 crc kubenswrapper[4790]: I1124 14:33:52.748684 4790 scope.go:117] "RemoveContainer" containerID="222e6323eefc74b29d9da853302e512f1fcf8f3d3cda65563dbc350271f4da6e" Nov 24 14:33:52 crc kubenswrapper[4790]: E1124 14:33:52.749118 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"222e6323eefc74b29d9da853302e512f1fcf8f3d3cda65563dbc350271f4da6e\": container with ID starting with 222e6323eefc74b29d9da853302e512f1fcf8f3d3cda65563dbc350271f4da6e not found: ID does not exist" containerID="222e6323eefc74b29d9da853302e512f1fcf8f3d3cda65563dbc350271f4da6e" Nov 24 14:33:52 crc kubenswrapper[4790]: I1124 14:33:52.749138 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"222e6323eefc74b29d9da853302e512f1fcf8f3d3cda65563dbc350271f4da6e"} err="failed to get container status \"222e6323eefc74b29d9da853302e512f1fcf8f3d3cda65563dbc350271f4da6e\": rpc error: code = NotFound desc = could not find container \"222e6323eefc74b29d9da853302e512f1fcf8f3d3cda65563dbc350271f4da6e\": container with ID starting with 222e6323eefc74b29d9da853302e512f1fcf8f3d3cda65563dbc350271f4da6e not found: ID does not exist" Nov 24 14:33:52 crc kubenswrapper[4790]: I1124 14:33:52.749155 4790 scope.go:117] "RemoveContainer" containerID="ffbf2bd0b5c199ec2c9fb87e3563dfd179ba42984badcf22ad62ba0bc352c077" Nov 24 14:33:52 crc kubenswrapper[4790]: E1124 14:33:52.749419 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffbf2bd0b5c199ec2c9fb87e3563dfd179ba42984badcf22ad62ba0bc352c077\": container with ID starting with ffbf2bd0b5c199ec2c9fb87e3563dfd179ba42984badcf22ad62ba0bc352c077 not found: ID does not exist" containerID="ffbf2bd0b5c199ec2c9fb87e3563dfd179ba42984badcf22ad62ba0bc352c077" Nov 24 14:33:52 crc kubenswrapper[4790]: I1124 14:33:52.749450 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffbf2bd0b5c199ec2c9fb87e3563dfd179ba42984badcf22ad62ba0bc352c077"} err="failed to get container status \"ffbf2bd0b5c199ec2c9fb87e3563dfd179ba42984badcf22ad62ba0bc352c077\": rpc error: code = NotFound desc = could not find container \"ffbf2bd0b5c199ec2c9fb87e3563dfd179ba42984badcf22ad62ba0bc352c077\": container with ID starting with ffbf2bd0b5c199ec2c9fb87e3563dfd179ba42984badcf22ad62ba0bc352c077 not found: ID does not exist" Nov 24 14:33:54 crc kubenswrapper[4790]: I1124 14:33:54.326254 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34330f45-b4e4-44a1-ae2c-58982b2f0ae7" path="/var/lib/kubelet/pods/34330f45-b4e4-44a1-ae2c-58982b2f0ae7/volumes" Nov 24 14:36:07 crc kubenswrapper[4790]: I1124 14:36:07.282758 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zwjvp"] Nov 24 14:36:07 crc kubenswrapper[4790]: E1124 14:36:07.284791 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34330f45-b4e4-44a1-ae2c-58982b2f0ae7" containerName="registry-server" Nov 24 14:36:07 crc kubenswrapper[4790]: I1124 14:36:07.284829 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="34330f45-b4e4-44a1-ae2c-58982b2f0ae7" containerName="registry-server" Nov 24 14:36:07 crc kubenswrapper[4790]: E1124 14:36:07.284919 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34330f45-b4e4-44a1-ae2c-58982b2f0ae7" containerName="extract-content" Nov 24 14:36:07 crc kubenswrapper[4790]: I1124 14:36:07.284940 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="34330f45-b4e4-44a1-ae2c-58982b2f0ae7" containerName="extract-content" Nov 24 14:36:07 crc kubenswrapper[4790]: E1124 14:36:07.284957 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34330f45-b4e4-44a1-ae2c-58982b2f0ae7" containerName="extract-utilities" Nov 24 14:36:07 crc kubenswrapper[4790]: I1124 14:36:07.284976 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="34330f45-b4e4-44a1-ae2c-58982b2f0ae7" containerName="extract-utilities" Nov 24 14:36:07 crc kubenswrapper[4790]: I1124 14:36:07.285391 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="34330f45-b4e4-44a1-ae2c-58982b2f0ae7" containerName="registry-server" Nov 24 14:36:07 crc kubenswrapper[4790]: I1124 14:36:07.288172 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zwjvp" Nov 24 14:36:07 crc kubenswrapper[4790]: I1124 14:36:07.306416 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zwjvp"] Nov 24 14:36:07 crc kubenswrapper[4790]: I1124 14:36:07.435185 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13d244de-09ce-4867-a07f-e31f4fe90857-catalog-content\") pod \"community-operators-zwjvp\" (UID: \"13d244de-09ce-4867-a07f-e31f4fe90857\") " pod="openshift-marketplace/community-operators-zwjvp" Nov 24 14:36:07 crc kubenswrapper[4790]: I1124 14:36:07.435254 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13d244de-09ce-4867-a07f-e31f4fe90857-utilities\") pod \"community-operators-zwjvp\" (UID: \"13d244de-09ce-4867-a07f-e31f4fe90857\") " pod="openshift-marketplace/community-operators-zwjvp" Nov 24 14:36:07 crc kubenswrapper[4790]: I1124 14:36:07.435615 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xz65w\" (UniqueName: \"kubernetes.io/projected/13d244de-09ce-4867-a07f-e31f4fe90857-kube-api-access-xz65w\") pod \"community-operators-zwjvp\" (UID: \"13d244de-09ce-4867-a07f-e31f4fe90857\") " pod="openshift-marketplace/community-operators-zwjvp" Nov 24 14:36:07 crc kubenswrapper[4790]: I1124 14:36:07.537721 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xz65w\" (UniqueName: \"kubernetes.io/projected/13d244de-09ce-4867-a07f-e31f4fe90857-kube-api-access-xz65w\") pod \"community-operators-zwjvp\" (UID: \"13d244de-09ce-4867-a07f-e31f4fe90857\") " pod="openshift-marketplace/community-operators-zwjvp" Nov 24 14:36:07 crc kubenswrapper[4790]: I1124 14:36:07.538190 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13d244de-09ce-4867-a07f-e31f4fe90857-catalog-content\") pod \"community-operators-zwjvp\" (UID: \"13d244de-09ce-4867-a07f-e31f4fe90857\") " pod="openshift-marketplace/community-operators-zwjvp" Nov 24 14:36:07 crc kubenswrapper[4790]: I1124 14:36:07.538233 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13d244de-09ce-4867-a07f-e31f4fe90857-utilities\") pod \"community-operators-zwjvp\" (UID: \"13d244de-09ce-4867-a07f-e31f4fe90857\") " pod="openshift-marketplace/community-operators-zwjvp" Nov 24 14:36:07 crc kubenswrapper[4790]: I1124 14:36:07.538650 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13d244de-09ce-4867-a07f-e31f4fe90857-catalog-content\") pod \"community-operators-zwjvp\" (UID: \"13d244de-09ce-4867-a07f-e31f4fe90857\") " pod="openshift-marketplace/community-operators-zwjvp" Nov 24 14:36:07 crc kubenswrapper[4790]: I1124 14:36:07.538772 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13d244de-09ce-4867-a07f-e31f4fe90857-utilities\") pod \"community-operators-zwjvp\" (UID: \"13d244de-09ce-4867-a07f-e31f4fe90857\") " pod="openshift-marketplace/community-operators-zwjvp" Nov 24 14:36:07 crc kubenswrapper[4790]: I1124 14:36:07.564325 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xz65w\" (UniqueName: \"kubernetes.io/projected/13d244de-09ce-4867-a07f-e31f4fe90857-kube-api-access-xz65w\") pod \"community-operators-zwjvp\" (UID: \"13d244de-09ce-4867-a07f-e31f4fe90857\") " pod="openshift-marketplace/community-operators-zwjvp" Nov 24 14:36:07 crc kubenswrapper[4790]: I1124 14:36:07.640620 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zwjvp" Nov 24 14:36:08 crc kubenswrapper[4790]: I1124 14:36:08.195176 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zwjvp"] Nov 24 14:36:08 crc kubenswrapper[4790]: I1124 14:36:08.225433 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwjvp" event={"ID":"13d244de-09ce-4867-a07f-e31f4fe90857","Type":"ContainerStarted","Data":"6b8fccde56cddec0e78f30f68dd4b829f429de07cd05bc62845d77ec6e323211"} Nov 24 14:36:09 crc kubenswrapper[4790]: I1124 14:36:09.239341 4790 generic.go:334] "Generic (PLEG): container finished" podID="13d244de-09ce-4867-a07f-e31f4fe90857" containerID="ecedbac62b382e0980e7f74bbc3f6bb9fca5fcde3afaec3a3103dc956363daca" exitCode=0 Nov 24 14:36:09 crc kubenswrapper[4790]: I1124 14:36:09.239627 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwjvp" event={"ID":"13d244de-09ce-4867-a07f-e31f4fe90857","Type":"ContainerDied","Data":"ecedbac62b382e0980e7f74bbc3f6bb9fca5fcde3afaec3a3103dc956363daca"} Nov 24 14:36:10 crc kubenswrapper[4790]: I1124 14:36:10.254407 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwjvp" event={"ID":"13d244de-09ce-4867-a07f-e31f4fe90857","Type":"ContainerStarted","Data":"db9dbd28986c0f053ebcc2afd276f69bab5375b61a3cb50fecc892bbb3ef7853"} Nov 24 14:36:11 crc kubenswrapper[4790]: I1124 14:36:11.269849 4790 generic.go:334] "Generic (PLEG): container finished" podID="13d244de-09ce-4867-a07f-e31f4fe90857" containerID="db9dbd28986c0f053ebcc2afd276f69bab5375b61a3cb50fecc892bbb3ef7853" exitCode=0 Nov 24 14:36:11 crc kubenswrapper[4790]: I1124 14:36:11.269983 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwjvp" event={"ID":"13d244de-09ce-4867-a07f-e31f4fe90857","Type":"ContainerDied","Data":"db9dbd28986c0f053ebcc2afd276f69bab5375b61a3cb50fecc892bbb3ef7853"} Nov 24 14:36:12 crc kubenswrapper[4790]: I1124 14:36:12.285407 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwjvp" event={"ID":"13d244de-09ce-4867-a07f-e31f4fe90857","Type":"ContainerStarted","Data":"f1d5ff6cc173a61349d77db71e8800ce236d75c6dcab837aa3c10a58fc2f97f2"} Nov 24 14:36:12 crc kubenswrapper[4790]: I1124 14:36:12.337528 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zwjvp" podStartSLOduration=2.828747495 podStartE2EDuration="5.337484887s" podCreationTimestamp="2025-11-24 14:36:07 +0000 UTC" firstStartedPulling="2025-11-24 14:36:09.242218692 +0000 UTC m=+5017.622112384" lastFinishedPulling="2025-11-24 14:36:11.750956084 +0000 UTC m=+5020.130849776" observedRunningTime="2025-11-24 14:36:12.320997614 +0000 UTC m=+5020.700891346" watchObservedRunningTime="2025-11-24 14:36:12.337484887 +0000 UTC m=+5020.717378589" Nov 24 14:36:13 crc kubenswrapper[4790]: I1124 14:36:13.938527 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:36:13 crc kubenswrapper[4790]: I1124 14:36:13.939258 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:36:17 crc kubenswrapper[4790]: I1124 14:36:17.641515 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zwjvp" Nov 24 14:36:17 crc kubenswrapper[4790]: I1124 14:36:17.642756 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zwjvp" Nov 24 14:36:17 crc kubenswrapper[4790]: I1124 14:36:17.723215 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zwjvp" Nov 24 14:36:18 crc kubenswrapper[4790]: I1124 14:36:18.436273 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zwjvp" Nov 24 14:36:18 crc kubenswrapper[4790]: I1124 14:36:18.519147 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zwjvp"] Nov 24 14:36:20 crc kubenswrapper[4790]: I1124 14:36:20.377125 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zwjvp" podUID="13d244de-09ce-4867-a07f-e31f4fe90857" containerName="registry-server" containerID="cri-o://f1d5ff6cc173a61349d77db71e8800ce236d75c6dcab837aa3c10a58fc2f97f2" gracePeriod=2 Nov 24 14:36:20 crc kubenswrapper[4790]: I1124 14:36:20.896386 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zwjvp" Nov 24 14:36:20 crc kubenswrapper[4790]: I1124 14:36:20.926052 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13d244de-09ce-4867-a07f-e31f4fe90857-catalog-content\") pod \"13d244de-09ce-4867-a07f-e31f4fe90857\" (UID: \"13d244de-09ce-4867-a07f-e31f4fe90857\") " Nov 24 14:36:20 crc kubenswrapper[4790]: I1124 14:36:20.926206 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xz65w\" (UniqueName: \"kubernetes.io/projected/13d244de-09ce-4867-a07f-e31f4fe90857-kube-api-access-xz65w\") pod \"13d244de-09ce-4867-a07f-e31f4fe90857\" (UID: \"13d244de-09ce-4867-a07f-e31f4fe90857\") " Nov 24 14:36:20 crc kubenswrapper[4790]: I1124 14:36:20.926262 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13d244de-09ce-4867-a07f-e31f4fe90857-utilities\") pod \"13d244de-09ce-4867-a07f-e31f4fe90857\" (UID: \"13d244de-09ce-4867-a07f-e31f4fe90857\") " Nov 24 14:36:20 crc kubenswrapper[4790]: I1124 14:36:20.928381 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13d244de-09ce-4867-a07f-e31f4fe90857-utilities" (OuterVolumeSpecName: "utilities") pod "13d244de-09ce-4867-a07f-e31f4fe90857" (UID: "13d244de-09ce-4867-a07f-e31f4fe90857"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:36:20 crc kubenswrapper[4790]: I1124 14:36:20.946493 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13d244de-09ce-4867-a07f-e31f4fe90857-kube-api-access-xz65w" (OuterVolumeSpecName: "kube-api-access-xz65w") pod "13d244de-09ce-4867-a07f-e31f4fe90857" (UID: "13d244de-09ce-4867-a07f-e31f4fe90857"). InnerVolumeSpecName "kube-api-access-xz65w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:36:21 crc kubenswrapper[4790]: I1124 14:36:21.029613 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xz65w\" (UniqueName: \"kubernetes.io/projected/13d244de-09ce-4867-a07f-e31f4fe90857-kube-api-access-xz65w\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:21 crc kubenswrapper[4790]: I1124 14:36:21.029709 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13d244de-09ce-4867-a07f-e31f4fe90857-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:21 crc kubenswrapper[4790]: I1124 14:36:21.387096 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13d244de-09ce-4867-a07f-e31f4fe90857-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "13d244de-09ce-4867-a07f-e31f4fe90857" (UID: "13d244de-09ce-4867-a07f-e31f4fe90857"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:36:21 crc kubenswrapper[4790]: I1124 14:36:21.391159 4790 generic.go:334] "Generic (PLEG): container finished" podID="13d244de-09ce-4867-a07f-e31f4fe90857" containerID="f1d5ff6cc173a61349d77db71e8800ce236d75c6dcab837aa3c10a58fc2f97f2" exitCode=0 Nov 24 14:36:21 crc kubenswrapper[4790]: I1124 14:36:21.391238 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zwjvp" Nov 24 14:36:21 crc kubenswrapper[4790]: I1124 14:36:21.391242 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwjvp" event={"ID":"13d244de-09ce-4867-a07f-e31f4fe90857","Type":"ContainerDied","Data":"f1d5ff6cc173a61349d77db71e8800ce236d75c6dcab837aa3c10a58fc2f97f2"} Nov 24 14:36:21 crc kubenswrapper[4790]: I1124 14:36:21.391441 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwjvp" event={"ID":"13d244de-09ce-4867-a07f-e31f4fe90857","Type":"ContainerDied","Data":"6b8fccde56cddec0e78f30f68dd4b829f429de07cd05bc62845d77ec6e323211"} Nov 24 14:36:21 crc kubenswrapper[4790]: I1124 14:36:21.391481 4790 scope.go:117] "RemoveContainer" containerID="f1d5ff6cc173a61349d77db71e8800ce236d75c6dcab837aa3c10a58fc2f97f2" Nov 24 14:36:21 crc kubenswrapper[4790]: I1124 14:36:21.419621 4790 scope.go:117] "RemoveContainer" containerID="db9dbd28986c0f053ebcc2afd276f69bab5375b61a3cb50fecc892bbb3ef7853" Nov 24 14:36:21 crc kubenswrapper[4790]: I1124 14:36:21.438135 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13d244de-09ce-4867-a07f-e31f4fe90857-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:36:21 crc kubenswrapper[4790]: I1124 14:36:21.440722 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zwjvp"] Nov 24 14:36:21 crc kubenswrapper[4790]: I1124 14:36:21.452120 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zwjvp"] Nov 24 14:36:21 crc kubenswrapper[4790]: I1124 14:36:21.467035 4790 scope.go:117] "RemoveContainer" containerID="ecedbac62b382e0980e7f74bbc3f6bb9fca5fcde3afaec3a3103dc956363daca" Nov 24 14:36:21 crc kubenswrapper[4790]: I1124 14:36:21.498146 4790 scope.go:117] "RemoveContainer" containerID="f1d5ff6cc173a61349d77db71e8800ce236d75c6dcab837aa3c10a58fc2f97f2" Nov 24 14:36:21 crc kubenswrapper[4790]: E1124 14:36:21.499033 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1d5ff6cc173a61349d77db71e8800ce236d75c6dcab837aa3c10a58fc2f97f2\": container with ID starting with f1d5ff6cc173a61349d77db71e8800ce236d75c6dcab837aa3c10a58fc2f97f2 not found: ID does not exist" containerID="f1d5ff6cc173a61349d77db71e8800ce236d75c6dcab837aa3c10a58fc2f97f2" Nov 24 14:36:21 crc kubenswrapper[4790]: I1124 14:36:21.499092 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1d5ff6cc173a61349d77db71e8800ce236d75c6dcab837aa3c10a58fc2f97f2"} err="failed to get container status \"f1d5ff6cc173a61349d77db71e8800ce236d75c6dcab837aa3c10a58fc2f97f2\": rpc error: code = NotFound desc = could not find container \"f1d5ff6cc173a61349d77db71e8800ce236d75c6dcab837aa3c10a58fc2f97f2\": container with ID starting with f1d5ff6cc173a61349d77db71e8800ce236d75c6dcab837aa3c10a58fc2f97f2 not found: ID does not exist" Nov 24 14:36:21 crc kubenswrapper[4790]: I1124 14:36:21.499125 4790 scope.go:117] "RemoveContainer" containerID="db9dbd28986c0f053ebcc2afd276f69bab5375b61a3cb50fecc892bbb3ef7853" Nov 24 14:36:21 crc kubenswrapper[4790]: E1124 14:36:21.500027 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db9dbd28986c0f053ebcc2afd276f69bab5375b61a3cb50fecc892bbb3ef7853\": container with ID starting with db9dbd28986c0f053ebcc2afd276f69bab5375b61a3cb50fecc892bbb3ef7853 not found: ID does not exist" containerID="db9dbd28986c0f053ebcc2afd276f69bab5375b61a3cb50fecc892bbb3ef7853" Nov 24 14:36:21 crc kubenswrapper[4790]: I1124 14:36:21.500064 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db9dbd28986c0f053ebcc2afd276f69bab5375b61a3cb50fecc892bbb3ef7853"} err="failed to get container status \"db9dbd28986c0f053ebcc2afd276f69bab5375b61a3cb50fecc892bbb3ef7853\": rpc error: code = NotFound desc = could not find container \"db9dbd28986c0f053ebcc2afd276f69bab5375b61a3cb50fecc892bbb3ef7853\": container with ID starting with db9dbd28986c0f053ebcc2afd276f69bab5375b61a3cb50fecc892bbb3ef7853 not found: ID does not exist" Nov 24 14:36:21 crc kubenswrapper[4790]: I1124 14:36:21.500084 4790 scope.go:117] "RemoveContainer" containerID="ecedbac62b382e0980e7f74bbc3f6bb9fca5fcde3afaec3a3103dc956363daca" Nov 24 14:36:21 crc kubenswrapper[4790]: E1124 14:36:21.500957 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecedbac62b382e0980e7f74bbc3f6bb9fca5fcde3afaec3a3103dc956363daca\": container with ID starting with ecedbac62b382e0980e7f74bbc3f6bb9fca5fcde3afaec3a3103dc956363daca not found: ID does not exist" containerID="ecedbac62b382e0980e7f74bbc3f6bb9fca5fcde3afaec3a3103dc956363daca" Nov 24 14:36:21 crc kubenswrapper[4790]: I1124 14:36:21.501040 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecedbac62b382e0980e7f74bbc3f6bb9fca5fcde3afaec3a3103dc956363daca"} err="failed to get container status \"ecedbac62b382e0980e7f74bbc3f6bb9fca5fcde3afaec3a3103dc956363daca\": rpc error: code = NotFound desc = could not find container \"ecedbac62b382e0980e7f74bbc3f6bb9fca5fcde3afaec3a3103dc956363daca\": container with ID starting with ecedbac62b382e0980e7f74bbc3f6bb9fca5fcde3afaec3a3103dc956363daca not found: ID does not exist" Nov 24 14:36:22 crc kubenswrapper[4790]: I1124 14:36:22.332927 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13d244de-09ce-4867-a07f-e31f4fe90857" path="/var/lib/kubelet/pods/13d244de-09ce-4867-a07f-e31f4fe90857/volumes" Nov 24 14:36:43 crc kubenswrapper[4790]: I1124 14:36:43.939408 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:36:43 crc kubenswrapper[4790]: I1124 14:36:43.940419 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:37:13 crc kubenswrapper[4790]: I1124 14:37:13.939254 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:37:13 crc kubenswrapper[4790]: I1124 14:37:13.940189 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:37:13 crc kubenswrapper[4790]: I1124 14:37:13.940276 4790 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 14:37:13 crc kubenswrapper[4790]: I1124 14:37:13.941710 4790 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1551c55c728005d2d8114bb5394315ba445d97c4e7ab35c743e2f23da5188013"} pod="openshift-machine-config-operator/machine-config-daemon-xz49t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:37:13 crc kubenswrapper[4790]: I1124 14:37:13.941823 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" containerID="cri-o://1551c55c728005d2d8114bb5394315ba445d97c4e7ab35c743e2f23da5188013" gracePeriod=600 Nov 24 14:37:14 crc kubenswrapper[4790]: I1124 14:37:14.224022 4790 generic.go:334] "Generic (PLEG): container finished" podID="0d73b133-48f0-455f-8f6a-742e633f631a" containerID="1551c55c728005d2d8114bb5394315ba445d97c4e7ab35c743e2f23da5188013" exitCode=0 Nov 24 14:37:14 crc kubenswrapper[4790]: I1124 14:37:14.224124 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerDied","Data":"1551c55c728005d2d8114bb5394315ba445d97c4e7ab35c743e2f23da5188013"} Nov 24 14:37:14 crc kubenswrapper[4790]: I1124 14:37:14.224260 4790 scope.go:117] "RemoveContainer" containerID="571bd0c27e6ad4756f7116ed8b5a40d06301f3d85436cea369594fdb322154a7" Nov 24 14:37:15 crc kubenswrapper[4790]: I1124 14:37:15.242308 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"92c086e44ccb28a62777d612e499209f7483422ca70b460ce6ec73c416e8e8b8"} Nov 24 14:37:47 crc kubenswrapper[4790]: I1124 14:37:47.580478 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-copy-data"] Nov 24 14:37:47 crc kubenswrapper[4790]: E1124 14:37:47.582522 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13d244de-09ce-4867-a07f-e31f4fe90857" containerName="registry-server" Nov 24 14:37:47 crc kubenswrapper[4790]: I1124 14:37:47.582564 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="13d244de-09ce-4867-a07f-e31f4fe90857" containerName="registry-server" Nov 24 14:37:47 crc kubenswrapper[4790]: E1124 14:37:47.582639 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13d244de-09ce-4867-a07f-e31f4fe90857" containerName="extract-content" Nov 24 14:37:47 crc kubenswrapper[4790]: I1124 14:37:47.582656 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="13d244de-09ce-4867-a07f-e31f4fe90857" containerName="extract-content" Nov 24 14:37:47 crc kubenswrapper[4790]: E1124 14:37:47.582704 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13d244de-09ce-4867-a07f-e31f4fe90857" containerName="extract-utilities" Nov 24 14:37:47 crc kubenswrapper[4790]: I1124 14:37:47.582724 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="13d244de-09ce-4867-a07f-e31f4fe90857" containerName="extract-utilities" Nov 24 14:37:47 crc kubenswrapper[4790]: I1124 14:37:47.583234 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="13d244de-09ce-4867-a07f-e31f4fe90857" containerName="registry-server" Nov 24 14:37:47 crc kubenswrapper[4790]: I1124 14:37:47.584776 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Nov 24 14:37:47 crc kubenswrapper[4790]: I1124 14:37:47.589386 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Nov 24 14:37:47 crc kubenswrapper[4790]: I1124 14:37:47.592633 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-fx2g4" Nov 24 14:37:47 crc kubenswrapper[4790]: I1124 14:37:47.772287 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f5e64ede-e98c-4007-9957-6a9735fe71e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f5e64ede-e98c-4007-9957-6a9735fe71e4\") pod \"mariadb-copy-data\" (UID: \"185395e2-de39-45a6-acd6-44f6a5405018\") " pod="openstack/mariadb-copy-data" Nov 24 14:37:47 crc kubenswrapper[4790]: I1124 14:37:47.772529 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-845wh\" (UniqueName: \"kubernetes.io/projected/185395e2-de39-45a6-acd6-44f6a5405018-kube-api-access-845wh\") pod \"mariadb-copy-data\" (UID: \"185395e2-de39-45a6-acd6-44f6a5405018\") " pod="openstack/mariadb-copy-data" Nov 24 14:37:47 crc kubenswrapper[4790]: I1124 14:37:47.874324 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-845wh\" (UniqueName: \"kubernetes.io/projected/185395e2-de39-45a6-acd6-44f6a5405018-kube-api-access-845wh\") pod \"mariadb-copy-data\" (UID: \"185395e2-de39-45a6-acd6-44f6a5405018\") " pod="openstack/mariadb-copy-data" Nov 24 14:37:47 crc kubenswrapper[4790]: I1124 14:37:47.874481 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-f5e64ede-e98c-4007-9957-6a9735fe71e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f5e64ede-e98c-4007-9957-6a9735fe71e4\") pod \"mariadb-copy-data\" (UID: \"185395e2-de39-45a6-acd6-44f6a5405018\") " pod="openstack/mariadb-copy-data" Nov 24 14:37:47 crc kubenswrapper[4790]: I1124 14:37:47.880030 4790 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:37:47 crc kubenswrapper[4790]: I1124 14:37:47.880107 4790 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-f5e64ede-e98c-4007-9957-6a9735fe71e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f5e64ede-e98c-4007-9957-6a9735fe71e4\") pod \"mariadb-copy-data\" (UID: \"185395e2-de39-45a6-acd6-44f6a5405018\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/bb3f317ce62734a34413d50c6268fccef1ea0eead7facd64efd219b365d49582/globalmount\"" pod="openstack/mariadb-copy-data" Nov 24 14:37:47 crc kubenswrapper[4790]: I1124 14:37:47.930855 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-845wh\" (UniqueName: \"kubernetes.io/projected/185395e2-de39-45a6-acd6-44f6a5405018-kube-api-access-845wh\") pod \"mariadb-copy-data\" (UID: \"185395e2-de39-45a6-acd6-44f6a5405018\") " pod="openstack/mariadb-copy-data" Nov 24 14:37:47 crc kubenswrapper[4790]: I1124 14:37:47.945729 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-f5e64ede-e98c-4007-9957-6a9735fe71e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f5e64ede-e98c-4007-9957-6a9735fe71e4\") pod \"mariadb-copy-data\" (UID: \"185395e2-de39-45a6-acd6-44f6a5405018\") " pod="openstack/mariadb-copy-data" Nov 24 14:37:48 crc kubenswrapper[4790]: I1124 14:37:48.229972 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Nov 24 14:37:48 crc kubenswrapper[4790]: I1124 14:37:48.898573 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Nov 24 14:37:48 crc kubenswrapper[4790]: W1124 14:37:48.903564 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod185395e2_de39_45a6_acd6_44f6a5405018.slice/crio-6f5bba76d4cf33212a2645d0cc6576e9f1373420e461875fa85d1bf278a95db9 WatchSource:0}: Error finding container 6f5bba76d4cf33212a2645d0cc6576e9f1373420e461875fa85d1bf278a95db9: Status 404 returned error can't find the container with id 6f5bba76d4cf33212a2645d0cc6576e9f1373420e461875fa85d1bf278a95db9 Nov 24 14:37:49 crc kubenswrapper[4790]: I1124 14:37:49.653090 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"185395e2-de39-45a6-acd6-44f6a5405018","Type":"ContainerStarted","Data":"972bb850604ae326cd1ba67e236b9e051e64189a7b2711475af7417ec61cdd17"} Nov 24 14:37:49 crc kubenswrapper[4790]: I1124 14:37:49.653595 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"185395e2-de39-45a6-acd6-44f6a5405018","Type":"ContainerStarted","Data":"6f5bba76d4cf33212a2645d0cc6576e9f1373420e461875fa85d1bf278a95db9"} Nov 24 14:37:49 crc kubenswrapper[4790]: I1124 14:37:49.687109 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-copy-data" podStartSLOduration=3.687083515 podStartE2EDuration="3.687083515s" podCreationTimestamp="2025-11-24 14:37:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:37:49.678963042 +0000 UTC m=+5118.058856744" watchObservedRunningTime="2025-11-24 14:37:49.687083515 +0000 UTC m=+5118.066977217" Nov 24 14:37:52 crc kubenswrapper[4790]: I1124 14:37:52.452326 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Nov 24 14:37:52 crc kubenswrapper[4790]: I1124 14:37:52.455710 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Nov 24 14:37:52 crc kubenswrapper[4790]: I1124 14:37:52.466358 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Nov 24 14:37:52 crc kubenswrapper[4790]: I1124 14:37:52.574382 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2hvl\" (UniqueName: \"kubernetes.io/projected/bcb13254-e917-4431-b9ed-ee1d65203860-kube-api-access-p2hvl\") pod \"mariadb-client\" (UID: \"bcb13254-e917-4431-b9ed-ee1d65203860\") " pod="openstack/mariadb-client" Nov 24 14:37:52 crc kubenswrapper[4790]: I1124 14:37:52.677064 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2hvl\" (UniqueName: \"kubernetes.io/projected/bcb13254-e917-4431-b9ed-ee1d65203860-kube-api-access-p2hvl\") pod \"mariadb-client\" (UID: \"bcb13254-e917-4431-b9ed-ee1d65203860\") " pod="openstack/mariadb-client" Nov 24 14:37:52 crc kubenswrapper[4790]: I1124 14:37:52.711952 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2hvl\" (UniqueName: \"kubernetes.io/projected/bcb13254-e917-4431-b9ed-ee1d65203860-kube-api-access-p2hvl\") pod \"mariadb-client\" (UID: \"bcb13254-e917-4431-b9ed-ee1d65203860\") " pod="openstack/mariadb-client" Nov 24 14:37:52 crc kubenswrapper[4790]: I1124 14:37:52.784181 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Nov 24 14:37:53 crc kubenswrapper[4790]: I1124 14:37:53.086011 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Nov 24 14:37:53 crc kubenswrapper[4790]: W1124 14:37:53.095869 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbcb13254_e917_4431_b9ed_ee1d65203860.slice/crio-03427bc20a6f58cd81aec5dfff6df486a15a027b7b1cd48f74e5218590ab466f WatchSource:0}: Error finding container 03427bc20a6f58cd81aec5dfff6df486a15a027b7b1cd48f74e5218590ab466f: Status 404 returned error can't find the container with id 03427bc20a6f58cd81aec5dfff6df486a15a027b7b1cd48f74e5218590ab466f Nov 24 14:37:53 crc kubenswrapper[4790]: I1124 14:37:53.704797 4790 generic.go:334] "Generic (PLEG): container finished" podID="bcb13254-e917-4431-b9ed-ee1d65203860" containerID="52d7add285f4fde8c621a513f658634507c8af9b6670f4357fba03a56395eb3e" exitCode=0 Nov 24 14:37:53 crc kubenswrapper[4790]: I1124 14:37:53.705071 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"bcb13254-e917-4431-b9ed-ee1d65203860","Type":"ContainerDied","Data":"52d7add285f4fde8c621a513f658634507c8af9b6670f4357fba03a56395eb3e"} Nov 24 14:37:53 crc kubenswrapper[4790]: I1124 14:37:53.705436 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"bcb13254-e917-4431-b9ed-ee1d65203860","Type":"ContainerStarted","Data":"03427bc20a6f58cd81aec5dfff6df486a15a027b7b1cd48f74e5218590ab466f"} Nov 24 14:37:55 crc kubenswrapper[4790]: I1124 14:37:55.142602 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Nov 24 14:37:55 crc kubenswrapper[4790]: I1124 14:37:55.167693 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_bcb13254-e917-4431-b9ed-ee1d65203860/mariadb-client/0.log" Nov 24 14:37:55 crc kubenswrapper[4790]: I1124 14:37:55.199045 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Nov 24 14:37:55 crc kubenswrapper[4790]: I1124 14:37:55.206353 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Nov 24 14:37:55 crc kubenswrapper[4790]: I1124 14:37:55.229803 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2hvl\" (UniqueName: \"kubernetes.io/projected/bcb13254-e917-4431-b9ed-ee1d65203860-kube-api-access-p2hvl\") pod \"bcb13254-e917-4431-b9ed-ee1d65203860\" (UID: \"bcb13254-e917-4431-b9ed-ee1d65203860\") " Nov 24 14:37:55 crc kubenswrapper[4790]: I1124 14:37:55.238100 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcb13254-e917-4431-b9ed-ee1d65203860-kube-api-access-p2hvl" (OuterVolumeSpecName: "kube-api-access-p2hvl") pod "bcb13254-e917-4431-b9ed-ee1d65203860" (UID: "bcb13254-e917-4431-b9ed-ee1d65203860"). InnerVolumeSpecName "kube-api-access-p2hvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:37:55 crc kubenswrapper[4790]: I1124 14:37:55.332306 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2hvl\" (UniqueName: \"kubernetes.io/projected/bcb13254-e917-4431-b9ed-ee1d65203860-kube-api-access-p2hvl\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:55 crc kubenswrapper[4790]: I1124 14:37:55.345415 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Nov 24 14:37:55 crc kubenswrapper[4790]: E1124 14:37:55.346153 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcb13254-e917-4431-b9ed-ee1d65203860" containerName="mariadb-client" Nov 24 14:37:55 crc kubenswrapper[4790]: I1124 14:37:55.346194 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcb13254-e917-4431-b9ed-ee1d65203860" containerName="mariadb-client" Nov 24 14:37:55 crc kubenswrapper[4790]: I1124 14:37:55.346533 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcb13254-e917-4431-b9ed-ee1d65203860" containerName="mariadb-client" Nov 24 14:37:55 crc kubenswrapper[4790]: I1124 14:37:55.347591 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Nov 24 14:37:55 crc kubenswrapper[4790]: I1124 14:37:55.353766 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Nov 24 14:37:55 crc kubenswrapper[4790]: I1124 14:37:55.433940 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7r2fc\" (UniqueName: \"kubernetes.io/projected/4db693c5-09a7-4e1f-a590-32cacd5bd784-kube-api-access-7r2fc\") pod \"mariadb-client\" (UID: \"4db693c5-09a7-4e1f-a590-32cacd5bd784\") " pod="openstack/mariadb-client" Nov 24 14:37:55 crc kubenswrapper[4790]: I1124 14:37:55.535583 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7r2fc\" (UniqueName: \"kubernetes.io/projected/4db693c5-09a7-4e1f-a590-32cacd5bd784-kube-api-access-7r2fc\") pod \"mariadb-client\" (UID: \"4db693c5-09a7-4e1f-a590-32cacd5bd784\") " pod="openstack/mariadb-client" Nov 24 14:37:55 crc kubenswrapper[4790]: I1124 14:37:55.571652 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7r2fc\" (UniqueName: \"kubernetes.io/projected/4db693c5-09a7-4e1f-a590-32cacd5bd784-kube-api-access-7r2fc\") pod \"mariadb-client\" (UID: \"4db693c5-09a7-4e1f-a590-32cacd5bd784\") " pod="openstack/mariadb-client" Nov 24 14:37:55 crc kubenswrapper[4790]: I1124 14:37:55.676612 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Nov 24 14:37:55 crc kubenswrapper[4790]: I1124 14:37:55.738344 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03427bc20a6f58cd81aec5dfff6df486a15a027b7b1cd48f74e5218590ab466f" Nov 24 14:37:55 crc kubenswrapper[4790]: I1124 14:37:55.738431 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Nov 24 14:37:55 crc kubenswrapper[4790]: I1124 14:37:55.771837 4790 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/mariadb-client" oldPodUID="bcb13254-e917-4431-b9ed-ee1d65203860" podUID="4db693c5-09a7-4e1f-a590-32cacd5bd784" Nov 24 14:37:55 crc kubenswrapper[4790]: I1124 14:37:55.998861 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Nov 24 14:37:56 crc kubenswrapper[4790]: W1124 14:37:56.003329 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4db693c5_09a7_4e1f_a590_32cacd5bd784.slice/crio-a62781a12d411e09fcf57b2df3a95b223c53ee7ac82406b1af1ea98ad407e5b1 WatchSource:0}: Error finding container a62781a12d411e09fcf57b2df3a95b223c53ee7ac82406b1af1ea98ad407e5b1: Status 404 returned error can't find the container with id a62781a12d411e09fcf57b2df3a95b223c53ee7ac82406b1af1ea98ad407e5b1 Nov 24 14:37:56 crc kubenswrapper[4790]: I1124 14:37:56.328685 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcb13254-e917-4431-b9ed-ee1d65203860" path="/var/lib/kubelet/pods/bcb13254-e917-4431-b9ed-ee1d65203860/volumes" Nov 24 14:37:56 crc kubenswrapper[4790]: I1124 14:37:56.754182 4790 generic.go:334] "Generic (PLEG): container finished" podID="4db693c5-09a7-4e1f-a590-32cacd5bd784" containerID="764634a785bc3b87e6b8c05109af63fdfe9732a3c7dee748919f23f75f043384" exitCode=0 Nov 24 14:37:56 crc kubenswrapper[4790]: I1124 14:37:56.754323 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"4db693c5-09a7-4e1f-a590-32cacd5bd784","Type":"ContainerDied","Data":"764634a785bc3b87e6b8c05109af63fdfe9732a3c7dee748919f23f75f043384"} Nov 24 14:37:56 crc kubenswrapper[4790]: I1124 14:37:56.754874 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"4db693c5-09a7-4e1f-a590-32cacd5bd784","Type":"ContainerStarted","Data":"a62781a12d411e09fcf57b2df3a95b223c53ee7ac82406b1af1ea98ad407e5b1"} Nov 24 14:37:58 crc kubenswrapper[4790]: I1124 14:37:58.178510 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Nov 24 14:37:58 crc kubenswrapper[4790]: I1124 14:37:58.208399 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_4db693c5-09a7-4e1f-a590-32cacd5bd784/mariadb-client/0.log" Nov 24 14:37:58 crc kubenswrapper[4790]: I1124 14:37:58.242318 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Nov 24 14:37:58 crc kubenswrapper[4790]: I1124 14:37:58.251554 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Nov 24 14:37:58 crc kubenswrapper[4790]: I1124 14:37:58.298548 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7r2fc\" (UniqueName: \"kubernetes.io/projected/4db693c5-09a7-4e1f-a590-32cacd5bd784-kube-api-access-7r2fc\") pod \"4db693c5-09a7-4e1f-a590-32cacd5bd784\" (UID: \"4db693c5-09a7-4e1f-a590-32cacd5bd784\") " Nov 24 14:37:58 crc kubenswrapper[4790]: I1124 14:37:58.309375 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4db693c5-09a7-4e1f-a590-32cacd5bd784-kube-api-access-7r2fc" (OuterVolumeSpecName: "kube-api-access-7r2fc") pod "4db693c5-09a7-4e1f-a590-32cacd5bd784" (UID: "4db693c5-09a7-4e1f-a590-32cacd5bd784"). InnerVolumeSpecName "kube-api-access-7r2fc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:37:58 crc kubenswrapper[4790]: I1124 14:37:58.331354 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4db693c5-09a7-4e1f-a590-32cacd5bd784" path="/var/lib/kubelet/pods/4db693c5-09a7-4e1f-a590-32cacd5bd784/volumes" Nov 24 14:37:58 crc kubenswrapper[4790]: I1124 14:37:58.401855 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7r2fc\" (UniqueName: \"kubernetes.io/projected/4db693c5-09a7-4e1f-a590-32cacd5bd784-kube-api-access-7r2fc\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:58 crc kubenswrapper[4790]: I1124 14:37:58.779321 4790 scope.go:117] "RemoveContainer" containerID="764634a785bc3b87e6b8c05109af63fdfe9732a3c7dee748919f23f75f043384" Nov 24 14:37:58 crc kubenswrapper[4790]: I1124 14:37:58.779406 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.458186 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 14:38:32 crc kubenswrapper[4790]: E1124 14:38:32.459646 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4db693c5-09a7-4e1f-a590-32cacd5bd784" containerName="mariadb-client" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.459674 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="4db693c5-09a7-4e1f-a590-32cacd5bd784" containerName="mariadb-client" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.460219 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="4db693c5-09a7-4e1f-a590-32cacd5bd784" containerName="mariadb-client" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.463601 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.471527 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.471748 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.471982 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-75qff" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.475131 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.479265 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-1"] Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.481088 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.485839 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-2"] Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.488826 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.489672 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35ca55b2-63a0-4c8e-b3f5-f4b645692137-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"35ca55b2-63a0-4c8e-b3f5-f4b645692137\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.489865 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/35ca55b2-63a0-4c8e-b3f5-f4b645692137-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"35ca55b2-63a0-4c8e-b3f5-f4b645692137\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.490027 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/35ca55b2-63a0-4c8e-b3f5-f4b645692137-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"35ca55b2-63a0-4c8e-b3f5-f4b645692137\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.490280 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4m5r\" (UniqueName: \"kubernetes.io/projected/35ca55b2-63a0-4c8e-b3f5-f4b645692137-kube-api-access-z4m5r\") pod \"ovsdbserver-nb-0\" (UID: \"35ca55b2-63a0-4c8e-b3f5-f4b645692137\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.490329 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-82c7babf-c3bd-4b56-a4fb-54c572ca90f8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-82c7babf-c3bd-4b56-a4fb-54c572ca90f8\") pod \"ovsdbserver-nb-0\" (UID: \"35ca55b2-63a0-4c8e-b3f5-f4b645692137\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.490432 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35ca55b2-63a0-4c8e-b3f5-f4b645692137-config\") pod \"ovsdbserver-nb-0\" (UID: \"35ca55b2-63a0-4c8e-b3f5-f4b645692137\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.519256 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.533753 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.592456 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35ca55b2-63a0-4c8e-b3f5-f4b645692137-config\") pod \"ovsdbserver-nb-0\" (UID: \"35ca55b2-63a0-4c8e-b3f5-f4b645692137\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.592580 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35ca55b2-63a0-4c8e-b3f5-f4b645692137-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"35ca55b2-63a0-4c8e-b3f5-f4b645692137\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.592623 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/35ca55b2-63a0-4c8e-b3f5-f4b645692137-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"35ca55b2-63a0-4c8e-b3f5-f4b645692137\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.592661 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/35ca55b2-63a0-4c8e-b3f5-f4b645692137-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"35ca55b2-63a0-4c8e-b3f5-f4b645692137\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.592726 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4m5r\" (UniqueName: \"kubernetes.io/projected/35ca55b2-63a0-4c8e-b3f5-f4b645692137-kube-api-access-z4m5r\") pod \"ovsdbserver-nb-0\" (UID: \"35ca55b2-63a0-4c8e-b3f5-f4b645692137\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.592755 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-82c7babf-c3bd-4b56-a4fb-54c572ca90f8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-82c7babf-c3bd-4b56-a4fb-54c572ca90f8\") pod \"ovsdbserver-nb-0\" (UID: \"35ca55b2-63a0-4c8e-b3f5-f4b645692137\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.594613 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35ca55b2-63a0-4c8e-b3f5-f4b645692137-config\") pod \"ovsdbserver-nb-0\" (UID: \"35ca55b2-63a0-4c8e-b3f5-f4b645692137\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.594921 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/35ca55b2-63a0-4c8e-b3f5-f4b645692137-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"35ca55b2-63a0-4c8e-b3f5-f4b645692137\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.595195 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/35ca55b2-63a0-4c8e-b3f5-f4b645692137-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"35ca55b2-63a0-4c8e-b3f5-f4b645692137\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.599571 4790 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.599616 4790 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-82c7babf-c3bd-4b56-a4fb-54c572ca90f8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-82c7babf-c3bd-4b56-a4fb-54c572ca90f8\") pod \"ovsdbserver-nb-0\" (UID: \"35ca55b2-63a0-4c8e-b3f5-f4b645692137\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c1c9fbff3525d5be598f39cc28ac3ce4662edd465f56ee2a5c8367d18703c644/globalmount\"" pod="openstack/ovsdbserver-nb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.602130 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35ca55b2-63a0-4c8e-b3f5-f4b645692137-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"35ca55b2-63a0-4c8e-b3f5-f4b645692137\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.611704 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4m5r\" (UniqueName: \"kubernetes.io/projected/35ca55b2-63a0-4c8e-b3f5-f4b645692137-kube-api-access-z4m5r\") pod \"ovsdbserver-nb-0\" (UID: \"35ca55b2-63a0-4c8e-b3f5-f4b645692137\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.630187 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-82c7babf-c3bd-4b56-a4fb-54c572ca90f8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-82c7babf-c3bd-4b56-a4fb-54c572ca90f8\") pod \"ovsdbserver-nb-0\" (UID: \"35ca55b2-63a0-4c8e-b3f5-f4b645692137\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.659102 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.660640 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.667189 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.667500 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.668079 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-lrhzd" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.679937 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-2"] Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.683209 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.694281 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/361b98d3-3b9b-4671-85b9-7b9c62ac72d7-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"361b98d3-3b9b-4671-85b9-7b9c62ac72d7\") " pod="openstack/ovsdbserver-nb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.694321 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d069694e-57dd-4c26-9dca-b063d74fb7ea-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"d069694e-57dd-4c26-9dca-b063d74fb7ea\") " pod="openstack/ovsdbserver-nb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.694348 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-87c70be0-4048-4100-ba02-a551afb84b55\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87c70be0-4048-4100-ba02-a551afb84b55\") pod \"ovsdbserver-nb-2\" (UID: \"d069694e-57dd-4c26-9dca-b063d74fb7ea\") " pod="openstack/ovsdbserver-nb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.694379 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/361b98d3-3b9b-4671-85b9-7b9c62ac72d7-config\") pod \"ovsdbserver-nb-1\" (UID: \"361b98d3-3b9b-4671-85b9-7b9c62ac72d7\") " pod="openstack/ovsdbserver-nb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.694406 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d069694e-57dd-4c26-9dca-b063d74fb7ea-config\") pod \"ovsdbserver-nb-2\" (UID: \"d069694e-57dd-4c26-9dca-b063d74fb7ea\") " pod="openstack/ovsdbserver-nb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.694435 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/361b98d3-3b9b-4671-85b9-7b9c62ac72d7-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"361b98d3-3b9b-4671-85b9-7b9c62ac72d7\") " pod="openstack/ovsdbserver-nb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.694471 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2f343fdf-a2d5-4adb-8285-2482799b0c68\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2f343fdf-a2d5-4adb-8285-2482799b0c68\") pod \"ovsdbserver-nb-1\" (UID: \"361b98d3-3b9b-4671-85b9-7b9c62ac72d7\") " pod="openstack/ovsdbserver-nb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.694513 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d069694e-57dd-4c26-9dca-b063d74fb7ea-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"d069694e-57dd-4c26-9dca-b063d74fb7ea\") " pod="openstack/ovsdbserver-nb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.694546 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzrks\" (UniqueName: \"kubernetes.io/projected/d069694e-57dd-4c26-9dca-b063d74fb7ea-kube-api-access-mzrks\") pod \"ovsdbserver-nb-2\" (UID: \"d069694e-57dd-4c26-9dca-b063d74fb7ea\") " pod="openstack/ovsdbserver-nb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.694582 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jhwg\" (UniqueName: \"kubernetes.io/projected/361b98d3-3b9b-4671-85b9-7b9c62ac72d7-kube-api-access-4jhwg\") pod \"ovsdbserver-nb-1\" (UID: \"361b98d3-3b9b-4671-85b9-7b9c62ac72d7\") " pod="openstack/ovsdbserver-nb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.694599 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d069694e-57dd-4c26-9dca-b063d74fb7ea-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"d069694e-57dd-4c26-9dca-b063d74fb7ea\") " pod="openstack/ovsdbserver-nb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.694629 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/361b98d3-3b9b-4671-85b9-7b9c62ac72d7-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"361b98d3-3b9b-4671-85b9-7b9c62ac72d7\") " pod="openstack/ovsdbserver-nb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.704099 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.714911 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-1"] Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.716788 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.722003 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.737008 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.796259 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a3b39ff3-6c4b-4fa3-8fe7-d7b5d3bce6e3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a3b39ff3-6c4b-4fa3-8fe7-d7b5d3bce6e3\") pod \"ovsdbserver-sb-2\" (UID: \"b3d12b48-59a8-45df-8ff3-7850796b5163\") " pod="openstack/ovsdbserver-sb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.796320 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04996e8f-e013-4af2-8890-ca8bd3c21974-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"04996e8f-e013-4af2-8890-ca8bd3c21974\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.796364 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2f343fdf-a2d5-4adb-8285-2482799b0c68\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2f343fdf-a2d5-4adb-8285-2482799b0c68\") pod \"ovsdbserver-nb-1\" (UID: \"361b98d3-3b9b-4671-85b9-7b9c62ac72d7\") " pod="openstack/ovsdbserver-nb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.796394 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-66302b23-966c-41fd-85d6-484c1f52b796\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-66302b23-966c-41fd-85d6-484c1f52b796\") pod \"ovsdbserver-sb-1\" (UID: \"bb538d03-ca9f-4629-935e-f586455cd4e5\") " pod="openstack/ovsdbserver-sb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.796417 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04996e8f-e013-4af2-8890-ca8bd3c21974-config\") pod \"ovsdbserver-sb-0\" (UID: \"04996e8f-e013-4af2-8890-ca8bd3c21974\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.796443 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d069694e-57dd-4c26-9dca-b063d74fb7ea-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"d069694e-57dd-4c26-9dca-b063d74fb7ea\") " pod="openstack/ovsdbserver-nb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.796464 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b3d12b48-59a8-45df-8ff3-7850796b5163-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"b3d12b48-59a8-45df-8ff3-7850796b5163\") " pod="openstack/ovsdbserver-sb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.796671 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzrks\" (UniqueName: \"kubernetes.io/projected/d069694e-57dd-4c26-9dca-b063d74fb7ea-kube-api-access-mzrks\") pod \"ovsdbserver-nb-2\" (UID: \"d069694e-57dd-4c26-9dca-b063d74fb7ea\") " pod="openstack/ovsdbserver-nb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.796711 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3d12b48-59a8-45df-8ff3-7850796b5163-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"b3d12b48-59a8-45df-8ff3-7850796b5163\") " pod="openstack/ovsdbserver-sb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.796741 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04996e8f-e013-4af2-8890-ca8bd3c21974-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"04996e8f-e013-4af2-8890-ca8bd3c21974\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.796774 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjkws\" (UniqueName: \"kubernetes.io/projected/b3d12b48-59a8-45df-8ff3-7850796b5163-kube-api-access-qjkws\") pod \"ovsdbserver-sb-2\" (UID: \"b3d12b48-59a8-45df-8ff3-7850796b5163\") " pod="openstack/ovsdbserver-sb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.796805 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d069694e-57dd-4c26-9dca-b063d74fb7ea-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"d069694e-57dd-4c26-9dca-b063d74fb7ea\") " pod="openstack/ovsdbserver-nb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.796826 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jhwg\" (UniqueName: \"kubernetes.io/projected/361b98d3-3b9b-4671-85b9-7b9c62ac72d7-kube-api-access-4jhwg\") pod \"ovsdbserver-nb-1\" (UID: \"361b98d3-3b9b-4671-85b9-7b9c62ac72d7\") " pod="openstack/ovsdbserver-nb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.796844 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/04996e8f-e013-4af2-8890-ca8bd3c21974-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"04996e8f-e013-4af2-8890-ca8bd3c21974\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.796866 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3d12b48-59a8-45df-8ff3-7850796b5163-config\") pod \"ovsdbserver-sb-2\" (UID: \"b3d12b48-59a8-45df-8ff3-7850796b5163\") " pod="openstack/ovsdbserver-sb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.796913 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/361b98d3-3b9b-4671-85b9-7b9c62ac72d7-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"361b98d3-3b9b-4671-85b9-7b9c62ac72d7\") " pod="openstack/ovsdbserver-nb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.796943 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-c0f94ef3-df34-495d-a8e3-38823bb9ec0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c0f94ef3-df34-495d-a8e3-38823bb9ec0e\") pod \"ovsdbserver-sb-0\" (UID: \"04996e8f-e013-4af2-8890-ca8bd3c21974\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.796961 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb538d03-ca9f-4629-935e-f586455cd4e5-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"bb538d03-ca9f-4629-935e-f586455cd4e5\") " pod="openstack/ovsdbserver-sb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.796981 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bb538d03-ca9f-4629-935e-f586455cd4e5-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"bb538d03-ca9f-4629-935e-f586455cd4e5\") " pod="openstack/ovsdbserver-sb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.797013 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/361b98d3-3b9b-4671-85b9-7b9c62ac72d7-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"361b98d3-3b9b-4671-85b9-7b9c62ac72d7\") " pod="openstack/ovsdbserver-nb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.797035 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d069694e-57dd-4c26-9dca-b063d74fb7ea-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"d069694e-57dd-4c26-9dca-b063d74fb7ea\") " pod="openstack/ovsdbserver-nb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.797065 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-87c70be0-4048-4100-ba02-a551afb84b55\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87c70be0-4048-4100-ba02-a551afb84b55\") pod \"ovsdbserver-nb-2\" (UID: \"d069694e-57dd-4c26-9dca-b063d74fb7ea\") " pod="openstack/ovsdbserver-nb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.797090 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/361b98d3-3b9b-4671-85b9-7b9c62ac72d7-config\") pod \"ovsdbserver-nb-1\" (UID: \"361b98d3-3b9b-4671-85b9-7b9c62ac72d7\") " pod="openstack/ovsdbserver-nb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.797110 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5x2v7\" (UniqueName: \"kubernetes.io/projected/bb538d03-ca9f-4629-935e-f586455cd4e5-kube-api-access-5x2v7\") pod \"ovsdbserver-sb-1\" (UID: \"bb538d03-ca9f-4629-935e-f586455cd4e5\") " pod="openstack/ovsdbserver-sb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.797135 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb538d03-ca9f-4629-935e-f586455cd4e5-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"bb538d03-ca9f-4629-935e-f586455cd4e5\") " pod="openstack/ovsdbserver-sb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.797156 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2v7v\" (UniqueName: \"kubernetes.io/projected/04996e8f-e013-4af2-8890-ca8bd3c21974-kube-api-access-m2v7v\") pod \"ovsdbserver-sb-0\" (UID: \"04996e8f-e013-4af2-8890-ca8bd3c21974\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.797177 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb538d03-ca9f-4629-935e-f586455cd4e5-config\") pod \"ovsdbserver-sb-1\" (UID: \"bb538d03-ca9f-4629-935e-f586455cd4e5\") " pod="openstack/ovsdbserver-sb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.797204 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d069694e-57dd-4c26-9dca-b063d74fb7ea-config\") pod \"ovsdbserver-nb-2\" (UID: \"d069694e-57dd-4c26-9dca-b063d74fb7ea\") " pod="openstack/ovsdbserver-nb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.797225 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/361b98d3-3b9b-4671-85b9-7b9c62ac72d7-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"361b98d3-3b9b-4671-85b9-7b9c62ac72d7\") " pod="openstack/ovsdbserver-nb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.798030 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b3d12b48-59a8-45df-8ff3-7850796b5163-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"b3d12b48-59a8-45df-8ff3-7850796b5163\") " pod="openstack/ovsdbserver-sb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.798237 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/361b98d3-3b9b-4671-85b9-7b9c62ac72d7-config\") pod \"ovsdbserver-nb-1\" (UID: \"361b98d3-3b9b-4671-85b9-7b9c62ac72d7\") " pod="openstack/ovsdbserver-nb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.798264 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/361b98d3-3b9b-4671-85b9-7b9c62ac72d7-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"361b98d3-3b9b-4671-85b9-7b9c62ac72d7\") " pod="openstack/ovsdbserver-nb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.798273 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d069694e-57dd-4c26-9dca-b063d74fb7ea-config\") pod \"ovsdbserver-nb-2\" (UID: \"d069694e-57dd-4c26-9dca-b063d74fb7ea\") " pod="openstack/ovsdbserver-nb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.798843 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d069694e-57dd-4c26-9dca-b063d74fb7ea-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"d069694e-57dd-4c26-9dca-b063d74fb7ea\") " pod="openstack/ovsdbserver-nb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.799005 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/361b98d3-3b9b-4671-85b9-7b9c62ac72d7-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"361b98d3-3b9b-4671-85b9-7b9c62ac72d7\") " pod="openstack/ovsdbserver-nb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.799369 4790 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.799402 4790 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-87c70be0-4048-4100-ba02-a551afb84b55\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87c70be0-4048-4100-ba02-a551afb84b55\") pod \"ovsdbserver-nb-2\" (UID: \"d069694e-57dd-4c26-9dca-b063d74fb7ea\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/aa783e6d9cd5a41a46613b9189872f15841fb15e9c9fbba33384851c4f05e554/globalmount\"" pod="openstack/ovsdbserver-nb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.799437 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d069694e-57dd-4c26-9dca-b063d74fb7ea-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"d069694e-57dd-4c26-9dca-b063d74fb7ea\") " pod="openstack/ovsdbserver-nb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.801871 4790 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.802041 4790 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2f343fdf-a2d5-4adb-8285-2482799b0c68\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2f343fdf-a2d5-4adb-8285-2482799b0c68\") pod \"ovsdbserver-nb-1\" (UID: \"361b98d3-3b9b-4671-85b9-7b9c62ac72d7\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4b55b48936ba5959cd6b3c7240178a52b946eb7fc253922c67c71e726d34c7f4/globalmount\"" pod="openstack/ovsdbserver-nb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.802270 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/361b98d3-3b9b-4671-85b9-7b9c62ac72d7-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"361b98d3-3b9b-4671-85b9-7b9c62ac72d7\") " pod="openstack/ovsdbserver-nb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.804393 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d069694e-57dd-4c26-9dca-b063d74fb7ea-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"d069694e-57dd-4c26-9dca-b063d74fb7ea\") " pod="openstack/ovsdbserver-nb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.805440 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.816512 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jhwg\" (UniqueName: \"kubernetes.io/projected/361b98d3-3b9b-4671-85b9-7b9c62ac72d7-kube-api-access-4jhwg\") pod \"ovsdbserver-nb-1\" (UID: \"361b98d3-3b9b-4671-85b9-7b9c62ac72d7\") " pod="openstack/ovsdbserver-nb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.818654 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzrks\" (UniqueName: \"kubernetes.io/projected/d069694e-57dd-4c26-9dca-b063d74fb7ea-kube-api-access-mzrks\") pod \"ovsdbserver-nb-2\" (UID: \"d069694e-57dd-4c26-9dca-b063d74fb7ea\") " pod="openstack/ovsdbserver-nb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.835271 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-87c70be0-4048-4100-ba02-a551afb84b55\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87c70be0-4048-4100-ba02-a551afb84b55\") pod \"ovsdbserver-nb-2\" (UID: \"d069694e-57dd-4c26-9dca-b063d74fb7ea\") " pod="openstack/ovsdbserver-nb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.840763 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.844535 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2f343fdf-a2d5-4adb-8285-2482799b0c68\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2f343fdf-a2d5-4adb-8285-2482799b0c68\") pod \"ovsdbserver-nb-1\" (UID: \"361b98d3-3b9b-4671-85b9-7b9c62ac72d7\") " pod="openstack/ovsdbserver-nb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.900226 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a3b39ff3-6c4b-4fa3-8fe7-d7b5d3bce6e3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a3b39ff3-6c4b-4fa3-8fe7-d7b5d3bce6e3\") pod \"ovsdbserver-sb-2\" (UID: \"b3d12b48-59a8-45df-8ff3-7850796b5163\") " pod="openstack/ovsdbserver-sb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.900291 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04996e8f-e013-4af2-8890-ca8bd3c21974-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"04996e8f-e013-4af2-8890-ca8bd3c21974\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.900323 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-66302b23-966c-41fd-85d6-484c1f52b796\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-66302b23-966c-41fd-85d6-484c1f52b796\") pod \"ovsdbserver-sb-1\" (UID: \"bb538d03-ca9f-4629-935e-f586455cd4e5\") " pod="openstack/ovsdbserver-sb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.900346 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04996e8f-e013-4af2-8890-ca8bd3c21974-config\") pod \"ovsdbserver-sb-0\" (UID: \"04996e8f-e013-4af2-8890-ca8bd3c21974\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.900365 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b3d12b48-59a8-45df-8ff3-7850796b5163-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"b3d12b48-59a8-45df-8ff3-7850796b5163\") " pod="openstack/ovsdbserver-sb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.900403 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3d12b48-59a8-45df-8ff3-7850796b5163-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"b3d12b48-59a8-45df-8ff3-7850796b5163\") " pod="openstack/ovsdbserver-sb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.900432 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04996e8f-e013-4af2-8890-ca8bd3c21974-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"04996e8f-e013-4af2-8890-ca8bd3c21974\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.900452 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjkws\" (UniqueName: \"kubernetes.io/projected/b3d12b48-59a8-45df-8ff3-7850796b5163-kube-api-access-qjkws\") pod \"ovsdbserver-sb-2\" (UID: \"b3d12b48-59a8-45df-8ff3-7850796b5163\") " pod="openstack/ovsdbserver-sb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.900480 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/04996e8f-e013-4af2-8890-ca8bd3c21974-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"04996e8f-e013-4af2-8890-ca8bd3c21974\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.900503 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3d12b48-59a8-45df-8ff3-7850796b5163-config\") pod \"ovsdbserver-sb-2\" (UID: \"b3d12b48-59a8-45df-8ff3-7850796b5163\") " pod="openstack/ovsdbserver-sb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.900540 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-c0f94ef3-df34-495d-a8e3-38823bb9ec0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c0f94ef3-df34-495d-a8e3-38823bb9ec0e\") pod \"ovsdbserver-sb-0\" (UID: \"04996e8f-e013-4af2-8890-ca8bd3c21974\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.900561 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb538d03-ca9f-4629-935e-f586455cd4e5-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"bb538d03-ca9f-4629-935e-f586455cd4e5\") " pod="openstack/ovsdbserver-sb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.900580 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bb538d03-ca9f-4629-935e-f586455cd4e5-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"bb538d03-ca9f-4629-935e-f586455cd4e5\") " pod="openstack/ovsdbserver-sb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.900627 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5x2v7\" (UniqueName: \"kubernetes.io/projected/bb538d03-ca9f-4629-935e-f586455cd4e5-kube-api-access-5x2v7\") pod \"ovsdbserver-sb-1\" (UID: \"bb538d03-ca9f-4629-935e-f586455cd4e5\") " pod="openstack/ovsdbserver-sb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.900651 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb538d03-ca9f-4629-935e-f586455cd4e5-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"bb538d03-ca9f-4629-935e-f586455cd4e5\") " pod="openstack/ovsdbserver-sb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.900672 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2v7v\" (UniqueName: \"kubernetes.io/projected/04996e8f-e013-4af2-8890-ca8bd3c21974-kube-api-access-m2v7v\") pod \"ovsdbserver-sb-0\" (UID: \"04996e8f-e013-4af2-8890-ca8bd3c21974\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.900690 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb538d03-ca9f-4629-935e-f586455cd4e5-config\") pod \"ovsdbserver-sb-1\" (UID: \"bb538d03-ca9f-4629-935e-f586455cd4e5\") " pod="openstack/ovsdbserver-sb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.900723 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b3d12b48-59a8-45df-8ff3-7850796b5163-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"b3d12b48-59a8-45df-8ff3-7850796b5163\") " pod="openstack/ovsdbserver-sb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.901872 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b3d12b48-59a8-45df-8ff3-7850796b5163-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"b3d12b48-59a8-45df-8ff3-7850796b5163\") " pod="openstack/ovsdbserver-sb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.903098 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3d12b48-59a8-45df-8ff3-7850796b5163-config\") pod \"ovsdbserver-sb-2\" (UID: \"b3d12b48-59a8-45df-8ff3-7850796b5163\") " pod="openstack/ovsdbserver-sb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.903267 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/04996e8f-e013-4af2-8890-ca8bd3c21974-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"04996e8f-e013-4af2-8890-ca8bd3c21974\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.904941 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04996e8f-e013-4af2-8890-ca8bd3c21974-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"04996e8f-e013-4af2-8890-ca8bd3c21974\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.904947 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04996e8f-e013-4af2-8890-ca8bd3c21974-config\") pod \"ovsdbserver-sb-0\" (UID: \"04996e8f-e013-4af2-8890-ca8bd3c21974\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.905615 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb538d03-ca9f-4629-935e-f586455cd4e5-config\") pod \"ovsdbserver-sb-1\" (UID: \"bb538d03-ca9f-4629-935e-f586455cd4e5\") " pod="openstack/ovsdbserver-sb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.905919 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb538d03-ca9f-4629-935e-f586455cd4e5-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"bb538d03-ca9f-4629-935e-f586455cd4e5\") " pod="openstack/ovsdbserver-sb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.906291 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bb538d03-ca9f-4629-935e-f586455cd4e5-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"bb538d03-ca9f-4629-935e-f586455cd4e5\") " pod="openstack/ovsdbserver-sb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.909724 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b3d12b48-59a8-45df-8ff3-7850796b5163-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"b3d12b48-59a8-45df-8ff3-7850796b5163\") " pod="openstack/ovsdbserver-sb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.911662 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3d12b48-59a8-45df-8ff3-7850796b5163-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"b3d12b48-59a8-45df-8ff3-7850796b5163\") " pod="openstack/ovsdbserver-sb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.911954 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04996e8f-e013-4af2-8890-ca8bd3c21974-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"04996e8f-e013-4af2-8890-ca8bd3c21974\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.913286 4790 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.913321 4790 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a3b39ff3-6c4b-4fa3-8fe7-d7b5d3bce6e3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a3b39ff3-6c4b-4fa3-8fe7-d7b5d3bce6e3\") pod \"ovsdbserver-sb-2\" (UID: \"b3d12b48-59a8-45df-8ff3-7850796b5163\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/250b9866883596861569eb11eefe4f99a9eab4d535705d831cdea3f64555e0dd/globalmount\"" pod="openstack/ovsdbserver-sb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.914033 4790 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.914064 4790 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-c0f94ef3-df34-495d-a8e3-38823bb9ec0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c0f94ef3-df34-495d-a8e3-38823bb9ec0e\") pod \"ovsdbserver-sb-0\" (UID: \"04996e8f-e013-4af2-8890-ca8bd3c21974\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/9b21229c2c12f168a05d39a798272a926010a58c6ef591d67661273e79910008/globalmount\"" pod="openstack/ovsdbserver-sb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.914428 4790 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.914482 4790 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-66302b23-966c-41fd-85d6-484c1f52b796\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-66302b23-966c-41fd-85d6-484c1f52b796\") pod \"ovsdbserver-sb-1\" (UID: \"bb538d03-ca9f-4629-935e-f586455cd4e5\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/08dcdccb579c56022e88cd612bf8b1a19647bf9de6416b35a499cec92cbc9a8e/globalmount\"" pod="openstack/ovsdbserver-sb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.918821 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb538d03-ca9f-4629-935e-f586455cd4e5-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"bb538d03-ca9f-4629-935e-f586455cd4e5\") " pod="openstack/ovsdbserver-sb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.928142 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjkws\" (UniqueName: \"kubernetes.io/projected/b3d12b48-59a8-45df-8ff3-7850796b5163-kube-api-access-qjkws\") pod \"ovsdbserver-sb-2\" (UID: \"b3d12b48-59a8-45df-8ff3-7850796b5163\") " pod="openstack/ovsdbserver-sb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.929863 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2v7v\" (UniqueName: \"kubernetes.io/projected/04996e8f-e013-4af2-8890-ca8bd3c21974-kube-api-access-m2v7v\") pod \"ovsdbserver-sb-0\" (UID: \"04996e8f-e013-4af2-8890-ca8bd3c21974\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.932414 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5x2v7\" (UniqueName: \"kubernetes.io/projected/bb538d03-ca9f-4629-935e-f586455cd4e5-kube-api-access-5x2v7\") pod \"ovsdbserver-sb-1\" (UID: \"bb538d03-ca9f-4629-935e-f586455cd4e5\") " pod="openstack/ovsdbserver-sb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.950283 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-66302b23-966c-41fd-85d6-484c1f52b796\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-66302b23-966c-41fd-85d6-484c1f52b796\") pod \"ovsdbserver-sb-1\" (UID: \"bb538d03-ca9f-4629-935e-f586455cd4e5\") " pod="openstack/ovsdbserver-sb-1" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.956872 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a3b39ff3-6c4b-4fa3-8fe7-d7b5d3bce6e3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a3b39ff3-6c4b-4fa3-8fe7-d7b5d3bce6e3\") pod \"ovsdbserver-sb-2\" (UID: \"b3d12b48-59a8-45df-8ff3-7850796b5163\") " pod="openstack/ovsdbserver-sb-2" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.958452 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-c0f94ef3-df34-495d-a8e3-38823bb9ec0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c0f94ef3-df34-495d-a8e3-38823bb9ec0e\") pod \"ovsdbserver-sb-0\" (UID: \"04996e8f-e013-4af2-8890-ca8bd3c21974\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:38:32 crc kubenswrapper[4790]: I1124 14:38:32.986899 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 14:38:33 crc kubenswrapper[4790]: I1124 14:38:33.004148 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Nov 24 14:38:33 crc kubenswrapper[4790]: I1124 14:38:33.039854 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Nov 24 14:38:33 crc kubenswrapper[4790]: I1124 14:38:33.125042 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Nov 24 14:38:33 crc kubenswrapper[4790]: I1124 14:38:33.256259 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 14:38:33 crc kubenswrapper[4790]: W1124 14:38:33.260982 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35ca55b2_63a0_4c8e_b3f5_f4b645692137.slice/crio-976d1fdbd3bd59f3f109bbebb7fd9c1fbe51d75c1e051ca5fb067a233bc7c18e WatchSource:0}: Error finding container 976d1fdbd3bd59f3f109bbebb7fd9c1fbe51d75c1e051ca5fb067a233bc7c18e: Status 404 returned error can't find the container with id 976d1fdbd3bd59f3f109bbebb7fd9c1fbe51d75c1e051ca5fb067a233bc7c18e Nov 24 14:38:33 crc kubenswrapper[4790]: I1124 14:38:33.368859 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Nov 24 14:38:33 crc kubenswrapper[4790]: W1124 14:38:33.386114 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd069694e_57dd_4c26_9dca_b063d74fb7ea.slice/crio-8f7080e77a8e0664b61d84fd002da2ad7b6c9a7e44b3a1362c006ea74f1df753 WatchSource:0}: Error finding container 8f7080e77a8e0664b61d84fd002da2ad7b6c9a7e44b3a1362c006ea74f1df753: Status 404 returned error can't find the container with id 8f7080e77a8e0664b61d84fd002da2ad7b6c9a7e44b3a1362c006ea74f1df753 Nov 24 14:38:33 crc kubenswrapper[4790]: I1124 14:38:33.560533 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Nov 24 14:38:33 crc kubenswrapper[4790]: I1124 14:38:33.666707 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 14:38:33 crc kubenswrapper[4790]: I1124 14:38:33.767187 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Nov 24 14:38:33 crc kubenswrapper[4790]: W1124 14:38:33.773705 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod361b98d3_3b9b_4671_85b9_7b9c62ac72d7.slice/crio-2ed4f7875a113cbb353eb7176719b662808cb94378515eef0415e5b0da4458d5 WatchSource:0}: Error finding container 2ed4f7875a113cbb353eb7176719b662808cb94378515eef0415e5b0da4458d5: Status 404 returned error can't find the container with id 2ed4f7875a113cbb353eb7176719b662808cb94378515eef0415e5b0da4458d5 Nov 24 14:38:34 crc kubenswrapper[4790]: I1124 14:38:34.159909 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"35ca55b2-63a0-4c8e-b3f5-f4b645692137","Type":"ContainerStarted","Data":"9aafe4c890c1a0abf5cfda4264a54851299391aa07d33f40036129c5e5278adf"} Nov 24 14:38:34 crc kubenswrapper[4790]: I1124 14:38:34.159972 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"35ca55b2-63a0-4c8e-b3f5-f4b645692137","Type":"ContainerStarted","Data":"3539898f63d2038979b76c41ecacac394c4748a8aab4f751a1acb057219dc1ef"} Nov 24 14:38:34 crc kubenswrapper[4790]: I1124 14:38:34.159983 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"35ca55b2-63a0-4c8e-b3f5-f4b645692137","Type":"ContainerStarted","Data":"976d1fdbd3bd59f3f109bbebb7fd9c1fbe51d75c1e051ca5fb067a233bc7c18e"} Nov 24 14:38:34 crc kubenswrapper[4790]: I1124 14:38:34.162468 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"d069694e-57dd-4c26-9dca-b063d74fb7ea","Type":"ContainerStarted","Data":"d09c5befadbd6243341e5942b7ad7c1b9f2c3b6c5e04d31e60ec64e06fcccb0c"} Nov 24 14:38:34 crc kubenswrapper[4790]: I1124 14:38:34.162516 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"d069694e-57dd-4c26-9dca-b063d74fb7ea","Type":"ContainerStarted","Data":"2991c051561d4aae85785c649548aa03d041e906a1292c9c0da9ec258be96b8a"} Nov 24 14:38:34 crc kubenswrapper[4790]: I1124 14:38:34.162534 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"d069694e-57dd-4c26-9dca-b063d74fb7ea","Type":"ContainerStarted","Data":"8f7080e77a8e0664b61d84fd002da2ad7b6c9a7e44b3a1362c006ea74f1df753"} Nov 24 14:38:34 crc kubenswrapper[4790]: I1124 14:38:34.166025 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"04996e8f-e013-4af2-8890-ca8bd3c21974","Type":"ContainerStarted","Data":"f53dd91f97dc8d3f3ea134c7179d42981e88050d72ee095fd9d2255eb6af06fc"} Nov 24 14:38:34 crc kubenswrapper[4790]: I1124 14:38:34.166094 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"04996e8f-e013-4af2-8890-ca8bd3c21974","Type":"ContainerStarted","Data":"e96ffc7a6dc21d60c5737f0230bf253c24f4760bf234c1033cea969b84f2ae60"} Nov 24 14:38:34 crc kubenswrapper[4790]: I1124 14:38:34.166107 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"04996e8f-e013-4af2-8890-ca8bd3c21974","Type":"ContainerStarted","Data":"cd2ec203a59a1dcbe29137608566057f10c479192d353be762560dab8dd9fa8e"} Nov 24 14:38:34 crc kubenswrapper[4790]: I1124 14:38:34.168225 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"361b98d3-3b9b-4671-85b9-7b9c62ac72d7","Type":"ContainerStarted","Data":"79bc21502d1111b646fb684647eb657aecae1129b14fb3ff319b984b620ce806"} Nov 24 14:38:34 crc kubenswrapper[4790]: I1124 14:38:34.168255 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"361b98d3-3b9b-4671-85b9-7b9c62ac72d7","Type":"ContainerStarted","Data":"8dca4af88741950f79d88373208baaa234ae0ff32e4d7fa3e496f06427b4e79c"} Nov 24 14:38:34 crc kubenswrapper[4790]: I1124 14:38:34.168279 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"361b98d3-3b9b-4671-85b9-7b9c62ac72d7","Type":"ContainerStarted","Data":"2ed4f7875a113cbb353eb7176719b662808cb94378515eef0415e5b0da4458d5"} Nov 24 14:38:34 crc kubenswrapper[4790]: I1124 14:38:34.170723 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"b3d12b48-59a8-45df-8ff3-7850796b5163","Type":"ContainerStarted","Data":"3f870ecf49325e9df16d857c85252763eac0131e1dc244504ec581c8f1548341"} Nov 24 14:38:34 crc kubenswrapper[4790]: I1124 14:38:34.170779 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"b3d12b48-59a8-45df-8ff3-7850796b5163","Type":"ContainerStarted","Data":"21f6d328a03865f76b7d54525190fe4a9adb912790448d2cd81c913c2fe74853"} Nov 24 14:38:34 crc kubenswrapper[4790]: I1124 14:38:34.170792 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"b3d12b48-59a8-45df-8ff3-7850796b5163","Type":"ContainerStarted","Data":"3cdf6ebe99b5dd6fd1f53fc2b6d29ed83ccf0ee02312b682443c851adc2e1dc0"} Nov 24 14:38:34 crc kubenswrapper[4790]: I1124 14:38:34.188280 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=3.188255846 podStartE2EDuration="3.188255846s" podCreationTimestamp="2025-11-24 14:38:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:38:34.181981768 +0000 UTC m=+5162.561875430" watchObservedRunningTime="2025-11-24 14:38:34.188255846 +0000 UTC m=+5162.568149518" Nov 24 14:38:34 crc kubenswrapper[4790]: I1124 14:38:34.209649 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=3.209616194 podStartE2EDuration="3.209616194s" podCreationTimestamp="2025-11-24 14:38:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:38:34.200045248 +0000 UTC m=+5162.579938950" watchObservedRunningTime="2025-11-24 14:38:34.209616194 +0000 UTC m=+5162.589509856" Nov 24 14:38:34 crc kubenswrapper[4790]: I1124 14:38:34.234697 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-1" podStartSLOduration=3.234676111 podStartE2EDuration="3.234676111s" podCreationTimestamp="2025-11-24 14:38:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:38:34.233435734 +0000 UTC m=+5162.613329396" watchObservedRunningTime="2025-11-24 14:38:34.234676111 +0000 UTC m=+5162.614569773" Nov 24 14:38:34 crc kubenswrapper[4790]: I1124 14:38:34.255227 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-2" podStartSLOduration=3.255204144 podStartE2EDuration="3.255204144s" podCreationTimestamp="2025-11-24 14:38:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:38:34.254487112 +0000 UTC m=+5162.634380774" watchObservedRunningTime="2025-11-24 14:38:34.255204144 +0000 UTC m=+5162.635097806" Nov 24 14:38:34 crc kubenswrapper[4790]: I1124 14:38:34.273808 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-2" podStartSLOduration=3.273789869 podStartE2EDuration="3.273789869s" podCreationTimestamp="2025-11-24 14:38:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:38:34.270329215 +0000 UTC m=+5162.650222887" watchObservedRunningTime="2025-11-24 14:38:34.273789869 +0000 UTC m=+5162.653683531" Nov 24 14:38:34 crc kubenswrapper[4790]: I1124 14:38:34.311451 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Nov 24 14:38:35 crc kubenswrapper[4790]: I1124 14:38:35.186722 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"bb538d03-ca9f-4629-935e-f586455cd4e5","Type":"ContainerStarted","Data":"d1629e020ebbee05537c48ec5575a595ffee0e6dbbd21c834644f8a7e29daafa"} Nov 24 14:38:35 crc kubenswrapper[4790]: I1124 14:38:35.186835 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"bb538d03-ca9f-4629-935e-f586455cd4e5","Type":"ContainerStarted","Data":"09d386cec0aafa9624399e05930e8acd232dee26cc17d1d6c619121c841a256b"} Nov 24 14:38:35 crc kubenswrapper[4790]: I1124 14:38:35.186848 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"bb538d03-ca9f-4629-935e-f586455cd4e5","Type":"ContainerStarted","Data":"0813c9b9e3a066feaf817291839497362b7d58597a30c444aed1e2caf1421355"} Nov 24 14:38:35 crc kubenswrapper[4790]: I1124 14:38:35.219593 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-1" podStartSLOduration=4.219562737 podStartE2EDuration="4.219562737s" podCreationTimestamp="2025-11-24 14:38:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:38:35.213992311 +0000 UTC m=+5163.593885993" watchObservedRunningTime="2025-11-24 14:38:35.219562737 +0000 UTC m=+5163.599456429" Nov 24 14:38:35 crc kubenswrapper[4790]: I1124 14:38:35.805565 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 24 14:38:35 crc kubenswrapper[4790]: I1124 14:38:35.841196 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-2" Nov 24 14:38:35 crc kubenswrapper[4790]: I1124 14:38:35.987549 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 24 14:38:36 crc kubenswrapper[4790]: I1124 14:38:36.005134 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-2" Nov 24 14:38:36 crc kubenswrapper[4790]: I1124 14:38:36.040328 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-1" Nov 24 14:38:36 crc kubenswrapper[4790]: I1124 14:38:36.079468 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-2" Nov 24 14:38:36 crc kubenswrapper[4790]: I1124 14:38:36.125840 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-1" Nov 24 14:38:36 crc kubenswrapper[4790]: I1124 14:38:36.188549 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-1" Nov 24 14:38:36 crc kubenswrapper[4790]: I1124 14:38:36.199473 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-1" Nov 24 14:38:36 crc kubenswrapper[4790]: I1124 14:38:36.199523 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-2" Nov 24 14:38:37 crc kubenswrapper[4790]: I1124 14:38:37.805821 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 24 14:38:37 crc kubenswrapper[4790]: I1124 14:38:37.840929 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-2" Nov 24 14:38:37 crc kubenswrapper[4790]: I1124 14:38:37.987137 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.040589 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-1" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.082521 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-2" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.206711 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-1" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.380800 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c9c4b4dfc-h6f2z"] Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.387845 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c9c4b4dfc-h6f2z" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.392904 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.407270 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c9c4b4dfc-h6f2z"] Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.519453 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f3537c77-1bb1-4595-92c4-a9298a5df3a9-ovsdbserver-sb\") pod \"dnsmasq-dns-6c9c4b4dfc-h6f2z\" (UID: \"f3537c77-1bb1-4595-92c4-a9298a5df3a9\") " pod="openstack/dnsmasq-dns-6c9c4b4dfc-h6f2z" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.519540 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3537c77-1bb1-4595-92c4-a9298a5df3a9-config\") pod \"dnsmasq-dns-6c9c4b4dfc-h6f2z\" (UID: \"f3537c77-1bb1-4595-92c4-a9298a5df3a9\") " pod="openstack/dnsmasq-dns-6c9c4b4dfc-h6f2z" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.519714 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfndl\" (UniqueName: \"kubernetes.io/projected/f3537c77-1bb1-4595-92c4-a9298a5df3a9-kube-api-access-jfndl\") pod \"dnsmasq-dns-6c9c4b4dfc-h6f2z\" (UID: \"f3537c77-1bb1-4595-92c4-a9298a5df3a9\") " pod="openstack/dnsmasq-dns-6c9c4b4dfc-h6f2z" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.519811 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f3537c77-1bb1-4595-92c4-a9298a5df3a9-dns-svc\") pod \"dnsmasq-dns-6c9c4b4dfc-h6f2z\" (UID: \"f3537c77-1bb1-4595-92c4-a9298a5df3a9\") " pod="openstack/dnsmasq-dns-6c9c4b4dfc-h6f2z" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.592622 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c9c4b4dfc-h6f2z"] Nov 24 14:38:38 crc kubenswrapper[4790]: E1124 14:38:38.593509 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-jfndl ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-6c9c4b4dfc-h6f2z" podUID="f3537c77-1bb1-4595-92c4-a9298a5df3a9" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.622138 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f3537c77-1bb1-4595-92c4-a9298a5df3a9-ovsdbserver-sb\") pod \"dnsmasq-dns-6c9c4b4dfc-h6f2z\" (UID: \"f3537c77-1bb1-4595-92c4-a9298a5df3a9\") " pod="openstack/dnsmasq-dns-6c9c4b4dfc-h6f2z" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.622226 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3537c77-1bb1-4595-92c4-a9298a5df3a9-config\") pod \"dnsmasq-dns-6c9c4b4dfc-h6f2z\" (UID: \"f3537c77-1bb1-4595-92c4-a9298a5df3a9\") " pod="openstack/dnsmasq-dns-6c9c4b4dfc-h6f2z" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.622272 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfndl\" (UniqueName: \"kubernetes.io/projected/f3537c77-1bb1-4595-92c4-a9298a5df3a9-kube-api-access-jfndl\") pod \"dnsmasq-dns-6c9c4b4dfc-h6f2z\" (UID: \"f3537c77-1bb1-4595-92c4-a9298a5df3a9\") " pod="openstack/dnsmasq-dns-6c9c4b4dfc-h6f2z" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.622303 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f3537c77-1bb1-4595-92c4-a9298a5df3a9-dns-svc\") pod \"dnsmasq-dns-6c9c4b4dfc-h6f2z\" (UID: \"f3537c77-1bb1-4595-92c4-a9298a5df3a9\") " pod="openstack/dnsmasq-dns-6c9c4b4dfc-h6f2z" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.623194 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f3537c77-1bb1-4595-92c4-a9298a5df3a9-dns-svc\") pod \"dnsmasq-dns-6c9c4b4dfc-h6f2z\" (UID: \"f3537c77-1bb1-4595-92c4-a9298a5df3a9\") " pod="openstack/dnsmasq-dns-6c9c4b4dfc-h6f2z" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.623302 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f3537c77-1bb1-4595-92c4-a9298a5df3a9-ovsdbserver-sb\") pod \"dnsmasq-dns-6c9c4b4dfc-h6f2z\" (UID: \"f3537c77-1bb1-4595-92c4-a9298a5df3a9\") " pod="openstack/dnsmasq-dns-6c9c4b4dfc-h6f2z" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.623377 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3537c77-1bb1-4595-92c4-a9298a5df3a9-config\") pod \"dnsmasq-dns-6c9c4b4dfc-h6f2z\" (UID: \"f3537c77-1bb1-4595-92c4-a9298a5df3a9\") " pod="openstack/dnsmasq-dns-6c9c4b4dfc-h6f2z" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.623729 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77bf5858df-7s64q"] Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.625360 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77bf5858df-7s64q" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.627997 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.651006 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfndl\" (UniqueName: \"kubernetes.io/projected/f3537c77-1bb1-4595-92c4-a9298a5df3a9-kube-api-access-jfndl\") pod \"dnsmasq-dns-6c9c4b4dfc-h6f2z\" (UID: \"f3537c77-1bb1-4595-92c4-a9298a5df3a9\") " pod="openstack/dnsmasq-dns-6c9c4b4dfc-h6f2z" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.657369 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77bf5858df-7s64q"] Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.724592 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/42529236-56e6-400f-a34f-8d55dcff6a77-ovsdbserver-sb\") pod \"dnsmasq-dns-77bf5858df-7s64q\" (UID: \"42529236-56e6-400f-a34f-8d55dcff6a77\") " pod="openstack/dnsmasq-dns-77bf5858df-7s64q" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.724646 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/42529236-56e6-400f-a34f-8d55dcff6a77-dns-svc\") pod \"dnsmasq-dns-77bf5858df-7s64q\" (UID: \"42529236-56e6-400f-a34f-8d55dcff6a77\") " pod="openstack/dnsmasq-dns-77bf5858df-7s64q" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.725692 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/42529236-56e6-400f-a34f-8d55dcff6a77-ovsdbserver-nb\") pod \"dnsmasq-dns-77bf5858df-7s64q\" (UID: \"42529236-56e6-400f-a34f-8d55dcff6a77\") " pod="openstack/dnsmasq-dns-77bf5858df-7s64q" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.725808 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42529236-56e6-400f-a34f-8d55dcff6a77-config\") pod \"dnsmasq-dns-77bf5858df-7s64q\" (UID: \"42529236-56e6-400f-a34f-8d55dcff6a77\") " pod="openstack/dnsmasq-dns-77bf5858df-7s64q" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.725989 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2c4fl\" (UniqueName: \"kubernetes.io/projected/42529236-56e6-400f-a34f-8d55dcff6a77-kube-api-access-2c4fl\") pod \"dnsmasq-dns-77bf5858df-7s64q\" (UID: \"42529236-56e6-400f-a34f-8d55dcff6a77\") " pod="openstack/dnsmasq-dns-77bf5858df-7s64q" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.827652 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42529236-56e6-400f-a34f-8d55dcff6a77-config\") pod \"dnsmasq-dns-77bf5858df-7s64q\" (UID: \"42529236-56e6-400f-a34f-8d55dcff6a77\") " pod="openstack/dnsmasq-dns-77bf5858df-7s64q" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.827738 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2c4fl\" (UniqueName: \"kubernetes.io/projected/42529236-56e6-400f-a34f-8d55dcff6a77-kube-api-access-2c4fl\") pod \"dnsmasq-dns-77bf5858df-7s64q\" (UID: \"42529236-56e6-400f-a34f-8d55dcff6a77\") " pod="openstack/dnsmasq-dns-77bf5858df-7s64q" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.827781 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/42529236-56e6-400f-a34f-8d55dcff6a77-ovsdbserver-sb\") pod \"dnsmasq-dns-77bf5858df-7s64q\" (UID: \"42529236-56e6-400f-a34f-8d55dcff6a77\") " pod="openstack/dnsmasq-dns-77bf5858df-7s64q" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.827806 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/42529236-56e6-400f-a34f-8d55dcff6a77-dns-svc\") pod \"dnsmasq-dns-77bf5858df-7s64q\" (UID: \"42529236-56e6-400f-a34f-8d55dcff6a77\") " pod="openstack/dnsmasq-dns-77bf5858df-7s64q" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.827920 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/42529236-56e6-400f-a34f-8d55dcff6a77-ovsdbserver-nb\") pod \"dnsmasq-dns-77bf5858df-7s64q\" (UID: \"42529236-56e6-400f-a34f-8d55dcff6a77\") " pod="openstack/dnsmasq-dns-77bf5858df-7s64q" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.828625 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/42529236-56e6-400f-a34f-8d55dcff6a77-ovsdbserver-nb\") pod \"dnsmasq-dns-77bf5858df-7s64q\" (UID: \"42529236-56e6-400f-a34f-8d55dcff6a77\") " pod="openstack/dnsmasq-dns-77bf5858df-7s64q" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.828627 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42529236-56e6-400f-a34f-8d55dcff6a77-config\") pod \"dnsmasq-dns-77bf5858df-7s64q\" (UID: \"42529236-56e6-400f-a34f-8d55dcff6a77\") " pod="openstack/dnsmasq-dns-77bf5858df-7s64q" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.829234 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/42529236-56e6-400f-a34f-8d55dcff6a77-ovsdbserver-sb\") pod \"dnsmasq-dns-77bf5858df-7s64q\" (UID: \"42529236-56e6-400f-a34f-8d55dcff6a77\") " pod="openstack/dnsmasq-dns-77bf5858df-7s64q" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.829485 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/42529236-56e6-400f-a34f-8d55dcff6a77-dns-svc\") pod \"dnsmasq-dns-77bf5858df-7s64q\" (UID: \"42529236-56e6-400f-a34f-8d55dcff6a77\") " pod="openstack/dnsmasq-dns-77bf5858df-7s64q" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.848177 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2c4fl\" (UniqueName: \"kubernetes.io/projected/42529236-56e6-400f-a34f-8d55dcff6a77-kube-api-access-2c4fl\") pod \"dnsmasq-dns-77bf5858df-7s64q\" (UID: \"42529236-56e6-400f-a34f-8d55dcff6a77\") " pod="openstack/dnsmasq-dns-77bf5858df-7s64q" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.871032 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.880368 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-2" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.917152 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.930087 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-2" Nov 24 14:38:38 crc kubenswrapper[4790]: I1124 14:38:38.944545 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77bf5858df-7s64q" Nov 24 14:38:39 crc kubenswrapper[4790]: I1124 14:38:39.038532 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 24 14:38:39 crc kubenswrapper[4790]: I1124 14:38:39.090866 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 24 14:38:39 crc kubenswrapper[4790]: I1124 14:38:39.117407 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-1" Nov 24 14:38:39 crc kubenswrapper[4790]: I1124 14:38:39.240843 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c9c4b4dfc-h6f2z" Nov 24 14:38:39 crc kubenswrapper[4790]: I1124 14:38:39.258192 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c9c4b4dfc-h6f2z" Nov 24 14:38:39 crc kubenswrapper[4790]: I1124 14:38:39.280548 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-1" Nov 24 14:38:39 crc kubenswrapper[4790]: I1124 14:38:39.440781 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f3537c77-1bb1-4595-92c4-a9298a5df3a9-dns-svc\") pod \"f3537c77-1bb1-4595-92c4-a9298a5df3a9\" (UID: \"f3537c77-1bb1-4595-92c4-a9298a5df3a9\") " Nov 24 14:38:39 crc kubenswrapper[4790]: I1124 14:38:39.440929 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f3537c77-1bb1-4595-92c4-a9298a5df3a9-ovsdbserver-sb\") pod \"f3537c77-1bb1-4595-92c4-a9298a5df3a9\" (UID: \"f3537c77-1bb1-4595-92c4-a9298a5df3a9\") " Nov 24 14:38:39 crc kubenswrapper[4790]: I1124 14:38:39.441034 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jfndl\" (UniqueName: \"kubernetes.io/projected/f3537c77-1bb1-4595-92c4-a9298a5df3a9-kube-api-access-jfndl\") pod \"f3537c77-1bb1-4595-92c4-a9298a5df3a9\" (UID: \"f3537c77-1bb1-4595-92c4-a9298a5df3a9\") " Nov 24 14:38:39 crc kubenswrapper[4790]: I1124 14:38:39.441062 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3537c77-1bb1-4595-92c4-a9298a5df3a9-config\") pod \"f3537c77-1bb1-4595-92c4-a9298a5df3a9\" (UID: \"f3537c77-1bb1-4595-92c4-a9298a5df3a9\") " Nov 24 14:38:39 crc kubenswrapper[4790]: I1124 14:38:39.441544 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3537c77-1bb1-4595-92c4-a9298a5df3a9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f3537c77-1bb1-4595-92c4-a9298a5df3a9" (UID: "f3537c77-1bb1-4595-92c4-a9298a5df3a9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:38:39 crc kubenswrapper[4790]: I1124 14:38:39.441655 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3537c77-1bb1-4595-92c4-a9298a5df3a9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f3537c77-1bb1-4595-92c4-a9298a5df3a9" (UID: "f3537c77-1bb1-4595-92c4-a9298a5df3a9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:38:39 crc kubenswrapper[4790]: I1124 14:38:39.441852 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3537c77-1bb1-4595-92c4-a9298a5df3a9-config" (OuterVolumeSpecName: "config") pod "f3537c77-1bb1-4595-92c4-a9298a5df3a9" (UID: "f3537c77-1bb1-4595-92c4-a9298a5df3a9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:38:39 crc kubenswrapper[4790]: I1124 14:38:39.441960 4790 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f3537c77-1bb1-4595-92c4-a9298a5df3a9-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:39 crc kubenswrapper[4790]: I1124 14:38:39.441985 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f3537c77-1bb1-4595-92c4-a9298a5df3a9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:39 crc kubenswrapper[4790]: I1124 14:38:39.447491 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3537c77-1bb1-4595-92c4-a9298a5df3a9-kube-api-access-jfndl" (OuterVolumeSpecName: "kube-api-access-jfndl") pod "f3537c77-1bb1-4595-92c4-a9298a5df3a9" (UID: "f3537c77-1bb1-4595-92c4-a9298a5df3a9"). InnerVolumeSpecName "kube-api-access-jfndl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:38:39 crc kubenswrapper[4790]: I1124 14:38:39.466279 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77bf5858df-7s64q"] Nov 24 14:38:39 crc kubenswrapper[4790]: W1124 14:38:39.485441 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod42529236_56e6_400f_a34f_8d55dcff6a77.slice/crio-f9b31068df07659591dc5b5a0011a3e1ff68e3cb958378cb4ff2139e9c344990 WatchSource:0}: Error finding container f9b31068df07659591dc5b5a0011a3e1ff68e3cb958378cb4ff2139e9c344990: Status 404 returned error can't find the container with id f9b31068df07659591dc5b5a0011a3e1ff68e3cb958378cb4ff2139e9c344990 Nov 24 14:38:39 crc kubenswrapper[4790]: I1124 14:38:39.544336 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jfndl\" (UniqueName: \"kubernetes.io/projected/f3537c77-1bb1-4595-92c4-a9298a5df3a9-kube-api-access-jfndl\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:39 crc kubenswrapper[4790]: I1124 14:38:39.544380 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3537c77-1bb1-4595-92c4-a9298a5df3a9-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:40 crc kubenswrapper[4790]: I1124 14:38:40.257293 4790 generic.go:334] "Generic (PLEG): container finished" podID="42529236-56e6-400f-a34f-8d55dcff6a77" containerID="0f8c45aaa6207a956a3591838b8194dfe4de88d9247e7a138640cd4a140fb06c" exitCode=0 Nov 24 14:38:40 crc kubenswrapper[4790]: I1124 14:38:40.257407 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77bf5858df-7s64q" event={"ID":"42529236-56e6-400f-a34f-8d55dcff6a77","Type":"ContainerDied","Data":"0f8c45aaa6207a956a3591838b8194dfe4de88d9247e7a138640cd4a140fb06c"} Nov 24 14:38:40 crc kubenswrapper[4790]: I1124 14:38:40.257553 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77bf5858df-7s64q" event={"ID":"42529236-56e6-400f-a34f-8d55dcff6a77","Type":"ContainerStarted","Data":"f9b31068df07659591dc5b5a0011a3e1ff68e3cb958378cb4ff2139e9c344990"} Nov 24 14:38:40 crc kubenswrapper[4790]: I1124 14:38:40.257574 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c9c4b4dfc-h6f2z" Nov 24 14:38:41 crc kubenswrapper[4790]: I1124 14:38:41.274059 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77bf5858df-7s64q" event={"ID":"42529236-56e6-400f-a34f-8d55dcff6a77","Type":"ContainerStarted","Data":"a24444ba4dbbc1ab8e1990bd03a669237261e7dd92debd2bc53b97198fc0f7e7"} Nov 24 14:38:41 crc kubenswrapper[4790]: I1124 14:38:41.274686 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77bf5858df-7s64q" Nov 24 14:38:41 crc kubenswrapper[4790]: I1124 14:38:41.314952 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77bf5858df-7s64q" podStartSLOduration=3.314920036 podStartE2EDuration="3.314920036s" podCreationTimestamp="2025-11-24 14:38:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:38:41.309930057 +0000 UTC m=+5169.689823759" watchObservedRunningTime="2025-11-24 14:38:41.314920036 +0000 UTC m=+5169.694813738" Nov 24 14:38:41 crc kubenswrapper[4790]: I1124 14:38:41.884789 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-copy-data"] Nov 24 14:38:41 crc kubenswrapper[4790]: I1124 14:38:41.887027 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Nov 24 14:38:41 crc kubenswrapper[4790]: I1124 14:38:41.890852 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovn-data-cert" Nov 24 14:38:41 crc kubenswrapper[4790]: I1124 14:38:41.897857 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Nov 24 14:38:42 crc kubenswrapper[4790]: I1124 14:38:42.001171 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/67fabd6a-992b-49ab-b90f-98fddba09ce0-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"67fabd6a-992b-49ab-b90f-98fddba09ce0\") " pod="openstack/ovn-copy-data" Nov 24 14:38:42 crc kubenswrapper[4790]: I1124 14:38:42.001319 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-012d64e2-c94b-454b-b5bb-6a6a047812a5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-012d64e2-c94b-454b-b5bb-6a6a047812a5\") pod \"ovn-copy-data\" (UID: \"67fabd6a-992b-49ab-b90f-98fddba09ce0\") " pod="openstack/ovn-copy-data" Nov 24 14:38:42 crc kubenswrapper[4790]: I1124 14:38:42.001389 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmmp9\" (UniqueName: \"kubernetes.io/projected/67fabd6a-992b-49ab-b90f-98fddba09ce0-kube-api-access-bmmp9\") pod \"ovn-copy-data\" (UID: \"67fabd6a-992b-49ab-b90f-98fddba09ce0\") " pod="openstack/ovn-copy-data" Nov 24 14:38:42 crc kubenswrapper[4790]: I1124 14:38:42.103419 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/67fabd6a-992b-49ab-b90f-98fddba09ce0-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"67fabd6a-992b-49ab-b90f-98fddba09ce0\") " pod="openstack/ovn-copy-data" Nov 24 14:38:42 crc kubenswrapper[4790]: I1124 14:38:42.103593 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-012d64e2-c94b-454b-b5bb-6a6a047812a5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-012d64e2-c94b-454b-b5bb-6a6a047812a5\") pod \"ovn-copy-data\" (UID: \"67fabd6a-992b-49ab-b90f-98fddba09ce0\") " pod="openstack/ovn-copy-data" Nov 24 14:38:42 crc kubenswrapper[4790]: I1124 14:38:42.103689 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmmp9\" (UniqueName: \"kubernetes.io/projected/67fabd6a-992b-49ab-b90f-98fddba09ce0-kube-api-access-bmmp9\") pod \"ovn-copy-data\" (UID: \"67fabd6a-992b-49ab-b90f-98fddba09ce0\") " pod="openstack/ovn-copy-data" Nov 24 14:38:42 crc kubenswrapper[4790]: I1124 14:38:42.108399 4790 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:38:42 crc kubenswrapper[4790]: I1124 14:38:42.108719 4790 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-012d64e2-c94b-454b-b5bb-6a6a047812a5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-012d64e2-c94b-454b-b5bb-6a6a047812a5\") pod \"ovn-copy-data\" (UID: \"67fabd6a-992b-49ab-b90f-98fddba09ce0\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/40909fcf81baff3357f1aa0ebaf5326b986e4f8d8299f4569675b30997fbd6b2/globalmount\"" pod="openstack/ovn-copy-data" Nov 24 14:38:42 crc kubenswrapper[4790]: I1124 14:38:42.111592 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/67fabd6a-992b-49ab-b90f-98fddba09ce0-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"67fabd6a-992b-49ab-b90f-98fddba09ce0\") " pod="openstack/ovn-copy-data" Nov 24 14:38:42 crc kubenswrapper[4790]: I1124 14:38:42.136733 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmmp9\" (UniqueName: \"kubernetes.io/projected/67fabd6a-992b-49ab-b90f-98fddba09ce0-kube-api-access-bmmp9\") pod \"ovn-copy-data\" (UID: \"67fabd6a-992b-49ab-b90f-98fddba09ce0\") " pod="openstack/ovn-copy-data" Nov 24 14:38:42 crc kubenswrapper[4790]: I1124 14:38:42.153687 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-012d64e2-c94b-454b-b5bb-6a6a047812a5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-012d64e2-c94b-454b-b5bb-6a6a047812a5\") pod \"ovn-copy-data\" (UID: \"67fabd6a-992b-49ab-b90f-98fddba09ce0\") " pod="openstack/ovn-copy-data" Nov 24 14:38:42 crc kubenswrapper[4790]: I1124 14:38:42.216756 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Nov 24 14:38:42 crc kubenswrapper[4790]: I1124 14:38:42.775116 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Nov 24 14:38:42 crc kubenswrapper[4790]: W1124 14:38:42.781660 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67fabd6a_992b_49ab_b90f_98fddba09ce0.slice/crio-6064a7e91df03fe9ebf63359ffb45c3b8aafeb51c330ddea8d62ad2a96147db8 WatchSource:0}: Error finding container 6064a7e91df03fe9ebf63359ffb45c3b8aafeb51c330ddea8d62ad2a96147db8: Status 404 returned error can't find the container with id 6064a7e91df03fe9ebf63359ffb45c3b8aafeb51c330ddea8d62ad2a96147db8 Nov 24 14:38:43 crc kubenswrapper[4790]: I1124 14:38:43.304324 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"67fabd6a-992b-49ab-b90f-98fddba09ce0","Type":"ContainerStarted","Data":"ec835f53de660a2c046a34875abb04980f7575d9c55dcb8fc367cafb608533c0"} Nov 24 14:38:43 crc kubenswrapper[4790]: I1124 14:38:43.304942 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"67fabd6a-992b-49ab-b90f-98fddba09ce0","Type":"ContainerStarted","Data":"6064a7e91df03fe9ebf63359ffb45c3b8aafeb51c330ddea8d62ad2a96147db8"} Nov 24 14:38:43 crc kubenswrapper[4790]: I1124 14:38:43.338369 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-copy-data" podStartSLOduration=3.338336649 podStartE2EDuration="3.338336649s" podCreationTimestamp="2025-11-24 14:38:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:38:43.326050662 +0000 UTC m=+5171.705944384" watchObservedRunningTime="2025-11-24 14:38:43.338336649 +0000 UTC m=+5171.718230351" Nov 24 14:38:48 crc kubenswrapper[4790]: I1124 14:38:48.600416 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 24 14:38:48 crc kubenswrapper[4790]: I1124 14:38:48.604345 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 14:38:48 crc kubenswrapper[4790]: I1124 14:38:48.607351 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 24 14:38:48 crc kubenswrapper[4790]: I1124 14:38:48.613748 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 24 14:38:48 crc kubenswrapper[4790]: I1124 14:38:48.617734 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-kc678" Nov 24 14:38:48 crc kubenswrapper[4790]: I1124 14:38:48.626233 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 14:38:48 crc kubenswrapper[4790]: I1124 14:38:48.754375 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5364dcd2-a902-4a92-a96c-a7ff91d0b50e-scripts\") pod \"ovn-northd-0\" (UID: \"5364dcd2-a902-4a92-a96c-a7ff91d0b50e\") " pod="openstack/ovn-northd-0" Nov 24 14:38:48 crc kubenswrapper[4790]: I1124 14:38:48.754787 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbzgp\" (UniqueName: \"kubernetes.io/projected/5364dcd2-a902-4a92-a96c-a7ff91d0b50e-kube-api-access-pbzgp\") pod \"ovn-northd-0\" (UID: \"5364dcd2-a902-4a92-a96c-a7ff91d0b50e\") " pod="openstack/ovn-northd-0" Nov 24 14:38:48 crc kubenswrapper[4790]: I1124 14:38:48.754825 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5364dcd2-a902-4a92-a96c-a7ff91d0b50e-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"5364dcd2-a902-4a92-a96c-a7ff91d0b50e\") " pod="openstack/ovn-northd-0" Nov 24 14:38:48 crc kubenswrapper[4790]: I1124 14:38:48.754870 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5364dcd2-a902-4a92-a96c-a7ff91d0b50e-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"5364dcd2-a902-4a92-a96c-a7ff91d0b50e\") " pod="openstack/ovn-northd-0" Nov 24 14:38:48 crc kubenswrapper[4790]: I1124 14:38:48.754918 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5364dcd2-a902-4a92-a96c-a7ff91d0b50e-config\") pod \"ovn-northd-0\" (UID: \"5364dcd2-a902-4a92-a96c-a7ff91d0b50e\") " pod="openstack/ovn-northd-0" Nov 24 14:38:48 crc kubenswrapper[4790]: I1124 14:38:48.857308 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5364dcd2-a902-4a92-a96c-a7ff91d0b50e-scripts\") pod \"ovn-northd-0\" (UID: \"5364dcd2-a902-4a92-a96c-a7ff91d0b50e\") " pod="openstack/ovn-northd-0" Nov 24 14:38:48 crc kubenswrapper[4790]: I1124 14:38:48.857378 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbzgp\" (UniqueName: \"kubernetes.io/projected/5364dcd2-a902-4a92-a96c-a7ff91d0b50e-kube-api-access-pbzgp\") pod \"ovn-northd-0\" (UID: \"5364dcd2-a902-4a92-a96c-a7ff91d0b50e\") " pod="openstack/ovn-northd-0" Nov 24 14:38:48 crc kubenswrapper[4790]: I1124 14:38:48.857417 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5364dcd2-a902-4a92-a96c-a7ff91d0b50e-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"5364dcd2-a902-4a92-a96c-a7ff91d0b50e\") " pod="openstack/ovn-northd-0" Nov 24 14:38:48 crc kubenswrapper[4790]: I1124 14:38:48.857462 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5364dcd2-a902-4a92-a96c-a7ff91d0b50e-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"5364dcd2-a902-4a92-a96c-a7ff91d0b50e\") " pod="openstack/ovn-northd-0" Nov 24 14:38:48 crc kubenswrapper[4790]: I1124 14:38:48.857501 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5364dcd2-a902-4a92-a96c-a7ff91d0b50e-config\") pod \"ovn-northd-0\" (UID: \"5364dcd2-a902-4a92-a96c-a7ff91d0b50e\") " pod="openstack/ovn-northd-0" Nov 24 14:38:48 crc kubenswrapper[4790]: I1124 14:38:48.858684 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5364dcd2-a902-4a92-a96c-a7ff91d0b50e-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"5364dcd2-a902-4a92-a96c-a7ff91d0b50e\") " pod="openstack/ovn-northd-0" Nov 24 14:38:48 crc kubenswrapper[4790]: I1124 14:38:48.858998 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5364dcd2-a902-4a92-a96c-a7ff91d0b50e-config\") pod \"ovn-northd-0\" (UID: \"5364dcd2-a902-4a92-a96c-a7ff91d0b50e\") " pod="openstack/ovn-northd-0" Nov 24 14:38:48 crc kubenswrapper[4790]: I1124 14:38:48.859002 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5364dcd2-a902-4a92-a96c-a7ff91d0b50e-scripts\") pod \"ovn-northd-0\" (UID: \"5364dcd2-a902-4a92-a96c-a7ff91d0b50e\") " pod="openstack/ovn-northd-0" Nov 24 14:38:48 crc kubenswrapper[4790]: I1124 14:38:48.867621 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5364dcd2-a902-4a92-a96c-a7ff91d0b50e-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"5364dcd2-a902-4a92-a96c-a7ff91d0b50e\") " pod="openstack/ovn-northd-0" Nov 24 14:38:48 crc kubenswrapper[4790]: I1124 14:38:48.881128 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbzgp\" (UniqueName: \"kubernetes.io/projected/5364dcd2-a902-4a92-a96c-a7ff91d0b50e-kube-api-access-pbzgp\") pod \"ovn-northd-0\" (UID: \"5364dcd2-a902-4a92-a96c-a7ff91d0b50e\") " pod="openstack/ovn-northd-0" Nov 24 14:38:48 crc kubenswrapper[4790]: I1124 14:38:48.934036 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 14:38:48 crc kubenswrapper[4790]: I1124 14:38:48.949106 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77bf5858df-7s64q" Nov 24 14:38:49 crc kubenswrapper[4790]: I1124 14:38:49.018989 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-wqrxr"] Nov 24 14:38:49 crc kubenswrapper[4790]: I1124 14:38:49.019309 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b7946d7b9-wqrxr" podUID="ecff746c-6546-41be-bee5-c07166f192dc" containerName="dnsmasq-dns" containerID="cri-o://fbc6287eaed7e81d9101554856b82bca2c140175830a008c338b860df97ff048" gracePeriod=10 Nov 24 14:38:49 crc kubenswrapper[4790]: I1124 14:38:49.372821 4790 generic.go:334] "Generic (PLEG): container finished" podID="ecff746c-6546-41be-bee5-c07166f192dc" containerID="fbc6287eaed7e81d9101554856b82bca2c140175830a008c338b860df97ff048" exitCode=0 Nov 24 14:38:49 crc kubenswrapper[4790]: I1124 14:38:49.372926 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-wqrxr" event={"ID":"ecff746c-6546-41be-bee5-c07166f192dc","Type":"ContainerDied","Data":"fbc6287eaed7e81d9101554856b82bca2c140175830a008c338b860df97ff048"} Nov 24 14:38:49 crc kubenswrapper[4790]: I1124 14:38:49.407361 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-wqrxr" Nov 24 14:38:49 crc kubenswrapper[4790]: I1124 14:38:49.462078 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 14:38:49 crc kubenswrapper[4790]: I1124 14:38:49.580276 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xm4c\" (UniqueName: \"kubernetes.io/projected/ecff746c-6546-41be-bee5-c07166f192dc-kube-api-access-2xm4c\") pod \"ecff746c-6546-41be-bee5-c07166f192dc\" (UID: \"ecff746c-6546-41be-bee5-c07166f192dc\") " Nov 24 14:38:49 crc kubenswrapper[4790]: I1124 14:38:49.582347 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ecff746c-6546-41be-bee5-c07166f192dc-dns-svc\") pod \"ecff746c-6546-41be-bee5-c07166f192dc\" (UID: \"ecff746c-6546-41be-bee5-c07166f192dc\") " Nov 24 14:38:49 crc kubenswrapper[4790]: I1124 14:38:49.582633 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecff746c-6546-41be-bee5-c07166f192dc-config\") pod \"ecff746c-6546-41be-bee5-c07166f192dc\" (UID: \"ecff746c-6546-41be-bee5-c07166f192dc\") " Nov 24 14:38:49 crc kubenswrapper[4790]: I1124 14:38:49.633605 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecff746c-6546-41be-bee5-c07166f192dc-kube-api-access-2xm4c" (OuterVolumeSpecName: "kube-api-access-2xm4c") pod "ecff746c-6546-41be-bee5-c07166f192dc" (UID: "ecff746c-6546-41be-bee5-c07166f192dc"). InnerVolumeSpecName "kube-api-access-2xm4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:38:49 crc kubenswrapper[4790]: I1124 14:38:49.683238 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecff746c-6546-41be-bee5-c07166f192dc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ecff746c-6546-41be-bee5-c07166f192dc" (UID: "ecff746c-6546-41be-bee5-c07166f192dc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:38:49 crc kubenswrapper[4790]: I1124 14:38:49.690473 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecff746c-6546-41be-bee5-c07166f192dc-config" (OuterVolumeSpecName: "config") pod "ecff746c-6546-41be-bee5-c07166f192dc" (UID: "ecff746c-6546-41be-bee5-c07166f192dc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:38:49 crc kubenswrapper[4790]: I1124 14:38:49.691310 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xm4c\" (UniqueName: \"kubernetes.io/projected/ecff746c-6546-41be-bee5-c07166f192dc-kube-api-access-2xm4c\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:49 crc kubenswrapper[4790]: I1124 14:38:49.691343 4790 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ecff746c-6546-41be-bee5-c07166f192dc-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:49 crc kubenswrapper[4790]: I1124 14:38:49.691353 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecff746c-6546-41be-bee5-c07166f192dc-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:50 crc kubenswrapper[4790]: I1124 14:38:50.389796 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-wqrxr" event={"ID":"ecff746c-6546-41be-bee5-c07166f192dc","Type":"ContainerDied","Data":"08741f67c2124c4017b63d958a30ea6ec29c9fad754febdac567a16e7806b4c8"} Nov 24 14:38:50 crc kubenswrapper[4790]: I1124 14:38:50.389927 4790 scope.go:117] "RemoveContainer" containerID="fbc6287eaed7e81d9101554856b82bca2c140175830a008c338b860df97ff048" Nov 24 14:38:50 crc kubenswrapper[4790]: I1124 14:38:50.389945 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-wqrxr" Nov 24 14:38:50 crc kubenswrapper[4790]: I1124 14:38:50.396244 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"5364dcd2-a902-4a92-a96c-a7ff91d0b50e","Type":"ContainerStarted","Data":"ff65043ec418c41a74147df433e0c0449d2e79d4cbb354fcb6ff9e2e91c497a3"} Nov 24 14:38:50 crc kubenswrapper[4790]: I1124 14:38:50.396338 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"5364dcd2-a902-4a92-a96c-a7ff91d0b50e","Type":"ContainerStarted","Data":"b517033322981336f1c665fb1e91a154713d953bb7334b91653f941e7d311756"} Nov 24 14:38:50 crc kubenswrapper[4790]: I1124 14:38:50.396375 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"5364dcd2-a902-4a92-a96c-a7ff91d0b50e","Type":"ContainerStarted","Data":"37a2bbca939f01852cf92c69f51738ea0bbbef727d08d78859b9fd3b003cbde9"} Nov 24 14:38:50 crc kubenswrapper[4790]: I1124 14:38:50.396711 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 24 14:38:50 crc kubenswrapper[4790]: I1124 14:38:50.426349 4790 scope.go:117] "RemoveContainer" containerID="1dcdc3f4c7a2479d6bd089ba9359c7e2bebafaa5f2620c830395377a0db42989" Nov 24 14:38:50 crc kubenswrapper[4790]: I1124 14:38:50.435505 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.435470278 podStartE2EDuration="2.435470278s" podCreationTimestamp="2025-11-24 14:38:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:38:50.427397237 +0000 UTC m=+5178.807290939" watchObservedRunningTime="2025-11-24 14:38:50.435470278 +0000 UTC m=+5178.815363970" Nov 24 14:38:50 crc kubenswrapper[4790]: I1124 14:38:50.458806 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-wqrxr"] Nov 24 14:38:50 crc kubenswrapper[4790]: I1124 14:38:50.466065 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-wqrxr"] Nov 24 14:38:52 crc kubenswrapper[4790]: I1124 14:38:52.330398 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecff746c-6546-41be-bee5-c07166f192dc" path="/var/lib/kubelet/pods/ecff746c-6546-41be-bee5-c07166f192dc/volumes" Nov 24 14:38:54 crc kubenswrapper[4790]: I1124 14:38:54.113496 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-grlct"] Nov 24 14:38:54 crc kubenswrapper[4790]: E1124 14:38:54.114742 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecff746c-6546-41be-bee5-c07166f192dc" containerName="init" Nov 24 14:38:54 crc kubenswrapper[4790]: I1124 14:38:54.114756 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecff746c-6546-41be-bee5-c07166f192dc" containerName="init" Nov 24 14:38:54 crc kubenswrapper[4790]: E1124 14:38:54.114788 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecff746c-6546-41be-bee5-c07166f192dc" containerName="dnsmasq-dns" Nov 24 14:38:54 crc kubenswrapper[4790]: I1124 14:38:54.114795 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecff746c-6546-41be-bee5-c07166f192dc" containerName="dnsmasq-dns" Nov 24 14:38:54 crc kubenswrapper[4790]: I1124 14:38:54.115145 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecff746c-6546-41be-bee5-c07166f192dc" containerName="dnsmasq-dns" Nov 24 14:38:54 crc kubenswrapper[4790]: I1124 14:38:54.116023 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-grlct" Nov 24 14:38:54 crc kubenswrapper[4790]: I1124 14:38:54.143472 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-grlct"] Nov 24 14:38:54 crc kubenswrapper[4790]: I1124 14:38:54.207769 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-8ec9-account-create-hzpw7"] Nov 24 14:38:54 crc kubenswrapper[4790]: I1124 14:38:54.209451 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8ec9-account-create-hzpw7" Nov 24 14:38:54 crc kubenswrapper[4790]: I1124 14:38:54.212484 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 24 14:38:54 crc kubenswrapper[4790]: I1124 14:38:54.221547 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8ec9-account-create-hzpw7"] Nov 24 14:38:54 crc kubenswrapper[4790]: I1124 14:38:54.292217 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fl7qq\" (UniqueName: \"kubernetes.io/projected/be69b02c-42ef-437f-8ff3-47973dc25ccb-kube-api-access-fl7qq\") pod \"keystone-db-create-grlct\" (UID: \"be69b02c-42ef-437f-8ff3-47973dc25ccb\") " pod="openstack/keystone-db-create-grlct" Nov 24 14:38:54 crc kubenswrapper[4790]: I1124 14:38:54.292437 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be69b02c-42ef-437f-8ff3-47973dc25ccb-operator-scripts\") pod \"keystone-db-create-grlct\" (UID: \"be69b02c-42ef-437f-8ff3-47973dc25ccb\") " pod="openstack/keystone-db-create-grlct" Nov 24 14:38:54 crc kubenswrapper[4790]: I1124 14:38:54.394607 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnxwz\" (UniqueName: \"kubernetes.io/projected/c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44-kube-api-access-dnxwz\") pod \"keystone-8ec9-account-create-hzpw7\" (UID: \"c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44\") " pod="openstack/keystone-8ec9-account-create-hzpw7" Nov 24 14:38:54 crc kubenswrapper[4790]: I1124 14:38:54.394770 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44-operator-scripts\") pod \"keystone-8ec9-account-create-hzpw7\" (UID: \"c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44\") " pod="openstack/keystone-8ec9-account-create-hzpw7" Nov 24 14:38:54 crc kubenswrapper[4790]: I1124 14:38:54.394850 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fl7qq\" (UniqueName: \"kubernetes.io/projected/be69b02c-42ef-437f-8ff3-47973dc25ccb-kube-api-access-fl7qq\") pod \"keystone-db-create-grlct\" (UID: \"be69b02c-42ef-437f-8ff3-47973dc25ccb\") " pod="openstack/keystone-db-create-grlct" Nov 24 14:38:54 crc kubenswrapper[4790]: I1124 14:38:54.394902 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be69b02c-42ef-437f-8ff3-47973dc25ccb-operator-scripts\") pod \"keystone-db-create-grlct\" (UID: \"be69b02c-42ef-437f-8ff3-47973dc25ccb\") " pod="openstack/keystone-db-create-grlct" Nov 24 14:38:54 crc kubenswrapper[4790]: I1124 14:38:54.395694 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be69b02c-42ef-437f-8ff3-47973dc25ccb-operator-scripts\") pod \"keystone-db-create-grlct\" (UID: \"be69b02c-42ef-437f-8ff3-47973dc25ccb\") " pod="openstack/keystone-db-create-grlct" Nov 24 14:38:54 crc kubenswrapper[4790]: I1124 14:38:54.426934 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fl7qq\" (UniqueName: \"kubernetes.io/projected/be69b02c-42ef-437f-8ff3-47973dc25ccb-kube-api-access-fl7qq\") pod \"keystone-db-create-grlct\" (UID: \"be69b02c-42ef-437f-8ff3-47973dc25ccb\") " pod="openstack/keystone-db-create-grlct" Nov 24 14:38:54 crc kubenswrapper[4790]: I1124 14:38:54.445606 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-grlct" Nov 24 14:38:54 crc kubenswrapper[4790]: I1124 14:38:54.496980 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnxwz\" (UniqueName: \"kubernetes.io/projected/c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44-kube-api-access-dnxwz\") pod \"keystone-8ec9-account-create-hzpw7\" (UID: \"c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44\") " pod="openstack/keystone-8ec9-account-create-hzpw7" Nov 24 14:38:54 crc kubenswrapper[4790]: I1124 14:38:54.497089 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44-operator-scripts\") pod \"keystone-8ec9-account-create-hzpw7\" (UID: \"c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44\") " pod="openstack/keystone-8ec9-account-create-hzpw7" Nov 24 14:38:54 crc kubenswrapper[4790]: I1124 14:38:54.497940 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44-operator-scripts\") pod \"keystone-8ec9-account-create-hzpw7\" (UID: \"c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44\") " pod="openstack/keystone-8ec9-account-create-hzpw7" Nov 24 14:38:54 crc kubenswrapper[4790]: I1124 14:38:54.515684 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnxwz\" (UniqueName: \"kubernetes.io/projected/c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44-kube-api-access-dnxwz\") pod \"keystone-8ec9-account-create-hzpw7\" (UID: \"c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44\") " pod="openstack/keystone-8ec9-account-create-hzpw7" Nov 24 14:38:54 crc kubenswrapper[4790]: I1124 14:38:54.531047 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8ec9-account-create-hzpw7" Nov 24 14:38:54 crc kubenswrapper[4790]: I1124 14:38:54.917951 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-grlct"] Nov 24 14:38:54 crc kubenswrapper[4790]: W1124 14:38:54.921048 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbe69b02c_42ef_437f_8ff3_47973dc25ccb.slice/crio-45c15af53857d529c936c02d6e802c6f745a0174de3a4cce8e5385df8bba0836 WatchSource:0}: Error finding container 45c15af53857d529c936c02d6e802c6f745a0174de3a4cce8e5385df8bba0836: Status 404 returned error can't find the container with id 45c15af53857d529c936c02d6e802c6f745a0174de3a4cce8e5385df8bba0836 Nov 24 14:38:55 crc kubenswrapper[4790]: W1124 14:38:55.099263 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc939cf2c_81e7_4f9a_bb5b_08f04f6c1f44.slice/crio-211c7da672916163fe4b4ded3d698f394036229895c2d1ec49eabf861e1c62fd WatchSource:0}: Error finding container 211c7da672916163fe4b4ded3d698f394036229895c2d1ec49eabf861e1c62fd: Status 404 returned error can't find the container with id 211c7da672916163fe4b4ded3d698f394036229895c2d1ec49eabf861e1c62fd Nov 24 14:38:55 crc kubenswrapper[4790]: I1124 14:38:55.099854 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8ec9-account-create-hzpw7"] Nov 24 14:38:55 crc kubenswrapper[4790]: I1124 14:38:55.455266 4790 generic.go:334] "Generic (PLEG): container finished" podID="be69b02c-42ef-437f-8ff3-47973dc25ccb" containerID="e6f5cad26b70c28d261270b94a0b9ce4f97f19ee078677d537320e82e304e761" exitCode=0 Nov 24 14:38:55 crc kubenswrapper[4790]: I1124 14:38:55.455348 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-grlct" event={"ID":"be69b02c-42ef-437f-8ff3-47973dc25ccb","Type":"ContainerDied","Data":"e6f5cad26b70c28d261270b94a0b9ce4f97f19ee078677d537320e82e304e761"} Nov 24 14:38:55 crc kubenswrapper[4790]: I1124 14:38:55.455816 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-grlct" event={"ID":"be69b02c-42ef-437f-8ff3-47973dc25ccb","Type":"ContainerStarted","Data":"45c15af53857d529c936c02d6e802c6f745a0174de3a4cce8e5385df8bba0836"} Nov 24 14:38:55 crc kubenswrapper[4790]: I1124 14:38:55.459592 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8ec9-account-create-hzpw7" event={"ID":"c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44","Type":"ContainerStarted","Data":"ccb07244be050d105ef0c39518348f82fbaa508a6212856a2de14a89f191b49b"} Nov 24 14:38:55 crc kubenswrapper[4790]: I1124 14:38:55.459678 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8ec9-account-create-hzpw7" event={"ID":"c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44","Type":"ContainerStarted","Data":"211c7da672916163fe4b4ded3d698f394036229895c2d1ec49eabf861e1c62fd"} Nov 24 14:38:55 crc kubenswrapper[4790]: I1124 14:38:55.510325 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-8ec9-account-create-hzpw7" podStartSLOduration=1.510296487 podStartE2EDuration="1.510296487s" podCreationTimestamp="2025-11-24 14:38:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:38:55.496981939 +0000 UTC m=+5183.876875641" watchObservedRunningTime="2025-11-24 14:38:55.510296487 +0000 UTC m=+5183.890190189" Nov 24 14:38:56 crc kubenswrapper[4790]: I1124 14:38:56.476452 4790 generic.go:334] "Generic (PLEG): container finished" podID="c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44" containerID="ccb07244be050d105ef0c39518348f82fbaa508a6212856a2de14a89f191b49b" exitCode=0 Nov 24 14:38:56 crc kubenswrapper[4790]: I1124 14:38:56.476588 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8ec9-account-create-hzpw7" event={"ID":"c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44","Type":"ContainerDied","Data":"ccb07244be050d105ef0c39518348f82fbaa508a6212856a2de14a89f191b49b"} Nov 24 14:38:56 crc kubenswrapper[4790]: I1124 14:38:56.908824 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-grlct" Nov 24 14:38:56 crc kubenswrapper[4790]: I1124 14:38:56.955498 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fl7qq\" (UniqueName: \"kubernetes.io/projected/be69b02c-42ef-437f-8ff3-47973dc25ccb-kube-api-access-fl7qq\") pod \"be69b02c-42ef-437f-8ff3-47973dc25ccb\" (UID: \"be69b02c-42ef-437f-8ff3-47973dc25ccb\") " Nov 24 14:38:56 crc kubenswrapper[4790]: I1124 14:38:56.955716 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be69b02c-42ef-437f-8ff3-47973dc25ccb-operator-scripts\") pod \"be69b02c-42ef-437f-8ff3-47973dc25ccb\" (UID: \"be69b02c-42ef-437f-8ff3-47973dc25ccb\") " Nov 24 14:38:56 crc kubenswrapper[4790]: I1124 14:38:56.957086 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be69b02c-42ef-437f-8ff3-47973dc25ccb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "be69b02c-42ef-437f-8ff3-47973dc25ccb" (UID: "be69b02c-42ef-437f-8ff3-47973dc25ccb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:38:56 crc kubenswrapper[4790]: I1124 14:38:56.968490 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be69b02c-42ef-437f-8ff3-47973dc25ccb-kube-api-access-fl7qq" (OuterVolumeSpecName: "kube-api-access-fl7qq") pod "be69b02c-42ef-437f-8ff3-47973dc25ccb" (UID: "be69b02c-42ef-437f-8ff3-47973dc25ccb"). InnerVolumeSpecName "kube-api-access-fl7qq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:38:57 crc kubenswrapper[4790]: I1124 14:38:57.059529 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fl7qq\" (UniqueName: \"kubernetes.io/projected/be69b02c-42ef-437f-8ff3-47973dc25ccb-kube-api-access-fl7qq\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:57 crc kubenswrapper[4790]: I1124 14:38:57.059616 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be69b02c-42ef-437f-8ff3-47973dc25ccb-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:57 crc kubenswrapper[4790]: I1124 14:38:57.492091 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-grlct" event={"ID":"be69b02c-42ef-437f-8ff3-47973dc25ccb","Type":"ContainerDied","Data":"45c15af53857d529c936c02d6e802c6f745a0174de3a4cce8e5385df8bba0836"} Nov 24 14:38:57 crc kubenswrapper[4790]: I1124 14:38:57.492588 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45c15af53857d529c936c02d6e802c6f745a0174de3a4cce8e5385df8bba0836" Nov 24 14:38:57 crc kubenswrapper[4790]: I1124 14:38:57.492146 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-grlct" Nov 24 14:38:57 crc kubenswrapper[4790]: I1124 14:38:57.988865 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8ec9-account-create-hzpw7" Nov 24 14:38:58 crc kubenswrapper[4790]: I1124 14:38:58.084379 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44-operator-scripts\") pod \"c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44\" (UID: \"c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44\") " Nov 24 14:38:58 crc kubenswrapper[4790]: I1124 14:38:58.084560 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnxwz\" (UniqueName: \"kubernetes.io/projected/c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44-kube-api-access-dnxwz\") pod \"c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44\" (UID: \"c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44\") " Nov 24 14:38:58 crc kubenswrapper[4790]: I1124 14:38:58.085552 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44" (UID: "c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:38:58 crc kubenswrapper[4790]: I1124 14:38:58.086058 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:58 crc kubenswrapper[4790]: I1124 14:38:58.092845 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44-kube-api-access-dnxwz" (OuterVolumeSpecName: "kube-api-access-dnxwz") pod "c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44" (UID: "c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44"). InnerVolumeSpecName "kube-api-access-dnxwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:38:58 crc kubenswrapper[4790]: I1124 14:38:58.186797 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnxwz\" (UniqueName: \"kubernetes.io/projected/c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44-kube-api-access-dnxwz\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:58 crc kubenswrapper[4790]: I1124 14:38:58.502517 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8ec9-account-create-hzpw7" event={"ID":"c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44","Type":"ContainerDied","Data":"211c7da672916163fe4b4ded3d698f394036229895c2d1ec49eabf861e1c62fd"} Nov 24 14:38:58 crc kubenswrapper[4790]: I1124 14:38:58.502580 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8ec9-account-create-hzpw7" Nov 24 14:38:58 crc kubenswrapper[4790]: I1124 14:38:58.502583 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="211c7da672916163fe4b4ded3d698f394036229895c2d1ec49eabf861e1c62fd" Nov 24 14:38:59 crc kubenswrapper[4790]: I1124 14:38:59.867039 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-qmm47"] Nov 24 14:38:59 crc kubenswrapper[4790]: E1124 14:38:59.867483 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44" containerName="mariadb-account-create" Nov 24 14:38:59 crc kubenswrapper[4790]: I1124 14:38:59.867503 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44" containerName="mariadb-account-create" Nov 24 14:38:59 crc kubenswrapper[4790]: E1124 14:38:59.867529 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be69b02c-42ef-437f-8ff3-47973dc25ccb" containerName="mariadb-database-create" Nov 24 14:38:59 crc kubenswrapper[4790]: I1124 14:38:59.867538 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="be69b02c-42ef-437f-8ff3-47973dc25ccb" containerName="mariadb-database-create" Nov 24 14:38:59 crc kubenswrapper[4790]: I1124 14:38:59.867751 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44" containerName="mariadb-account-create" Nov 24 14:38:59 crc kubenswrapper[4790]: I1124 14:38:59.867772 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="be69b02c-42ef-437f-8ff3-47973dc25ccb" containerName="mariadb-database-create" Nov 24 14:38:59 crc kubenswrapper[4790]: I1124 14:38:59.868569 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-qmm47" Nov 24 14:38:59 crc kubenswrapper[4790]: I1124 14:38:59.876353 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 14:38:59 crc kubenswrapper[4790]: I1124 14:38:59.876363 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 14:38:59 crc kubenswrapper[4790]: I1124 14:38:59.876926 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 14:38:59 crc kubenswrapper[4790]: I1124 14:38:59.881537 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-58t98" Nov 24 14:38:59 crc kubenswrapper[4790]: I1124 14:38:59.888348 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-qmm47"] Nov 24 14:38:59 crc kubenswrapper[4790]: I1124 14:38:59.925764 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42e1cf5d-d1e0-4a6c-8607-f37dd401f552-combined-ca-bundle\") pod \"keystone-db-sync-qmm47\" (UID: \"42e1cf5d-d1e0-4a6c-8607-f37dd401f552\") " pod="openstack/keystone-db-sync-qmm47" Nov 24 14:38:59 crc kubenswrapper[4790]: I1124 14:38:59.925852 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6wtk\" (UniqueName: \"kubernetes.io/projected/42e1cf5d-d1e0-4a6c-8607-f37dd401f552-kube-api-access-f6wtk\") pod \"keystone-db-sync-qmm47\" (UID: \"42e1cf5d-d1e0-4a6c-8607-f37dd401f552\") " pod="openstack/keystone-db-sync-qmm47" Nov 24 14:38:59 crc kubenswrapper[4790]: I1124 14:38:59.926270 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42e1cf5d-d1e0-4a6c-8607-f37dd401f552-config-data\") pod \"keystone-db-sync-qmm47\" (UID: \"42e1cf5d-d1e0-4a6c-8607-f37dd401f552\") " pod="openstack/keystone-db-sync-qmm47" Nov 24 14:39:00 crc kubenswrapper[4790]: I1124 14:39:00.027727 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42e1cf5d-d1e0-4a6c-8607-f37dd401f552-config-data\") pod \"keystone-db-sync-qmm47\" (UID: \"42e1cf5d-d1e0-4a6c-8607-f37dd401f552\") " pod="openstack/keystone-db-sync-qmm47" Nov 24 14:39:00 crc kubenswrapper[4790]: I1124 14:39:00.027842 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42e1cf5d-d1e0-4a6c-8607-f37dd401f552-combined-ca-bundle\") pod \"keystone-db-sync-qmm47\" (UID: \"42e1cf5d-d1e0-4a6c-8607-f37dd401f552\") " pod="openstack/keystone-db-sync-qmm47" Nov 24 14:39:00 crc kubenswrapper[4790]: I1124 14:39:00.027907 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6wtk\" (UniqueName: \"kubernetes.io/projected/42e1cf5d-d1e0-4a6c-8607-f37dd401f552-kube-api-access-f6wtk\") pod \"keystone-db-sync-qmm47\" (UID: \"42e1cf5d-d1e0-4a6c-8607-f37dd401f552\") " pod="openstack/keystone-db-sync-qmm47" Nov 24 14:39:00 crc kubenswrapper[4790]: I1124 14:39:00.037175 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42e1cf5d-d1e0-4a6c-8607-f37dd401f552-config-data\") pod \"keystone-db-sync-qmm47\" (UID: \"42e1cf5d-d1e0-4a6c-8607-f37dd401f552\") " pod="openstack/keystone-db-sync-qmm47" Nov 24 14:39:00 crc kubenswrapper[4790]: I1124 14:39:00.037448 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42e1cf5d-d1e0-4a6c-8607-f37dd401f552-combined-ca-bundle\") pod \"keystone-db-sync-qmm47\" (UID: \"42e1cf5d-d1e0-4a6c-8607-f37dd401f552\") " pod="openstack/keystone-db-sync-qmm47" Nov 24 14:39:00 crc kubenswrapper[4790]: I1124 14:39:00.056616 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6wtk\" (UniqueName: \"kubernetes.io/projected/42e1cf5d-d1e0-4a6c-8607-f37dd401f552-kube-api-access-f6wtk\") pod \"keystone-db-sync-qmm47\" (UID: \"42e1cf5d-d1e0-4a6c-8607-f37dd401f552\") " pod="openstack/keystone-db-sync-qmm47" Nov 24 14:39:00 crc kubenswrapper[4790]: I1124 14:39:00.208771 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-qmm47" Nov 24 14:39:00 crc kubenswrapper[4790]: I1124 14:39:00.675469 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-qmm47"] Nov 24 14:39:00 crc kubenswrapper[4790]: W1124 14:39:00.683073 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod42e1cf5d_d1e0_4a6c_8607_f37dd401f552.slice/crio-6d82bfe373329960bb8da09f4dd1a027228f38a51f50cd571bcecbbdc44697f6 WatchSource:0}: Error finding container 6d82bfe373329960bb8da09f4dd1a027228f38a51f50cd571bcecbbdc44697f6: Status 404 returned error can't find the container with id 6d82bfe373329960bb8da09f4dd1a027228f38a51f50cd571bcecbbdc44697f6 Nov 24 14:39:01 crc kubenswrapper[4790]: I1124 14:39:01.559433 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-qmm47" event={"ID":"42e1cf5d-d1e0-4a6c-8607-f37dd401f552","Type":"ContainerStarted","Data":"7e81773b13d8d2c24f38bcc52d5bc1472270c6c363df339602ede7da9a78f8e5"} Nov 24 14:39:01 crc kubenswrapper[4790]: I1124 14:39:01.561975 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-qmm47" event={"ID":"42e1cf5d-d1e0-4a6c-8607-f37dd401f552","Type":"ContainerStarted","Data":"6d82bfe373329960bb8da09f4dd1a027228f38a51f50cd571bcecbbdc44697f6"} Nov 24 14:39:01 crc kubenswrapper[4790]: I1124 14:39:01.592756 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-qmm47" podStartSLOduration=2.592720229 podStartE2EDuration="2.592720229s" podCreationTimestamp="2025-11-24 14:38:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:39:01.588444362 +0000 UTC m=+5189.968338054" watchObservedRunningTime="2025-11-24 14:39:01.592720229 +0000 UTC m=+5189.972613921" Nov 24 14:39:03 crc kubenswrapper[4790]: I1124 14:39:03.603928 4790 generic.go:334] "Generic (PLEG): container finished" podID="42e1cf5d-d1e0-4a6c-8607-f37dd401f552" containerID="7e81773b13d8d2c24f38bcc52d5bc1472270c6c363df339602ede7da9a78f8e5" exitCode=0 Nov 24 14:39:03 crc kubenswrapper[4790]: I1124 14:39:03.604112 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-qmm47" event={"ID":"42e1cf5d-d1e0-4a6c-8607-f37dd401f552","Type":"ContainerDied","Data":"7e81773b13d8d2c24f38bcc52d5bc1472270c6c363df339602ede7da9a78f8e5"} Nov 24 14:39:04 crc kubenswrapper[4790]: I1124 14:39:04.038192 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 24 14:39:04 crc kubenswrapper[4790]: I1124 14:39:04.980110 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-qmm47" Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.037970 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42e1cf5d-d1e0-4a6c-8607-f37dd401f552-combined-ca-bundle\") pod \"42e1cf5d-d1e0-4a6c-8607-f37dd401f552\" (UID: \"42e1cf5d-d1e0-4a6c-8607-f37dd401f552\") " Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.038122 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6wtk\" (UniqueName: \"kubernetes.io/projected/42e1cf5d-d1e0-4a6c-8607-f37dd401f552-kube-api-access-f6wtk\") pod \"42e1cf5d-d1e0-4a6c-8607-f37dd401f552\" (UID: \"42e1cf5d-d1e0-4a6c-8607-f37dd401f552\") " Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.038181 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42e1cf5d-d1e0-4a6c-8607-f37dd401f552-config-data\") pod \"42e1cf5d-d1e0-4a6c-8607-f37dd401f552\" (UID: \"42e1cf5d-d1e0-4a6c-8607-f37dd401f552\") " Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.049182 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42e1cf5d-d1e0-4a6c-8607-f37dd401f552-kube-api-access-f6wtk" (OuterVolumeSpecName: "kube-api-access-f6wtk") pod "42e1cf5d-d1e0-4a6c-8607-f37dd401f552" (UID: "42e1cf5d-d1e0-4a6c-8607-f37dd401f552"). InnerVolumeSpecName "kube-api-access-f6wtk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.072419 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42e1cf5d-d1e0-4a6c-8607-f37dd401f552-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "42e1cf5d-d1e0-4a6c-8607-f37dd401f552" (UID: "42e1cf5d-d1e0-4a6c-8607-f37dd401f552"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.098591 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42e1cf5d-d1e0-4a6c-8607-f37dd401f552-config-data" (OuterVolumeSpecName: "config-data") pod "42e1cf5d-d1e0-4a6c-8607-f37dd401f552" (UID: "42e1cf5d-d1e0-4a6c-8607-f37dd401f552"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.140419 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42e1cf5d-d1e0-4a6c-8607-f37dd401f552-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.140647 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6wtk\" (UniqueName: \"kubernetes.io/projected/42e1cf5d-d1e0-4a6c-8607-f37dd401f552-kube-api-access-f6wtk\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.140738 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42e1cf5d-d1e0-4a6c-8607-f37dd401f552-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.632152 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-qmm47" event={"ID":"42e1cf5d-d1e0-4a6c-8607-f37dd401f552","Type":"ContainerDied","Data":"6d82bfe373329960bb8da09f4dd1a027228f38a51f50cd571bcecbbdc44697f6"} Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.632666 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d82bfe373329960bb8da09f4dd1a027228f38a51f50cd571bcecbbdc44697f6" Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.632358 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-qmm47" Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.899651 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d9f46b68f-hlsvx"] Nov 24 14:39:05 crc kubenswrapper[4790]: E1124 14:39:05.900186 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42e1cf5d-d1e0-4a6c-8607-f37dd401f552" containerName="keystone-db-sync" Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.900265 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="42e1cf5d-d1e0-4a6c-8607-f37dd401f552" containerName="keystone-db-sync" Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.900561 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="42e1cf5d-d1e0-4a6c-8607-f37dd401f552" containerName="keystone-db-sync" Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.903339 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d9f46b68f-hlsvx" Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.915075 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d9f46b68f-hlsvx"] Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.947055 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-jhx4s"] Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.948359 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jhx4s" Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.953312 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-config-data\") pod \"keystone-bootstrap-jhx4s\" (UID: \"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb\") " pod="openstack/keystone-bootstrap-jhx4s" Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.953390 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-combined-ca-bundle\") pod \"keystone-bootstrap-jhx4s\" (UID: \"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb\") " pod="openstack/keystone-bootstrap-jhx4s" Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.953419 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x488k\" (UniqueName: \"kubernetes.io/projected/1c1b93c1-d293-4c3e-b139-e030990dcba2-kube-api-access-x488k\") pod \"dnsmasq-dns-5d9f46b68f-hlsvx\" (UID: \"1c1b93c1-d293-4c3e-b139-e030990dcba2\") " pod="openstack/dnsmasq-dns-5d9f46b68f-hlsvx" Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.953439 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dspws\" (UniqueName: \"kubernetes.io/projected/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-kube-api-access-dspws\") pod \"keystone-bootstrap-jhx4s\" (UID: \"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb\") " pod="openstack/keystone-bootstrap-jhx4s" Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.953485 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-scripts\") pod \"keystone-bootstrap-jhx4s\" (UID: \"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb\") " pod="openstack/keystone-bootstrap-jhx4s" Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.953506 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-credential-keys\") pod \"keystone-bootstrap-jhx4s\" (UID: \"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb\") " pod="openstack/keystone-bootstrap-jhx4s" Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.953535 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c1b93c1-d293-4c3e-b139-e030990dcba2-dns-svc\") pod \"dnsmasq-dns-5d9f46b68f-hlsvx\" (UID: \"1c1b93c1-d293-4c3e-b139-e030990dcba2\") " pod="openstack/dnsmasq-dns-5d9f46b68f-hlsvx" Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.953547 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.953568 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c1b93c1-d293-4c3e-b139-e030990dcba2-ovsdbserver-nb\") pod \"dnsmasq-dns-5d9f46b68f-hlsvx\" (UID: \"1c1b93c1-d293-4c3e-b139-e030990dcba2\") " pod="openstack/dnsmasq-dns-5d9f46b68f-hlsvx" Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.953589 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-fernet-keys\") pod \"keystone-bootstrap-jhx4s\" (UID: \"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb\") " pod="openstack/keystone-bootstrap-jhx4s" Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.953611 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c1b93c1-d293-4c3e-b139-e030990dcba2-config\") pod \"dnsmasq-dns-5d9f46b68f-hlsvx\" (UID: \"1c1b93c1-d293-4c3e-b139-e030990dcba2\") " pod="openstack/dnsmasq-dns-5d9f46b68f-hlsvx" Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.953650 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c1b93c1-d293-4c3e-b139-e030990dcba2-ovsdbserver-sb\") pod \"dnsmasq-dns-5d9f46b68f-hlsvx\" (UID: \"1c1b93c1-d293-4c3e-b139-e030990dcba2\") " pod="openstack/dnsmasq-dns-5d9f46b68f-hlsvx" Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.953851 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-58t98" Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.953998 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.954019 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 14:39:05 crc kubenswrapper[4790]: I1124 14:39:05.954125 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 14:39:06 crc kubenswrapper[4790]: I1124 14:39:06.008516 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-jhx4s"] Nov 24 14:39:06 crc kubenswrapper[4790]: I1124 14:39:06.056196 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-config-data\") pod \"keystone-bootstrap-jhx4s\" (UID: \"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb\") " pod="openstack/keystone-bootstrap-jhx4s" Nov 24 14:39:06 crc kubenswrapper[4790]: I1124 14:39:06.056284 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-combined-ca-bundle\") pod \"keystone-bootstrap-jhx4s\" (UID: \"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb\") " pod="openstack/keystone-bootstrap-jhx4s" Nov 24 14:39:06 crc kubenswrapper[4790]: I1124 14:39:06.057016 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x488k\" (UniqueName: \"kubernetes.io/projected/1c1b93c1-d293-4c3e-b139-e030990dcba2-kube-api-access-x488k\") pod \"dnsmasq-dns-5d9f46b68f-hlsvx\" (UID: \"1c1b93c1-d293-4c3e-b139-e030990dcba2\") " pod="openstack/dnsmasq-dns-5d9f46b68f-hlsvx" Nov 24 14:39:06 crc kubenswrapper[4790]: I1124 14:39:06.057099 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dspws\" (UniqueName: \"kubernetes.io/projected/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-kube-api-access-dspws\") pod \"keystone-bootstrap-jhx4s\" (UID: \"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb\") " pod="openstack/keystone-bootstrap-jhx4s" Nov 24 14:39:06 crc kubenswrapper[4790]: I1124 14:39:06.057325 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-scripts\") pod \"keystone-bootstrap-jhx4s\" (UID: \"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb\") " pod="openstack/keystone-bootstrap-jhx4s" Nov 24 14:39:06 crc kubenswrapper[4790]: I1124 14:39:06.057375 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-credential-keys\") pod \"keystone-bootstrap-jhx4s\" (UID: \"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb\") " pod="openstack/keystone-bootstrap-jhx4s" Nov 24 14:39:06 crc kubenswrapper[4790]: I1124 14:39:06.057479 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c1b93c1-d293-4c3e-b139-e030990dcba2-dns-svc\") pod \"dnsmasq-dns-5d9f46b68f-hlsvx\" (UID: \"1c1b93c1-d293-4c3e-b139-e030990dcba2\") " pod="openstack/dnsmasq-dns-5d9f46b68f-hlsvx" Nov 24 14:39:06 crc kubenswrapper[4790]: I1124 14:39:06.057581 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c1b93c1-d293-4c3e-b139-e030990dcba2-ovsdbserver-nb\") pod \"dnsmasq-dns-5d9f46b68f-hlsvx\" (UID: \"1c1b93c1-d293-4c3e-b139-e030990dcba2\") " pod="openstack/dnsmasq-dns-5d9f46b68f-hlsvx" Nov 24 14:39:06 crc kubenswrapper[4790]: I1124 14:39:06.057617 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-fernet-keys\") pod \"keystone-bootstrap-jhx4s\" (UID: \"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb\") " pod="openstack/keystone-bootstrap-jhx4s" Nov 24 14:39:06 crc kubenswrapper[4790]: I1124 14:39:06.057661 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c1b93c1-d293-4c3e-b139-e030990dcba2-config\") pod \"dnsmasq-dns-5d9f46b68f-hlsvx\" (UID: \"1c1b93c1-d293-4c3e-b139-e030990dcba2\") " pod="openstack/dnsmasq-dns-5d9f46b68f-hlsvx" Nov 24 14:39:06 crc kubenswrapper[4790]: I1124 14:39:06.057760 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c1b93c1-d293-4c3e-b139-e030990dcba2-ovsdbserver-sb\") pod \"dnsmasq-dns-5d9f46b68f-hlsvx\" (UID: \"1c1b93c1-d293-4c3e-b139-e030990dcba2\") " pod="openstack/dnsmasq-dns-5d9f46b68f-hlsvx" Nov 24 14:39:06 crc kubenswrapper[4790]: I1124 14:39:06.060889 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c1b93c1-d293-4c3e-b139-e030990dcba2-config\") pod \"dnsmasq-dns-5d9f46b68f-hlsvx\" (UID: \"1c1b93c1-d293-4c3e-b139-e030990dcba2\") " pod="openstack/dnsmasq-dns-5d9f46b68f-hlsvx" Nov 24 14:39:06 crc kubenswrapper[4790]: I1124 14:39:06.061130 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c1b93c1-d293-4c3e-b139-e030990dcba2-ovsdbserver-sb\") pod \"dnsmasq-dns-5d9f46b68f-hlsvx\" (UID: \"1c1b93c1-d293-4c3e-b139-e030990dcba2\") " pod="openstack/dnsmasq-dns-5d9f46b68f-hlsvx" Nov 24 14:39:06 crc kubenswrapper[4790]: I1124 14:39:06.062146 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c1b93c1-d293-4c3e-b139-e030990dcba2-dns-svc\") pod \"dnsmasq-dns-5d9f46b68f-hlsvx\" (UID: \"1c1b93c1-d293-4c3e-b139-e030990dcba2\") " pod="openstack/dnsmasq-dns-5d9f46b68f-hlsvx" Nov 24 14:39:06 crc kubenswrapper[4790]: I1124 14:39:06.062299 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c1b93c1-d293-4c3e-b139-e030990dcba2-ovsdbserver-nb\") pod \"dnsmasq-dns-5d9f46b68f-hlsvx\" (UID: \"1c1b93c1-d293-4c3e-b139-e030990dcba2\") " pod="openstack/dnsmasq-dns-5d9f46b68f-hlsvx" Nov 24 14:39:06 crc kubenswrapper[4790]: I1124 14:39:06.064118 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-scripts\") pod \"keystone-bootstrap-jhx4s\" (UID: \"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb\") " pod="openstack/keystone-bootstrap-jhx4s" Nov 24 14:39:06 crc kubenswrapper[4790]: I1124 14:39:06.066082 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-config-data\") pod \"keystone-bootstrap-jhx4s\" (UID: \"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb\") " pod="openstack/keystone-bootstrap-jhx4s" Nov 24 14:39:06 crc kubenswrapper[4790]: I1124 14:39:06.066917 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-fernet-keys\") pod \"keystone-bootstrap-jhx4s\" (UID: \"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb\") " pod="openstack/keystone-bootstrap-jhx4s" Nov 24 14:39:06 crc kubenswrapper[4790]: I1124 14:39:06.069457 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-credential-keys\") pod \"keystone-bootstrap-jhx4s\" (UID: \"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb\") " pod="openstack/keystone-bootstrap-jhx4s" Nov 24 14:39:06 crc kubenswrapper[4790]: I1124 14:39:06.073648 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-combined-ca-bundle\") pod \"keystone-bootstrap-jhx4s\" (UID: \"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb\") " pod="openstack/keystone-bootstrap-jhx4s" Nov 24 14:39:06 crc kubenswrapper[4790]: I1124 14:39:06.082382 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x488k\" (UniqueName: \"kubernetes.io/projected/1c1b93c1-d293-4c3e-b139-e030990dcba2-kube-api-access-x488k\") pod \"dnsmasq-dns-5d9f46b68f-hlsvx\" (UID: \"1c1b93c1-d293-4c3e-b139-e030990dcba2\") " pod="openstack/dnsmasq-dns-5d9f46b68f-hlsvx" Nov 24 14:39:06 crc kubenswrapper[4790]: I1124 14:39:06.086129 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dspws\" (UniqueName: \"kubernetes.io/projected/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-kube-api-access-dspws\") pod \"keystone-bootstrap-jhx4s\" (UID: \"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb\") " pod="openstack/keystone-bootstrap-jhx4s" Nov 24 14:39:06 crc kubenswrapper[4790]: I1124 14:39:06.226913 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d9f46b68f-hlsvx" Nov 24 14:39:06 crc kubenswrapper[4790]: I1124 14:39:06.274356 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jhx4s" Nov 24 14:39:06 crc kubenswrapper[4790]: W1124 14:39:06.731248 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1c1b93c1_d293_4c3e_b139_e030990dcba2.slice/crio-3ea85bbedb9d74c7f7cbbd932af015d3daa6a28a8ef1a49eafbb441da31ccc97 WatchSource:0}: Error finding container 3ea85bbedb9d74c7f7cbbd932af015d3daa6a28a8ef1a49eafbb441da31ccc97: Status 404 returned error can't find the container with id 3ea85bbedb9d74c7f7cbbd932af015d3daa6a28a8ef1a49eafbb441da31ccc97 Nov 24 14:39:06 crc kubenswrapper[4790]: I1124 14:39:06.731684 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d9f46b68f-hlsvx"] Nov 24 14:39:06 crc kubenswrapper[4790]: I1124 14:39:06.739692 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-jhx4s"] Nov 24 14:39:06 crc kubenswrapper[4790]: W1124 14:39:06.743174 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67b13f6d_81b3_49c8_a1a5_aae1725dbbcb.slice/crio-be65b1bada55837435f37cdc9dd53ea0bd971926f351231d35f01bc9a06f6665 WatchSource:0}: Error finding container be65b1bada55837435f37cdc9dd53ea0bd971926f351231d35f01bc9a06f6665: Status 404 returned error can't find the container with id be65b1bada55837435f37cdc9dd53ea0bd971926f351231d35f01bc9a06f6665 Nov 24 14:39:07 crc kubenswrapper[4790]: I1124 14:39:07.660209 4790 generic.go:334] "Generic (PLEG): container finished" podID="1c1b93c1-d293-4c3e-b139-e030990dcba2" containerID="d1c2c0b6af2effcc8ddcc9bc41f66f2836726e7202c02bf5166bab98a6bcd290" exitCode=0 Nov 24 14:39:07 crc kubenswrapper[4790]: I1124 14:39:07.660587 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d9f46b68f-hlsvx" event={"ID":"1c1b93c1-d293-4c3e-b139-e030990dcba2","Type":"ContainerDied","Data":"d1c2c0b6af2effcc8ddcc9bc41f66f2836726e7202c02bf5166bab98a6bcd290"} Nov 24 14:39:07 crc kubenswrapper[4790]: I1124 14:39:07.660958 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d9f46b68f-hlsvx" event={"ID":"1c1b93c1-d293-4c3e-b139-e030990dcba2","Type":"ContainerStarted","Data":"3ea85bbedb9d74c7f7cbbd932af015d3daa6a28a8ef1a49eafbb441da31ccc97"} Nov 24 14:39:07 crc kubenswrapper[4790]: I1124 14:39:07.674537 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jhx4s" event={"ID":"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb","Type":"ContainerStarted","Data":"90eaf19c23289a7e2dcfee76d9001e56223190ed57614857ef7469ace4095ff6"} Nov 24 14:39:07 crc kubenswrapper[4790]: I1124 14:39:07.674610 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jhx4s" event={"ID":"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb","Type":"ContainerStarted","Data":"be65b1bada55837435f37cdc9dd53ea0bd971926f351231d35f01bc9a06f6665"} Nov 24 14:39:07 crc kubenswrapper[4790]: I1124 14:39:07.827214 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-jhx4s" podStartSLOduration=2.827184612 podStartE2EDuration="2.827184612s" podCreationTimestamp="2025-11-24 14:39:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:39:07.773186639 +0000 UTC m=+5196.153080311" watchObservedRunningTime="2025-11-24 14:39:07.827184612 +0000 UTC m=+5196.207078274" Nov 24 14:39:08 crc kubenswrapper[4790]: I1124 14:39:08.693244 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d9f46b68f-hlsvx" event={"ID":"1c1b93c1-d293-4c3e-b139-e030990dcba2","Type":"ContainerStarted","Data":"c662a0191d3751f33a241866f428c2666f14285bb9cbbe2310801f7049acaf42"} Nov 24 14:39:08 crc kubenswrapper[4790]: I1124 14:39:08.693740 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d9f46b68f-hlsvx" Nov 24 14:39:10 crc kubenswrapper[4790]: I1124 14:39:10.424364 4790 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","podf3537c77-1bb1-4595-92c4-a9298a5df3a9"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort podf3537c77-1bb1-4595-92c4-a9298a5df3a9] : Timed out while waiting for systemd to remove kubepods-besteffort-podf3537c77_1bb1_4595_92c4_a9298a5df3a9.slice" Nov 24 14:39:10 crc kubenswrapper[4790]: E1124 14:39:10.424955 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort podf3537c77-1bb1-4595-92c4-a9298a5df3a9] : unable to destroy cgroup paths for cgroup [kubepods besteffort podf3537c77-1bb1-4595-92c4-a9298a5df3a9] : Timed out while waiting for systemd to remove kubepods-besteffort-podf3537c77_1bb1_4595_92c4_a9298a5df3a9.slice" pod="openstack/dnsmasq-dns-6c9c4b4dfc-h6f2z" podUID="f3537c77-1bb1-4595-92c4-a9298a5df3a9" Nov 24 14:39:10 crc kubenswrapper[4790]: I1124 14:39:10.716715 4790 generic.go:334] "Generic (PLEG): container finished" podID="67b13f6d-81b3-49c8-a1a5-aae1725dbbcb" containerID="90eaf19c23289a7e2dcfee76d9001e56223190ed57614857ef7469ace4095ff6" exitCode=0 Nov 24 14:39:10 crc kubenswrapper[4790]: I1124 14:39:10.716919 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jhx4s" event={"ID":"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb","Type":"ContainerDied","Data":"90eaf19c23289a7e2dcfee76d9001e56223190ed57614857ef7469ace4095ff6"} Nov 24 14:39:10 crc kubenswrapper[4790]: I1124 14:39:10.717373 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c9c4b4dfc-h6f2z" Nov 24 14:39:10 crc kubenswrapper[4790]: I1124 14:39:10.753740 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d9f46b68f-hlsvx" podStartSLOduration=5.753704268 podStartE2EDuration="5.753704268s" podCreationTimestamp="2025-11-24 14:39:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:39:08.725360969 +0000 UTC m=+5197.105254671" watchObservedRunningTime="2025-11-24 14:39:10.753704268 +0000 UTC m=+5199.133597970" Nov 24 14:39:10 crc kubenswrapper[4790]: I1124 14:39:10.808431 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c9c4b4dfc-h6f2z"] Nov 24 14:39:10 crc kubenswrapper[4790]: I1124 14:39:10.818173 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c9c4b4dfc-h6f2z"] Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.197747 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jhx4s" Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.335810 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3537c77-1bb1-4595-92c4-a9298a5df3a9" path="/var/lib/kubelet/pods/f3537c77-1bb1-4595-92c4-a9298a5df3a9/volumes" Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.397009 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-config-data\") pod \"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb\" (UID: \"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb\") " Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.397097 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-combined-ca-bundle\") pod \"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb\" (UID: \"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb\") " Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.397178 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-fernet-keys\") pod \"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb\" (UID: \"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb\") " Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.397255 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dspws\" (UniqueName: \"kubernetes.io/projected/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-kube-api-access-dspws\") pod \"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb\" (UID: \"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb\") " Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.397397 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-scripts\") pod \"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb\" (UID: \"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb\") " Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.397551 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-credential-keys\") pod \"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb\" (UID: \"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb\") " Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.405336 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "67b13f6d-81b3-49c8-a1a5-aae1725dbbcb" (UID: "67b13f6d-81b3-49c8-a1a5-aae1725dbbcb"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.405979 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-scripts" (OuterVolumeSpecName: "scripts") pod "67b13f6d-81b3-49c8-a1a5-aae1725dbbcb" (UID: "67b13f6d-81b3-49c8-a1a5-aae1725dbbcb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.406560 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "67b13f6d-81b3-49c8-a1a5-aae1725dbbcb" (UID: "67b13f6d-81b3-49c8-a1a5-aae1725dbbcb"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.406966 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-kube-api-access-dspws" (OuterVolumeSpecName: "kube-api-access-dspws") pod "67b13f6d-81b3-49c8-a1a5-aae1725dbbcb" (UID: "67b13f6d-81b3-49c8-a1a5-aae1725dbbcb"). InnerVolumeSpecName "kube-api-access-dspws". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.427992 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67b13f6d-81b3-49c8-a1a5-aae1725dbbcb" (UID: "67b13f6d-81b3-49c8-a1a5-aae1725dbbcb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.445913 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-config-data" (OuterVolumeSpecName: "config-data") pod "67b13f6d-81b3-49c8-a1a5-aae1725dbbcb" (UID: "67b13f6d-81b3-49c8-a1a5-aae1725dbbcb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.500656 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.500697 4790 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.500712 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.500727 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.500742 4790 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.500755 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dspws\" (UniqueName: \"kubernetes.io/projected/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb-kube-api-access-dspws\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.742345 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jhx4s" event={"ID":"67b13f6d-81b3-49c8-a1a5-aae1725dbbcb","Type":"ContainerDied","Data":"be65b1bada55837435f37cdc9dd53ea0bd971926f351231d35f01bc9a06f6665"} Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.742426 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be65b1bada55837435f37cdc9dd53ea0bd971926f351231d35f01bc9a06f6665" Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.742479 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jhx4s" Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.856349 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-jhx4s"] Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.865372 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-jhx4s"] Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.936864 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-c7bhw"] Nov 24 14:39:12 crc kubenswrapper[4790]: E1124 14:39:12.937502 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67b13f6d-81b3-49c8-a1a5-aae1725dbbcb" containerName="keystone-bootstrap" Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.937524 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="67b13f6d-81b3-49c8-a1a5-aae1725dbbcb" containerName="keystone-bootstrap" Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.937725 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="67b13f6d-81b3-49c8-a1a5-aae1725dbbcb" containerName="keystone-bootstrap" Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.938597 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-c7bhw" Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.941474 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.943000 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.943182 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.943327 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.943812 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-58t98" Nov 24 14:39:12 crc kubenswrapper[4790]: I1124 14:39:12.952311 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-c7bhw"] Nov 24 14:39:13 crc kubenswrapper[4790]: I1124 14:39:13.132617 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgmtr\" (UniqueName: \"kubernetes.io/projected/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-kube-api-access-zgmtr\") pod \"keystone-bootstrap-c7bhw\" (UID: \"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79\") " pod="openstack/keystone-bootstrap-c7bhw" Nov 24 14:39:13 crc kubenswrapper[4790]: I1124 14:39:13.132753 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-scripts\") pod \"keystone-bootstrap-c7bhw\" (UID: \"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79\") " pod="openstack/keystone-bootstrap-c7bhw" Nov 24 14:39:13 crc kubenswrapper[4790]: I1124 14:39:13.133546 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-fernet-keys\") pod \"keystone-bootstrap-c7bhw\" (UID: \"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79\") " pod="openstack/keystone-bootstrap-c7bhw" Nov 24 14:39:13 crc kubenswrapper[4790]: I1124 14:39:13.134060 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-credential-keys\") pod \"keystone-bootstrap-c7bhw\" (UID: \"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79\") " pod="openstack/keystone-bootstrap-c7bhw" Nov 24 14:39:13 crc kubenswrapper[4790]: I1124 14:39:13.134224 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-config-data\") pod \"keystone-bootstrap-c7bhw\" (UID: \"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79\") " pod="openstack/keystone-bootstrap-c7bhw" Nov 24 14:39:13 crc kubenswrapper[4790]: I1124 14:39:13.134547 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-combined-ca-bundle\") pod \"keystone-bootstrap-c7bhw\" (UID: \"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79\") " pod="openstack/keystone-bootstrap-c7bhw" Nov 24 14:39:13 crc kubenswrapper[4790]: I1124 14:39:13.236486 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-credential-keys\") pod \"keystone-bootstrap-c7bhw\" (UID: \"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79\") " pod="openstack/keystone-bootstrap-c7bhw" Nov 24 14:39:13 crc kubenswrapper[4790]: I1124 14:39:13.237002 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-config-data\") pod \"keystone-bootstrap-c7bhw\" (UID: \"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79\") " pod="openstack/keystone-bootstrap-c7bhw" Nov 24 14:39:13 crc kubenswrapper[4790]: I1124 14:39:13.237073 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-combined-ca-bundle\") pod \"keystone-bootstrap-c7bhw\" (UID: \"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79\") " pod="openstack/keystone-bootstrap-c7bhw" Nov 24 14:39:13 crc kubenswrapper[4790]: I1124 14:39:13.237118 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgmtr\" (UniqueName: \"kubernetes.io/projected/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-kube-api-access-zgmtr\") pod \"keystone-bootstrap-c7bhw\" (UID: \"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79\") " pod="openstack/keystone-bootstrap-c7bhw" Nov 24 14:39:13 crc kubenswrapper[4790]: I1124 14:39:13.237178 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-scripts\") pod \"keystone-bootstrap-c7bhw\" (UID: \"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79\") " pod="openstack/keystone-bootstrap-c7bhw" Nov 24 14:39:13 crc kubenswrapper[4790]: I1124 14:39:13.237211 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-fernet-keys\") pod \"keystone-bootstrap-c7bhw\" (UID: \"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79\") " pod="openstack/keystone-bootstrap-c7bhw" Nov 24 14:39:13 crc kubenswrapper[4790]: I1124 14:39:13.243098 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-combined-ca-bundle\") pod \"keystone-bootstrap-c7bhw\" (UID: \"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79\") " pod="openstack/keystone-bootstrap-c7bhw" Nov 24 14:39:13 crc kubenswrapper[4790]: I1124 14:39:13.243215 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-fernet-keys\") pod \"keystone-bootstrap-c7bhw\" (UID: \"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79\") " pod="openstack/keystone-bootstrap-c7bhw" Nov 24 14:39:13 crc kubenswrapper[4790]: I1124 14:39:13.244205 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-config-data\") pod \"keystone-bootstrap-c7bhw\" (UID: \"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79\") " pod="openstack/keystone-bootstrap-c7bhw" Nov 24 14:39:13 crc kubenswrapper[4790]: I1124 14:39:13.245478 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-credential-keys\") pod \"keystone-bootstrap-c7bhw\" (UID: \"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79\") " pod="openstack/keystone-bootstrap-c7bhw" Nov 24 14:39:13 crc kubenswrapper[4790]: I1124 14:39:13.248967 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-scripts\") pod \"keystone-bootstrap-c7bhw\" (UID: \"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79\") " pod="openstack/keystone-bootstrap-c7bhw" Nov 24 14:39:13 crc kubenswrapper[4790]: I1124 14:39:13.264518 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgmtr\" (UniqueName: \"kubernetes.io/projected/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-kube-api-access-zgmtr\") pod \"keystone-bootstrap-c7bhw\" (UID: \"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79\") " pod="openstack/keystone-bootstrap-c7bhw" Nov 24 14:39:13 crc kubenswrapper[4790]: I1124 14:39:13.563458 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-c7bhw" Nov 24 14:39:14 crc kubenswrapper[4790]: I1124 14:39:14.105317 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-c7bhw"] Nov 24 14:39:14 crc kubenswrapper[4790]: I1124 14:39:14.333606 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67b13f6d-81b3-49c8-a1a5-aae1725dbbcb" path="/var/lib/kubelet/pods/67b13f6d-81b3-49c8-a1a5-aae1725dbbcb/volumes" Nov 24 14:39:14 crc kubenswrapper[4790]: I1124 14:39:14.773945 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-c7bhw" event={"ID":"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79","Type":"ContainerStarted","Data":"0d23ee6e35232d5cc608a2b677baea86e7b31006ea39bad6c35a214535941fe4"} Nov 24 14:39:14 crc kubenswrapper[4790]: I1124 14:39:14.774045 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-c7bhw" event={"ID":"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79","Type":"ContainerStarted","Data":"e17350fb9906ac587d4ce385ea7694fc2d6c749766b3d62918c21fbb29e27464"} Nov 24 14:39:14 crc kubenswrapper[4790]: I1124 14:39:14.806353 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-c7bhw" podStartSLOduration=2.806322988 podStartE2EDuration="2.806322988s" podCreationTimestamp="2025-11-24 14:39:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:39:14.794963978 +0000 UTC m=+5203.174857670" watchObservedRunningTime="2025-11-24 14:39:14.806322988 +0000 UTC m=+5203.186216690" Nov 24 14:39:16 crc kubenswrapper[4790]: I1124 14:39:16.228194 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5d9f46b68f-hlsvx" Nov 24 14:39:16 crc kubenswrapper[4790]: I1124 14:39:16.334498 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77bf5858df-7s64q"] Nov 24 14:39:16 crc kubenswrapper[4790]: I1124 14:39:16.334833 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77bf5858df-7s64q" podUID="42529236-56e6-400f-a34f-8d55dcff6a77" containerName="dnsmasq-dns" containerID="cri-o://a24444ba4dbbc1ab8e1990bd03a669237261e7dd92debd2bc53b97198fc0f7e7" gracePeriod=10 Nov 24 14:39:16 crc kubenswrapper[4790]: I1124 14:39:16.804377 4790 generic.go:334] "Generic (PLEG): container finished" podID="42529236-56e6-400f-a34f-8d55dcff6a77" containerID="a24444ba4dbbc1ab8e1990bd03a669237261e7dd92debd2bc53b97198fc0f7e7" exitCode=0 Nov 24 14:39:16 crc kubenswrapper[4790]: I1124 14:39:16.805430 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77bf5858df-7s64q" event={"ID":"42529236-56e6-400f-a34f-8d55dcff6a77","Type":"ContainerDied","Data":"a24444ba4dbbc1ab8e1990bd03a669237261e7dd92debd2bc53b97198fc0f7e7"} Nov 24 14:39:16 crc kubenswrapper[4790]: I1124 14:39:16.807278 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77bf5858df-7s64q" event={"ID":"42529236-56e6-400f-a34f-8d55dcff6a77","Type":"ContainerDied","Data":"f9b31068df07659591dc5b5a0011a3e1ff68e3cb958378cb4ff2139e9c344990"} Nov 24 14:39:16 crc kubenswrapper[4790]: I1124 14:39:16.807344 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f9b31068df07659591dc5b5a0011a3e1ff68e3cb958378cb4ff2139e9c344990" Nov 24 14:39:16 crc kubenswrapper[4790]: I1124 14:39:16.837945 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77bf5858df-7s64q" Nov 24 14:39:17 crc kubenswrapper[4790]: I1124 14:39:17.026201 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/42529236-56e6-400f-a34f-8d55dcff6a77-ovsdbserver-nb\") pod \"42529236-56e6-400f-a34f-8d55dcff6a77\" (UID: \"42529236-56e6-400f-a34f-8d55dcff6a77\") " Nov 24 14:39:17 crc kubenswrapper[4790]: I1124 14:39:17.026310 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2c4fl\" (UniqueName: \"kubernetes.io/projected/42529236-56e6-400f-a34f-8d55dcff6a77-kube-api-access-2c4fl\") pod \"42529236-56e6-400f-a34f-8d55dcff6a77\" (UID: \"42529236-56e6-400f-a34f-8d55dcff6a77\") " Nov 24 14:39:17 crc kubenswrapper[4790]: I1124 14:39:17.026343 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/42529236-56e6-400f-a34f-8d55dcff6a77-ovsdbserver-sb\") pod \"42529236-56e6-400f-a34f-8d55dcff6a77\" (UID: \"42529236-56e6-400f-a34f-8d55dcff6a77\") " Nov 24 14:39:17 crc kubenswrapper[4790]: I1124 14:39:17.026445 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42529236-56e6-400f-a34f-8d55dcff6a77-config\") pod \"42529236-56e6-400f-a34f-8d55dcff6a77\" (UID: \"42529236-56e6-400f-a34f-8d55dcff6a77\") " Nov 24 14:39:17 crc kubenswrapper[4790]: I1124 14:39:17.026531 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/42529236-56e6-400f-a34f-8d55dcff6a77-dns-svc\") pod \"42529236-56e6-400f-a34f-8d55dcff6a77\" (UID: \"42529236-56e6-400f-a34f-8d55dcff6a77\") " Nov 24 14:39:17 crc kubenswrapper[4790]: I1124 14:39:17.034570 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42529236-56e6-400f-a34f-8d55dcff6a77-kube-api-access-2c4fl" (OuterVolumeSpecName: "kube-api-access-2c4fl") pod "42529236-56e6-400f-a34f-8d55dcff6a77" (UID: "42529236-56e6-400f-a34f-8d55dcff6a77"). InnerVolumeSpecName "kube-api-access-2c4fl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:39:17 crc kubenswrapper[4790]: I1124 14:39:17.077119 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42529236-56e6-400f-a34f-8d55dcff6a77-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "42529236-56e6-400f-a34f-8d55dcff6a77" (UID: "42529236-56e6-400f-a34f-8d55dcff6a77"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:39:17 crc kubenswrapper[4790]: I1124 14:39:17.081403 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42529236-56e6-400f-a34f-8d55dcff6a77-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "42529236-56e6-400f-a34f-8d55dcff6a77" (UID: "42529236-56e6-400f-a34f-8d55dcff6a77"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:39:17 crc kubenswrapper[4790]: I1124 14:39:17.084423 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42529236-56e6-400f-a34f-8d55dcff6a77-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "42529236-56e6-400f-a34f-8d55dcff6a77" (UID: "42529236-56e6-400f-a34f-8d55dcff6a77"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:39:17 crc kubenswrapper[4790]: I1124 14:39:17.095141 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42529236-56e6-400f-a34f-8d55dcff6a77-config" (OuterVolumeSpecName: "config") pod "42529236-56e6-400f-a34f-8d55dcff6a77" (UID: "42529236-56e6-400f-a34f-8d55dcff6a77"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:39:17 crc kubenswrapper[4790]: I1124 14:39:17.128675 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/42529236-56e6-400f-a34f-8d55dcff6a77-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:17 crc kubenswrapper[4790]: I1124 14:39:17.128706 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2c4fl\" (UniqueName: \"kubernetes.io/projected/42529236-56e6-400f-a34f-8d55dcff6a77-kube-api-access-2c4fl\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:17 crc kubenswrapper[4790]: I1124 14:39:17.128717 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/42529236-56e6-400f-a34f-8d55dcff6a77-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:17 crc kubenswrapper[4790]: I1124 14:39:17.128728 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42529236-56e6-400f-a34f-8d55dcff6a77-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:17 crc kubenswrapper[4790]: I1124 14:39:17.128738 4790 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/42529236-56e6-400f-a34f-8d55dcff6a77-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:17 crc kubenswrapper[4790]: I1124 14:39:17.823290 4790 generic.go:334] "Generic (PLEG): container finished" podID="c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79" containerID="0d23ee6e35232d5cc608a2b677baea86e7b31006ea39bad6c35a214535941fe4" exitCode=0 Nov 24 14:39:17 crc kubenswrapper[4790]: I1124 14:39:17.823772 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-c7bhw" event={"ID":"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79","Type":"ContainerDied","Data":"0d23ee6e35232d5cc608a2b677baea86e7b31006ea39bad6c35a214535941fe4"} Nov 24 14:39:17 crc kubenswrapper[4790]: I1124 14:39:17.824054 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77bf5858df-7s64q" Nov 24 14:39:17 crc kubenswrapper[4790]: I1124 14:39:17.901560 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77bf5858df-7s64q"] Nov 24 14:39:17 crc kubenswrapper[4790]: I1124 14:39:17.917629 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77bf5858df-7s64q"] Nov 24 14:39:18 crc kubenswrapper[4790]: I1124 14:39:18.334563 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42529236-56e6-400f-a34f-8d55dcff6a77" path="/var/lib/kubelet/pods/42529236-56e6-400f-a34f-8d55dcff6a77/volumes" Nov 24 14:39:19 crc kubenswrapper[4790]: I1124 14:39:19.235312 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-c7bhw" Nov 24 14:39:19 crc kubenswrapper[4790]: I1124 14:39:19.391277 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-credential-keys\") pod \"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79\" (UID: \"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79\") " Nov 24 14:39:19 crc kubenswrapper[4790]: I1124 14:39:19.391474 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-scripts\") pod \"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79\" (UID: \"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79\") " Nov 24 14:39:19 crc kubenswrapper[4790]: I1124 14:39:19.391721 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-fernet-keys\") pod \"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79\" (UID: \"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79\") " Nov 24 14:39:19 crc kubenswrapper[4790]: I1124 14:39:19.392144 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgmtr\" (UniqueName: \"kubernetes.io/projected/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-kube-api-access-zgmtr\") pod \"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79\" (UID: \"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79\") " Nov 24 14:39:19 crc kubenswrapper[4790]: I1124 14:39:19.393100 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-combined-ca-bundle\") pod \"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79\" (UID: \"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79\") " Nov 24 14:39:19 crc kubenswrapper[4790]: I1124 14:39:19.393325 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-config-data\") pod \"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79\" (UID: \"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79\") " Nov 24 14:39:19 crc kubenswrapper[4790]: I1124 14:39:19.399801 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79" (UID: "c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:19 crc kubenswrapper[4790]: I1124 14:39:19.399863 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-scripts" (OuterVolumeSpecName: "scripts") pod "c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79" (UID: "c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:19 crc kubenswrapper[4790]: I1124 14:39:19.399920 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79" (UID: "c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:19 crc kubenswrapper[4790]: I1124 14:39:19.401001 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-kube-api-access-zgmtr" (OuterVolumeSpecName: "kube-api-access-zgmtr") pod "c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79" (UID: "c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79"). InnerVolumeSpecName "kube-api-access-zgmtr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:39:19 crc kubenswrapper[4790]: I1124 14:39:19.424806 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79" (UID: "c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:19 crc kubenswrapper[4790]: I1124 14:39:19.436411 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-config-data" (OuterVolumeSpecName: "config-data") pod "c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79" (UID: "c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:19 crc kubenswrapper[4790]: I1124 14:39:19.496332 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgmtr\" (UniqueName: \"kubernetes.io/projected/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-kube-api-access-zgmtr\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:19 crc kubenswrapper[4790]: I1124 14:39:19.496389 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:19 crc kubenswrapper[4790]: I1124 14:39:19.496410 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:19 crc kubenswrapper[4790]: I1124 14:39:19.496448 4790 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:19 crc kubenswrapper[4790]: I1124 14:39:19.496468 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:19 crc kubenswrapper[4790]: I1124 14:39:19.496486 4790 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:19 crc kubenswrapper[4790]: I1124 14:39:19.852180 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-c7bhw" event={"ID":"c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79","Type":"ContainerDied","Data":"e17350fb9906ac587d4ce385ea7694fc2d6c749766b3d62918c21fbb29e27464"} Nov 24 14:39:19 crc kubenswrapper[4790]: I1124 14:39:19.852763 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e17350fb9906ac587d4ce385ea7694fc2d6c749766b3d62918c21fbb29e27464" Nov 24 14:39:19 crc kubenswrapper[4790]: I1124 14:39:19.852272 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-c7bhw" Nov 24 14:39:20 crc kubenswrapper[4790]: I1124 14:39:20.007337 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-787bf45df4-blpgw"] Nov 24 14:39:20 crc kubenswrapper[4790]: E1124 14:39:20.007970 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79" containerName="keystone-bootstrap" Nov 24 14:39:20 crc kubenswrapper[4790]: I1124 14:39:20.007989 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79" containerName="keystone-bootstrap" Nov 24 14:39:20 crc kubenswrapper[4790]: E1124 14:39:20.008010 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42529236-56e6-400f-a34f-8d55dcff6a77" containerName="init" Nov 24 14:39:20 crc kubenswrapper[4790]: I1124 14:39:20.008017 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="42529236-56e6-400f-a34f-8d55dcff6a77" containerName="init" Nov 24 14:39:20 crc kubenswrapper[4790]: E1124 14:39:20.008035 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42529236-56e6-400f-a34f-8d55dcff6a77" containerName="dnsmasq-dns" Nov 24 14:39:20 crc kubenswrapper[4790]: I1124 14:39:20.008044 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="42529236-56e6-400f-a34f-8d55dcff6a77" containerName="dnsmasq-dns" Nov 24 14:39:20 crc kubenswrapper[4790]: I1124 14:39:20.008215 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="42529236-56e6-400f-a34f-8d55dcff6a77" containerName="dnsmasq-dns" Nov 24 14:39:20 crc kubenswrapper[4790]: I1124 14:39:20.008235 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79" containerName="keystone-bootstrap" Nov 24 14:39:20 crc kubenswrapper[4790]: I1124 14:39:20.008903 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-787bf45df4-blpgw" Nov 24 14:39:20 crc kubenswrapper[4790]: I1124 14:39:20.020001 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 14:39:20 crc kubenswrapper[4790]: I1124 14:39:20.020118 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 14:39:20 crc kubenswrapper[4790]: I1124 14:39:20.020355 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 14:39:20 crc kubenswrapper[4790]: I1124 14:39:20.020750 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-58t98" Nov 24 14:39:20 crc kubenswrapper[4790]: I1124 14:39:20.028209 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-787bf45df4-blpgw"] Nov 24 14:39:20 crc kubenswrapper[4790]: I1124 14:39:20.108303 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dbb6d63-d31f-4951-ac0c-5ec7309c129a-scripts\") pod \"keystone-787bf45df4-blpgw\" (UID: \"7dbb6d63-d31f-4951-ac0c-5ec7309c129a\") " pod="openstack/keystone-787bf45df4-blpgw" Nov 24 14:39:20 crc kubenswrapper[4790]: I1124 14:39:20.108369 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7dbb6d63-d31f-4951-ac0c-5ec7309c129a-credential-keys\") pod \"keystone-787bf45df4-blpgw\" (UID: \"7dbb6d63-d31f-4951-ac0c-5ec7309c129a\") " pod="openstack/keystone-787bf45df4-blpgw" Nov 24 14:39:20 crc kubenswrapper[4790]: I1124 14:39:20.108424 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7dbb6d63-d31f-4951-ac0c-5ec7309c129a-fernet-keys\") pod \"keystone-787bf45df4-blpgw\" (UID: \"7dbb6d63-d31f-4951-ac0c-5ec7309c129a\") " pod="openstack/keystone-787bf45df4-blpgw" Nov 24 14:39:20 crc kubenswrapper[4790]: I1124 14:39:20.108465 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdmc5\" (UniqueName: \"kubernetes.io/projected/7dbb6d63-d31f-4951-ac0c-5ec7309c129a-kube-api-access-bdmc5\") pod \"keystone-787bf45df4-blpgw\" (UID: \"7dbb6d63-d31f-4951-ac0c-5ec7309c129a\") " pod="openstack/keystone-787bf45df4-blpgw" Nov 24 14:39:20 crc kubenswrapper[4790]: I1124 14:39:20.108490 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dbb6d63-d31f-4951-ac0c-5ec7309c129a-combined-ca-bundle\") pod \"keystone-787bf45df4-blpgw\" (UID: \"7dbb6d63-d31f-4951-ac0c-5ec7309c129a\") " pod="openstack/keystone-787bf45df4-blpgw" Nov 24 14:39:20 crc kubenswrapper[4790]: I1124 14:39:20.108520 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dbb6d63-d31f-4951-ac0c-5ec7309c129a-config-data\") pod \"keystone-787bf45df4-blpgw\" (UID: \"7dbb6d63-d31f-4951-ac0c-5ec7309c129a\") " pod="openstack/keystone-787bf45df4-blpgw" Nov 24 14:39:20 crc kubenswrapper[4790]: I1124 14:39:20.209298 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dbb6d63-d31f-4951-ac0c-5ec7309c129a-config-data\") pod \"keystone-787bf45df4-blpgw\" (UID: \"7dbb6d63-d31f-4951-ac0c-5ec7309c129a\") " pod="openstack/keystone-787bf45df4-blpgw" Nov 24 14:39:20 crc kubenswrapper[4790]: I1124 14:39:20.209392 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dbb6d63-d31f-4951-ac0c-5ec7309c129a-scripts\") pod \"keystone-787bf45df4-blpgw\" (UID: \"7dbb6d63-d31f-4951-ac0c-5ec7309c129a\") " pod="openstack/keystone-787bf45df4-blpgw" Nov 24 14:39:20 crc kubenswrapper[4790]: I1124 14:39:20.209420 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7dbb6d63-d31f-4951-ac0c-5ec7309c129a-credential-keys\") pod \"keystone-787bf45df4-blpgw\" (UID: \"7dbb6d63-d31f-4951-ac0c-5ec7309c129a\") " pod="openstack/keystone-787bf45df4-blpgw" Nov 24 14:39:20 crc kubenswrapper[4790]: I1124 14:39:20.209477 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7dbb6d63-d31f-4951-ac0c-5ec7309c129a-fernet-keys\") pod \"keystone-787bf45df4-blpgw\" (UID: \"7dbb6d63-d31f-4951-ac0c-5ec7309c129a\") " pod="openstack/keystone-787bf45df4-blpgw" Nov 24 14:39:20 crc kubenswrapper[4790]: I1124 14:39:20.209530 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdmc5\" (UniqueName: \"kubernetes.io/projected/7dbb6d63-d31f-4951-ac0c-5ec7309c129a-kube-api-access-bdmc5\") pod \"keystone-787bf45df4-blpgw\" (UID: \"7dbb6d63-d31f-4951-ac0c-5ec7309c129a\") " pod="openstack/keystone-787bf45df4-blpgw" Nov 24 14:39:20 crc kubenswrapper[4790]: I1124 14:39:20.209551 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dbb6d63-d31f-4951-ac0c-5ec7309c129a-combined-ca-bundle\") pod \"keystone-787bf45df4-blpgw\" (UID: \"7dbb6d63-d31f-4951-ac0c-5ec7309c129a\") " pod="openstack/keystone-787bf45df4-blpgw" Nov 24 14:39:20 crc kubenswrapper[4790]: I1124 14:39:20.215047 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dbb6d63-d31f-4951-ac0c-5ec7309c129a-scripts\") pod \"keystone-787bf45df4-blpgw\" (UID: \"7dbb6d63-d31f-4951-ac0c-5ec7309c129a\") " pod="openstack/keystone-787bf45df4-blpgw" Nov 24 14:39:20 crc kubenswrapper[4790]: I1124 14:39:20.215317 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dbb6d63-d31f-4951-ac0c-5ec7309c129a-combined-ca-bundle\") pod \"keystone-787bf45df4-blpgw\" (UID: \"7dbb6d63-d31f-4951-ac0c-5ec7309c129a\") " pod="openstack/keystone-787bf45df4-blpgw" Nov 24 14:39:20 crc kubenswrapper[4790]: I1124 14:39:20.215371 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7dbb6d63-d31f-4951-ac0c-5ec7309c129a-credential-keys\") pod \"keystone-787bf45df4-blpgw\" (UID: \"7dbb6d63-d31f-4951-ac0c-5ec7309c129a\") " pod="openstack/keystone-787bf45df4-blpgw" Nov 24 14:39:20 crc kubenswrapper[4790]: I1124 14:39:20.215786 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dbb6d63-d31f-4951-ac0c-5ec7309c129a-config-data\") pod \"keystone-787bf45df4-blpgw\" (UID: \"7dbb6d63-d31f-4951-ac0c-5ec7309c129a\") " pod="openstack/keystone-787bf45df4-blpgw" Nov 24 14:39:20 crc kubenswrapper[4790]: I1124 14:39:20.216614 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7dbb6d63-d31f-4951-ac0c-5ec7309c129a-fernet-keys\") pod \"keystone-787bf45df4-blpgw\" (UID: \"7dbb6d63-d31f-4951-ac0c-5ec7309c129a\") " pod="openstack/keystone-787bf45df4-blpgw" Nov 24 14:39:20 crc kubenswrapper[4790]: I1124 14:39:20.231406 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdmc5\" (UniqueName: \"kubernetes.io/projected/7dbb6d63-d31f-4951-ac0c-5ec7309c129a-kube-api-access-bdmc5\") pod \"keystone-787bf45df4-blpgw\" (UID: \"7dbb6d63-d31f-4951-ac0c-5ec7309c129a\") " pod="openstack/keystone-787bf45df4-blpgw" Nov 24 14:39:20 crc kubenswrapper[4790]: I1124 14:39:20.329120 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-787bf45df4-blpgw" Nov 24 14:39:20 crc kubenswrapper[4790]: I1124 14:39:20.837020 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-787bf45df4-blpgw"] Nov 24 14:39:20 crc kubenswrapper[4790]: W1124 14:39:20.847690 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7dbb6d63_d31f_4951_ac0c_5ec7309c129a.slice/crio-add405ba7ad3f5adfd390ca38f0f555cfdec4bf01541e3090125099cd262bbe1 WatchSource:0}: Error finding container add405ba7ad3f5adfd390ca38f0f555cfdec4bf01541e3090125099cd262bbe1: Status 404 returned error can't find the container with id add405ba7ad3f5adfd390ca38f0f555cfdec4bf01541e3090125099cd262bbe1 Nov 24 14:39:20 crc kubenswrapper[4790]: I1124 14:39:20.873091 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-787bf45df4-blpgw" event={"ID":"7dbb6d63-d31f-4951-ac0c-5ec7309c129a","Type":"ContainerStarted","Data":"add405ba7ad3f5adfd390ca38f0f555cfdec4bf01541e3090125099cd262bbe1"} Nov 24 14:39:21 crc kubenswrapper[4790]: I1124 14:39:21.887197 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-787bf45df4-blpgw" event={"ID":"7dbb6d63-d31f-4951-ac0c-5ec7309c129a","Type":"ContainerStarted","Data":"32f97b16a026355a09348aebf798faf03f89452836a8d3e7c27bdef41321c0f8"} Nov 24 14:39:21 crc kubenswrapper[4790]: I1124 14:39:21.888156 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-787bf45df4-blpgw" Nov 24 14:39:21 crc kubenswrapper[4790]: I1124 14:39:21.910580 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-787bf45df4-blpgw" podStartSLOduration=2.910556159 podStartE2EDuration="2.910556159s" podCreationTimestamp="2025-11-24 14:39:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:39:21.905739645 +0000 UTC m=+5210.285633307" watchObservedRunningTime="2025-11-24 14:39:21.910556159 +0000 UTC m=+5210.290449821" Nov 24 14:39:43 crc kubenswrapper[4790]: I1124 14:39:43.939301 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:39:43 crc kubenswrapper[4790]: I1124 14:39:43.939958 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:39:47 crc kubenswrapper[4790]: I1124 14:39:47.018619 4790 scope.go:117] "RemoveContainer" containerID="798103c1fdecc46c8858e184adbd64a086eafc76c05f970f4526cde26b4c9719" Nov 24 14:39:47 crc kubenswrapper[4790]: I1124 14:39:47.057753 4790 scope.go:117] "RemoveContainer" containerID="37e69b339f97ae0520e33f147dfb96088843ee20fb88a78f589302602baa82d1" Nov 24 14:39:47 crc kubenswrapper[4790]: I1124 14:39:47.107766 4790 scope.go:117] "RemoveContainer" containerID="01b7eddcd71f1f402684449f51f27b0a55c1785a6727bc6d135cff897cba5157" Nov 24 14:39:47 crc kubenswrapper[4790]: I1124 14:39:47.169659 4790 scope.go:117] "RemoveContainer" containerID="ca0870a13339f5a612e870cbdb9d7e0b9dfe2acfb69a8e31c89d5756866515c3" Nov 24 14:39:47 crc kubenswrapper[4790]: I1124 14:39:47.212910 4790 scope.go:117] "RemoveContainer" containerID="aa3082ba93f1d59782e9be84026e161036aff7db70b05cb7ff37f9566a5819c3" Nov 24 14:39:47 crc kubenswrapper[4790]: I1124 14:39:47.257387 4790 scope.go:117] "RemoveContainer" containerID="37e005f44c60754f8dcfc8dda04f4ecba44c5a6391db55078ba6521473c59745" Nov 24 14:39:47 crc kubenswrapper[4790]: I1124 14:39:47.289813 4790 scope.go:117] "RemoveContainer" containerID="c091dbbfabd40597bf770f01f653a630e1c315e029f65d1cf4747dde8a4a16a8" Nov 24 14:39:51 crc kubenswrapper[4790]: I1124 14:39:51.761677 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-787bf45df4-blpgw" Nov 24 14:39:54 crc kubenswrapper[4790]: I1124 14:39:54.303914 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 24 14:39:54 crc kubenswrapper[4790]: I1124 14:39:54.306434 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 14:39:54 crc kubenswrapper[4790]: I1124 14:39:54.310617 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 24 14:39:54 crc kubenswrapper[4790]: I1124 14:39:54.310931 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 24 14:39:54 crc kubenswrapper[4790]: I1124 14:39:54.311024 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-2sqf9" Nov 24 14:39:54 crc kubenswrapper[4790]: I1124 14:39:54.337724 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 14:39:54 crc kubenswrapper[4790]: I1124 14:39:54.365116 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 24 14:39:54 crc kubenswrapper[4790]: E1124 14:39:54.366197 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-sxhxd openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[kube-api-access-sxhxd openstack-config openstack-config-secret]: context canceled" pod="openstack/openstackclient" podUID="d0c8eb9a-8bc0-4765-a6b0-bc914666202b" Nov 24 14:39:54 crc kubenswrapper[4790]: I1124 14:39:54.375219 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 24 14:39:54 crc kubenswrapper[4790]: I1124 14:39:54.418414 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 24 14:39:54 crc kubenswrapper[4790]: I1124 14:39:54.419733 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 14:39:54 crc kubenswrapper[4790]: I1124 14:39:54.438225 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 14:39:54 crc kubenswrapper[4790]: I1124 14:39:54.503294 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rm9l9\" (UniqueName: \"kubernetes.io/projected/b5399fd5-dda4-4a13-954f-76a99b1577a3-kube-api-access-rm9l9\") pod \"openstackclient\" (UID: \"b5399fd5-dda4-4a13-954f-76a99b1577a3\") " pod="openstack/openstackclient" Nov 24 14:39:54 crc kubenswrapper[4790]: I1124 14:39:54.503354 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b5399fd5-dda4-4a13-954f-76a99b1577a3-openstack-config\") pod \"openstackclient\" (UID: \"b5399fd5-dda4-4a13-954f-76a99b1577a3\") " pod="openstack/openstackclient" Nov 24 14:39:54 crc kubenswrapper[4790]: I1124 14:39:54.503389 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b5399fd5-dda4-4a13-954f-76a99b1577a3-openstack-config-secret\") pod \"openstackclient\" (UID: \"b5399fd5-dda4-4a13-954f-76a99b1577a3\") " pod="openstack/openstackclient" Nov 24 14:39:54 crc kubenswrapper[4790]: I1124 14:39:54.605813 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rm9l9\" (UniqueName: \"kubernetes.io/projected/b5399fd5-dda4-4a13-954f-76a99b1577a3-kube-api-access-rm9l9\") pod \"openstackclient\" (UID: \"b5399fd5-dda4-4a13-954f-76a99b1577a3\") " pod="openstack/openstackclient" Nov 24 14:39:54 crc kubenswrapper[4790]: I1124 14:39:54.606432 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b5399fd5-dda4-4a13-954f-76a99b1577a3-openstack-config\") pod \"openstackclient\" (UID: \"b5399fd5-dda4-4a13-954f-76a99b1577a3\") " pod="openstack/openstackclient" Nov 24 14:39:54 crc kubenswrapper[4790]: I1124 14:39:54.606609 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b5399fd5-dda4-4a13-954f-76a99b1577a3-openstack-config-secret\") pod \"openstackclient\" (UID: \"b5399fd5-dda4-4a13-954f-76a99b1577a3\") " pod="openstack/openstackclient" Nov 24 14:39:54 crc kubenswrapper[4790]: I1124 14:39:54.608013 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b5399fd5-dda4-4a13-954f-76a99b1577a3-openstack-config\") pod \"openstackclient\" (UID: \"b5399fd5-dda4-4a13-954f-76a99b1577a3\") " pod="openstack/openstackclient" Nov 24 14:39:54 crc kubenswrapper[4790]: I1124 14:39:54.614161 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b5399fd5-dda4-4a13-954f-76a99b1577a3-openstack-config-secret\") pod \"openstackclient\" (UID: \"b5399fd5-dda4-4a13-954f-76a99b1577a3\") " pod="openstack/openstackclient" Nov 24 14:39:54 crc kubenswrapper[4790]: I1124 14:39:54.629835 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rm9l9\" (UniqueName: \"kubernetes.io/projected/b5399fd5-dda4-4a13-954f-76a99b1577a3-kube-api-access-rm9l9\") pod \"openstackclient\" (UID: \"b5399fd5-dda4-4a13-954f-76a99b1577a3\") " pod="openstack/openstackclient" Nov 24 14:39:54 crc kubenswrapper[4790]: I1124 14:39:54.747336 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 14:39:55 crc kubenswrapper[4790]: I1124 14:39:55.265825 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 14:39:55 crc kubenswrapper[4790]: I1124 14:39:55.324163 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"b5399fd5-dda4-4a13-954f-76a99b1577a3","Type":"ContainerStarted","Data":"72ae5d4b1d34d08f0e7991d81f4e94ab7feaefe95d875636bc630acdfc51c01d"} Nov 24 14:39:55 crc kubenswrapper[4790]: I1124 14:39:55.324168 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 14:39:55 crc kubenswrapper[4790]: I1124 14:39:55.329771 4790 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="d0c8eb9a-8bc0-4765-a6b0-bc914666202b" podUID="b5399fd5-dda4-4a13-954f-76a99b1577a3" Nov 24 14:39:55 crc kubenswrapper[4790]: I1124 14:39:55.338118 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 14:39:56 crc kubenswrapper[4790]: I1124 14:39:56.341710 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0c8eb9a-8bc0-4765-a6b0-bc914666202b" path="/var/lib/kubelet/pods/d0c8eb9a-8bc0-4765-a6b0-bc914666202b/volumes" Nov 24 14:39:56 crc kubenswrapper[4790]: I1124 14:39:56.355379 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 14:39:56 crc kubenswrapper[4790]: I1124 14:39:56.357827 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"b5399fd5-dda4-4a13-954f-76a99b1577a3","Type":"ContainerStarted","Data":"51ce6763398e66fcf80424e574cd3875e846b87aaa373c269f452945a5f9b9ce"} Nov 24 14:39:56 crc kubenswrapper[4790]: I1124 14:39:56.392238 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.3922048289999998 podStartE2EDuration="2.392204829s" podCreationTimestamp="2025-11-24 14:39:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:39:56.387360354 +0000 UTC m=+5244.767254116" watchObservedRunningTime="2025-11-24 14:39:56.392204829 +0000 UTC m=+5244.772098521" Nov 24 14:39:56 crc kubenswrapper[4790]: I1124 14:39:56.393780 4790 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="d0c8eb9a-8bc0-4765-a6b0-bc914666202b" podUID="b5399fd5-dda4-4a13-954f-76a99b1577a3" Nov 24 14:40:13 crc kubenswrapper[4790]: I1124 14:40:13.939265 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:40:13 crc kubenswrapper[4790]: I1124 14:40:13.940439 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:40:43 crc kubenswrapper[4790]: I1124 14:40:43.938802 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:40:43 crc kubenswrapper[4790]: I1124 14:40:43.940000 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:40:43 crc kubenswrapper[4790]: I1124 14:40:43.940102 4790 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 14:40:43 crc kubenswrapper[4790]: I1124 14:40:43.941254 4790 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"92c086e44ccb28a62777d612e499209f7483422ca70b460ce6ec73c416e8e8b8"} pod="openshift-machine-config-operator/machine-config-daemon-xz49t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:40:43 crc kubenswrapper[4790]: I1124 14:40:43.941375 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" containerID="cri-o://92c086e44ccb28a62777d612e499209f7483422ca70b460ce6ec73c416e8e8b8" gracePeriod=600 Nov 24 14:40:44 crc kubenswrapper[4790]: E1124 14:40:44.074743 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:40:44 crc kubenswrapper[4790]: I1124 14:40:44.948727 4790 generic.go:334] "Generic (PLEG): container finished" podID="0d73b133-48f0-455f-8f6a-742e633f631a" containerID="92c086e44ccb28a62777d612e499209f7483422ca70b460ce6ec73c416e8e8b8" exitCode=0 Nov 24 14:40:44 crc kubenswrapper[4790]: I1124 14:40:44.948800 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerDied","Data":"92c086e44ccb28a62777d612e499209f7483422ca70b460ce6ec73c416e8e8b8"} Nov 24 14:40:44 crc kubenswrapper[4790]: I1124 14:40:44.948859 4790 scope.go:117] "RemoveContainer" containerID="1551c55c728005d2d8114bb5394315ba445d97c4e7ab35c743e2f23da5188013" Nov 24 14:40:44 crc kubenswrapper[4790]: I1124 14:40:44.952286 4790 scope.go:117] "RemoveContainer" containerID="92c086e44ccb28a62777d612e499209f7483422ca70b460ce6ec73c416e8e8b8" Nov 24 14:40:44 crc kubenswrapper[4790]: E1124 14:40:44.959729 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:40:58 crc kubenswrapper[4790]: I1124 14:40:58.315445 4790 scope.go:117] "RemoveContainer" containerID="92c086e44ccb28a62777d612e499209f7483422ca70b460ce6ec73c416e8e8b8" Nov 24 14:40:58 crc kubenswrapper[4790]: E1124 14:40:58.317965 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:41:09 crc kubenswrapper[4790]: I1124 14:41:09.315181 4790 scope.go:117] "RemoveContainer" containerID="92c086e44ccb28a62777d612e499209f7483422ca70b460ce6ec73c416e8e8b8" Nov 24 14:41:09 crc kubenswrapper[4790]: E1124 14:41:09.316482 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:41:22 crc kubenswrapper[4790]: I1124 14:41:22.323091 4790 scope.go:117] "RemoveContainer" containerID="92c086e44ccb28a62777d612e499209f7483422ca70b460ce6ec73c416e8e8b8" Nov 24 14:41:22 crc kubenswrapper[4790]: E1124 14:41:22.324424 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:41:30 crc kubenswrapper[4790]: E1124 14:41:30.557857 4790 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.69:38566->38.129.56.69:36625: write tcp 38.129.56.69:38566->38.129.56.69:36625: write: connection reset by peer Nov 24 14:41:34 crc kubenswrapper[4790]: I1124 14:41:34.314944 4790 scope.go:117] "RemoveContainer" containerID="92c086e44ccb28a62777d612e499209f7483422ca70b460ce6ec73c416e8e8b8" Nov 24 14:41:34 crc kubenswrapper[4790]: E1124 14:41:34.316140 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:41:36 crc kubenswrapper[4790]: I1124 14:41:36.693303 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-60fd-account-create-h4ql2"] Nov 24 14:41:36 crc kubenswrapper[4790]: I1124 14:41:36.695466 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-60fd-account-create-h4ql2" Nov 24 14:41:36 crc kubenswrapper[4790]: I1124 14:41:36.697914 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 24 14:41:36 crc kubenswrapper[4790]: I1124 14:41:36.702509 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-qvmkw"] Nov 24 14:41:36 crc kubenswrapper[4790]: I1124 14:41:36.704441 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-qvmkw" Nov 24 14:41:36 crc kubenswrapper[4790]: I1124 14:41:36.711789 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-60fd-account-create-h4ql2"] Nov 24 14:41:36 crc kubenswrapper[4790]: I1124 14:41:36.720941 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-qvmkw"] Nov 24 14:41:36 crc kubenswrapper[4790]: I1124 14:41:36.725842 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5kvs\" (UniqueName: \"kubernetes.io/projected/b7e7c18a-1b25-4d46-8cc5-77110ecc453d-kube-api-access-z5kvs\") pod \"barbican-60fd-account-create-h4ql2\" (UID: \"b7e7c18a-1b25-4d46-8cc5-77110ecc453d\") " pod="openstack/barbican-60fd-account-create-h4ql2" Nov 24 14:41:36 crc kubenswrapper[4790]: I1124 14:41:36.726023 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8sk7\" (UniqueName: \"kubernetes.io/projected/e6dbd003-4923-4aee-a240-7debd90a4a5b-kube-api-access-j8sk7\") pod \"barbican-db-create-qvmkw\" (UID: \"e6dbd003-4923-4aee-a240-7debd90a4a5b\") " pod="openstack/barbican-db-create-qvmkw" Nov 24 14:41:36 crc kubenswrapper[4790]: I1124 14:41:36.726075 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6dbd003-4923-4aee-a240-7debd90a4a5b-operator-scripts\") pod \"barbican-db-create-qvmkw\" (UID: \"e6dbd003-4923-4aee-a240-7debd90a4a5b\") " pod="openstack/barbican-db-create-qvmkw" Nov 24 14:41:36 crc kubenswrapper[4790]: I1124 14:41:36.726106 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7e7c18a-1b25-4d46-8cc5-77110ecc453d-operator-scripts\") pod \"barbican-60fd-account-create-h4ql2\" (UID: \"b7e7c18a-1b25-4d46-8cc5-77110ecc453d\") " pod="openstack/barbican-60fd-account-create-h4ql2" Nov 24 14:41:36 crc kubenswrapper[4790]: I1124 14:41:36.828330 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8sk7\" (UniqueName: \"kubernetes.io/projected/e6dbd003-4923-4aee-a240-7debd90a4a5b-kube-api-access-j8sk7\") pod \"barbican-db-create-qvmkw\" (UID: \"e6dbd003-4923-4aee-a240-7debd90a4a5b\") " pod="openstack/barbican-db-create-qvmkw" Nov 24 14:41:36 crc kubenswrapper[4790]: I1124 14:41:36.828429 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6dbd003-4923-4aee-a240-7debd90a4a5b-operator-scripts\") pod \"barbican-db-create-qvmkw\" (UID: \"e6dbd003-4923-4aee-a240-7debd90a4a5b\") " pod="openstack/barbican-db-create-qvmkw" Nov 24 14:41:36 crc kubenswrapper[4790]: I1124 14:41:36.828497 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7e7c18a-1b25-4d46-8cc5-77110ecc453d-operator-scripts\") pod \"barbican-60fd-account-create-h4ql2\" (UID: \"b7e7c18a-1b25-4d46-8cc5-77110ecc453d\") " pod="openstack/barbican-60fd-account-create-h4ql2" Nov 24 14:41:36 crc kubenswrapper[4790]: I1124 14:41:36.828554 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5kvs\" (UniqueName: \"kubernetes.io/projected/b7e7c18a-1b25-4d46-8cc5-77110ecc453d-kube-api-access-z5kvs\") pod \"barbican-60fd-account-create-h4ql2\" (UID: \"b7e7c18a-1b25-4d46-8cc5-77110ecc453d\") " pod="openstack/barbican-60fd-account-create-h4ql2" Nov 24 14:41:36 crc kubenswrapper[4790]: I1124 14:41:36.829744 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6dbd003-4923-4aee-a240-7debd90a4a5b-operator-scripts\") pod \"barbican-db-create-qvmkw\" (UID: \"e6dbd003-4923-4aee-a240-7debd90a4a5b\") " pod="openstack/barbican-db-create-qvmkw" Nov 24 14:41:36 crc kubenswrapper[4790]: I1124 14:41:36.830144 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7e7c18a-1b25-4d46-8cc5-77110ecc453d-operator-scripts\") pod \"barbican-60fd-account-create-h4ql2\" (UID: \"b7e7c18a-1b25-4d46-8cc5-77110ecc453d\") " pod="openstack/barbican-60fd-account-create-h4ql2" Nov 24 14:41:36 crc kubenswrapper[4790]: I1124 14:41:36.852941 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5kvs\" (UniqueName: \"kubernetes.io/projected/b7e7c18a-1b25-4d46-8cc5-77110ecc453d-kube-api-access-z5kvs\") pod \"barbican-60fd-account-create-h4ql2\" (UID: \"b7e7c18a-1b25-4d46-8cc5-77110ecc453d\") " pod="openstack/barbican-60fd-account-create-h4ql2" Nov 24 14:41:36 crc kubenswrapper[4790]: I1124 14:41:36.853429 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8sk7\" (UniqueName: \"kubernetes.io/projected/e6dbd003-4923-4aee-a240-7debd90a4a5b-kube-api-access-j8sk7\") pod \"barbican-db-create-qvmkw\" (UID: \"e6dbd003-4923-4aee-a240-7debd90a4a5b\") " pod="openstack/barbican-db-create-qvmkw" Nov 24 14:41:37 crc kubenswrapper[4790]: I1124 14:41:37.025333 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-60fd-account-create-h4ql2" Nov 24 14:41:37 crc kubenswrapper[4790]: I1124 14:41:37.038603 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-qvmkw" Nov 24 14:41:37 crc kubenswrapper[4790]: I1124 14:41:37.314103 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-60fd-account-create-h4ql2"] Nov 24 14:41:37 crc kubenswrapper[4790]: I1124 14:41:37.429594 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-60fd-account-create-h4ql2" event={"ID":"b7e7c18a-1b25-4d46-8cc5-77110ecc453d","Type":"ContainerStarted","Data":"c8df3ca4a993d7b2e2b58e34f887a62e536bb72a8c65238e016a7dc7fb0e2b6a"} Nov 24 14:41:37 crc kubenswrapper[4790]: I1124 14:41:37.585932 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-qvmkw"] Nov 24 14:41:37 crc kubenswrapper[4790]: W1124 14:41:37.591838 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode6dbd003_4923_4aee_a240_7debd90a4a5b.slice/crio-2fe97fa354374ce8cb1481509121b6f011c6077e1410d5fa4ad637f39a2c2c28 WatchSource:0}: Error finding container 2fe97fa354374ce8cb1481509121b6f011c6077e1410d5fa4ad637f39a2c2c28: Status 404 returned error can't find the container with id 2fe97fa354374ce8cb1481509121b6f011c6077e1410d5fa4ad637f39a2c2c28 Nov 24 14:41:38 crc kubenswrapper[4790]: I1124 14:41:38.456637 4790 generic.go:334] "Generic (PLEG): container finished" podID="e6dbd003-4923-4aee-a240-7debd90a4a5b" containerID="355778f4672909263ed17cc3bf2bd12f174b827d9d95c64ad4a0d4542fa66ea1" exitCode=0 Nov 24 14:41:38 crc kubenswrapper[4790]: I1124 14:41:38.456770 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-qvmkw" event={"ID":"e6dbd003-4923-4aee-a240-7debd90a4a5b","Type":"ContainerDied","Data":"355778f4672909263ed17cc3bf2bd12f174b827d9d95c64ad4a0d4542fa66ea1"} Nov 24 14:41:38 crc kubenswrapper[4790]: I1124 14:41:38.457270 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-qvmkw" event={"ID":"e6dbd003-4923-4aee-a240-7debd90a4a5b","Type":"ContainerStarted","Data":"2fe97fa354374ce8cb1481509121b6f011c6077e1410d5fa4ad637f39a2c2c28"} Nov 24 14:41:38 crc kubenswrapper[4790]: I1124 14:41:38.466019 4790 generic.go:334] "Generic (PLEG): container finished" podID="b7e7c18a-1b25-4d46-8cc5-77110ecc453d" containerID="6588be2e81cf6ac91d6145cc3944b2885e9e86a11a03dbf6ad7632b91ca0821d" exitCode=0 Nov 24 14:41:38 crc kubenswrapper[4790]: I1124 14:41:38.466103 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-60fd-account-create-h4ql2" event={"ID":"b7e7c18a-1b25-4d46-8cc5-77110ecc453d","Type":"ContainerDied","Data":"6588be2e81cf6ac91d6145cc3944b2885e9e86a11a03dbf6ad7632b91ca0821d"} Nov 24 14:41:39 crc kubenswrapper[4790]: I1124 14:41:39.915094 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-60fd-account-create-h4ql2" Nov 24 14:41:39 crc kubenswrapper[4790]: I1124 14:41:39.922000 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-qvmkw" Nov 24 14:41:40 crc kubenswrapper[4790]: I1124 14:41:40.098109 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7e7c18a-1b25-4d46-8cc5-77110ecc453d-operator-scripts\") pod \"b7e7c18a-1b25-4d46-8cc5-77110ecc453d\" (UID: \"b7e7c18a-1b25-4d46-8cc5-77110ecc453d\") " Nov 24 14:41:40 crc kubenswrapper[4790]: I1124 14:41:40.098251 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5kvs\" (UniqueName: \"kubernetes.io/projected/b7e7c18a-1b25-4d46-8cc5-77110ecc453d-kube-api-access-z5kvs\") pod \"b7e7c18a-1b25-4d46-8cc5-77110ecc453d\" (UID: \"b7e7c18a-1b25-4d46-8cc5-77110ecc453d\") " Nov 24 14:41:40 crc kubenswrapper[4790]: I1124 14:41:40.098340 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6dbd003-4923-4aee-a240-7debd90a4a5b-operator-scripts\") pod \"e6dbd003-4923-4aee-a240-7debd90a4a5b\" (UID: \"e6dbd003-4923-4aee-a240-7debd90a4a5b\") " Nov 24 14:41:40 crc kubenswrapper[4790]: I1124 14:41:40.098421 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8sk7\" (UniqueName: \"kubernetes.io/projected/e6dbd003-4923-4aee-a240-7debd90a4a5b-kube-api-access-j8sk7\") pod \"e6dbd003-4923-4aee-a240-7debd90a4a5b\" (UID: \"e6dbd003-4923-4aee-a240-7debd90a4a5b\") " Nov 24 14:41:40 crc kubenswrapper[4790]: I1124 14:41:40.099497 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6dbd003-4923-4aee-a240-7debd90a4a5b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e6dbd003-4923-4aee-a240-7debd90a4a5b" (UID: "e6dbd003-4923-4aee-a240-7debd90a4a5b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:41:40 crc kubenswrapper[4790]: I1124 14:41:40.099538 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7e7c18a-1b25-4d46-8cc5-77110ecc453d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b7e7c18a-1b25-4d46-8cc5-77110ecc453d" (UID: "b7e7c18a-1b25-4d46-8cc5-77110ecc453d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:41:40 crc kubenswrapper[4790]: I1124 14:41:40.110489 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6dbd003-4923-4aee-a240-7debd90a4a5b-kube-api-access-j8sk7" (OuterVolumeSpecName: "kube-api-access-j8sk7") pod "e6dbd003-4923-4aee-a240-7debd90a4a5b" (UID: "e6dbd003-4923-4aee-a240-7debd90a4a5b"). InnerVolumeSpecName "kube-api-access-j8sk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:41:40 crc kubenswrapper[4790]: I1124 14:41:40.110761 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7e7c18a-1b25-4d46-8cc5-77110ecc453d-kube-api-access-z5kvs" (OuterVolumeSpecName: "kube-api-access-z5kvs") pod "b7e7c18a-1b25-4d46-8cc5-77110ecc453d" (UID: "b7e7c18a-1b25-4d46-8cc5-77110ecc453d"). InnerVolumeSpecName "kube-api-access-z5kvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:41:40 crc kubenswrapper[4790]: I1124 14:41:40.202683 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7e7c18a-1b25-4d46-8cc5-77110ecc453d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:40 crc kubenswrapper[4790]: I1124 14:41:40.202762 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5kvs\" (UniqueName: \"kubernetes.io/projected/b7e7c18a-1b25-4d46-8cc5-77110ecc453d-kube-api-access-z5kvs\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:40 crc kubenswrapper[4790]: I1124 14:41:40.202793 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6dbd003-4923-4aee-a240-7debd90a4a5b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:40 crc kubenswrapper[4790]: I1124 14:41:40.202823 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8sk7\" (UniqueName: \"kubernetes.io/projected/e6dbd003-4923-4aee-a240-7debd90a4a5b-kube-api-access-j8sk7\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:40 crc kubenswrapper[4790]: I1124 14:41:40.494634 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-qvmkw" event={"ID":"e6dbd003-4923-4aee-a240-7debd90a4a5b","Type":"ContainerDied","Data":"2fe97fa354374ce8cb1481509121b6f011c6077e1410d5fa4ad637f39a2c2c28"} Nov 24 14:41:40 crc kubenswrapper[4790]: I1124 14:41:40.494714 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2fe97fa354374ce8cb1481509121b6f011c6077e1410d5fa4ad637f39a2c2c28" Nov 24 14:41:40 crc kubenswrapper[4790]: I1124 14:41:40.495246 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-qvmkw" Nov 24 14:41:40 crc kubenswrapper[4790]: I1124 14:41:40.497815 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-60fd-account-create-h4ql2" event={"ID":"b7e7c18a-1b25-4d46-8cc5-77110ecc453d","Type":"ContainerDied","Data":"c8df3ca4a993d7b2e2b58e34f887a62e536bb72a8c65238e016a7dc7fb0e2b6a"} Nov 24 14:41:40 crc kubenswrapper[4790]: I1124 14:41:40.497929 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8df3ca4a993d7b2e2b58e34f887a62e536bb72a8c65238e016a7dc7fb0e2b6a" Nov 24 14:41:40 crc kubenswrapper[4790]: I1124 14:41:40.498173 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-60fd-account-create-h4ql2" Nov 24 14:41:42 crc kubenswrapper[4790]: I1124 14:41:42.065653 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-bqxsj"] Nov 24 14:41:42 crc kubenswrapper[4790]: E1124 14:41:42.066475 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6dbd003-4923-4aee-a240-7debd90a4a5b" containerName="mariadb-database-create" Nov 24 14:41:42 crc kubenswrapper[4790]: I1124 14:41:42.066495 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6dbd003-4923-4aee-a240-7debd90a4a5b" containerName="mariadb-database-create" Nov 24 14:41:42 crc kubenswrapper[4790]: E1124 14:41:42.066518 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7e7c18a-1b25-4d46-8cc5-77110ecc453d" containerName="mariadb-account-create" Nov 24 14:41:42 crc kubenswrapper[4790]: I1124 14:41:42.066526 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7e7c18a-1b25-4d46-8cc5-77110ecc453d" containerName="mariadb-account-create" Nov 24 14:41:42 crc kubenswrapper[4790]: I1124 14:41:42.066744 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7e7c18a-1b25-4d46-8cc5-77110ecc453d" containerName="mariadb-account-create" Nov 24 14:41:42 crc kubenswrapper[4790]: I1124 14:41:42.066782 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6dbd003-4923-4aee-a240-7debd90a4a5b" containerName="mariadb-database-create" Nov 24 14:41:42 crc kubenswrapper[4790]: I1124 14:41:42.068083 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-bqxsj" Nov 24 14:41:42 crc kubenswrapper[4790]: I1124 14:41:42.071143 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-hjfnb" Nov 24 14:41:42 crc kubenswrapper[4790]: I1124 14:41:42.072486 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 24 14:41:42 crc kubenswrapper[4790]: I1124 14:41:42.075371 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-bqxsj"] Nov 24 14:41:42 crc kubenswrapper[4790]: I1124 14:41:42.252139 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb099850-58c3-4280-8c66-865463ca19c8-combined-ca-bundle\") pod \"barbican-db-sync-bqxsj\" (UID: \"bb099850-58c3-4280-8c66-865463ca19c8\") " pod="openstack/barbican-db-sync-bqxsj" Nov 24 14:41:42 crc kubenswrapper[4790]: I1124 14:41:42.252194 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7n7v6\" (UniqueName: \"kubernetes.io/projected/bb099850-58c3-4280-8c66-865463ca19c8-kube-api-access-7n7v6\") pod \"barbican-db-sync-bqxsj\" (UID: \"bb099850-58c3-4280-8c66-865463ca19c8\") " pod="openstack/barbican-db-sync-bqxsj" Nov 24 14:41:42 crc kubenswrapper[4790]: I1124 14:41:42.252218 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bb099850-58c3-4280-8c66-865463ca19c8-db-sync-config-data\") pod \"barbican-db-sync-bqxsj\" (UID: \"bb099850-58c3-4280-8c66-865463ca19c8\") " pod="openstack/barbican-db-sync-bqxsj" Nov 24 14:41:42 crc kubenswrapper[4790]: I1124 14:41:42.354621 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb099850-58c3-4280-8c66-865463ca19c8-combined-ca-bundle\") pod \"barbican-db-sync-bqxsj\" (UID: \"bb099850-58c3-4280-8c66-865463ca19c8\") " pod="openstack/barbican-db-sync-bqxsj" Nov 24 14:41:42 crc kubenswrapper[4790]: I1124 14:41:42.354727 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7n7v6\" (UniqueName: \"kubernetes.io/projected/bb099850-58c3-4280-8c66-865463ca19c8-kube-api-access-7n7v6\") pod \"barbican-db-sync-bqxsj\" (UID: \"bb099850-58c3-4280-8c66-865463ca19c8\") " pod="openstack/barbican-db-sync-bqxsj" Nov 24 14:41:42 crc kubenswrapper[4790]: I1124 14:41:42.354794 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bb099850-58c3-4280-8c66-865463ca19c8-db-sync-config-data\") pod \"barbican-db-sync-bqxsj\" (UID: \"bb099850-58c3-4280-8c66-865463ca19c8\") " pod="openstack/barbican-db-sync-bqxsj" Nov 24 14:41:42 crc kubenswrapper[4790]: I1124 14:41:42.365385 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb099850-58c3-4280-8c66-865463ca19c8-combined-ca-bundle\") pod \"barbican-db-sync-bqxsj\" (UID: \"bb099850-58c3-4280-8c66-865463ca19c8\") " pod="openstack/barbican-db-sync-bqxsj" Nov 24 14:41:42 crc kubenswrapper[4790]: I1124 14:41:42.365489 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bb099850-58c3-4280-8c66-865463ca19c8-db-sync-config-data\") pod \"barbican-db-sync-bqxsj\" (UID: \"bb099850-58c3-4280-8c66-865463ca19c8\") " pod="openstack/barbican-db-sync-bqxsj" Nov 24 14:41:42 crc kubenswrapper[4790]: I1124 14:41:42.390855 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7n7v6\" (UniqueName: \"kubernetes.io/projected/bb099850-58c3-4280-8c66-865463ca19c8-kube-api-access-7n7v6\") pod \"barbican-db-sync-bqxsj\" (UID: \"bb099850-58c3-4280-8c66-865463ca19c8\") " pod="openstack/barbican-db-sync-bqxsj" Nov 24 14:41:42 crc kubenswrapper[4790]: I1124 14:41:42.687631 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-bqxsj" Nov 24 14:41:43 crc kubenswrapper[4790]: I1124 14:41:43.267506 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-bqxsj"] Nov 24 14:41:43 crc kubenswrapper[4790]: W1124 14:41:43.276268 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb099850_58c3_4280_8c66_865463ca19c8.slice/crio-8183d5261e66034a3e2447bfffbc7ea7c98c91d6c4edcb5ab2e686a7952ea1f6 WatchSource:0}: Error finding container 8183d5261e66034a3e2447bfffbc7ea7c98c91d6c4edcb5ab2e686a7952ea1f6: Status 404 returned error can't find the container with id 8183d5261e66034a3e2447bfffbc7ea7c98c91d6c4edcb5ab2e686a7952ea1f6 Nov 24 14:41:43 crc kubenswrapper[4790]: I1124 14:41:43.533500 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-bqxsj" event={"ID":"bb099850-58c3-4280-8c66-865463ca19c8","Type":"ContainerStarted","Data":"8183d5261e66034a3e2447bfffbc7ea7c98c91d6c4edcb5ab2e686a7952ea1f6"} Nov 24 14:41:44 crc kubenswrapper[4790]: I1124 14:41:44.554087 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-bqxsj" event={"ID":"bb099850-58c3-4280-8c66-865463ca19c8","Type":"ContainerStarted","Data":"b84fe988a3b86185ef928788e240cae9609c3919645f171e1dc828daefa3244d"} Nov 24 14:41:44 crc kubenswrapper[4790]: I1124 14:41:44.581444 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-bqxsj" podStartSLOduration=2.581403579 podStartE2EDuration="2.581403579s" podCreationTimestamp="2025-11-24 14:41:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:41:44.576728899 +0000 UTC m=+5352.956622561" watchObservedRunningTime="2025-11-24 14:41:44.581403579 +0000 UTC m=+5352.961297291" Nov 24 14:41:45 crc kubenswrapper[4790]: I1124 14:41:45.570687 4790 generic.go:334] "Generic (PLEG): container finished" podID="bb099850-58c3-4280-8c66-865463ca19c8" containerID="b84fe988a3b86185ef928788e240cae9609c3919645f171e1dc828daefa3244d" exitCode=0 Nov 24 14:41:45 crc kubenswrapper[4790]: I1124 14:41:45.570768 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-bqxsj" event={"ID":"bb099850-58c3-4280-8c66-865463ca19c8","Type":"ContainerDied","Data":"b84fe988a3b86185ef928788e240cae9609c3919645f171e1dc828daefa3244d"} Nov 24 14:41:46 crc kubenswrapper[4790]: I1124 14:41:46.970436 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-bqxsj" Nov 24 14:41:47 crc kubenswrapper[4790]: I1124 14:41:47.165389 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7n7v6\" (UniqueName: \"kubernetes.io/projected/bb099850-58c3-4280-8c66-865463ca19c8-kube-api-access-7n7v6\") pod \"bb099850-58c3-4280-8c66-865463ca19c8\" (UID: \"bb099850-58c3-4280-8c66-865463ca19c8\") " Nov 24 14:41:47 crc kubenswrapper[4790]: I1124 14:41:47.165777 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb099850-58c3-4280-8c66-865463ca19c8-combined-ca-bundle\") pod \"bb099850-58c3-4280-8c66-865463ca19c8\" (UID: \"bb099850-58c3-4280-8c66-865463ca19c8\") " Nov 24 14:41:47 crc kubenswrapper[4790]: I1124 14:41:47.166709 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bb099850-58c3-4280-8c66-865463ca19c8-db-sync-config-data\") pod \"bb099850-58c3-4280-8c66-865463ca19c8\" (UID: \"bb099850-58c3-4280-8c66-865463ca19c8\") " Nov 24 14:41:47 crc kubenswrapper[4790]: I1124 14:41:47.176620 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb099850-58c3-4280-8c66-865463ca19c8-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "bb099850-58c3-4280-8c66-865463ca19c8" (UID: "bb099850-58c3-4280-8c66-865463ca19c8"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:41:47 crc kubenswrapper[4790]: I1124 14:41:47.181087 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb099850-58c3-4280-8c66-865463ca19c8-kube-api-access-7n7v6" (OuterVolumeSpecName: "kube-api-access-7n7v6") pod "bb099850-58c3-4280-8c66-865463ca19c8" (UID: "bb099850-58c3-4280-8c66-865463ca19c8"). InnerVolumeSpecName "kube-api-access-7n7v6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:41:47 crc kubenswrapper[4790]: I1124 14:41:47.213221 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb099850-58c3-4280-8c66-865463ca19c8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bb099850-58c3-4280-8c66-865463ca19c8" (UID: "bb099850-58c3-4280-8c66-865463ca19c8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:41:47 crc kubenswrapper[4790]: I1124 14:41:47.271227 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb099850-58c3-4280-8c66-865463ca19c8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:47 crc kubenswrapper[4790]: I1124 14:41:47.271283 4790 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bb099850-58c3-4280-8c66-865463ca19c8-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:47 crc kubenswrapper[4790]: I1124 14:41:47.271303 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7n7v6\" (UniqueName: \"kubernetes.io/projected/bb099850-58c3-4280-8c66-865463ca19c8-kube-api-access-7n7v6\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:47 crc kubenswrapper[4790]: I1124 14:41:47.315201 4790 scope.go:117] "RemoveContainer" containerID="92c086e44ccb28a62777d612e499209f7483422ca70b460ce6ec73c416e8e8b8" Nov 24 14:41:47 crc kubenswrapper[4790]: E1124 14:41:47.315849 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:41:47 crc kubenswrapper[4790]: I1124 14:41:47.596341 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-bqxsj" event={"ID":"bb099850-58c3-4280-8c66-865463ca19c8","Type":"ContainerDied","Data":"8183d5261e66034a3e2447bfffbc7ea7c98c91d6c4edcb5ab2e686a7952ea1f6"} Nov 24 14:41:47 crc kubenswrapper[4790]: I1124 14:41:47.596725 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8183d5261e66034a3e2447bfffbc7ea7c98c91d6c4edcb5ab2e686a7952ea1f6" Nov 24 14:41:47 crc kubenswrapper[4790]: I1124 14:41:47.596455 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-bqxsj" Nov 24 14:41:47 crc kubenswrapper[4790]: I1124 14:41:47.879613 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-846d64984d-vjkzr"] Nov 24 14:41:47 crc kubenswrapper[4790]: E1124 14:41:47.880052 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb099850-58c3-4280-8c66-865463ca19c8" containerName="barbican-db-sync" Nov 24 14:41:47 crc kubenswrapper[4790]: I1124 14:41:47.880067 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb099850-58c3-4280-8c66-865463ca19c8" containerName="barbican-db-sync" Nov 24 14:41:47 crc kubenswrapper[4790]: I1124 14:41:47.880264 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb099850-58c3-4280-8c66-865463ca19c8" containerName="barbican-db-sync" Nov 24 14:41:47 crc kubenswrapper[4790]: I1124 14:41:47.883285 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-846d64984d-vjkzr" Nov 24 14:41:47 crc kubenswrapper[4790]: I1124 14:41:47.891518 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 24 14:41:47 crc kubenswrapper[4790]: I1124 14:41:47.891791 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-hjfnb" Nov 24 14:41:47 crc kubenswrapper[4790]: I1124 14:41:47.896028 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 24 14:41:47 crc kubenswrapper[4790]: I1124 14:41:47.903184 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-6f48884455-sk4db"] Nov 24 14:41:47 crc kubenswrapper[4790]: I1124 14:41:47.916177 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6f48884455-sk4db" Nov 24 14:41:47 crc kubenswrapper[4790]: I1124 14:41:47.921378 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 24 14:41:47 crc kubenswrapper[4790]: I1124 14:41:47.926063 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-846d64984d-vjkzr"] Nov 24 14:41:47 crc kubenswrapper[4790]: I1124 14:41:47.937965 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6f48884455-sk4db"] Nov 24 14:41:47 crc kubenswrapper[4790]: I1124 14:41:47.988325 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkb4v\" (UniqueName: \"kubernetes.io/projected/18179e0b-5942-4780-8240-ea2496142b5c-kube-api-access-wkb4v\") pod \"barbican-keystone-listener-846d64984d-vjkzr\" (UID: \"18179e0b-5942-4780-8240-ea2496142b5c\") " pod="openstack/barbican-keystone-listener-846d64984d-vjkzr" Nov 24 14:41:47 crc kubenswrapper[4790]: I1124 14:41:47.988392 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18179e0b-5942-4780-8240-ea2496142b5c-config-data\") pod \"barbican-keystone-listener-846d64984d-vjkzr\" (UID: \"18179e0b-5942-4780-8240-ea2496142b5c\") " pod="openstack/barbican-keystone-listener-846d64984d-vjkzr" Nov 24 14:41:47 crc kubenswrapper[4790]: I1124 14:41:47.988426 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18179e0b-5942-4780-8240-ea2496142b5c-logs\") pod \"barbican-keystone-listener-846d64984d-vjkzr\" (UID: \"18179e0b-5942-4780-8240-ea2496142b5c\") " pod="openstack/barbican-keystone-listener-846d64984d-vjkzr" Nov 24 14:41:47 crc kubenswrapper[4790]: I1124 14:41:47.988462 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/18179e0b-5942-4780-8240-ea2496142b5c-config-data-custom\") pod \"barbican-keystone-listener-846d64984d-vjkzr\" (UID: \"18179e0b-5942-4780-8240-ea2496142b5c\") " pod="openstack/barbican-keystone-listener-846d64984d-vjkzr" Nov 24 14:41:47 crc kubenswrapper[4790]: I1124 14:41:47.988529 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18179e0b-5942-4780-8240-ea2496142b5c-combined-ca-bundle\") pod \"barbican-keystone-listener-846d64984d-vjkzr\" (UID: \"18179e0b-5942-4780-8240-ea2496142b5c\") " pod="openstack/barbican-keystone-listener-846d64984d-vjkzr" Nov 24 14:41:47 crc kubenswrapper[4790]: I1124 14:41:47.989729 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-c845456b5-5r449"] Nov 24 14:41:47 crc kubenswrapper[4790]: I1124 14:41:47.991517 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c845456b5-5r449" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.017813 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c845456b5-5r449"] Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.091090 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/18179e0b-5942-4780-8240-ea2496142b5c-config-data-custom\") pod \"barbican-keystone-listener-846d64984d-vjkzr\" (UID: \"18179e0b-5942-4780-8240-ea2496142b5c\") " pod="openstack/barbican-keystone-listener-846d64984d-vjkzr" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.091198 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f00df2c8-01c0-4cd2-a699-97fadd168c78-logs\") pod \"barbican-worker-6f48884455-sk4db\" (UID: \"f00df2c8-01c0-4cd2-a699-97fadd168c78\") " pod="openstack/barbican-worker-6f48884455-sk4db" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.091237 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgprw\" (UniqueName: \"kubernetes.io/projected/f00df2c8-01c0-4cd2-a699-97fadd168c78-kube-api-access-fgprw\") pod \"barbican-worker-6f48884455-sk4db\" (UID: \"f00df2c8-01c0-4cd2-a699-97fadd168c78\") " pod="openstack/barbican-worker-6f48884455-sk4db" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.091288 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18179e0b-5942-4780-8240-ea2496142b5c-combined-ca-bundle\") pod \"barbican-keystone-listener-846d64984d-vjkzr\" (UID: \"18179e0b-5942-4780-8240-ea2496142b5c\") " pod="openstack/barbican-keystone-listener-846d64984d-vjkzr" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.091333 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f00df2c8-01c0-4cd2-a699-97fadd168c78-config-data-custom\") pod \"barbican-worker-6f48884455-sk4db\" (UID: \"f00df2c8-01c0-4cd2-a699-97fadd168c78\") " pod="openstack/barbican-worker-6f48884455-sk4db" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.091362 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkb4v\" (UniqueName: \"kubernetes.io/projected/18179e0b-5942-4780-8240-ea2496142b5c-kube-api-access-wkb4v\") pod \"barbican-keystone-listener-846d64984d-vjkzr\" (UID: \"18179e0b-5942-4780-8240-ea2496142b5c\") " pod="openstack/barbican-keystone-listener-846d64984d-vjkzr" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.091406 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18179e0b-5942-4780-8240-ea2496142b5c-config-data\") pod \"barbican-keystone-listener-846d64984d-vjkzr\" (UID: \"18179e0b-5942-4780-8240-ea2496142b5c\") " pod="openstack/barbican-keystone-listener-846d64984d-vjkzr" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.091427 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f00df2c8-01c0-4cd2-a699-97fadd168c78-config-data\") pod \"barbican-worker-6f48884455-sk4db\" (UID: \"f00df2c8-01c0-4cd2-a699-97fadd168c78\") " pod="openstack/barbican-worker-6f48884455-sk4db" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.091466 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18179e0b-5942-4780-8240-ea2496142b5c-logs\") pod \"barbican-keystone-listener-846d64984d-vjkzr\" (UID: \"18179e0b-5942-4780-8240-ea2496142b5c\") " pod="openstack/barbican-keystone-listener-846d64984d-vjkzr" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.091485 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f00df2c8-01c0-4cd2-a699-97fadd168c78-combined-ca-bundle\") pod \"barbican-worker-6f48884455-sk4db\" (UID: \"f00df2c8-01c0-4cd2-a699-97fadd168c78\") " pod="openstack/barbican-worker-6f48884455-sk4db" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.096536 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18179e0b-5942-4780-8240-ea2496142b5c-logs\") pod \"barbican-keystone-listener-846d64984d-vjkzr\" (UID: \"18179e0b-5942-4780-8240-ea2496142b5c\") " pod="openstack/barbican-keystone-listener-846d64984d-vjkzr" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.114767 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/18179e0b-5942-4780-8240-ea2496142b5c-config-data-custom\") pod \"barbican-keystone-listener-846d64984d-vjkzr\" (UID: \"18179e0b-5942-4780-8240-ea2496142b5c\") " pod="openstack/barbican-keystone-listener-846d64984d-vjkzr" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.123221 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18179e0b-5942-4780-8240-ea2496142b5c-combined-ca-bundle\") pod \"barbican-keystone-listener-846d64984d-vjkzr\" (UID: \"18179e0b-5942-4780-8240-ea2496142b5c\") " pod="openstack/barbican-keystone-listener-846d64984d-vjkzr" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.127866 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkb4v\" (UniqueName: \"kubernetes.io/projected/18179e0b-5942-4780-8240-ea2496142b5c-kube-api-access-wkb4v\") pod \"barbican-keystone-listener-846d64984d-vjkzr\" (UID: \"18179e0b-5942-4780-8240-ea2496142b5c\") " pod="openstack/barbican-keystone-listener-846d64984d-vjkzr" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.131682 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18179e0b-5942-4780-8240-ea2496142b5c-config-data\") pod \"barbican-keystone-listener-846d64984d-vjkzr\" (UID: \"18179e0b-5942-4780-8240-ea2496142b5c\") " pod="openstack/barbican-keystone-listener-846d64984d-vjkzr" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.152155 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6c6d75b7d-nbj94"] Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.155727 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c6d75b7d-nbj94" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.162454 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6c6d75b7d-nbj94"] Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.162649 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.193730 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/276b7ec3-db22-4125-b19b-1f73ece2a2e0-ovsdbserver-sb\") pod \"dnsmasq-dns-c845456b5-5r449\" (UID: \"276b7ec3-db22-4125-b19b-1f73ece2a2e0\") " pod="openstack/dnsmasq-dns-c845456b5-5r449" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.193801 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f00df2c8-01c0-4cd2-a699-97fadd168c78-config-data\") pod \"barbican-worker-6f48884455-sk4db\" (UID: \"f00df2c8-01c0-4cd2-a699-97fadd168c78\") " pod="openstack/barbican-worker-6f48884455-sk4db" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.193841 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/276b7ec3-db22-4125-b19b-1f73ece2a2e0-config\") pod \"dnsmasq-dns-c845456b5-5r449\" (UID: \"276b7ec3-db22-4125-b19b-1f73ece2a2e0\") " pod="openstack/dnsmasq-dns-c845456b5-5r449" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.193897 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f00df2c8-01c0-4cd2-a699-97fadd168c78-combined-ca-bundle\") pod \"barbican-worker-6f48884455-sk4db\" (UID: \"f00df2c8-01c0-4cd2-a699-97fadd168c78\") " pod="openstack/barbican-worker-6f48884455-sk4db" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.193924 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/276b7ec3-db22-4125-b19b-1f73ece2a2e0-ovsdbserver-nb\") pod \"dnsmasq-dns-c845456b5-5r449\" (UID: \"276b7ec3-db22-4125-b19b-1f73ece2a2e0\") " pod="openstack/dnsmasq-dns-c845456b5-5r449" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.193991 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f00df2c8-01c0-4cd2-a699-97fadd168c78-logs\") pod \"barbican-worker-6f48884455-sk4db\" (UID: \"f00df2c8-01c0-4cd2-a699-97fadd168c78\") " pod="openstack/barbican-worker-6f48884455-sk4db" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.194020 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgprw\" (UniqueName: \"kubernetes.io/projected/f00df2c8-01c0-4cd2-a699-97fadd168c78-kube-api-access-fgprw\") pod \"barbican-worker-6f48884455-sk4db\" (UID: \"f00df2c8-01c0-4cd2-a699-97fadd168c78\") " pod="openstack/barbican-worker-6f48884455-sk4db" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.194056 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mktbd\" (UniqueName: \"kubernetes.io/projected/276b7ec3-db22-4125-b19b-1f73ece2a2e0-kube-api-access-mktbd\") pod \"dnsmasq-dns-c845456b5-5r449\" (UID: \"276b7ec3-db22-4125-b19b-1f73ece2a2e0\") " pod="openstack/dnsmasq-dns-c845456b5-5r449" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.194104 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f00df2c8-01c0-4cd2-a699-97fadd168c78-config-data-custom\") pod \"barbican-worker-6f48884455-sk4db\" (UID: \"f00df2c8-01c0-4cd2-a699-97fadd168c78\") " pod="openstack/barbican-worker-6f48884455-sk4db" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.194136 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/276b7ec3-db22-4125-b19b-1f73ece2a2e0-dns-svc\") pod \"dnsmasq-dns-c845456b5-5r449\" (UID: \"276b7ec3-db22-4125-b19b-1f73ece2a2e0\") " pod="openstack/dnsmasq-dns-c845456b5-5r449" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.194571 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f00df2c8-01c0-4cd2-a699-97fadd168c78-logs\") pod \"barbican-worker-6f48884455-sk4db\" (UID: \"f00df2c8-01c0-4cd2-a699-97fadd168c78\") " pod="openstack/barbican-worker-6f48884455-sk4db" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.201561 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f00df2c8-01c0-4cd2-a699-97fadd168c78-config-data-custom\") pod \"barbican-worker-6f48884455-sk4db\" (UID: \"f00df2c8-01c0-4cd2-a699-97fadd168c78\") " pod="openstack/barbican-worker-6f48884455-sk4db" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.201962 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f00df2c8-01c0-4cd2-a699-97fadd168c78-config-data\") pod \"barbican-worker-6f48884455-sk4db\" (UID: \"f00df2c8-01c0-4cd2-a699-97fadd168c78\") " pod="openstack/barbican-worker-6f48884455-sk4db" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.210872 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-846d64984d-vjkzr" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.212898 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f00df2c8-01c0-4cd2-a699-97fadd168c78-combined-ca-bundle\") pod \"barbican-worker-6f48884455-sk4db\" (UID: \"f00df2c8-01c0-4cd2-a699-97fadd168c78\") " pod="openstack/barbican-worker-6f48884455-sk4db" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.213525 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgprw\" (UniqueName: \"kubernetes.io/projected/f00df2c8-01c0-4cd2-a699-97fadd168c78-kube-api-access-fgprw\") pod \"barbican-worker-6f48884455-sk4db\" (UID: \"f00df2c8-01c0-4cd2-a699-97fadd168c78\") " pod="openstack/barbican-worker-6f48884455-sk4db" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.243820 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6f48884455-sk4db" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.296691 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/276b7ec3-db22-4125-b19b-1f73ece2a2e0-dns-svc\") pod \"dnsmasq-dns-c845456b5-5r449\" (UID: \"276b7ec3-db22-4125-b19b-1f73ece2a2e0\") " pod="openstack/dnsmasq-dns-c845456b5-5r449" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.296767 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0e05357-2caf-4e9c-b18d-d0fc75a067d7-config-data\") pod \"barbican-api-6c6d75b7d-nbj94\" (UID: \"b0e05357-2caf-4e9c-b18d-d0fc75a067d7\") " pod="openstack/barbican-api-6c6d75b7d-nbj94" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.296792 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/276b7ec3-db22-4125-b19b-1f73ece2a2e0-ovsdbserver-sb\") pod \"dnsmasq-dns-c845456b5-5r449\" (UID: \"276b7ec3-db22-4125-b19b-1f73ece2a2e0\") " pod="openstack/dnsmasq-dns-c845456b5-5r449" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.296814 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnjkp\" (UniqueName: \"kubernetes.io/projected/b0e05357-2caf-4e9c-b18d-d0fc75a067d7-kube-api-access-jnjkp\") pod \"barbican-api-6c6d75b7d-nbj94\" (UID: \"b0e05357-2caf-4e9c-b18d-d0fc75a067d7\") " pod="openstack/barbican-api-6c6d75b7d-nbj94" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.296977 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/276b7ec3-db22-4125-b19b-1f73ece2a2e0-config\") pod \"dnsmasq-dns-c845456b5-5r449\" (UID: \"276b7ec3-db22-4125-b19b-1f73ece2a2e0\") " pod="openstack/dnsmasq-dns-c845456b5-5r449" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.297023 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/276b7ec3-db22-4125-b19b-1f73ece2a2e0-ovsdbserver-nb\") pod \"dnsmasq-dns-c845456b5-5r449\" (UID: \"276b7ec3-db22-4125-b19b-1f73ece2a2e0\") " pod="openstack/dnsmasq-dns-c845456b5-5r449" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.297050 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b0e05357-2caf-4e9c-b18d-d0fc75a067d7-config-data-custom\") pod \"barbican-api-6c6d75b7d-nbj94\" (UID: \"b0e05357-2caf-4e9c-b18d-d0fc75a067d7\") " pod="openstack/barbican-api-6c6d75b7d-nbj94" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.297097 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0e05357-2caf-4e9c-b18d-d0fc75a067d7-logs\") pod \"barbican-api-6c6d75b7d-nbj94\" (UID: \"b0e05357-2caf-4e9c-b18d-d0fc75a067d7\") " pod="openstack/barbican-api-6c6d75b7d-nbj94" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.297138 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mktbd\" (UniqueName: \"kubernetes.io/projected/276b7ec3-db22-4125-b19b-1f73ece2a2e0-kube-api-access-mktbd\") pod \"dnsmasq-dns-c845456b5-5r449\" (UID: \"276b7ec3-db22-4125-b19b-1f73ece2a2e0\") " pod="openstack/dnsmasq-dns-c845456b5-5r449" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.297184 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0e05357-2caf-4e9c-b18d-d0fc75a067d7-combined-ca-bundle\") pod \"barbican-api-6c6d75b7d-nbj94\" (UID: \"b0e05357-2caf-4e9c-b18d-d0fc75a067d7\") " pod="openstack/barbican-api-6c6d75b7d-nbj94" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.298014 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/276b7ec3-db22-4125-b19b-1f73ece2a2e0-dns-svc\") pod \"dnsmasq-dns-c845456b5-5r449\" (UID: \"276b7ec3-db22-4125-b19b-1f73ece2a2e0\") " pod="openstack/dnsmasq-dns-c845456b5-5r449" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.299033 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/276b7ec3-db22-4125-b19b-1f73ece2a2e0-ovsdbserver-nb\") pod \"dnsmasq-dns-c845456b5-5r449\" (UID: \"276b7ec3-db22-4125-b19b-1f73ece2a2e0\") " pod="openstack/dnsmasq-dns-c845456b5-5r449" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.299067 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/276b7ec3-db22-4125-b19b-1f73ece2a2e0-ovsdbserver-sb\") pod \"dnsmasq-dns-c845456b5-5r449\" (UID: \"276b7ec3-db22-4125-b19b-1f73ece2a2e0\") " pod="openstack/dnsmasq-dns-c845456b5-5r449" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.299214 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/276b7ec3-db22-4125-b19b-1f73ece2a2e0-config\") pod \"dnsmasq-dns-c845456b5-5r449\" (UID: \"276b7ec3-db22-4125-b19b-1f73ece2a2e0\") " pod="openstack/dnsmasq-dns-c845456b5-5r449" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.318369 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mktbd\" (UniqueName: \"kubernetes.io/projected/276b7ec3-db22-4125-b19b-1f73ece2a2e0-kube-api-access-mktbd\") pod \"dnsmasq-dns-c845456b5-5r449\" (UID: \"276b7ec3-db22-4125-b19b-1f73ece2a2e0\") " pod="openstack/dnsmasq-dns-c845456b5-5r449" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.323327 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c845456b5-5r449" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.399336 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b0e05357-2caf-4e9c-b18d-d0fc75a067d7-config-data-custom\") pod \"barbican-api-6c6d75b7d-nbj94\" (UID: \"b0e05357-2caf-4e9c-b18d-d0fc75a067d7\") " pod="openstack/barbican-api-6c6d75b7d-nbj94" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.399707 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0e05357-2caf-4e9c-b18d-d0fc75a067d7-logs\") pod \"barbican-api-6c6d75b7d-nbj94\" (UID: \"b0e05357-2caf-4e9c-b18d-d0fc75a067d7\") " pod="openstack/barbican-api-6c6d75b7d-nbj94" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.399760 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0e05357-2caf-4e9c-b18d-d0fc75a067d7-combined-ca-bundle\") pod \"barbican-api-6c6d75b7d-nbj94\" (UID: \"b0e05357-2caf-4e9c-b18d-d0fc75a067d7\") " pod="openstack/barbican-api-6c6d75b7d-nbj94" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.399798 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0e05357-2caf-4e9c-b18d-d0fc75a067d7-config-data\") pod \"barbican-api-6c6d75b7d-nbj94\" (UID: \"b0e05357-2caf-4e9c-b18d-d0fc75a067d7\") " pod="openstack/barbican-api-6c6d75b7d-nbj94" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.399824 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnjkp\" (UniqueName: \"kubernetes.io/projected/b0e05357-2caf-4e9c-b18d-d0fc75a067d7-kube-api-access-jnjkp\") pod \"barbican-api-6c6d75b7d-nbj94\" (UID: \"b0e05357-2caf-4e9c-b18d-d0fc75a067d7\") " pod="openstack/barbican-api-6c6d75b7d-nbj94" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.400691 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0e05357-2caf-4e9c-b18d-d0fc75a067d7-logs\") pod \"barbican-api-6c6d75b7d-nbj94\" (UID: \"b0e05357-2caf-4e9c-b18d-d0fc75a067d7\") " pod="openstack/barbican-api-6c6d75b7d-nbj94" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.406111 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0e05357-2caf-4e9c-b18d-d0fc75a067d7-config-data\") pod \"barbican-api-6c6d75b7d-nbj94\" (UID: \"b0e05357-2caf-4e9c-b18d-d0fc75a067d7\") " pod="openstack/barbican-api-6c6d75b7d-nbj94" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.406985 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0e05357-2caf-4e9c-b18d-d0fc75a067d7-combined-ca-bundle\") pod \"barbican-api-6c6d75b7d-nbj94\" (UID: \"b0e05357-2caf-4e9c-b18d-d0fc75a067d7\") " pod="openstack/barbican-api-6c6d75b7d-nbj94" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.407102 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b0e05357-2caf-4e9c-b18d-d0fc75a067d7-config-data-custom\") pod \"barbican-api-6c6d75b7d-nbj94\" (UID: \"b0e05357-2caf-4e9c-b18d-d0fc75a067d7\") " pod="openstack/barbican-api-6c6d75b7d-nbj94" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.422430 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnjkp\" (UniqueName: \"kubernetes.io/projected/b0e05357-2caf-4e9c-b18d-d0fc75a067d7-kube-api-access-jnjkp\") pod \"barbican-api-6c6d75b7d-nbj94\" (UID: \"b0e05357-2caf-4e9c-b18d-d0fc75a067d7\") " pod="openstack/barbican-api-6c6d75b7d-nbj94" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.496557 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c6d75b7d-nbj94" Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.695964 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-846d64984d-vjkzr"] Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.810047 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6f48884455-sk4db"] Nov 24 14:41:48 crc kubenswrapper[4790]: W1124 14:41:48.815309 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf00df2c8_01c0_4cd2_a699_97fadd168c78.slice/crio-dcf0fa4e55b0b3b9aee44c7139a2f066e3a454def764c8f6f35e22c22fb05563 WatchSource:0}: Error finding container dcf0fa4e55b0b3b9aee44c7139a2f066e3a454def764c8f6f35e22c22fb05563: Status 404 returned error can't find the container with id dcf0fa4e55b0b3b9aee44c7139a2f066e3a454def764c8f6f35e22c22fb05563 Nov 24 14:41:48 crc kubenswrapper[4790]: W1124 14:41:48.921662 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod276b7ec3_db22_4125_b19b_1f73ece2a2e0.slice/crio-53bb2981249675cf261478980a1e6e327aeeb5d1b17d3f6f4b3d20c30c68d1f7 WatchSource:0}: Error finding container 53bb2981249675cf261478980a1e6e327aeeb5d1b17d3f6f4b3d20c30c68d1f7: Status 404 returned error can't find the container with id 53bb2981249675cf261478980a1e6e327aeeb5d1b17d3f6f4b3d20c30c68d1f7 Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.921719 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c845456b5-5r449"] Nov 24 14:41:48 crc kubenswrapper[4790]: I1124 14:41:48.988645 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6c6d75b7d-nbj94"] Nov 24 14:41:48 crc kubenswrapper[4790]: W1124 14:41:48.998682 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0e05357_2caf_4e9c_b18d_d0fc75a067d7.slice/crio-f68eca8a49237d7893eddf134fdd157c88ae734bce02bbe42a141048c9fc8983 WatchSource:0}: Error finding container f68eca8a49237d7893eddf134fdd157c88ae734bce02bbe42a141048c9fc8983: Status 404 returned error can't find the container with id f68eca8a49237d7893eddf134fdd157c88ae734bce02bbe42a141048c9fc8983 Nov 24 14:41:49 crc kubenswrapper[4790]: I1124 14:41:49.615933 4790 generic.go:334] "Generic (PLEG): container finished" podID="276b7ec3-db22-4125-b19b-1f73ece2a2e0" containerID="daf9d675b23c85e2de5d85cbea38d6209f07c940044927ee961c988eb2fbab6e" exitCode=0 Nov 24 14:41:49 crc kubenswrapper[4790]: I1124 14:41:49.615983 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c845456b5-5r449" event={"ID":"276b7ec3-db22-4125-b19b-1f73ece2a2e0","Type":"ContainerDied","Data":"daf9d675b23c85e2de5d85cbea38d6209f07c940044927ee961c988eb2fbab6e"} Nov 24 14:41:49 crc kubenswrapper[4790]: I1124 14:41:49.616497 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c845456b5-5r449" event={"ID":"276b7ec3-db22-4125-b19b-1f73ece2a2e0","Type":"ContainerStarted","Data":"53bb2981249675cf261478980a1e6e327aeeb5d1b17d3f6f4b3d20c30c68d1f7"} Nov 24 14:41:49 crc kubenswrapper[4790]: I1124 14:41:49.620401 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c6d75b7d-nbj94" event={"ID":"b0e05357-2caf-4e9c-b18d-d0fc75a067d7","Type":"ContainerStarted","Data":"696f95f1c6edc2c2af765a2c414c7dfb0b5eb8d76cd83770a17ca446255c28c8"} Nov 24 14:41:49 crc kubenswrapper[4790]: I1124 14:41:49.620453 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c6d75b7d-nbj94" event={"ID":"b0e05357-2caf-4e9c-b18d-d0fc75a067d7","Type":"ContainerStarted","Data":"2839ffb10ef7efa691cdac30a8c60562ad06c6ab6f7c31248b7229aedac988cd"} Nov 24 14:41:49 crc kubenswrapper[4790]: I1124 14:41:49.620466 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c6d75b7d-nbj94" event={"ID":"b0e05357-2caf-4e9c-b18d-d0fc75a067d7","Type":"ContainerStarted","Data":"f68eca8a49237d7893eddf134fdd157c88ae734bce02bbe42a141048c9fc8983"} Nov 24 14:41:49 crc kubenswrapper[4790]: I1124 14:41:49.620728 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6c6d75b7d-nbj94" Nov 24 14:41:49 crc kubenswrapper[4790]: I1124 14:41:49.620757 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6c6d75b7d-nbj94" Nov 24 14:41:49 crc kubenswrapper[4790]: I1124 14:41:49.623281 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-846d64984d-vjkzr" event={"ID":"18179e0b-5942-4780-8240-ea2496142b5c","Type":"ContainerStarted","Data":"368a55da186fbe942cf260087f22bebbb2d598a2212d7eb12379b9e841d5f887"} Nov 24 14:41:49 crc kubenswrapper[4790]: I1124 14:41:49.623350 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-846d64984d-vjkzr" event={"ID":"18179e0b-5942-4780-8240-ea2496142b5c","Type":"ContainerStarted","Data":"dff75d5b2f91e8bc18f36b57950425bf61c6aabb8eb5250cabd80ac2433aa46d"} Nov 24 14:41:49 crc kubenswrapper[4790]: I1124 14:41:49.623379 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-846d64984d-vjkzr" event={"ID":"18179e0b-5942-4780-8240-ea2496142b5c","Type":"ContainerStarted","Data":"51be52fbfda20ba5049f85d3f6afcef7d935336fc91dc286f4f61a4a2e888866"} Nov 24 14:41:49 crc kubenswrapper[4790]: I1124 14:41:49.644402 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6f48884455-sk4db" event={"ID":"f00df2c8-01c0-4cd2-a699-97fadd168c78","Type":"ContainerStarted","Data":"d114cbc572a4dd6d9256dd9425da1d3f9a47fbb5fafa1b094e901f6262fe5518"} Nov 24 14:41:49 crc kubenswrapper[4790]: I1124 14:41:49.645451 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6f48884455-sk4db" event={"ID":"f00df2c8-01c0-4cd2-a699-97fadd168c78","Type":"ContainerStarted","Data":"9c183035971f6db60eab479abc6eca4a054337c4a4e3267b45555bca6ce304d5"} Nov 24 14:41:49 crc kubenswrapper[4790]: I1124 14:41:49.645538 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6f48884455-sk4db" event={"ID":"f00df2c8-01c0-4cd2-a699-97fadd168c78","Type":"ContainerStarted","Data":"dcf0fa4e55b0b3b9aee44c7139a2f066e3a454def764c8f6f35e22c22fb05563"} Nov 24 14:41:49 crc kubenswrapper[4790]: I1124 14:41:49.690300 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6c6d75b7d-nbj94" podStartSLOduration=1.690278226 podStartE2EDuration="1.690278226s" podCreationTimestamp="2025-11-24 14:41:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:41:49.666144825 +0000 UTC m=+5358.046038497" watchObservedRunningTime="2025-11-24 14:41:49.690278226 +0000 UTC m=+5358.070171888" Nov 24 14:41:49 crc kubenswrapper[4790]: I1124 14:41:49.696078 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-846d64984d-vjkzr" podStartSLOduration=2.696046758 podStartE2EDuration="2.696046758s" podCreationTimestamp="2025-11-24 14:41:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:41:49.683239696 +0000 UTC m=+5358.063133358" watchObservedRunningTime="2025-11-24 14:41:49.696046758 +0000 UTC m=+5358.075940410" Nov 24 14:41:49 crc kubenswrapper[4790]: I1124 14:41:49.716895 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-6f48884455-sk4db" podStartSLOduration=2.716856639 podStartE2EDuration="2.716856639s" podCreationTimestamp="2025-11-24 14:41:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:41:49.70884545 +0000 UTC m=+5358.088739122" watchObservedRunningTime="2025-11-24 14:41:49.716856639 +0000 UTC m=+5358.096750301" Nov 24 14:41:50 crc kubenswrapper[4790]: I1124 14:41:50.661354 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c845456b5-5r449" event={"ID":"276b7ec3-db22-4125-b19b-1f73ece2a2e0","Type":"ContainerStarted","Data":"1fc3a2c78cd3375cf56cfb1606313decff81fe791f7637254f9855a07a8902d6"} Nov 24 14:41:50 crc kubenswrapper[4790]: I1124 14:41:50.698904 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-c845456b5-5r449" podStartSLOduration=3.698807428 podStartE2EDuration="3.698807428s" podCreationTimestamp="2025-11-24 14:41:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:41:50.689958203 +0000 UTC m=+5359.069851865" watchObservedRunningTime="2025-11-24 14:41:50.698807428 +0000 UTC m=+5359.078701130" Nov 24 14:41:51 crc kubenswrapper[4790]: I1124 14:41:51.673278 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-c845456b5-5r449" Nov 24 14:41:58 crc kubenswrapper[4790]: I1124 14:41:58.329437 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-c845456b5-5r449" Nov 24 14:41:58 crc kubenswrapper[4790]: I1124 14:41:58.418860 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d9f46b68f-hlsvx"] Nov 24 14:41:58 crc kubenswrapper[4790]: I1124 14:41:58.419146 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5d9f46b68f-hlsvx" podUID="1c1b93c1-d293-4c3e-b139-e030990dcba2" containerName="dnsmasq-dns" containerID="cri-o://c662a0191d3751f33a241866f428c2666f14285bb9cbbe2310801f7049acaf42" gracePeriod=10 Nov 24 14:41:58 crc kubenswrapper[4790]: I1124 14:41:58.764910 4790 generic.go:334] "Generic (PLEG): container finished" podID="1c1b93c1-d293-4c3e-b139-e030990dcba2" containerID="c662a0191d3751f33a241866f428c2666f14285bb9cbbe2310801f7049acaf42" exitCode=0 Nov 24 14:41:58 crc kubenswrapper[4790]: I1124 14:41:58.765223 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d9f46b68f-hlsvx" event={"ID":"1c1b93c1-d293-4c3e-b139-e030990dcba2","Type":"ContainerDied","Data":"c662a0191d3751f33a241866f428c2666f14285bb9cbbe2310801f7049acaf42"} Nov 24 14:41:58 crc kubenswrapper[4790]: I1124 14:41:58.892327 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d9f46b68f-hlsvx" Nov 24 14:41:59 crc kubenswrapper[4790]: I1124 14:41:59.028711 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c1b93c1-d293-4c3e-b139-e030990dcba2-ovsdbserver-nb\") pod \"1c1b93c1-d293-4c3e-b139-e030990dcba2\" (UID: \"1c1b93c1-d293-4c3e-b139-e030990dcba2\") " Nov 24 14:41:59 crc kubenswrapper[4790]: I1124 14:41:59.029173 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x488k\" (UniqueName: \"kubernetes.io/projected/1c1b93c1-d293-4c3e-b139-e030990dcba2-kube-api-access-x488k\") pod \"1c1b93c1-d293-4c3e-b139-e030990dcba2\" (UID: \"1c1b93c1-d293-4c3e-b139-e030990dcba2\") " Nov 24 14:41:59 crc kubenswrapper[4790]: I1124 14:41:59.029194 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c1b93c1-d293-4c3e-b139-e030990dcba2-dns-svc\") pod \"1c1b93c1-d293-4c3e-b139-e030990dcba2\" (UID: \"1c1b93c1-d293-4c3e-b139-e030990dcba2\") " Nov 24 14:41:59 crc kubenswrapper[4790]: I1124 14:41:59.029241 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c1b93c1-d293-4c3e-b139-e030990dcba2-config\") pod \"1c1b93c1-d293-4c3e-b139-e030990dcba2\" (UID: \"1c1b93c1-d293-4c3e-b139-e030990dcba2\") " Nov 24 14:41:59 crc kubenswrapper[4790]: I1124 14:41:59.029354 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c1b93c1-d293-4c3e-b139-e030990dcba2-ovsdbserver-sb\") pod \"1c1b93c1-d293-4c3e-b139-e030990dcba2\" (UID: \"1c1b93c1-d293-4c3e-b139-e030990dcba2\") " Nov 24 14:41:59 crc kubenswrapper[4790]: I1124 14:41:59.049406 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c1b93c1-d293-4c3e-b139-e030990dcba2-kube-api-access-x488k" (OuterVolumeSpecName: "kube-api-access-x488k") pod "1c1b93c1-d293-4c3e-b139-e030990dcba2" (UID: "1c1b93c1-d293-4c3e-b139-e030990dcba2"). InnerVolumeSpecName "kube-api-access-x488k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:41:59 crc kubenswrapper[4790]: I1124 14:41:59.078074 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c1b93c1-d293-4c3e-b139-e030990dcba2-config" (OuterVolumeSpecName: "config") pod "1c1b93c1-d293-4c3e-b139-e030990dcba2" (UID: "1c1b93c1-d293-4c3e-b139-e030990dcba2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:41:59 crc kubenswrapper[4790]: I1124 14:41:59.081029 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c1b93c1-d293-4c3e-b139-e030990dcba2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1c1b93c1-d293-4c3e-b139-e030990dcba2" (UID: "1c1b93c1-d293-4c3e-b139-e030990dcba2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:41:59 crc kubenswrapper[4790]: I1124 14:41:59.084912 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c1b93c1-d293-4c3e-b139-e030990dcba2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1c1b93c1-d293-4c3e-b139-e030990dcba2" (UID: "1c1b93c1-d293-4c3e-b139-e030990dcba2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:41:59 crc kubenswrapper[4790]: I1124 14:41:59.089764 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c1b93c1-d293-4c3e-b139-e030990dcba2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1c1b93c1-d293-4c3e-b139-e030990dcba2" (UID: "1c1b93c1-d293-4c3e-b139-e030990dcba2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:41:59 crc kubenswrapper[4790]: I1124 14:41:59.131132 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x488k\" (UniqueName: \"kubernetes.io/projected/1c1b93c1-d293-4c3e-b139-e030990dcba2-kube-api-access-x488k\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:59 crc kubenswrapper[4790]: I1124 14:41:59.131170 4790 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c1b93c1-d293-4c3e-b139-e030990dcba2-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:59 crc kubenswrapper[4790]: I1124 14:41:59.131184 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c1b93c1-d293-4c3e-b139-e030990dcba2-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:59 crc kubenswrapper[4790]: I1124 14:41:59.131193 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c1b93c1-d293-4c3e-b139-e030990dcba2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:59 crc kubenswrapper[4790]: I1124 14:41:59.131203 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c1b93c1-d293-4c3e-b139-e030990dcba2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:59 crc kubenswrapper[4790]: I1124 14:41:59.783049 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d9f46b68f-hlsvx" event={"ID":"1c1b93c1-d293-4c3e-b139-e030990dcba2","Type":"ContainerDied","Data":"3ea85bbedb9d74c7f7cbbd932af015d3daa6a28a8ef1a49eafbb441da31ccc97"} Nov 24 14:41:59 crc kubenswrapper[4790]: I1124 14:41:59.783136 4790 scope.go:117] "RemoveContainer" containerID="c662a0191d3751f33a241866f428c2666f14285bb9cbbe2310801f7049acaf42" Nov 24 14:41:59 crc kubenswrapper[4790]: I1124 14:41:59.785048 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d9f46b68f-hlsvx" Nov 24 14:41:59 crc kubenswrapper[4790]: I1124 14:41:59.827944 4790 scope.go:117] "RemoveContainer" containerID="d1c2c0b6af2effcc8ddcc9bc41f66f2836726e7202c02bf5166bab98a6bcd290" Nov 24 14:41:59 crc kubenswrapper[4790]: I1124 14:41:59.840028 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6c6d75b7d-nbj94" Nov 24 14:41:59 crc kubenswrapper[4790]: I1124 14:41:59.857936 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d9f46b68f-hlsvx"] Nov 24 14:41:59 crc kubenswrapper[4790]: I1124 14:41:59.873630 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d9f46b68f-hlsvx"] Nov 24 14:41:59 crc kubenswrapper[4790]: I1124 14:41:59.904372 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6c6d75b7d-nbj94" Nov 24 14:42:00 crc kubenswrapper[4790]: I1124 14:42:00.332705 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c1b93c1-d293-4c3e-b139-e030990dcba2" path="/var/lib/kubelet/pods/1c1b93c1-d293-4c3e-b139-e030990dcba2/volumes" Nov 24 14:42:02 crc kubenswrapper[4790]: I1124 14:42:02.324695 4790 scope.go:117] "RemoveContainer" containerID="92c086e44ccb28a62777d612e499209f7483422ca70b460ce6ec73c416e8e8b8" Nov 24 14:42:02 crc kubenswrapper[4790]: E1124 14:42:02.326026 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:42:12 crc kubenswrapper[4790]: I1124 14:42:12.329420 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-7zn9c"] Nov 24 14:42:12 crc kubenswrapper[4790]: E1124 14:42:12.330414 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c1b93c1-d293-4c3e-b139-e030990dcba2" containerName="init" Nov 24 14:42:12 crc kubenswrapper[4790]: I1124 14:42:12.330431 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c1b93c1-d293-4c3e-b139-e030990dcba2" containerName="init" Nov 24 14:42:12 crc kubenswrapper[4790]: E1124 14:42:12.330453 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c1b93c1-d293-4c3e-b139-e030990dcba2" containerName="dnsmasq-dns" Nov 24 14:42:12 crc kubenswrapper[4790]: I1124 14:42:12.330461 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c1b93c1-d293-4c3e-b139-e030990dcba2" containerName="dnsmasq-dns" Nov 24 14:42:12 crc kubenswrapper[4790]: I1124 14:42:12.330712 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c1b93c1-d293-4c3e-b139-e030990dcba2" containerName="dnsmasq-dns" Nov 24 14:42:12 crc kubenswrapper[4790]: I1124 14:42:12.331488 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-7zn9c" Nov 24 14:42:12 crc kubenswrapper[4790]: I1124 14:42:12.343398 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-7zn9c"] Nov 24 14:42:12 crc kubenswrapper[4790]: I1124 14:42:12.428834 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5df6-account-create-j89gz"] Nov 24 14:42:12 crc kubenswrapper[4790]: I1124 14:42:12.430224 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5df6-account-create-j89gz" Nov 24 14:42:12 crc kubenswrapper[4790]: I1124 14:42:12.432389 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 24 14:42:12 crc kubenswrapper[4790]: I1124 14:42:12.438926 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5df6-account-create-j89gz"] Nov 24 14:42:12 crc kubenswrapper[4790]: I1124 14:42:12.522752 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2c200d0-9126-40d1-8b7b-d15e646a0f5f-operator-scripts\") pod \"neutron-db-create-7zn9c\" (UID: \"c2c200d0-9126-40d1-8b7b-d15e646a0f5f\") " pod="openstack/neutron-db-create-7zn9c" Nov 24 14:42:12 crc kubenswrapper[4790]: I1124 14:42:12.522857 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92m2k\" (UniqueName: \"kubernetes.io/projected/c2c200d0-9126-40d1-8b7b-d15e646a0f5f-kube-api-access-92m2k\") pod \"neutron-db-create-7zn9c\" (UID: \"c2c200d0-9126-40d1-8b7b-d15e646a0f5f\") " pod="openstack/neutron-db-create-7zn9c" Nov 24 14:42:12 crc kubenswrapper[4790]: I1124 14:42:12.625631 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cm58n\" (UniqueName: \"kubernetes.io/projected/71846a64-7dcc-41ca-907b-5481e438a3e9-kube-api-access-cm58n\") pod \"neutron-5df6-account-create-j89gz\" (UID: \"71846a64-7dcc-41ca-907b-5481e438a3e9\") " pod="openstack/neutron-5df6-account-create-j89gz" Nov 24 14:42:12 crc kubenswrapper[4790]: I1124 14:42:12.625715 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/71846a64-7dcc-41ca-907b-5481e438a3e9-operator-scripts\") pod \"neutron-5df6-account-create-j89gz\" (UID: \"71846a64-7dcc-41ca-907b-5481e438a3e9\") " pod="openstack/neutron-5df6-account-create-j89gz" Nov 24 14:42:12 crc kubenswrapper[4790]: I1124 14:42:12.625769 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2c200d0-9126-40d1-8b7b-d15e646a0f5f-operator-scripts\") pod \"neutron-db-create-7zn9c\" (UID: \"c2c200d0-9126-40d1-8b7b-d15e646a0f5f\") " pod="openstack/neutron-db-create-7zn9c" Nov 24 14:42:12 crc kubenswrapper[4790]: I1124 14:42:12.625974 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92m2k\" (UniqueName: \"kubernetes.io/projected/c2c200d0-9126-40d1-8b7b-d15e646a0f5f-kube-api-access-92m2k\") pod \"neutron-db-create-7zn9c\" (UID: \"c2c200d0-9126-40d1-8b7b-d15e646a0f5f\") " pod="openstack/neutron-db-create-7zn9c" Nov 24 14:42:12 crc kubenswrapper[4790]: I1124 14:42:12.626714 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2c200d0-9126-40d1-8b7b-d15e646a0f5f-operator-scripts\") pod \"neutron-db-create-7zn9c\" (UID: \"c2c200d0-9126-40d1-8b7b-d15e646a0f5f\") " pod="openstack/neutron-db-create-7zn9c" Nov 24 14:42:12 crc kubenswrapper[4790]: I1124 14:42:12.659519 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92m2k\" (UniqueName: \"kubernetes.io/projected/c2c200d0-9126-40d1-8b7b-d15e646a0f5f-kube-api-access-92m2k\") pod \"neutron-db-create-7zn9c\" (UID: \"c2c200d0-9126-40d1-8b7b-d15e646a0f5f\") " pod="openstack/neutron-db-create-7zn9c" Nov 24 14:42:12 crc kubenswrapper[4790]: I1124 14:42:12.728136 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/71846a64-7dcc-41ca-907b-5481e438a3e9-operator-scripts\") pod \"neutron-5df6-account-create-j89gz\" (UID: \"71846a64-7dcc-41ca-907b-5481e438a3e9\") " pod="openstack/neutron-5df6-account-create-j89gz" Nov 24 14:42:12 crc kubenswrapper[4790]: I1124 14:42:12.728357 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cm58n\" (UniqueName: \"kubernetes.io/projected/71846a64-7dcc-41ca-907b-5481e438a3e9-kube-api-access-cm58n\") pod \"neutron-5df6-account-create-j89gz\" (UID: \"71846a64-7dcc-41ca-907b-5481e438a3e9\") " pod="openstack/neutron-5df6-account-create-j89gz" Nov 24 14:42:12 crc kubenswrapper[4790]: I1124 14:42:12.729154 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/71846a64-7dcc-41ca-907b-5481e438a3e9-operator-scripts\") pod \"neutron-5df6-account-create-j89gz\" (UID: \"71846a64-7dcc-41ca-907b-5481e438a3e9\") " pod="openstack/neutron-5df6-account-create-j89gz" Nov 24 14:42:12 crc kubenswrapper[4790]: I1124 14:42:12.751614 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cm58n\" (UniqueName: \"kubernetes.io/projected/71846a64-7dcc-41ca-907b-5481e438a3e9-kube-api-access-cm58n\") pod \"neutron-5df6-account-create-j89gz\" (UID: \"71846a64-7dcc-41ca-907b-5481e438a3e9\") " pod="openstack/neutron-5df6-account-create-j89gz" Nov 24 14:42:12 crc kubenswrapper[4790]: I1124 14:42:12.755462 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5df6-account-create-j89gz" Nov 24 14:42:12 crc kubenswrapper[4790]: I1124 14:42:12.953654 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-7zn9c" Nov 24 14:42:13 crc kubenswrapper[4790]: I1124 14:42:13.237727 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5df6-account-create-j89gz"] Nov 24 14:42:13 crc kubenswrapper[4790]: I1124 14:42:13.363955 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-7zn9c"] Nov 24 14:42:13 crc kubenswrapper[4790]: W1124 14:42:13.375120 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2c200d0_9126_40d1_8b7b_d15e646a0f5f.slice/crio-46c128c792fd8ad87a909d1d946ec6e36938d1e828f07dee30978ba9d9402134 WatchSource:0}: Error finding container 46c128c792fd8ad87a909d1d946ec6e36938d1e828f07dee30978ba9d9402134: Status 404 returned error can't find the container with id 46c128c792fd8ad87a909d1d946ec6e36938d1e828f07dee30978ba9d9402134 Nov 24 14:42:13 crc kubenswrapper[4790]: E1124 14:42:13.771411 4790 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod71846a64_7dcc_41ca_907b_5481e438a3e9.slice/crio-029996d1ea7dc030dc5a485d8ab4c25956d0a826ee1f959513f3b02fd3a01cdb.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod71846a64_7dcc_41ca_907b_5481e438a3e9.slice/crio-conmon-029996d1ea7dc030dc5a485d8ab4c25956d0a826ee1f959513f3b02fd3a01cdb.scope\": RecentStats: unable to find data in memory cache]" Nov 24 14:42:13 crc kubenswrapper[4790]: I1124 14:42:13.941939 4790 generic.go:334] "Generic (PLEG): container finished" podID="c2c200d0-9126-40d1-8b7b-d15e646a0f5f" containerID="fe4c4c352cb5e5a2c377b66b28d211014956cf029ce8935dbacb519b6f7f42b7" exitCode=0 Nov 24 14:42:13 crc kubenswrapper[4790]: I1124 14:42:13.942026 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-7zn9c" event={"ID":"c2c200d0-9126-40d1-8b7b-d15e646a0f5f","Type":"ContainerDied","Data":"fe4c4c352cb5e5a2c377b66b28d211014956cf029ce8935dbacb519b6f7f42b7"} Nov 24 14:42:13 crc kubenswrapper[4790]: I1124 14:42:13.942085 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-7zn9c" event={"ID":"c2c200d0-9126-40d1-8b7b-d15e646a0f5f","Type":"ContainerStarted","Data":"46c128c792fd8ad87a909d1d946ec6e36938d1e828f07dee30978ba9d9402134"} Nov 24 14:42:13 crc kubenswrapper[4790]: I1124 14:42:13.945308 4790 generic.go:334] "Generic (PLEG): container finished" podID="71846a64-7dcc-41ca-907b-5481e438a3e9" containerID="029996d1ea7dc030dc5a485d8ab4c25956d0a826ee1f959513f3b02fd3a01cdb" exitCode=0 Nov 24 14:42:13 crc kubenswrapper[4790]: I1124 14:42:13.945378 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5df6-account-create-j89gz" event={"ID":"71846a64-7dcc-41ca-907b-5481e438a3e9","Type":"ContainerDied","Data":"029996d1ea7dc030dc5a485d8ab4c25956d0a826ee1f959513f3b02fd3a01cdb"} Nov 24 14:42:13 crc kubenswrapper[4790]: I1124 14:42:13.945435 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5df6-account-create-j89gz" event={"ID":"71846a64-7dcc-41ca-907b-5481e438a3e9","Type":"ContainerStarted","Data":"d3d386ed3e99d6c916cd50d51e043cb13aeb3cae87d8539cbd19c0f60c4fffe9"} Nov 24 14:42:15 crc kubenswrapper[4790]: I1124 14:42:15.378400 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5df6-account-create-j89gz" Nov 24 14:42:15 crc kubenswrapper[4790]: I1124 14:42:15.383320 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-7zn9c" Nov 24 14:42:15 crc kubenswrapper[4790]: I1124 14:42:15.493247 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92m2k\" (UniqueName: \"kubernetes.io/projected/c2c200d0-9126-40d1-8b7b-d15e646a0f5f-kube-api-access-92m2k\") pod \"c2c200d0-9126-40d1-8b7b-d15e646a0f5f\" (UID: \"c2c200d0-9126-40d1-8b7b-d15e646a0f5f\") " Nov 24 14:42:15 crc kubenswrapper[4790]: I1124 14:42:15.493352 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/71846a64-7dcc-41ca-907b-5481e438a3e9-operator-scripts\") pod \"71846a64-7dcc-41ca-907b-5481e438a3e9\" (UID: \"71846a64-7dcc-41ca-907b-5481e438a3e9\") " Nov 24 14:42:15 crc kubenswrapper[4790]: I1124 14:42:15.493447 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2c200d0-9126-40d1-8b7b-d15e646a0f5f-operator-scripts\") pod \"c2c200d0-9126-40d1-8b7b-d15e646a0f5f\" (UID: \"c2c200d0-9126-40d1-8b7b-d15e646a0f5f\") " Nov 24 14:42:15 crc kubenswrapper[4790]: I1124 14:42:15.493649 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cm58n\" (UniqueName: \"kubernetes.io/projected/71846a64-7dcc-41ca-907b-5481e438a3e9-kube-api-access-cm58n\") pod \"71846a64-7dcc-41ca-907b-5481e438a3e9\" (UID: \"71846a64-7dcc-41ca-907b-5481e438a3e9\") " Nov 24 14:42:15 crc kubenswrapper[4790]: I1124 14:42:15.494249 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2c200d0-9126-40d1-8b7b-d15e646a0f5f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c2c200d0-9126-40d1-8b7b-d15e646a0f5f" (UID: "c2c200d0-9126-40d1-8b7b-d15e646a0f5f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:42:15 crc kubenswrapper[4790]: I1124 14:42:15.494314 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71846a64-7dcc-41ca-907b-5481e438a3e9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "71846a64-7dcc-41ca-907b-5481e438a3e9" (UID: "71846a64-7dcc-41ca-907b-5481e438a3e9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:42:15 crc kubenswrapper[4790]: I1124 14:42:15.494702 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2c200d0-9126-40d1-8b7b-d15e646a0f5f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:42:15 crc kubenswrapper[4790]: I1124 14:42:15.503051 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2c200d0-9126-40d1-8b7b-d15e646a0f5f-kube-api-access-92m2k" (OuterVolumeSpecName: "kube-api-access-92m2k") pod "c2c200d0-9126-40d1-8b7b-d15e646a0f5f" (UID: "c2c200d0-9126-40d1-8b7b-d15e646a0f5f"). InnerVolumeSpecName "kube-api-access-92m2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:42:15 crc kubenswrapper[4790]: I1124 14:42:15.503861 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71846a64-7dcc-41ca-907b-5481e438a3e9-kube-api-access-cm58n" (OuterVolumeSpecName: "kube-api-access-cm58n") pod "71846a64-7dcc-41ca-907b-5481e438a3e9" (UID: "71846a64-7dcc-41ca-907b-5481e438a3e9"). InnerVolumeSpecName "kube-api-access-cm58n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:42:15 crc kubenswrapper[4790]: I1124 14:42:15.597450 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/71846a64-7dcc-41ca-907b-5481e438a3e9-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:42:15 crc kubenswrapper[4790]: I1124 14:42:15.597504 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cm58n\" (UniqueName: \"kubernetes.io/projected/71846a64-7dcc-41ca-907b-5481e438a3e9-kube-api-access-cm58n\") on node \"crc\" DevicePath \"\"" Nov 24 14:42:15 crc kubenswrapper[4790]: I1124 14:42:15.597528 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92m2k\" (UniqueName: \"kubernetes.io/projected/c2c200d0-9126-40d1-8b7b-d15e646a0f5f-kube-api-access-92m2k\") on node \"crc\" DevicePath \"\"" Nov 24 14:42:15 crc kubenswrapper[4790]: I1124 14:42:15.969583 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5df6-account-create-j89gz" event={"ID":"71846a64-7dcc-41ca-907b-5481e438a3e9","Type":"ContainerDied","Data":"d3d386ed3e99d6c916cd50d51e043cb13aeb3cae87d8539cbd19c0f60c4fffe9"} Nov 24 14:42:15 crc kubenswrapper[4790]: I1124 14:42:15.969636 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3d386ed3e99d6c916cd50d51e043cb13aeb3cae87d8539cbd19c0f60c4fffe9" Nov 24 14:42:15 crc kubenswrapper[4790]: I1124 14:42:15.969656 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5df6-account-create-j89gz" Nov 24 14:42:15 crc kubenswrapper[4790]: I1124 14:42:15.972768 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-7zn9c" event={"ID":"c2c200d0-9126-40d1-8b7b-d15e646a0f5f","Type":"ContainerDied","Data":"46c128c792fd8ad87a909d1d946ec6e36938d1e828f07dee30978ba9d9402134"} Nov 24 14:42:15 crc kubenswrapper[4790]: I1124 14:42:15.972834 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46c128c792fd8ad87a909d1d946ec6e36938d1e828f07dee30978ba9d9402134" Nov 24 14:42:15 crc kubenswrapper[4790]: I1124 14:42:15.973008 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-7zn9c" Nov 24 14:42:16 crc kubenswrapper[4790]: I1124 14:42:16.315877 4790 scope.go:117] "RemoveContainer" containerID="92c086e44ccb28a62777d612e499209f7483422ca70b460ce6ec73c416e8e8b8" Nov 24 14:42:16 crc kubenswrapper[4790]: E1124 14:42:16.316487 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:42:17 crc kubenswrapper[4790]: I1124 14:42:17.610722 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-rw99k"] Nov 24 14:42:17 crc kubenswrapper[4790]: E1124 14:42:17.611536 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71846a64-7dcc-41ca-907b-5481e438a3e9" containerName="mariadb-account-create" Nov 24 14:42:17 crc kubenswrapper[4790]: I1124 14:42:17.611572 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="71846a64-7dcc-41ca-907b-5481e438a3e9" containerName="mariadb-account-create" Nov 24 14:42:17 crc kubenswrapper[4790]: E1124 14:42:17.611634 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2c200d0-9126-40d1-8b7b-d15e646a0f5f" containerName="mariadb-database-create" Nov 24 14:42:17 crc kubenswrapper[4790]: I1124 14:42:17.611653 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2c200d0-9126-40d1-8b7b-d15e646a0f5f" containerName="mariadb-database-create" Nov 24 14:42:17 crc kubenswrapper[4790]: I1124 14:42:17.612094 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="71846a64-7dcc-41ca-907b-5481e438a3e9" containerName="mariadb-account-create" Nov 24 14:42:17 crc kubenswrapper[4790]: I1124 14:42:17.612166 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2c200d0-9126-40d1-8b7b-d15e646a0f5f" containerName="mariadb-database-create" Nov 24 14:42:17 crc kubenswrapper[4790]: I1124 14:42:17.613369 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-rw99k" Nov 24 14:42:17 crc kubenswrapper[4790]: I1124 14:42:17.616626 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-z8ftl" Nov 24 14:42:17 crc kubenswrapper[4790]: I1124 14:42:17.617070 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 24 14:42:17 crc kubenswrapper[4790]: I1124 14:42:17.617351 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 24 14:42:17 crc kubenswrapper[4790]: I1124 14:42:17.619171 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-rw99k"] Nov 24 14:42:17 crc kubenswrapper[4790]: I1124 14:42:17.744526 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08-combined-ca-bundle\") pod \"neutron-db-sync-rw99k\" (UID: \"2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08\") " pod="openstack/neutron-db-sync-rw99k" Nov 24 14:42:17 crc kubenswrapper[4790]: I1124 14:42:17.744856 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8sjf\" (UniqueName: \"kubernetes.io/projected/2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08-kube-api-access-m8sjf\") pod \"neutron-db-sync-rw99k\" (UID: \"2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08\") " pod="openstack/neutron-db-sync-rw99k" Nov 24 14:42:17 crc kubenswrapper[4790]: I1124 14:42:17.744948 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08-config\") pod \"neutron-db-sync-rw99k\" (UID: \"2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08\") " pod="openstack/neutron-db-sync-rw99k" Nov 24 14:42:17 crc kubenswrapper[4790]: I1124 14:42:17.848112 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08-combined-ca-bundle\") pod \"neutron-db-sync-rw99k\" (UID: \"2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08\") " pod="openstack/neutron-db-sync-rw99k" Nov 24 14:42:17 crc kubenswrapper[4790]: I1124 14:42:17.848174 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8sjf\" (UniqueName: \"kubernetes.io/projected/2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08-kube-api-access-m8sjf\") pod \"neutron-db-sync-rw99k\" (UID: \"2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08\") " pod="openstack/neutron-db-sync-rw99k" Nov 24 14:42:17 crc kubenswrapper[4790]: I1124 14:42:17.848219 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08-config\") pod \"neutron-db-sync-rw99k\" (UID: \"2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08\") " pod="openstack/neutron-db-sync-rw99k" Nov 24 14:42:17 crc kubenswrapper[4790]: I1124 14:42:17.864105 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08-config\") pod \"neutron-db-sync-rw99k\" (UID: \"2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08\") " pod="openstack/neutron-db-sync-rw99k" Nov 24 14:42:17 crc kubenswrapper[4790]: I1124 14:42:17.864887 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08-combined-ca-bundle\") pod \"neutron-db-sync-rw99k\" (UID: \"2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08\") " pod="openstack/neutron-db-sync-rw99k" Nov 24 14:42:17 crc kubenswrapper[4790]: I1124 14:42:17.871303 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8sjf\" (UniqueName: \"kubernetes.io/projected/2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08-kube-api-access-m8sjf\") pod \"neutron-db-sync-rw99k\" (UID: \"2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08\") " pod="openstack/neutron-db-sync-rw99k" Nov 24 14:42:17 crc kubenswrapper[4790]: I1124 14:42:17.947146 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-rw99k" Nov 24 14:42:18 crc kubenswrapper[4790]: I1124 14:42:18.410596 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-rw99k"] Nov 24 14:42:19 crc kubenswrapper[4790]: I1124 14:42:19.028244 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-rw99k" event={"ID":"2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08","Type":"ContainerStarted","Data":"32ee5300164f12d5dd1b26b4aa0bf0fc4ed9444a1482a2967fdead9ddccc1570"} Nov 24 14:42:19 crc kubenswrapper[4790]: I1124 14:42:19.028724 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-rw99k" event={"ID":"2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08","Type":"ContainerStarted","Data":"23d8cc3071badc86e6c1c35aaf7b78ace543d6d026cef3192ccb1cce6e80718b"} Nov 24 14:42:19 crc kubenswrapper[4790]: I1124 14:42:19.071942 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-rw99k" podStartSLOduration=2.071889294 podStartE2EDuration="2.071889294s" podCreationTimestamp="2025-11-24 14:42:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:42:19.054962118 +0000 UTC m=+5387.434855800" watchObservedRunningTime="2025-11-24 14:42:19.071889294 +0000 UTC m=+5387.451782966" Nov 24 14:42:23 crc kubenswrapper[4790]: I1124 14:42:23.078590 4790 generic.go:334] "Generic (PLEG): container finished" podID="2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08" containerID="32ee5300164f12d5dd1b26b4aa0bf0fc4ed9444a1482a2967fdead9ddccc1570" exitCode=0 Nov 24 14:42:23 crc kubenswrapper[4790]: I1124 14:42:23.078703 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-rw99k" event={"ID":"2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08","Type":"ContainerDied","Data":"32ee5300164f12d5dd1b26b4aa0bf0fc4ed9444a1482a2967fdead9ddccc1570"} Nov 24 14:42:24 crc kubenswrapper[4790]: I1124 14:42:24.421845 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-rw99k" Nov 24 14:42:24 crc kubenswrapper[4790]: I1124 14:42:24.616702 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08-combined-ca-bundle\") pod \"2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08\" (UID: \"2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08\") " Nov 24 14:42:24 crc kubenswrapper[4790]: I1124 14:42:24.616793 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8sjf\" (UniqueName: \"kubernetes.io/projected/2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08-kube-api-access-m8sjf\") pod \"2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08\" (UID: \"2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08\") " Nov 24 14:42:24 crc kubenswrapper[4790]: I1124 14:42:24.616894 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08-config\") pod \"2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08\" (UID: \"2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08\") " Nov 24 14:42:24 crc kubenswrapper[4790]: I1124 14:42:24.623882 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08-kube-api-access-m8sjf" (OuterVolumeSpecName: "kube-api-access-m8sjf") pod "2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08" (UID: "2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08"). InnerVolumeSpecName "kube-api-access-m8sjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:42:24 crc kubenswrapper[4790]: I1124 14:42:24.661198 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08-config" (OuterVolumeSpecName: "config") pod "2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08" (UID: "2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:42:24 crc kubenswrapper[4790]: I1124 14:42:24.668310 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08" (UID: "2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:42:24 crc kubenswrapper[4790]: I1124 14:42:24.718747 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:42:24 crc kubenswrapper[4790]: I1124 14:42:24.718777 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8sjf\" (UniqueName: \"kubernetes.io/projected/2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08-kube-api-access-m8sjf\") on node \"crc\" DevicePath \"\"" Nov 24 14:42:24 crc kubenswrapper[4790]: I1124 14:42:24.718788 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.101303 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-rw99k" event={"ID":"2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08","Type":"ContainerDied","Data":"23d8cc3071badc86e6c1c35aaf7b78ace543d6d026cef3192ccb1cce6e80718b"} Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.101372 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23d8cc3071badc86e6c1c35aaf7b78ace543d6d026cef3192ccb1cce6e80718b" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.101447 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-rw99k" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.377424 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5688cb67b5-sfbn9"] Nov 24 14:42:25 crc kubenswrapper[4790]: E1124 14:42:25.377945 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08" containerName="neutron-db-sync" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.377961 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08" containerName="neutron-db-sync" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.378181 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08" containerName="neutron-db-sync" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.379407 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5688cb67b5-sfbn9" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.399157 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5688cb67b5-sfbn9"] Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.453806 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-bc5df678c-79bx5"] Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.459550 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-bc5df678c-79bx5" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.465909 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.468045 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-z8ftl" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.470744 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-bc5df678c-79bx5"] Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.470772 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.533063 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f65fadf5-f57c-44a8-9f38-0b8298add279-dns-svc\") pod \"dnsmasq-dns-5688cb67b5-sfbn9\" (UID: \"f65fadf5-f57c-44a8-9f38-0b8298add279\") " pod="openstack/dnsmasq-dns-5688cb67b5-sfbn9" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.533146 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hm5x\" (UniqueName: \"kubernetes.io/projected/f65fadf5-f57c-44a8-9f38-0b8298add279-kube-api-access-4hm5x\") pod \"dnsmasq-dns-5688cb67b5-sfbn9\" (UID: \"f65fadf5-f57c-44a8-9f38-0b8298add279\") " pod="openstack/dnsmasq-dns-5688cb67b5-sfbn9" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.533201 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f65fadf5-f57c-44a8-9f38-0b8298add279-config\") pod \"dnsmasq-dns-5688cb67b5-sfbn9\" (UID: \"f65fadf5-f57c-44a8-9f38-0b8298add279\") " pod="openstack/dnsmasq-dns-5688cb67b5-sfbn9" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.533222 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f65fadf5-f57c-44a8-9f38-0b8298add279-ovsdbserver-nb\") pod \"dnsmasq-dns-5688cb67b5-sfbn9\" (UID: \"f65fadf5-f57c-44a8-9f38-0b8298add279\") " pod="openstack/dnsmasq-dns-5688cb67b5-sfbn9" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.533296 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f65fadf5-f57c-44a8-9f38-0b8298add279-ovsdbserver-sb\") pod \"dnsmasq-dns-5688cb67b5-sfbn9\" (UID: \"f65fadf5-f57c-44a8-9f38-0b8298add279\") " pod="openstack/dnsmasq-dns-5688cb67b5-sfbn9" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.635036 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f65fadf5-f57c-44a8-9f38-0b8298add279-config\") pod \"dnsmasq-dns-5688cb67b5-sfbn9\" (UID: \"f65fadf5-f57c-44a8-9f38-0b8298add279\") " pod="openstack/dnsmasq-dns-5688cb67b5-sfbn9" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.635093 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f65fadf5-f57c-44a8-9f38-0b8298add279-ovsdbserver-nb\") pod \"dnsmasq-dns-5688cb67b5-sfbn9\" (UID: \"f65fadf5-f57c-44a8-9f38-0b8298add279\") " pod="openstack/dnsmasq-dns-5688cb67b5-sfbn9" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.635130 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f65fadf5-f57c-44a8-9f38-0b8298add279-ovsdbserver-sb\") pod \"dnsmasq-dns-5688cb67b5-sfbn9\" (UID: \"f65fadf5-f57c-44a8-9f38-0b8298add279\") " pod="openstack/dnsmasq-dns-5688cb67b5-sfbn9" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.635171 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/15b5bece-4f0a-41fb-8473-44d31fd383b6-config\") pod \"neutron-bc5df678c-79bx5\" (UID: \"15b5bece-4f0a-41fb-8473-44d31fd383b6\") " pod="openstack/neutron-bc5df678c-79bx5" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.635201 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f65fadf5-f57c-44a8-9f38-0b8298add279-dns-svc\") pod \"dnsmasq-dns-5688cb67b5-sfbn9\" (UID: \"f65fadf5-f57c-44a8-9f38-0b8298add279\") " pod="openstack/dnsmasq-dns-5688cb67b5-sfbn9" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.636011 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f65fadf5-f57c-44a8-9f38-0b8298add279-dns-svc\") pod \"dnsmasq-dns-5688cb67b5-sfbn9\" (UID: \"f65fadf5-f57c-44a8-9f38-0b8298add279\") " pod="openstack/dnsmasq-dns-5688cb67b5-sfbn9" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.636011 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f65fadf5-f57c-44a8-9f38-0b8298add279-ovsdbserver-sb\") pod \"dnsmasq-dns-5688cb67b5-sfbn9\" (UID: \"f65fadf5-f57c-44a8-9f38-0b8298add279\") " pod="openstack/dnsmasq-dns-5688cb67b5-sfbn9" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.635231 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92rdh\" (UniqueName: \"kubernetes.io/projected/15b5bece-4f0a-41fb-8473-44d31fd383b6-kube-api-access-92rdh\") pod \"neutron-bc5df678c-79bx5\" (UID: \"15b5bece-4f0a-41fb-8473-44d31fd383b6\") " pod="openstack/neutron-bc5df678c-79bx5" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.636326 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15b5bece-4f0a-41fb-8473-44d31fd383b6-combined-ca-bundle\") pod \"neutron-bc5df678c-79bx5\" (UID: \"15b5bece-4f0a-41fb-8473-44d31fd383b6\") " pod="openstack/neutron-bc5df678c-79bx5" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.636390 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f65fadf5-f57c-44a8-9f38-0b8298add279-config\") pod \"dnsmasq-dns-5688cb67b5-sfbn9\" (UID: \"f65fadf5-f57c-44a8-9f38-0b8298add279\") " pod="openstack/dnsmasq-dns-5688cb67b5-sfbn9" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.636562 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hm5x\" (UniqueName: \"kubernetes.io/projected/f65fadf5-f57c-44a8-9f38-0b8298add279-kube-api-access-4hm5x\") pod \"dnsmasq-dns-5688cb67b5-sfbn9\" (UID: \"f65fadf5-f57c-44a8-9f38-0b8298add279\") " pod="openstack/dnsmasq-dns-5688cb67b5-sfbn9" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.636716 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/15b5bece-4f0a-41fb-8473-44d31fd383b6-httpd-config\") pod \"neutron-bc5df678c-79bx5\" (UID: \"15b5bece-4f0a-41fb-8473-44d31fd383b6\") " pod="openstack/neutron-bc5df678c-79bx5" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.636862 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f65fadf5-f57c-44a8-9f38-0b8298add279-ovsdbserver-nb\") pod \"dnsmasq-dns-5688cb67b5-sfbn9\" (UID: \"f65fadf5-f57c-44a8-9f38-0b8298add279\") " pod="openstack/dnsmasq-dns-5688cb67b5-sfbn9" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.660230 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hm5x\" (UniqueName: \"kubernetes.io/projected/f65fadf5-f57c-44a8-9f38-0b8298add279-kube-api-access-4hm5x\") pod \"dnsmasq-dns-5688cb67b5-sfbn9\" (UID: \"f65fadf5-f57c-44a8-9f38-0b8298add279\") " pod="openstack/dnsmasq-dns-5688cb67b5-sfbn9" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.700084 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5688cb67b5-sfbn9" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.738442 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/15b5bece-4f0a-41fb-8473-44d31fd383b6-httpd-config\") pod \"neutron-bc5df678c-79bx5\" (UID: \"15b5bece-4f0a-41fb-8473-44d31fd383b6\") " pod="openstack/neutron-bc5df678c-79bx5" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.738528 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/15b5bece-4f0a-41fb-8473-44d31fd383b6-config\") pod \"neutron-bc5df678c-79bx5\" (UID: \"15b5bece-4f0a-41fb-8473-44d31fd383b6\") " pod="openstack/neutron-bc5df678c-79bx5" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.738573 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92rdh\" (UniqueName: \"kubernetes.io/projected/15b5bece-4f0a-41fb-8473-44d31fd383b6-kube-api-access-92rdh\") pod \"neutron-bc5df678c-79bx5\" (UID: \"15b5bece-4f0a-41fb-8473-44d31fd383b6\") " pod="openstack/neutron-bc5df678c-79bx5" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.738594 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15b5bece-4f0a-41fb-8473-44d31fd383b6-combined-ca-bundle\") pod \"neutron-bc5df678c-79bx5\" (UID: \"15b5bece-4f0a-41fb-8473-44d31fd383b6\") " pod="openstack/neutron-bc5df678c-79bx5" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.742560 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15b5bece-4f0a-41fb-8473-44d31fd383b6-combined-ca-bundle\") pod \"neutron-bc5df678c-79bx5\" (UID: \"15b5bece-4f0a-41fb-8473-44d31fd383b6\") " pod="openstack/neutron-bc5df678c-79bx5" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.743161 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/15b5bece-4f0a-41fb-8473-44d31fd383b6-config\") pod \"neutron-bc5df678c-79bx5\" (UID: \"15b5bece-4f0a-41fb-8473-44d31fd383b6\") " pod="openstack/neutron-bc5df678c-79bx5" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.743879 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/15b5bece-4f0a-41fb-8473-44d31fd383b6-httpd-config\") pod \"neutron-bc5df678c-79bx5\" (UID: \"15b5bece-4f0a-41fb-8473-44d31fd383b6\") " pod="openstack/neutron-bc5df678c-79bx5" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.755490 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92rdh\" (UniqueName: \"kubernetes.io/projected/15b5bece-4f0a-41fb-8473-44d31fd383b6-kube-api-access-92rdh\") pod \"neutron-bc5df678c-79bx5\" (UID: \"15b5bece-4f0a-41fb-8473-44d31fd383b6\") " pod="openstack/neutron-bc5df678c-79bx5" Nov 24 14:42:25 crc kubenswrapper[4790]: I1124 14:42:25.774576 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-bc5df678c-79bx5" Nov 24 14:42:26 crc kubenswrapper[4790]: I1124 14:42:26.157627 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5688cb67b5-sfbn9"] Nov 24 14:42:26 crc kubenswrapper[4790]: I1124 14:42:26.332992 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-bc5df678c-79bx5"] Nov 24 14:42:26 crc kubenswrapper[4790]: W1124 14:42:26.336473 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod15b5bece_4f0a_41fb_8473_44d31fd383b6.slice/crio-7367c1129b82aa089240f82ee40e86845ef2a0c5f4c6637273f44edc65d06257 WatchSource:0}: Error finding container 7367c1129b82aa089240f82ee40e86845ef2a0c5f4c6637273f44edc65d06257: Status 404 returned error can't find the container with id 7367c1129b82aa089240f82ee40e86845ef2a0c5f4c6637273f44edc65d06257 Nov 24 14:42:27 crc kubenswrapper[4790]: I1124 14:42:27.119166 4790 generic.go:334] "Generic (PLEG): container finished" podID="f65fadf5-f57c-44a8-9f38-0b8298add279" containerID="c1a9a9423008222f6a4bdbe517e15a3db74df3b99a3146347a86bc07adb28778" exitCode=0 Nov 24 14:42:27 crc kubenswrapper[4790]: I1124 14:42:27.119551 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5688cb67b5-sfbn9" event={"ID":"f65fadf5-f57c-44a8-9f38-0b8298add279","Type":"ContainerDied","Data":"c1a9a9423008222f6a4bdbe517e15a3db74df3b99a3146347a86bc07adb28778"} Nov 24 14:42:27 crc kubenswrapper[4790]: I1124 14:42:27.119599 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5688cb67b5-sfbn9" event={"ID":"f65fadf5-f57c-44a8-9f38-0b8298add279","Type":"ContainerStarted","Data":"331648eb92825368130e3d0017bb220fe8a6c407596dcb7e40e52e3ae07419cd"} Nov 24 14:42:27 crc kubenswrapper[4790]: I1124 14:42:27.123771 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-bc5df678c-79bx5" event={"ID":"15b5bece-4f0a-41fb-8473-44d31fd383b6","Type":"ContainerStarted","Data":"9cb0440cc66f81d41880bef2c2517399c2854f13b841b05c3c05b4a52c3ba34b"} Nov 24 14:42:27 crc kubenswrapper[4790]: I1124 14:42:27.123817 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-bc5df678c-79bx5" event={"ID":"15b5bece-4f0a-41fb-8473-44d31fd383b6","Type":"ContainerStarted","Data":"931c195a0da75caf80cf52da09e68b0cbbeccef6f27050aa0c11d1abc4c1b5c1"} Nov 24 14:42:27 crc kubenswrapper[4790]: I1124 14:42:27.123828 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-bc5df678c-79bx5" event={"ID":"15b5bece-4f0a-41fb-8473-44d31fd383b6","Type":"ContainerStarted","Data":"7367c1129b82aa089240f82ee40e86845ef2a0c5f4c6637273f44edc65d06257"} Nov 24 14:42:27 crc kubenswrapper[4790]: I1124 14:42:27.123920 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-bc5df678c-79bx5" Nov 24 14:42:27 crc kubenswrapper[4790]: I1124 14:42:27.174602 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-bc5df678c-79bx5" podStartSLOduration=2.174578255 podStartE2EDuration="2.174578255s" podCreationTimestamp="2025-11-24 14:42:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:42:27.167235015 +0000 UTC m=+5395.547128677" watchObservedRunningTime="2025-11-24 14:42:27.174578255 +0000 UTC m=+5395.554471918" Nov 24 14:42:28 crc kubenswrapper[4790]: I1124 14:42:28.137041 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5688cb67b5-sfbn9" event={"ID":"f65fadf5-f57c-44a8-9f38-0b8298add279","Type":"ContainerStarted","Data":"117f92bb49cd70bb69470788b5d891afeb4610c25b55e9215dc309366b270d59"} Nov 24 14:42:28 crc kubenswrapper[4790]: I1124 14:42:28.137572 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5688cb67b5-sfbn9" Nov 24 14:42:28 crc kubenswrapper[4790]: I1124 14:42:28.167006 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5688cb67b5-sfbn9" podStartSLOduration=3.1669784500000002 podStartE2EDuration="3.16697845s" podCreationTimestamp="2025-11-24 14:42:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:42:28.165546248 +0000 UTC m=+5396.545439930" watchObservedRunningTime="2025-11-24 14:42:28.16697845 +0000 UTC m=+5396.546872122" Nov 24 14:42:30 crc kubenswrapper[4790]: I1124 14:42:30.315009 4790 scope.go:117] "RemoveContainer" containerID="92c086e44ccb28a62777d612e499209f7483422ca70b460ce6ec73c416e8e8b8" Nov 24 14:42:30 crc kubenswrapper[4790]: E1124 14:42:30.315511 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:42:35 crc kubenswrapper[4790]: I1124 14:42:35.702174 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5688cb67b5-sfbn9" Nov 24 14:42:35 crc kubenswrapper[4790]: I1124 14:42:35.776944 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c845456b5-5r449"] Nov 24 14:42:35 crc kubenswrapper[4790]: I1124 14:42:35.777713 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-c845456b5-5r449" podUID="276b7ec3-db22-4125-b19b-1f73ece2a2e0" containerName="dnsmasq-dns" containerID="cri-o://1fc3a2c78cd3375cf56cfb1606313decff81fe791f7637254f9855a07a8902d6" gracePeriod=10 Nov 24 14:42:36 crc kubenswrapper[4790]: I1124 14:42:36.211542 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c845456b5-5r449" Nov 24 14:42:36 crc kubenswrapper[4790]: I1124 14:42:36.228105 4790 generic.go:334] "Generic (PLEG): container finished" podID="276b7ec3-db22-4125-b19b-1f73ece2a2e0" containerID="1fc3a2c78cd3375cf56cfb1606313decff81fe791f7637254f9855a07a8902d6" exitCode=0 Nov 24 14:42:36 crc kubenswrapper[4790]: I1124 14:42:36.228154 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c845456b5-5r449" event={"ID":"276b7ec3-db22-4125-b19b-1f73ece2a2e0","Type":"ContainerDied","Data":"1fc3a2c78cd3375cf56cfb1606313decff81fe791f7637254f9855a07a8902d6"} Nov 24 14:42:36 crc kubenswrapper[4790]: I1124 14:42:36.228187 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c845456b5-5r449" event={"ID":"276b7ec3-db22-4125-b19b-1f73ece2a2e0","Type":"ContainerDied","Data":"53bb2981249675cf261478980a1e6e327aeeb5d1b17d3f6f4b3d20c30c68d1f7"} Nov 24 14:42:36 crc kubenswrapper[4790]: I1124 14:42:36.228211 4790 scope.go:117] "RemoveContainer" containerID="1fc3a2c78cd3375cf56cfb1606313decff81fe791f7637254f9855a07a8902d6" Nov 24 14:42:36 crc kubenswrapper[4790]: I1124 14:42:36.228365 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c845456b5-5r449" Nov 24 14:42:36 crc kubenswrapper[4790]: I1124 14:42:36.247987 4790 scope.go:117] "RemoveContainer" containerID="daf9d675b23c85e2de5d85cbea38d6209f07c940044927ee961c988eb2fbab6e" Nov 24 14:42:36 crc kubenswrapper[4790]: I1124 14:42:36.266019 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/276b7ec3-db22-4125-b19b-1f73ece2a2e0-ovsdbserver-nb\") pod \"276b7ec3-db22-4125-b19b-1f73ece2a2e0\" (UID: \"276b7ec3-db22-4125-b19b-1f73ece2a2e0\") " Nov 24 14:42:36 crc kubenswrapper[4790]: I1124 14:42:36.266145 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/276b7ec3-db22-4125-b19b-1f73ece2a2e0-dns-svc\") pod \"276b7ec3-db22-4125-b19b-1f73ece2a2e0\" (UID: \"276b7ec3-db22-4125-b19b-1f73ece2a2e0\") " Nov 24 14:42:36 crc kubenswrapper[4790]: I1124 14:42:36.266462 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/276b7ec3-db22-4125-b19b-1f73ece2a2e0-ovsdbserver-sb\") pod \"276b7ec3-db22-4125-b19b-1f73ece2a2e0\" (UID: \"276b7ec3-db22-4125-b19b-1f73ece2a2e0\") " Nov 24 14:42:36 crc kubenswrapper[4790]: I1124 14:42:36.266625 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mktbd\" (UniqueName: \"kubernetes.io/projected/276b7ec3-db22-4125-b19b-1f73ece2a2e0-kube-api-access-mktbd\") pod \"276b7ec3-db22-4125-b19b-1f73ece2a2e0\" (UID: \"276b7ec3-db22-4125-b19b-1f73ece2a2e0\") " Nov 24 14:42:36 crc kubenswrapper[4790]: I1124 14:42:36.266778 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/276b7ec3-db22-4125-b19b-1f73ece2a2e0-config\") pod \"276b7ec3-db22-4125-b19b-1f73ece2a2e0\" (UID: \"276b7ec3-db22-4125-b19b-1f73ece2a2e0\") " Nov 24 14:42:36 crc kubenswrapper[4790]: I1124 14:42:36.276329 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/276b7ec3-db22-4125-b19b-1f73ece2a2e0-kube-api-access-mktbd" (OuterVolumeSpecName: "kube-api-access-mktbd") pod "276b7ec3-db22-4125-b19b-1f73ece2a2e0" (UID: "276b7ec3-db22-4125-b19b-1f73ece2a2e0"). InnerVolumeSpecName "kube-api-access-mktbd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:42:36 crc kubenswrapper[4790]: I1124 14:42:36.279260 4790 scope.go:117] "RemoveContainer" containerID="1fc3a2c78cd3375cf56cfb1606313decff81fe791f7637254f9855a07a8902d6" Nov 24 14:42:36 crc kubenswrapper[4790]: E1124 14:42:36.280511 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fc3a2c78cd3375cf56cfb1606313decff81fe791f7637254f9855a07a8902d6\": container with ID starting with 1fc3a2c78cd3375cf56cfb1606313decff81fe791f7637254f9855a07a8902d6 not found: ID does not exist" containerID="1fc3a2c78cd3375cf56cfb1606313decff81fe791f7637254f9855a07a8902d6" Nov 24 14:42:36 crc kubenswrapper[4790]: I1124 14:42:36.280554 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fc3a2c78cd3375cf56cfb1606313decff81fe791f7637254f9855a07a8902d6"} err="failed to get container status \"1fc3a2c78cd3375cf56cfb1606313decff81fe791f7637254f9855a07a8902d6\": rpc error: code = NotFound desc = could not find container \"1fc3a2c78cd3375cf56cfb1606313decff81fe791f7637254f9855a07a8902d6\": container with ID starting with 1fc3a2c78cd3375cf56cfb1606313decff81fe791f7637254f9855a07a8902d6 not found: ID does not exist" Nov 24 14:42:36 crc kubenswrapper[4790]: I1124 14:42:36.280579 4790 scope.go:117] "RemoveContainer" containerID="daf9d675b23c85e2de5d85cbea38d6209f07c940044927ee961c988eb2fbab6e" Nov 24 14:42:36 crc kubenswrapper[4790]: E1124 14:42:36.281038 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"daf9d675b23c85e2de5d85cbea38d6209f07c940044927ee961c988eb2fbab6e\": container with ID starting with daf9d675b23c85e2de5d85cbea38d6209f07c940044927ee961c988eb2fbab6e not found: ID does not exist" containerID="daf9d675b23c85e2de5d85cbea38d6209f07c940044927ee961c988eb2fbab6e" Nov 24 14:42:36 crc kubenswrapper[4790]: I1124 14:42:36.281068 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"daf9d675b23c85e2de5d85cbea38d6209f07c940044927ee961c988eb2fbab6e"} err="failed to get container status \"daf9d675b23c85e2de5d85cbea38d6209f07c940044927ee961c988eb2fbab6e\": rpc error: code = NotFound desc = could not find container \"daf9d675b23c85e2de5d85cbea38d6209f07c940044927ee961c988eb2fbab6e\": container with ID starting with daf9d675b23c85e2de5d85cbea38d6209f07c940044927ee961c988eb2fbab6e not found: ID does not exist" Nov 24 14:42:36 crc kubenswrapper[4790]: I1124 14:42:36.316760 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/276b7ec3-db22-4125-b19b-1f73ece2a2e0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "276b7ec3-db22-4125-b19b-1f73ece2a2e0" (UID: "276b7ec3-db22-4125-b19b-1f73ece2a2e0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:42:36 crc kubenswrapper[4790]: I1124 14:42:36.316788 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/276b7ec3-db22-4125-b19b-1f73ece2a2e0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "276b7ec3-db22-4125-b19b-1f73ece2a2e0" (UID: "276b7ec3-db22-4125-b19b-1f73ece2a2e0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:42:36 crc kubenswrapper[4790]: I1124 14:42:36.318840 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/276b7ec3-db22-4125-b19b-1f73ece2a2e0-config" (OuterVolumeSpecName: "config") pod "276b7ec3-db22-4125-b19b-1f73ece2a2e0" (UID: "276b7ec3-db22-4125-b19b-1f73ece2a2e0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:42:36 crc kubenswrapper[4790]: I1124 14:42:36.334263 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/276b7ec3-db22-4125-b19b-1f73ece2a2e0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "276b7ec3-db22-4125-b19b-1f73ece2a2e0" (UID: "276b7ec3-db22-4125-b19b-1f73ece2a2e0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:42:36 crc kubenswrapper[4790]: I1124 14:42:36.370100 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/276b7ec3-db22-4125-b19b-1f73ece2a2e0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:42:36 crc kubenswrapper[4790]: I1124 14:42:36.370138 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mktbd\" (UniqueName: \"kubernetes.io/projected/276b7ec3-db22-4125-b19b-1f73ece2a2e0-kube-api-access-mktbd\") on node \"crc\" DevicePath \"\"" Nov 24 14:42:36 crc kubenswrapper[4790]: I1124 14:42:36.370154 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/276b7ec3-db22-4125-b19b-1f73ece2a2e0-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:42:36 crc kubenswrapper[4790]: I1124 14:42:36.370166 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/276b7ec3-db22-4125-b19b-1f73ece2a2e0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:42:36 crc kubenswrapper[4790]: I1124 14:42:36.370181 4790 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/276b7ec3-db22-4125-b19b-1f73ece2a2e0-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:42:36 crc kubenswrapper[4790]: I1124 14:42:36.573236 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c845456b5-5r449"] Nov 24 14:42:36 crc kubenswrapper[4790]: I1124 14:42:36.581380 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-c845456b5-5r449"] Nov 24 14:42:38 crc kubenswrapper[4790]: I1124 14:42:38.340695 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="276b7ec3-db22-4125-b19b-1f73ece2a2e0" path="/var/lib/kubelet/pods/276b7ec3-db22-4125-b19b-1f73ece2a2e0/volumes" Nov 24 14:42:45 crc kubenswrapper[4790]: I1124 14:42:45.316210 4790 scope.go:117] "RemoveContainer" containerID="92c086e44ccb28a62777d612e499209f7483422ca70b460ce6ec73c416e8e8b8" Nov 24 14:42:45 crc kubenswrapper[4790]: E1124 14:42:45.316991 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:42:52 crc kubenswrapper[4790]: I1124 14:42:52.739888 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tkh75"] Nov 24 14:42:52 crc kubenswrapper[4790]: E1124 14:42:52.741433 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="276b7ec3-db22-4125-b19b-1f73ece2a2e0" containerName="dnsmasq-dns" Nov 24 14:42:52 crc kubenswrapper[4790]: I1124 14:42:52.741461 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="276b7ec3-db22-4125-b19b-1f73ece2a2e0" containerName="dnsmasq-dns" Nov 24 14:42:52 crc kubenswrapper[4790]: E1124 14:42:52.741517 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="276b7ec3-db22-4125-b19b-1f73ece2a2e0" containerName="init" Nov 24 14:42:52 crc kubenswrapper[4790]: I1124 14:42:52.741530 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="276b7ec3-db22-4125-b19b-1f73ece2a2e0" containerName="init" Nov 24 14:42:52 crc kubenswrapper[4790]: I1124 14:42:52.741898 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="276b7ec3-db22-4125-b19b-1f73ece2a2e0" containerName="dnsmasq-dns" Nov 24 14:42:52 crc kubenswrapper[4790]: I1124 14:42:52.746237 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tkh75" Nov 24 14:42:52 crc kubenswrapper[4790]: I1124 14:42:52.749160 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tkh75"] Nov 24 14:42:52 crc kubenswrapper[4790]: I1124 14:42:52.915567 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b95fdadb-f8e3-44c9-94cd-50a7becc12dc-utilities\") pod \"certified-operators-tkh75\" (UID: \"b95fdadb-f8e3-44c9-94cd-50a7becc12dc\") " pod="openshift-marketplace/certified-operators-tkh75" Nov 24 14:42:52 crc kubenswrapper[4790]: I1124 14:42:52.915711 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnqx8\" (UniqueName: \"kubernetes.io/projected/b95fdadb-f8e3-44c9-94cd-50a7becc12dc-kube-api-access-jnqx8\") pod \"certified-operators-tkh75\" (UID: \"b95fdadb-f8e3-44c9-94cd-50a7becc12dc\") " pod="openshift-marketplace/certified-operators-tkh75" Nov 24 14:42:52 crc kubenswrapper[4790]: I1124 14:42:52.915956 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b95fdadb-f8e3-44c9-94cd-50a7becc12dc-catalog-content\") pod \"certified-operators-tkh75\" (UID: \"b95fdadb-f8e3-44c9-94cd-50a7becc12dc\") " pod="openshift-marketplace/certified-operators-tkh75" Nov 24 14:42:53 crc kubenswrapper[4790]: I1124 14:42:53.020640 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnqx8\" (UniqueName: \"kubernetes.io/projected/b95fdadb-f8e3-44c9-94cd-50a7becc12dc-kube-api-access-jnqx8\") pod \"certified-operators-tkh75\" (UID: \"b95fdadb-f8e3-44c9-94cd-50a7becc12dc\") " pod="openshift-marketplace/certified-operators-tkh75" Nov 24 14:42:53 crc kubenswrapper[4790]: I1124 14:42:53.020844 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b95fdadb-f8e3-44c9-94cd-50a7becc12dc-catalog-content\") pod \"certified-operators-tkh75\" (UID: \"b95fdadb-f8e3-44c9-94cd-50a7becc12dc\") " pod="openshift-marketplace/certified-operators-tkh75" Nov 24 14:42:53 crc kubenswrapper[4790]: I1124 14:42:53.020931 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b95fdadb-f8e3-44c9-94cd-50a7becc12dc-utilities\") pod \"certified-operators-tkh75\" (UID: \"b95fdadb-f8e3-44c9-94cd-50a7becc12dc\") " pod="openshift-marketplace/certified-operators-tkh75" Nov 24 14:42:53 crc kubenswrapper[4790]: I1124 14:42:53.021606 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b95fdadb-f8e3-44c9-94cd-50a7becc12dc-utilities\") pod \"certified-operators-tkh75\" (UID: \"b95fdadb-f8e3-44c9-94cd-50a7becc12dc\") " pod="openshift-marketplace/certified-operators-tkh75" Nov 24 14:42:53 crc kubenswrapper[4790]: I1124 14:42:53.021691 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b95fdadb-f8e3-44c9-94cd-50a7becc12dc-catalog-content\") pod \"certified-operators-tkh75\" (UID: \"b95fdadb-f8e3-44c9-94cd-50a7becc12dc\") " pod="openshift-marketplace/certified-operators-tkh75" Nov 24 14:42:53 crc kubenswrapper[4790]: I1124 14:42:53.064524 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnqx8\" (UniqueName: \"kubernetes.io/projected/b95fdadb-f8e3-44c9-94cd-50a7becc12dc-kube-api-access-jnqx8\") pod \"certified-operators-tkh75\" (UID: \"b95fdadb-f8e3-44c9-94cd-50a7becc12dc\") " pod="openshift-marketplace/certified-operators-tkh75" Nov 24 14:42:53 crc kubenswrapper[4790]: I1124 14:42:53.071277 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tkh75" Nov 24 14:42:53 crc kubenswrapper[4790]: I1124 14:42:53.590752 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tkh75"] Nov 24 14:42:54 crc kubenswrapper[4790]: I1124 14:42:54.429399 4790 generic.go:334] "Generic (PLEG): container finished" podID="b95fdadb-f8e3-44c9-94cd-50a7becc12dc" containerID="6c674a892571bfa34983a2733749b3349f56bfd6e1dad4f921ec6ba9dae8d875" exitCode=0 Nov 24 14:42:54 crc kubenswrapper[4790]: I1124 14:42:54.429524 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tkh75" event={"ID":"b95fdadb-f8e3-44c9-94cd-50a7becc12dc","Type":"ContainerDied","Data":"6c674a892571bfa34983a2733749b3349f56bfd6e1dad4f921ec6ba9dae8d875"} Nov 24 14:42:54 crc kubenswrapper[4790]: I1124 14:42:54.430089 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tkh75" event={"ID":"b95fdadb-f8e3-44c9-94cd-50a7becc12dc","Type":"ContainerStarted","Data":"3195eefcc6d4226154f73efa18fca7d184d50536973b3d9529cf8613d9e984fc"} Nov 24 14:42:54 crc kubenswrapper[4790]: I1124 14:42:54.433307 4790 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 14:42:55 crc kubenswrapper[4790]: I1124 14:42:55.443543 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tkh75" event={"ID":"b95fdadb-f8e3-44c9-94cd-50a7becc12dc","Type":"ContainerStarted","Data":"f2d122bd24a5f91e529e9a28c1003014679f8b83742b7a08ac274d58c027794e"} Nov 24 14:42:55 crc kubenswrapper[4790]: I1124 14:42:55.789833 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-bc5df678c-79bx5" Nov 24 14:42:56 crc kubenswrapper[4790]: I1124 14:42:56.463184 4790 generic.go:334] "Generic (PLEG): container finished" podID="b95fdadb-f8e3-44c9-94cd-50a7becc12dc" containerID="f2d122bd24a5f91e529e9a28c1003014679f8b83742b7a08ac274d58c027794e" exitCode=0 Nov 24 14:42:56 crc kubenswrapper[4790]: I1124 14:42:56.463296 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tkh75" event={"ID":"b95fdadb-f8e3-44c9-94cd-50a7becc12dc","Type":"ContainerDied","Data":"f2d122bd24a5f91e529e9a28c1003014679f8b83742b7a08ac274d58c027794e"} Nov 24 14:42:58 crc kubenswrapper[4790]: I1124 14:42:58.517478 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tkh75" event={"ID":"b95fdadb-f8e3-44c9-94cd-50a7becc12dc","Type":"ContainerStarted","Data":"960b41998e00ebe32b9d3ff1772d66bb3336989ad91f23625316549196d64daf"} Nov 24 14:42:59 crc kubenswrapper[4790]: I1124 14:42:59.315265 4790 scope.go:117] "RemoveContainer" containerID="92c086e44ccb28a62777d612e499209f7483422ca70b460ce6ec73c416e8e8b8" Nov 24 14:42:59 crc kubenswrapper[4790]: E1124 14:42:59.316201 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:43:03 crc kubenswrapper[4790]: I1124 14:43:03.041860 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tkh75" podStartSLOduration=7.321048502 podStartE2EDuration="11.041836783s" podCreationTimestamp="2025-11-24 14:42:52 +0000 UTC" firstStartedPulling="2025-11-24 14:42:54.432679978 +0000 UTC m=+5422.812573680" lastFinishedPulling="2025-11-24 14:42:58.153468269 +0000 UTC m=+5426.533361961" observedRunningTime="2025-11-24 14:42:58.540872132 +0000 UTC m=+5426.920765804" watchObservedRunningTime="2025-11-24 14:43:03.041836783 +0000 UTC m=+5431.421730465" Nov 24 14:43:03 crc kubenswrapper[4790]: I1124 14:43:03.050448 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-dpfn5"] Nov 24 14:43:03 crc kubenswrapper[4790]: I1124 14:43:03.052005 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dpfn5" Nov 24 14:43:03 crc kubenswrapper[4790]: I1124 14:43:03.058756 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-dpfn5"] Nov 24 14:43:03 crc kubenswrapper[4790]: I1124 14:43:03.072497 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tkh75" Nov 24 14:43:03 crc kubenswrapper[4790]: I1124 14:43:03.072551 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tkh75" Nov 24 14:43:03 crc kubenswrapper[4790]: I1124 14:43:03.125375 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tkh75" Nov 24 14:43:03 crc kubenswrapper[4790]: I1124 14:43:03.128504 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-6b9f-account-create-wb2xr"] Nov 24 14:43:03 crc kubenswrapper[4790]: I1124 14:43:03.130012 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6b9f-account-create-wb2xr" Nov 24 14:43:03 crc kubenswrapper[4790]: I1124 14:43:03.132037 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 24 14:43:03 crc kubenswrapper[4790]: I1124 14:43:03.140815 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-6b9f-account-create-wb2xr"] Nov 24 14:43:03 crc kubenswrapper[4790]: I1124 14:43:03.231794 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fda25cdc-1a37-44a2-8314-461f0f1de55a-operator-scripts\") pod \"glance-6b9f-account-create-wb2xr\" (UID: \"fda25cdc-1a37-44a2-8314-461f0f1de55a\") " pod="openstack/glance-6b9f-account-create-wb2xr" Nov 24 14:43:03 crc kubenswrapper[4790]: I1124 14:43:03.231834 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e5cfb13-d03c-4981-b260-2ae89c41f0bf-operator-scripts\") pod \"glance-db-create-dpfn5\" (UID: \"2e5cfb13-d03c-4981-b260-2ae89c41f0bf\") " pod="openstack/glance-db-create-dpfn5" Nov 24 14:43:03 crc kubenswrapper[4790]: I1124 14:43:03.232025 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5xkz\" (UniqueName: \"kubernetes.io/projected/fda25cdc-1a37-44a2-8314-461f0f1de55a-kube-api-access-b5xkz\") pod \"glance-6b9f-account-create-wb2xr\" (UID: \"fda25cdc-1a37-44a2-8314-461f0f1de55a\") " pod="openstack/glance-6b9f-account-create-wb2xr" Nov 24 14:43:03 crc kubenswrapper[4790]: I1124 14:43:03.232071 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95qjb\" (UniqueName: \"kubernetes.io/projected/2e5cfb13-d03c-4981-b260-2ae89c41f0bf-kube-api-access-95qjb\") pod \"glance-db-create-dpfn5\" (UID: \"2e5cfb13-d03c-4981-b260-2ae89c41f0bf\") " pod="openstack/glance-db-create-dpfn5" Nov 24 14:43:03 crc kubenswrapper[4790]: I1124 14:43:03.333437 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5xkz\" (UniqueName: \"kubernetes.io/projected/fda25cdc-1a37-44a2-8314-461f0f1de55a-kube-api-access-b5xkz\") pod \"glance-6b9f-account-create-wb2xr\" (UID: \"fda25cdc-1a37-44a2-8314-461f0f1de55a\") " pod="openstack/glance-6b9f-account-create-wb2xr" Nov 24 14:43:03 crc kubenswrapper[4790]: I1124 14:43:03.333512 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95qjb\" (UniqueName: \"kubernetes.io/projected/2e5cfb13-d03c-4981-b260-2ae89c41f0bf-kube-api-access-95qjb\") pod \"glance-db-create-dpfn5\" (UID: \"2e5cfb13-d03c-4981-b260-2ae89c41f0bf\") " pod="openstack/glance-db-create-dpfn5" Nov 24 14:43:03 crc kubenswrapper[4790]: I1124 14:43:03.333546 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fda25cdc-1a37-44a2-8314-461f0f1de55a-operator-scripts\") pod \"glance-6b9f-account-create-wb2xr\" (UID: \"fda25cdc-1a37-44a2-8314-461f0f1de55a\") " pod="openstack/glance-6b9f-account-create-wb2xr" Nov 24 14:43:03 crc kubenswrapper[4790]: I1124 14:43:03.333567 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e5cfb13-d03c-4981-b260-2ae89c41f0bf-operator-scripts\") pod \"glance-db-create-dpfn5\" (UID: \"2e5cfb13-d03c-4981-b260-2ae89c41f0bf\") " pod="openstack/glance-db-create-dpfn5" Nov 24 14:43:03 crc kubenswrapper[4790]: I1124 14:43:03.334561 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e5cfb13-d03c-4981-b260-2ae89c41f0bf-operator-scripts\") pod \"glance-db-create-dpfn5\" (UID: \"2e5cfb13-d03c-4981-b260-2ae89c41f0bf\") " pod="openstack/glance-db-create-dpfn5" Nov 24 14:43:03 crc kubenswrapper[4790]: I1124 14:43:03.334610 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fda25cdc-1a37-44a2-8314-461f0f1de55a-operator-scripts\") pod \"glance-6b9f-account-create-wb2xr\" (UID: \"fda25cdc-1a37-44a2-8314-461f0f1de55a\") " pod="openstack/glance-6b9f-account-create-wb2xr" Nov 24 14:43:03 crc kubenswrapper[4790]: I1124 14:43:03.351937 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95qjb\" (UniqueName: \"kubernetes.io/projected/2e5cfb13-d03c-4981-b260-2ae89c41f0bf-kube-api-access-95qjb\") pod \"glance-db-create-dpfn5\" (UID: \"2e5cfb13-d03c-4981-b260-2ae89c41f0bf\") " pod="openstack/glance-db-create-dpfn5" Nov 24 14:43:03 crc kubenswrapper[4790]: I1124 14:43:03.360419 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5xkz\" (UniqueName: \"kubernetes.io/projected/fda25cdc-1a37-44a2-8314-461f0f1de55a-kube-api-access-b5xkz\") pod \"glance-6b9f-account-create-wb2xr\" (UID: \"fda25cdc-1a37-44a2-8314-461f0f1de55a\") " pod="openstack/glance-6b9f-account-create-wb2xr" Nov 24 14:43:03 crc kubenswrapper[4790]: I1124 14:43:03.383258 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dpfn5" Nov 24 14:43:03 crc kubenswrapper[4790]: I1124 14:43:03.455350 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6b9f-account-create-wb2xr" Nov 24 14:43:03 crc kubenswrapper[4790]: I1124 14:43:03.611568 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-dpfn5"] Nov 24 14:43:03 crc kubenswrapper[4790]: I1124 14:43:03.642079 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tkh75" Nov 24 14:43:03 crc kubenswrapper[4790]: I1124 14:43:03.690151 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tkh75"] Nov 24 14:43:03 crc kubenswrapper[4790]: I1124 14:43:03.913861 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-6b9f-account-create-wb2xr"] Nov 24 14:43:04 crc kubenswrapper[4790]: I1124 14:43:04.597012 4790 generic.go:334] "Generic (PLEG): container finished" podID="2e5cfb13-d03c-4981-b260-2ae89c41f0bf" containerID="b979bb516a5898b201f6b42009b4d7a638f44ab9c7b9e0e4bfef7783795882eb" exitCode=0 Nov 24 14:43:04 crc kubenswrapper[4790]: I1124 14:43:04.597600 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-dpfn5" event={"ID":"2e5cfb13-d03c-4981-b260-2ae89c41f0bf","Type":"ContainerDied","Data":"b979bb516a5898b201f6b42009b4d7a638f44ab9c7b9e0e4bfef7783795882eb"} Nov 24 14:43:04 crc kubenswrapper[4790]: I1124 14:43:04.597753 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-dpfn5" event={"ID":"2e5cfb13-d03c-4981-b260-2ae89c41f0bf","Type":"ContainerStarted","Data":"f3f4b7a2a73693a42c0632fa544c17bbb77d92c43b045f83edb517ee0a2c13cd"} Nov 24 14:43:04 crc kubenswrapper[4790]: I1124 14:43:04.599528 4790 generic.go:334] "Generic (PLEG): container finished" podID="fda25cdc-1a37-44a2-8314-461f0f1de55a" containerID="a87adc1cd91559216c434df0a21f1568e0d2f3b6b62c1ed294ed2e61f3e21d60" exitCode=0 Nov 24 14:43:04 crc kubenswrapper[4790]: I1124 14:43:04.599596 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6b9f-account-create-wb2xr" event={"ID":"fda25cdc-1a37-44a2-8314-461f0f1de55a","Type":"ContainerDied","Data":"a87adc1cd91559216c434df0a21f1568e0d2f3b6b62c1ed294ed2e61f3e21d60"} Nov 24 14:43:04 crc kubenswrapper[4790]: I1124 14:43:04.599656 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6b9f-account-create-wb2xr" event={"ID":"fda25cdc-1a37-44a2-8314-461f0f1de55a","Type":"ContainerStarted","Data":"b953930831d8612881b0b14d2d1b8be1779d9b867c754486304b4a9ff1e4e410"} Nov 24 14:43:05 crc kubenswrapper[4790]: I1124 14:43:05.612702 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tkh75" podUID="b95fdadb-f8e3-44c9-94cd-50a7becc12dc" containerName="registry-server" containerID="cri-o://960b41998e00ebe32b9d3ff1772d66bb3336989ad91f23625316549196d64daf" gracePeriod=2 Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.060721 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6b9f-account-create-wb2xr" Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.161983 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tkh75" Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.167243 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dpfn5" Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.193751 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5xkz\" (UniqueName: \"kubernetes.io/projected/fda25cdc-1a37-44a2-8314-461f0f1de55a-kube-api-access-b5xkz\") pod \"fda25cdc-1a37-44a2-8314-461f0f1de55a\" (UID: \"fda25cdc-1a37-44a2-8314-461f0f1de55a\") " Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.193838 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fda25cdc-1a37-44a2-8314-461f0f1de55a-operator-scripts\") pod \"fda25cdc-1a37-44a2-8314-461f0f1de55a\" (UID: \"fda25cdc-1a37-44a2-8314-461f0f1de55a\") " Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.194832 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda25cdc-1a37-44a2-8314-461f0f1de55a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fda25cdc-1a37-44a2-8314-461f0f1de55a" (UID: "fda25cdc-1a37-44a2-8314-461f0f1de55a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.204152 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda25cdc-1a37-44a2-8314-461f0f1de55a-kube-api-access-b5xkz" (OuterVolumeSpecName: "kube-api-access-b5xkz") pod "fda25cdc-1a37-44a2-8314-461f0f1de55a" (UID: "fda25cdc-1a37-44a2-8314-461f0f1de55a"). InnerVolumeSpecName "kube-api-access-b5xkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.295683 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95qjb\" (UniqueName: \"kubernetes.io/projected/2e5cfb13-d03c-4981-b260-2ae89c41f0bf-kube-api-access-95qjb\") pod \"2e5cfb13-d03c-4981-b260-2ae89c41f0bf\" (UID: \"2e5cfb13-d03c-4981-b260-2ae89c41f0bf\") " Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.295792 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e5cfb13-d03c-4981-b260-2ae89c41f0bf-operator-scripts\") pod \"2e5cfb13-d03c-4981-b260-2ae89c41f0bf\" (UID: \"2e5cfb13-d03c-4981-b260-2ae89c41f0bf\") " Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.295922 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b95fdadb-f8e3-44c9-94cd-50a7becc12dc-catalog-content\") pod \"b95fdadb-f8e3-44c9-94cd-50a7becc12dc\" (UID: \"b95fdadb-f8e3-44c9-94cd-50a7becc12dc\") " Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.295956 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b95fdadb-f8e3-44c9-94cd-50a7becc12dc-utilities\") pod \"b95fdadb-f8e3-44c9-94cd-50a7becc12dc\" (UID: \"b95fdadb-f8e3-44c9-94cd-50a7becc12dc\") " Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.295989 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnqx8\" (UniqueName: \"kubernetes.io/projected/b95fdadb-f8e3-44c9-94cd-50a7becc12dc-kube-api-access-jnqx8\") pod \"b95fdadb-f8e3-44c9-94cd-50a7becc12dc\" (UID: \"b95fdadb-f8e3-44c9-94cd-50a7becc12dc\") " Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.296425 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5xkz\" (UniqueName: \"kubernetes.io/projected/fda25cdc-1a37-44a2-8314-461f0f1de55a-kube-api-access-b5xkz\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.296443 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fda25cdc-1a37-44a2-8314-461f0f1de55a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.296770 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e5cfb13-d03c-4981-b260-2ae89c41f0bf-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2e5cfb13-d03c-4981-b260-2ae89c41f0bf" (UID: "2e5cfb13-d03c-4981-b260-2ae89c41f0bf"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.298053 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b95fdadb-f8e3-44c9-94cd-50a7becc12dc-utilities" (OuterVolumeSpecName: "utilities") pod "b95fdadb-f8e3-44c9-94cd-50a7becc12dc" (UID: "b95fdadb-f8e3-44c9-94cd-50a7becc12dc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.300055 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b95fdadb-f8e3-44c9-94cd-50a7becc12dc-kube-api-access-jnqx8" (OuterVolumeSpecName: "kube-api-access-jnqx8") pod "b95fdadb-f8e3-44c9-94cd-50a7becc12dc" (UID: "b95fdadb-f8e3-44c9-94cd-50a7becc12dc"). InnerVolumeSpecName "kube-api-access-jnqx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.301096 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e5cfb13-d03c-4981-b260-2ae89c41f0bf-kube-api-access-95qjb" (OuterVolumeSpecName: "kube-api-access-95qjb") pod "2e5cfb13-d03c-4981-b260-2ae89c41f0bf" (UID: "2e5cfb13-d03c-4981-b260-2ae89c41f0bf"). InnerVolumeSpecName "kube-api-access-95qjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.344459 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b95fdadb-f8e3-44c9-94cd-50a7becc12dc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b95fdadb-f8e3-44c9-94cd-50a7becc12dc" (UID: "b95fdadb-f8e3-44c9-94cd-50a7becc12dc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.398511 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e5cfb13-d03c-4981-b260-2ae89c41f0bf-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.398734 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b95fdadb-f8e3-44c9-94cd-50a7becc12dc-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.398832 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b95fdadb-f8e3-44c9-94cd-50a7becc12dc-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.398939 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnqx8\" (UniqueName: \"kubernetes.io/projected/b95fdadb-f8e3-44c9-94cd-50a7becc12dc-kube-api-access-jnqx8\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.399033 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95qjb\" (UniqueName: \"kubernetes.io/projected/2e5cfb13-d03c-4981-b260-2ae89c41f0bf-kube-api-access-95qjb\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.624256 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6b9f-account-create-wb2xr" event={"ID":"fda25cdc-1a37-44a2-8314-461f0f1de55a","Type":"ContainerDied","Data":"b953930831d8612881b0b14d2d1b8be1779d9b867c754486304b4a9ff1e4e410"} Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.624339 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b953930831d8612881b0b14d2d1b8be1779d9b867c754486304b4a9ff1e4e410" Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.624287 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6b9f-account-create-wb2xr" Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.626458 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-dpfn5" event={"ID":"2e5cfb13-d03c-4981-b260-2ae89c41f0bf","Type":"ContainerDied","Data":"f3f4b7a2a73693a42c0632fa544c17bbb77d92c43b045f83edb517ee0a2c13cd"} Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.626503 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3f4b7a2a73693a42c0632fa544c17bbb77d92c43b045f83edb517ee0a2c13cd" Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.626529 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dpfn5" Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.629484 4790 generic.go:334] "Generic (PLEG): container finished" podID="b95fdadb-f8e3-44c9-94cd-50a7becc12dc" containerID="960b41998e00ebe32b9d3ff1772d66bb3336989ad91f23625316549196d64daf" exitCode=0 Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.629515 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tkh75" event={"ID":"b95fdadb-f8e3-44c9-94cd-50a7becc12dc","Type":"ContainerDied","Data":"960b41998e00ebe32b9d3ff1772d66bb3336989ad91f23625316549196d64daf"} Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.629542 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tkh75" event={"ID":"b95fdadb-f8e3-44c9-94cd-50a7becc12dc","Type":"ContainerDied","Data":"3195eefcc6d4226154f73efa18fca7d184d50536973b3d9529cf8613d9e984fc"} Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.629562 4790 scope.go:117] "RemoveContainer" containerID="960b41998e00ebe32b9d3ff1772d66bb3336989ad91f23625316549196d64daf" Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.629578 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tkh75" Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.667872 4790 scope.go:117] "RemoveContainer" containerID="f2d122bd24a5f91e529e9a28c1003014679f8b83742b7a08ac274d58c027794e" Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.684829 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tkh75"] Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.692704 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tkh75"] Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.705288 4790 scope.go:117] "RemoveContainer" containerID="6c674a892571bfa34983a2733749b3349f56bfd6e1dad4f921ec6ba9dae8d875" Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.730556 4790 scope.go:117] "RemoveContainer" containerID="960b41998e00ebe32b9d3ff1772d66bb3336989ad91f23625316549196d64daf" Nov 24 14:43:06 crc kubenswrapper[4790]: E1124 14:43:06.731450 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"960b41998e00ebe32b9d3ff1772d66bb3336989ad91f23625316549196d64daf\": container with ID starting with 960b41998e00ebe32b9d3ff1772d66bb3336989ad91f23625316549196d64daf not found: ID does not exist" containerID="960b41998e00ebe32b9d3ff1772d66bb3336989ad91f23625316549196d64daf" Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.731510 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"960b41998e00ebe32b9d3ff1772d66bb3336989ad91f23625316549196d64daf"} err="failed to get container status \"960b41998e00ebe32b9d3ff1772d66bb3336989ad91f23625316549196d64daf\": rpc error: code = NotFound desc = could not find container \"960b41998e00ebe32b9d3ff1772d66bb3336989ad91f23625316549196d64daf\": container with ID starting with 960b41998e00ebe32b9d3ff1772d66bb3336989ad91f23625316549196d64daf not found: ID does not exist" Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.731592 4790 scope.go:117] "RemoveContainer" containerID="f2d122bd24a5f91e529e9a28c1003014679f8b83742b7a08ac274d58c027794e" Nov 24 14:43:06 crc kubenswrapper[4790]: E1124 14:43:06.732209 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2d122bd24a5f91e529e9a28c1003014679f8b83742b7a08ac274d58c027794e\": container with ID starting with f2d122bd24a5f91e529e9a28c1003014679f8b83742b7a08ac274d58c027794e not found: ID does not exist" containerID="f2d122bd24a5f91e529e9a28c1003014679f8b83742b7a08ac274d58c027794e" Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.732275 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2d122bd24a5f91e529e9a28c1003014679f8b83742b7a08ac274d58c027794e"} err="failed to get container status \"f2d122bd24a5f91e529e9a28c1003014679f8b83742b7a08ac274d58c027794e\": rpc error: code = NotFound desc = could not find container \"f2d122bd24a5f91e529e9a28c1003014679f8b83742b7a08ac274d58c027794e\": container with ID starting with f2d122bd24a5f91e529e9a28c1003014679f8b83742b7a08ac274d58c027794e not found: ID does not exist" Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.732316 4790 scope.go:117] "RemoveContainer" containerID="6c674a892571bfa34983a2733749b3349f56bfd6e1dad4f921ec6ba9dae8d875" Nov 24 14:43:06 crc kubenswrapper[4790]: E1124 14:43:06.732989 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c674a892571bfa34983a2733749b3349f56bfd6e1dad4f921ec6ba9dae8d875\": container with ID starting with 6c674a892571bfa34983a2733749b3349f56bfd6e1dad4f921ec6ba9dae8d875 not found: ID does not exist" containerID="6c674a892571bfa34983a2733749b3349f56bfd6e1dad4f921ec6ba9dae8d875" Nov 24 14:43:06 crc kubenswrapper[4790]: I1124 14:43:06.733031 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c674a892571bfa34983a2733749b3349f56bfd6e1dad4f921ec6ba9dae8d875"} err="failed to get container status \"6c674a892571bfa34983a2733749b3349f56bfd6e1dad4f921ec6ba9dae8d875\": rpc error: code = NotFound desc = could not find container \"6c674a892571bfa34983a2733749b3349f56bfd6e1dad4f921ec6ba9dae8d875\": container with ID starting with 6c674a892571bfa34983a2733749b3349f56bfd6e1dad4f921ec6ba9dae8d875 not found: ID does not exist" Nov 24 14:43:08 crc kubenswrapper[4790]: I1124 14:43:08.331731 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b95fdadb-f8e3-44c9-94cd-50a7becc12dc" path="/var/lib/kubelet/pods/b95fdadb-f8e3-44c9-94cd-50a7becc12dc/volumes" Nov 24 14:43:08 crc kubenswrapper[4790]: I1124 14:43:08.357007 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-tslzm"] Nov 24 14:43:08 crc kubenswrapper[4790]: E1124 14:43:08.357599 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b95fdadb-f8e3-44c9-94cd-50a7becc12dc" containerName="extract-content" Nov 24 14:43:08 crc kubenswrapper[4790]: I1124 14:43:08.357634 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="b95fdadb-f8e3-44c9-94cd-50a7becc12dc" containerName="extract-content" Nov 24 14:43:08 crc kubenswrapper[4790]: E1124 14:43:08.357668 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e5cfb13-d03c-4981-b260-2ae89c41f0bf" containerName="mariadb-database-create" Nov 24 14:43:08 crc kubenswrapper[4790]: I1124 14:43:08.357682 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e5cfb13-d03c-4981-b260-2ae89c41f0bf" containerName="mariadb-database-create" Nov 24 14:43:08 crc kubenswrapper[4790]: E1124 14:43:08.357712 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b95fdadb-f8e3-44c9-94cd-50a7becc12dc" containerName="registry-server" Nov 24 14:43:08 crc kubenswrapper[4790]: I1124 14:43:08.357728 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="b95fdadb-f8e3-44c9-94cd-50a7becc12dc" containerName="registry-server" Nov 24 14:43:08 crc kubenswrapper[4790]: E1124 14:43:08.357754 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fda25cdc-1a37-44a2-8314-461f0f1de55a" containerName="mariadb-account-create" Nov 24 14:43:08 crc kubenswrapper[4790]: I1124 14:43:08.357767 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="fda25cdc-1a37-44a2-8314-461f0f1de55a" containerName="mariadb-account-create" Nov 24 14:43:08 crc kubenswrapper[4790]: E1124 14:43:08.357791 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b95fdadb-f8e3-44c9-94cd-50a7becc12dc" containerName="extract-utilities" Nov 24 14:43:08 crc kubenswrapper[4790]: I1124 14:43:08.357803 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="b95fdadb-f8e3-44c9-94cd-50a7becc12dc" containerName="extract-utilities" Nov 24 14:43:08 crc kubenswrapper[4790]: I1124 14:43:08.358168 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="fda25cdc-1a37-44a2-8314-461f0f1de55a" containerName="mariadb-account-create" Nov 24 14:43:08 crc kubenswrapper[4790]: I1124 14:43:08.358225 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e5cfb13-d03c-4981-b260-2ae89c41f0bf" containerName="mariadb-database-create" Nov 24 14:43:08 crc kubenswrapper[4790]: I1124 14:43:08.358251 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="b95fdadb-f8e3-44c9-94cd-50a7becc12dc" containerName="registry-server" Nov 24 14:43:08 crc kubenswrapper[4790]: I1124 14:43:08.359226 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tslzm" Nov 24 14:43:08 crc kubenswrapper[4790]: I1124 14:43:08.362217 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 24 14:43:08 crc kubenswrapper[4790]: I1124 14:43:08.362294 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-kjz5s" Nov 24 14:43:08 crc kubenswrapper[4790]: I1124 14:43:08.372653 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-tslzm"] Nov 24 14:43:08 crc kubenswrapper[4790]: I1124 14:43:08.542622 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b7cfed5-a39c-4bee-b427-f74b2af51e73-combined-ca-bundle\") pod \"glance-db-sync-tslzm\" (UID: \"0b7cfed5-a39c-4bee-b427-f74b2af51e73\") " pod="openstack/glance-db-sync-tslzm" Nov 24 14:43:08 crc kubenswrapper[4790]: I1124 14:43:08.542703 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b7cfed5-a39c-4bee-b427-f74b2af51e73-config-data\") pod \"glance-db-sync-tslzm\" (UID: \"0b7cfed5-a39c-4bee-b427-f74b2af51e73\") " pod="openstack/glance-db-sync-tslzm" Nov 24 14:43:08 crc kubenswrapper[4790]: I1124 14:43:08.542971 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0b7cfed5-a39c-4bee-b427-f74b2af51e73-db-sync-config-data\") pod \"glance-db-sync-tslzm\" (UID: \"0b7cfed5-a39c-4bee-b427-f74b2af51e73\") " pod="openstack/glance-db-sync-tslzm" Nov 24 14:43:08 crc kubenswrapper[4790]: I1124 14:43:08.543025 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9f4s4\" (UniqueName: \"kubernetes.io/projected/0b7cfed5-a39c-4bee-b427-f74b2af51e73-kube-api-access-9f4s4\") pod \"glance-db-sync-tslzm\" (UID: \"0b7cfed5-a39c-4bee-b427-f74b2af51e73\") " pod="openstack/glance-db-sync-tslzm" Nov 24 14:43:08 crc kubenswrapper[4790]: I1124 14:43:08.644818 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0b7cfed5-a39c-4bee-b427-f74b2af51e73-db-sync-config-data\") pod \"glance-db-sync-tslzm\" (UID: \"0b7cfed5-a39c-4bee-b427-f74b2af51e73\") " pod="openstack/glance-db-sync-tslzm" Nov 24 14:43:08 crc kubenswrapper[4790]: I1124 14:43:08.645008 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9f4s4\" (UniqueName: \"kubernetes.io/projected/0b7cfed5-a39c-4bee-b427-f74b2af51e73-kube-api-access-9f4s4\") pod \"glance-db-sync-tslzm\" (UID: \"0b7cfed5-a39c-4bee-b427-f74b2af51e73\") " pod="openstack/glance-db-sync-tslzm" Nov 24 14:43:08 crc kubenswrapper[4790]: I1124 14:43:08.645152 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b7cfed5-a39c-4bee-b427-f74b2af51e73-combined-ca-bundle\") pod \"glance-db-sync-tslzm\" (UID: \"0b7cfed5-a39c-4bee-b427-f74b2af51e73\") " pod="openstack/glance-db-sync-tslzm" Nov 24 14:43:08 crc kubenswrapper[4790]: I1124 14:43:08.645298 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b7cfed5-a39c-4bee-b427-f74b2af51e73-config-data\") pod \"glance-db-sync-tslzm\" (UID: \"0b7cfed5-a39c-4bee-b427-f74b2af51e73\") " pod="openstack/glance-db-sync-tslzm" Nov 24 14:43:08 crc kubenswrapper[4790]: I1124 14:43:08.655427 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b7cfed5-a39c-4bee-b427-f74b2af51e73-config-data\") pod \"glance-db-sync-tslzm\" (UID: \"0b7cfed5-a39c-4bee-b427-f74b2af51e73\") " pod="openstack/glance-db-sync-tslzm" Nov 24 14:43:08 crc kubenswrapper[4790]: I1124 14:43:08.655429 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0b7cfed5-a39c-4bee-b427-f74b2af51e73-db-sync-config-data\") pod \"glance-db-sync-tslzm\" (UID: \"0b7cfed5-a39c-4bee-b427-f74b2af51e73\") " pod="openstack/glance-db-sync-tslzm" Nov 24 14:43:08 crc kubenswrapper[4790]: I1124 14:43:08.657641 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b7cfed5-a39c-4bee-b427-f74b2af51e73-combined-ca-bundle\") pod \"glance-db-sync-tslzm\" (UID: \"0b7cfed5-a39c-4bee-b427-f74b2af51e73\") " pod="openstack/glance-db-sync-tslzm" Nov 24 14:43:08 crc kubenswrapper[4790]: I1124 14:43:08.672291 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9f4s4\" (UniqueName: \"kubernetes.io/projected/0b7cfed5-a39c-4bee-b427-f74b2af51e73-kube-api-access-9f4s4\") pod \"glance-db-sync-tslzm\" (UID: \"0b7cfed5-a39c-4bee-b427-f74b2af51e73\") " pod="openstack/glance-db-sync-tslzm" Nov 24 14:43:08 crc kubenswrapper[4790]: I1124 14:43:08.694142 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tslzm" Nov 24 14:43:09 crc kubenswrapper[4790]: I1124 14:43:09.238415 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-tslzm"] Nov 24 14:43:09 crc kubenswrapper[4790]: I1124 14:43:09.666391 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tslzm" event={"ID":"0b7cfed5-a39c-4bee-b427-f74b2af51e73","Type":"ContainerStarted","Data":"f5ce55303820eac6858ac02cd20584937dcea7ef9fe2366e4c27e8ddf6e3aa72"} Nov 24 14:43:10 crc kubenswrapper[4790]: I1124 14:43:10.681836 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tslzm" event={"ID":"0b7cfed5-a39c-4bee-b427-f74b2af51e73","Type":"ContainerStarted","Data":"0a63b354c131cec2adc6fe916b132e0b8847ff1811f52a961791676583a7da0e"} Nov 24 14:43:10 crc kubenswrapper[4790]: I1124 14:43:10.716555 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-tslzm" podStartSLOduration=2.716470477 podStartE2EDuration="2.716470477s" podCreationTimestamp="2025-11-24 14:43:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:43:10.703614068 +0000 UTC m=+5439.083507730" watchObservedRunningTime="2025-11-24 14:43:10.716470477 +0000 UTC m=+5439.096364179" Nov 24 14:43:12 crc kubenswrapper[4790]: I1124 14:43:12.322748 4790 scope.go:117] "RemoveContainer" containerID="92c086e44ccb28a62777d612e499209f7483422ca70b460ce6ec73c416e8e8b8" Nov 24 14:43:12 crc kubenswrapper[4790]: E1124 14:43:12.323346 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:43:13 crc kubenswrapper[4790]: I1124 14:43:13.713836 4790 generic.go:334] "Generic (PLEG): container finished" podID="0b7cfed5-a39c-4bee-b427-f74b2af51e73" containerID="0a63b354c131cec2adc6fe916b132e0b8847ff1811f52a961791676583a7da0e" exitCode=0 Nov 24 14:43:13 crc kubenswrapper[4790]: I1124 14:43:13.713913 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tslzm" event={"ID":"0b7cfed5-a39c-4bee-b427-f74b2af51e73","Type":"ContainerDied","Data":"0a63b354c131cec2adc6fe916b132e0b8847ff1811f52a961791676583a7da0e"} Nov 24 14:43:15 crc kubenswrapper[4790]: I1124 14:43:15.164202 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tslzm" Nov 24 14:43:15 crc kubenswrapper[4790]: I1124 14:43:15.266113 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b7cfed5-a39c-4bee-b427-f74b2af51e73-combined-ca-bundle\") pod \"0b7cfed5-a39c-4bee-b427-f74b2af51e73\" (UID: \"0b7cfed5-a39c-4bee-b427-f74b2af51e73\") " Nov 24 14:43:15 crc kubenswrapper[4790]: I1124 14:43:15.266198 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b7cfed5-a39c-4bee-b427-f74b2af51e73-config-data\") pod \"0b7cfed5-a39c-4bee-b427-f74b2af51e73\" (UID: \"0b7cfed5-a39c-4bee-b427-f74b2af51e73\") " Nov 24 14:43:15 crc kubenswrapper[4790]: I1124 14:43:15.266291 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0b7cfed5-a39c-4bee-b427-f74b2af51e73-db-sync-config-data\") pod \"0b7cfed5-a39c-4bee-b427-f74b2af51e73\" (UID: \"0b7cfed5-a39c-4bee-b427-f74b2af51e73\") " Nov 24 14:43:15 crc kubenswrapper[4790]: I1124 14:43:15.266407 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9f4s4\" (UniqueName: \"kubernetes.io/projected/0b7cfed5-a39c-4bee-b427-f74b2af51e73-kube-api-access-9f4s4\") pod \"0b7cfed5-a39c-4bee-b427-f74b2af51e73\" (UID: \"0b7cfed5-a39c-4bee-b427-f74b2af51e73\") " Nov 24 14:43:15 crc kubenswrapper[4790]: I1124 14:43:15.273784 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b7cfed5-a39c-4bee-b427-f74b2af51e73-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "0b7cfed5-a39c-4bee-b427-f74b2af51e73" (UID: "0b7cfed5-a39c-4bee-b427-f74b2af51e73"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:43:15 crc kubenswrapper[4790]: I1124 14:43:15.274281 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b7cfed5-a39c-4bee-b427-f74b2af51e73-kube-api-access-9f4s4" (OuterVolumeSpecName: "kube-api-access-9f4s4") pod "0b7cfed5-a39c-4bee-b427-f74b2af51e73" (UID: "0b7cfed5-a39c-4bee-b427-f74b2af51e73"). InnerVolumeSpecName "kube-api-access-9f4s4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:43:15 crc kubenswrapper[4790]: I1124 14:43:15.307410 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b7cfed5-a39c-4bee-b427-f74b2af51e73-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0b7cfed5-a39c-4bee-b427-f74b2af51e73" (UID: "0b7cfed5-a39c-4bee-b427-f74b2af51e73"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:43:15 crc kubenswrapper[4790]: I1124 14:43:15.316101 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b7cfed5-a39c-4bee-b427-f74b2af51e73-config-data" (OuterVolumeSpecName: "config-data") pod "0b7cfed5-a39c-4bee-b427-f74b2af51e73" (UID: "0b7cfed5-a39c-4bee-b427-f74b2af51e73"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:43:15 crc kubenswrapper[4790]: I1124 14:43:15.368066 4790 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0b7cfed5-a39c-4bee-b427-f74b2af51e73-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:15 crc kubenswrapper[4790]: I1124 14:43:15.368101 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9f4s4\" (UniqueName: \"kubernetes.io/projected/0b7cfed5-a39c-4bee-b427-f74b2af51e73-kube-api-access-9f4s4\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:15 crc kubenswrapper[4790]: I1124 14:43:15.368112 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b7cfed5-a39c-4bee-b427-f74b2af51e73-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:15 crc kubenswrapper[4790]: I1124 14:43:15.368120 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b7cfed5-a39c-4bee-b427-f74b2af51e73-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:15 crc kubenswrapper[4790]: I1124 14:43:15.738819 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tslzm" event={"ID":"0b7cfed5-a39c-4bee-b427-f74b2af51e73","Type":"ContainerDied","Data":"f5ce55303820eac6858ac02cd20584937dcea7ef9fe2366e4c27e8ddf6e3aa72"} Nov 24 14:43:15 crc kubenswrapper[4790]: I1124 14:43:15.738982 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f5ce55303820eac6858ac02cd20584937dcea7ef9fe2366e4c27e8ddf6e3aa72" Nov 24 14:43:15 crc kubenswrapper[4790]: I1124 14:43:15.739082 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tslzm" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.109549 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:43:16 crc kubenswrapper[4790]: E1124 14:43:16.110036 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b7cfed5-a39c-4bee-b427-f74b2af51e73" containerName="glance-db-sync" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.110059 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b7cfed5-a39c-4bee-b427-f74b2af51e73" containerName="glance-db-sync" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.110259 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b7cfed5-a39c-4bee-b427-f74b2af51e73" containerName="glance-db-sync" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.113981 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.118475 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-kjz5s" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.118494 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.118969 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.120632 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.128868 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.184723 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e4e598de-a07c-4442-95be-4e4581ccfecc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e4e598de-a07c-4442-95be-4e4581ccfecc\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.185117 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4e598de-a07c-4442-95be-4e4581ccfecc-config-data\") pod \"glance-default-external-api-0\" (UID: \"e4e598de-a07c-4442-95be-4e4581ccfecc\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.185165 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6t9h\" (UniqueName: \"kubernetes.io/projected/e4e598de-a07c-4442-95be-4e4581ccfecc-kube-api-access-h6t9h\") pod \"glance-default-external-api-0\" (UID: \"e4e598de-a07c-4442-95be-4e4581ccfecc\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.185195 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4e598de-a07c-4442-95be-4e4581ccfecc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e4e598de-a07c-4442-95be-4e4581ccfecc\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.185249 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4e598de-a07c-4442-95be-4e4581ccfecc-logs\") pod \"glance-default-external-api-0\" (UID: \"e4e598de-a07c-4442-95be-4e4581ccfecc\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.185450 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4e598de-a07c-4442-95be-4e4581ccfecc-scripts\") pod \"glance-default-external-api-0\" (UID: \"e4e598de-a07c-4442-95be-4e4581ccfecc\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.185473 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e4e598de-a07c-4442-95be-4e4581ccfecc-ceph\") pod \"glance-default-external-api-0\" (UID: \"e4e598de-a07c-4442-95be-4e4581ccfecc\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.234109 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5dd6c96d9c-4mllx"] Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.235494 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dd6c96d9c-4mllx" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.256038 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5dd6c96d9c-4mllx"] Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.287624 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4e598de-a07c-4442-95be-4e4581ccfecc-scripts\") pod \"glance-default-external-api-0\" (UID: \"e4e598de-a07c-4442-95be-4e4581ccfecc\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.287685 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e4e598de-a07c-4442-95be-4e4581ccfecc-ceph\") pod \"glance-default-external-api-0\" (UID: \"e4e598de-a07c-4442-95be-4e4581ccfecc\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.287792 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e4e598de-a07c-4442-95be-4e4581ccfecc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e4e598de-a07c-4442-95be-4e4581ccfecc\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.287818 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4e598de-a07c-4442-95be-4e4581ccfecc-config-data\") pod \"glance-default-external-api-0\" (UID: \"e4e598de-a07c-4442-95be-4e4581ccfecc\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.287869 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6t9h\" (UniqueName: \"kubernetes.io/projected/e4e598de-a07c-4442-95be-4e4581ccfecc-kube-api-access-h6t9h\") pod \"glance-default-external-api-0\" (UID: \"e4e598de-a07c-4442-95be-4e4581ccfecc\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.287917 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4e598de-a07c-4442-95be-4e4581ccfecc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e4e598de-a07c-4442-95be-4e4581ccfecc\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.287977 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4e598de-a07c-4442-95be-4e4581ccfecc-logs\") pod \"glance-default-external-api-0\" (UID: \"e4e598de-a07c-4442-95be-4e4581ccfecc\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.288517 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e4e598de-a07c-4442-95be-4e4581ccfecc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e4e598de-a07c-4442-95be-4e4581ccfecc\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.288595 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4e598de-a07c-4442-95be-4e4581ccfecc-logs\") pod \"glance-default-external-api-0\" (UID: \"e4e598de-a07c-4442-95be-4e4581ccfecc\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.292433 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4e598de-a07c-4442-95be-4e4581ccfecc-config-data\") pod \"glance-default-external-api-0\" (UID: \"e4e598de-a07c-4442-95be-4e4581ccfecc\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.293040 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e4e598de-a07c-4442-95be-4e4581ccfecc-ceph\") pod \"glance-default-external-api-0\" (UID: \"e4e598de-a07c-4442-95be-4e4581ccfecc\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.295165 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4e598de-a07c-4442-95be-4e4581ccfecc-scripts\") pod \"glance-default-external-api-0\" (UID: \"e4e598de-a07c-4442-95be-4e4581ccfecc\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.296317 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4e598de-a07c-4442-95be-4e4581ccfecc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e4e598de-a07c-4442-95be-4e4581ccfecc\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.309819 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6t9h\" (UniqueName: \"kubernetes.io/projected/e4e598de-a07c-4442-95be-4e4581ccfecc-kube-api-access-h6t9h\") pod \"glance-default-external-api-0\" (UID: \"e4e598de-a07c-4442-95be-4e4581ccfecc\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.328843 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.331422 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.334259 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.351694 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.389267 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a903782f-4ab2-4439-92b0-bd0ddef96176-ovsdbserver-nb\") pod \"dnsmasq-dns-5dd6c96d9c-4mllx\" (UID: \"a903782f-4ab2-4439-92b0-bd0ddef96176\") " pod="openstack/dnsmasq-dns-5dd6c96d9c-4mllx" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.389309 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a903782f-4ab2-4439-92b0-bd0ddef96176-ovsdbserver-sb\") pod \"dnsmasq-dns-5dd6c96d9c-4mllx\" (UID: \"a903782f-4ab2-4439-92b0-bd0ddef96176\") " pod="openstack/dnsmasq-dns-5dd6c96d9c-4mllx" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.389364 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a903782f-4ab2-4439-92b0-bd0ddef96176-config\") pod \"dnsmasq-dns-5dd6c96d9c-4mllx\" (UID: \"a903782f-4ab2-4439-92b0-bd0ddef96176\") " pod="openstack/dnsmasq-dns-5dd6c96d9c-4mllx" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.389384 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fwlc\" (UniqueName: \"kubernetes.io/projected/a903782f-4ab2-4439-92b0-bd0ddef96176-kube-api-access-7fwlc\") pod \"dnsmasq-dns-5dd6c96d9c-4mllx\" (UID: \"a903782f-4ab2-4439-92b0-bd0ddef96176\") " pod="openstack/dnsmasq-dns-5dd6c96d9c-4mllx" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.389469 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a903782f-4ab2-4439-92b0-bd0ddef96176-dns-svc\") pod \"dnsmasq-dns-5dd6c96d9c-4mllx\" (UID: \"a903782f-4ab2-4439-92b0-bd0ddef96176\") " pod="openstack/dnsmasq-dns-5dd6c96d9c-4mllx" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.446174 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.491155 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/876e67dc-9b8e-4be5-8540-1016be2b320b-logs\") pod \"glance-default-internal-api-0\" (UID: \"876e67dc-9b8e-4be5-8540-1016be2b320b\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.491233 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a903782f-4ab2-4439-92b0-bd0ddef96176-ovsdbserver-sb\") pod \"dnsmasq-dns-5dd6c96d9c-4mllx\" (UID: \"a903782f-4ab2-4439-92b0-bd0ddef96176\") " pod="openstack/dnsmasq-dns-5dd6c96d9c-4mllx" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.491341 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a903782f-4ab2-4439-92b0-bd0ddef96176-ovsdbserver-nb\") pod \"dnsmasq-dns-5dd6c96d9c-4mllx\" (UID: \"a903782f-4ab2-4439-92b0-bd0ddef96176\") " pod="openstack/dnsmasq-dns-5dd6c96d9c-4mllx" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.491443 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/876e67dc-9b8e-4be5-8540-1016be2b320b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"876e67dc-9b8e-4be5-8540-1016be2b320b\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.491477 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a903782f-4ab2-4439-92b0-bd0ddef96176-config\") pod \"dnsmasq-dns-5dd6c96d9c-4mllx\" (UID: \"a903782f-4ab2-4439-92b0-bd0ddef96176\") " pod="openstack/dnsmasq-dns-5dd6c96d9c-4mllx" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.491499 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fwlc\" (UniqueName: \"kubernetes.io/projected/a903782f-4ab2-4439-92b0-bd0ddef96176-kube-api-access-7fwlc\") pod \"dnsmasq-dns-5dd6c96d9c-4mllx\" (UID: \"a903782f-4ab2-4439-92b0-bd0ddef96176\") " pod="openstack/dnsmasq-dns-5dd6c96d9c-4mllx" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.491667 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/876e67dc-9b8e-4be5-8540-1016be2b320b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"876e67dc-9b8e-4be5-8540-1016be2b320b\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.491771 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/876e67dc-9b8e-4be5-8540-1016be2b320b-ceph\") pod \"glance-default-internal-api-0\" (UID: \"876e67dc-9b8e-4be5-8540-1016be2b320b\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.491963 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a903782f-4ab2-4439-92b0-bd0ddef96176-dns-svc\") pod \"dnsmasq-dns-5dd6c96d9c-4mllx\" (UID: \"a903782f-4ab2-4439-92b0-bd0ddef96176\") " pod="openstack/dnsmasq-dns-5dd6c96d9c-4mllx" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.491995 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/876e67dc-9b8e-4be5-8540-1016be2b320b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"876e67dc-9b8e-4be5-8540-1016be2b320b\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.492047 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/876e67dc-9b8e-4be5-8540-1016be2b320b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"876e67dc-9b8e-4be5-8540-1016be2b320b\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.492072 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfkrk\" (UniqueName: \"kubernetes.io/projected/876e67dc-9b8e-4be5-8540-1016be2b320b-kube-api-access-sfkrk\") pod \"glance-default-internal-api-0\" (UID: \"876e67dc-9b8e-4be5-8540-1016be2b320b\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.492258 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a903782f-4ab2-4439-92b0-bd0ddef96176-ovsdbserver-nb\") pod \"dnsmasq-dns-5dd6c96d9c-4mllx\" (UID: \"a903782f-4ab2-4439-92b0-bd0ddef96176\") " pod="openstack/dnsmasq-dns-5dd6c96d9c-4mllx" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.492269 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a903782f-4ab2-4439-92b0-bd0ddef96176-ovsdbserver-sb\") pod \"dnsmasq-dns-5dd6c96d9c-4mllx\" (UID: \"a903782f-4ab2-4439-92b0-bd0ddef96176\") " pod="openstack/dnsmasq-dns-5dd6c96d9c-4mllx" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.492541 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a903782f-4ab2-4439-92b0-bd0ddef96176-config\") pod \"dnsmasq-dns-5dd6c96d9c-4mllx\" (UID: \"a903782f-4ab2-4439-92b0-bd0ddef96176\") " pod="openstack/dnsmasq-dns-5dd6c96d9c-4mllx" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.492941 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a903782f-4ab2-4439-92b0-bd0ddef96176-dns-svc\") pod \"dnsmasq-dns-5dd6c96d9c-4mllx\" (UID: \"a903782f-4ab2-4439-92b0-bd0ddef96176\") " pod="openstack/dnsmasq-dns-5dd6c96d9c-4mllx" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.511170 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fwlc\" (UniqueName: \"kubernetes.io/projected/a903782f-4ab2-4439-92b0-bd0ddef96176-kube-api-access-7fwlc\") pod \"dnsmasq-dns-5dd6c96d9c-4mllx\" (UID: \"a903782f-4ab2-4439-92b0-bd0ddef96176\") " pod="openstack/dnsmasq-dns-5dd6c96d9c-4mllx" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.554802 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dd6c96d9c-4mllx" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.594282 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/876e67dc-9b8e-4be5-8540-1016be2b320b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"876e67dc-9b8e-4be5-8540-1016be2b320b\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.594580 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/876e67dc-9b8e-4be5-8540-1016be2b320b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"876e67dc-9b8e-4be5-8540-1016be2b320b\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.594610 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/876e67dc-9b8e-4be5-8540-1016be2b320b-ceph\") pod \"glance-default-internal-api-0\" (UID: \"876e67dc-9b8e-4be5-8540-1016be2b320b\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.594670 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/876e67dc-9b8e-4be5-8540-1016be2b320b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"876e67dc-9b8e-4be5-8540-1016be2b320b\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.594699 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/876e67dc-9b8e-4be5-8540-1016be2b320b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"876e67dc-9b8e-4be5-8540-1016be2b320b\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.594718 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfkrk\" (UniqueName: \"kubernetes.io/projected/876e67dc-9b8e-4be5-8540-1016be2b320b-kube-api-access-sfkrk\") pod \"glance-default-internal-api-0\" (UID: \"876e67dc-9b8e-4be5-8540-1016be2b320b\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.594762 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/876e67dc-9b8e-4be5-8540-1016be2b320b-logs\") pod \"glance-default-internal-api-0\" (UID: \"876e67dc-9b8e-4be5-8540-1016be2b320b\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.594810 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/876e67dc-9b8e-4be5-8540-1016be2b320b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"876e67dc-9b8e-4be5-8540-1016be2b320b\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.595186 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/876e67dc-9b8e-4be5-8540-1016be2b320b-logs\") pod \"glance-default-internal-api-0\" (UID: \"876e67dc-9b8e-4be5-8540-1016be2b320b\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.600974 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/876e67dc-9b8e-4be5-8540-1016be2b320b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"876e67dc-9b8e-4be5-8540-1016be2b320b\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.601406 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/876e67dc-9b8e-4be5-8540-1016be2b320b-ceph\") pod \"glance-default-internal-api-0\" (UID: \"876e67dc-9b8e-4be5-8540-1016be2b320b\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.601431 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/876e67dc-9b8e-4be5-8540-1016be2b320b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"876e67dc-9b8e-4be5-8540-1016be2b320b\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.604967 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/876e67dc-9b8e-4be5-8540-1016be2b320b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"876e67dc-9b8e-4be5-8540-1016be2b320b\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.611174 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfkrk\" (UniqueName: \"kubernetes.io/projected/876e67dc-9b8e-4be5-8540-1016be2b320b-kube-api-access-sfkrk\") pod \"glance-default-internal-api-0\" (UID: \"876e67dc-9b8e-4be5-8540-1016be2b320b\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.697273 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:43:16 crc kubenswrapper[4790]: I1124 14:43:16.961114 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:43:17 crc kubenswrapper[4790]: I1124 14:43:17.046139 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5dd6c96d9c-4mllx"] Nov 24 14:43:17 crc kubenswrapper[4790]: W1124 14:43:17.055230 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda903782f_4ab2_4439_92b0_bd0ddef96176.slice/crio-0ea4cc6c19f5b453ffb420a71677208b805a48818f1d31a3aa184267bc6f1ff2 WatchSource:0}: Error finding container 0ea4cc6c19f5b453ffb420a71677208b805a48818f1d31a3aa184267bc6f1ff2: Status 404 returned error can't find the container with id 0ea4cc6c19f5b453ffb420a71677208b805a48818f1d31a3aa184267bc6f1ff2 Nov 24 14:43:17 crc kubenswrapper[4790]: I1124 14:43:17.175856 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:43:17 crc kubenswrapper[4790]: I1124 14:43:17.246320 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:43:17 crc kubenswrapper[4790]: W1124 14:43:17.255694 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod876e67dc_9b8e_4be5_8540_1016be2b320b.slice/crio-c4b693566d300070b0e59273b796db3a31ba1688e4c61bc7db1264db1b99b3f4 WatchSource:0}: Error finding container c4b693566d300070b0e59273b796db3a31ba1688e4c61bc7db1264db1b99b3f4: Status 404 returned error can't find the container with id c4b693566d300070b0e59273b796db3a31ba1688e4c61bc7db1264db1b99b3f4 Nov 24 14:43:17 crc kubenswrapper[4790]: I1124 14:43:17.795281 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"876e67dc-9b8e-4be5-8540-1016be2b320b","Type":"ContainerStarted","Data":"c4b693566d300070b0e59273b796db3a31ba1688e4c61bc7db1264db1b99b3f4"} Nov 24 14:43:17 crc kubenswrapper[4790]: I1124 14:43:17.797186 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e4e598de-a07c-4442-95be-4e4581ccfecc","Type":"ContainerStarted","Data":"9b0a53cf24994feff22fa9e596463277a5fed73b33d48ac34110c61327797409"} Nov 24 14:43:17 crc kubenswrapper[4790]: I1124 14:43:17.797242 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e4e598de-a07c-4442-95be-4e4581ccfecc","Type":"ContainerStarted","Data":"0663fd424e343355438a85cd6cba93f84222d7065e6511f3c1c8dae912275ded"} Nov 24 14:43:17 crc kubenswrapper[4790]: I1124 14:43:17.804431 4790 generic.go:334] "Generic (PLEG): container finished" podID="a903782f-4ab2-4439-92b0-bd0ddef96176" containerID="b9218f5e28274a6d669b38671f94950773306acf16e25bbde80f20207a697021" exitCode=0 Nov 24 14:43:17 crc kubenswrapper[4790]: I1124 14:43:17.804501 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dd6c96d9c-4mllx" event={"ID":"a903782f-4ab2-4439-92b0-bd0ddef96176","Type":"ContainerDied","Data":"b9218f5e28274a6d669b38671f94950773306acf16e25bbde80f20207a697021"} Nov 24 14:43:17 crc kubenswrapper[4790]: I1124 14:43:17.804585 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dd6c96d9c-4mllx" event={"ID":"a903782f-4ab2-4439-92b0-bd0ddef96176","Type":"ContainerStarted","Data":"0ea4cc6c19f5b453ffb420a71677208b805a48818f1d31a3aa184267bc6f1ff2"} Nov 24 14:43:18 crc kubenswrapper[4790]: I1124 14:43:18.814410 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e4e598de-a07c-4442-95be-4e4581ccfecc","Type":"ContainerStarted","Data":"f1c86dc59c7e15be123331c76847778676fb8cb7b331a1654d6b1ccd5514004e"} Nov 24 14:43:18 crc kubenswrapper[4790]: I1124 14:43:18.814552 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e4e598de-a07c-4442-95be-4e4581ccfecc" containerName="glance-log" containerID="cri-o://9b0a53cf24994feff22fa9e596463277a5fed73b33d48ac34110c61327797409" gracePeriod=30 Nov 24 14:43:18 crc kubenswrapper[4790]: I1124 14:43:18.814659 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e4e598de-a07c-4442-95be-4e4581ccfecc" containerName="glance-httpd" containerID="cri-o://f1c86dc59c7e15be123331c76847778676fb8cb7b331a1654d6b1ccd5514004e" gracePeriod=30 Nov 24 14:43:18 crc kubenswrapper[4790]: I1124 14:43:18.818692 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dd6c96d9c-4mllx" event={"ID":"a903782f-4ab2-4439-92b0-bd0ddef96176","Type":"ContainerStarted","Data":"1f0586302e08eb2f9b2f20f43edd13990a2e9aa8f223fe3d1d3b60518be779ed"} Nov 24 14:43:18 crc kubenswrapper[4790]: I1124 14:43:18.819219 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5dd6c96d9c-4mllx" Nov 24 14:43:18 crc kubenswrapper[4790]: I1124 14:43:18.822600 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"876e67dc-9b8e-4be5-8540-1016be2b320b","Type":"ContainerStarted","Data":"7a4eaa8754c042c3e140e113fb7d8e11fea794f34887e23cfdf8227f4339e622"} Nov 24 14:43:18 crc kubenswrapper[4790]: I1124 14:43:18.822703 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"876e67dc-9b8e-4be5-8540-1016be2b320b","Type":"ContainerStarted","Data":"f849df65aaf49ab550a0e1b52f4272e5e72e4c702435be5e9bd86899280eb996"} Nov 24 14:43:18 crc kubenswrapper[4790]: I1124 14:43:18.869975 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=2.869936559 podStartE2EDuration="2.869936559s" podCreationTimestamp="2025-11-24 14:43:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:43:18.85535857 +0000 UTC m=+5447.235252232" watchObservedRunningTime="2025-11-24 14:43:18.869936559 +0000 UTC m=+5447.249830251" Nov 24 14:43:18 crc kubenswrapper[4790]: I1124 14:43:18.882693 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=2.882670695 podStartE2EDuration="2.882670695s" podCreationTimestamp="2025-11-24 14:43:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:43:18.877165367 +0000 UTC m=+5447.257059049" watchObservedRunningTime="2025-11-24 14:43:18.882670695 +0000 UTC m=+5447.262564357" Nov 24 14:43:18 crc kubenswrapper[4790]: I1124 14:43:18.905633 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5dd6c96d9c-4mllx" podStartSLOduration=2.905611863 podStartE2EDuration="2.905611863s" podCreationTimestamp="2025-11-24 14:43:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:43:18.902053631 +0000 UTC m=+5447.281947333" watchObservedRunningTime="2025-11-24 14:43:18.905611863 +0000 UTC m=+5447.285505525" Nov 24 14:43:18 crc kubenswrapper[4790]: I1124 14:43:18.911920 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.412307 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.558277 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6t9h\" (UniqueName: \"kubernetes.io/projected/e4e598de-a07c-4442-95be-4e4581ccfecc-kube-api-access-h6t9h\") pod \"e4e598de-a07c-4442-95be-4e4581ccfecc\" (UID: \"e4e598de-a07c-4442-95be-4e4581ccfecc\") " Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.558346 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e4e598de-a07c-4442-95be-4e4581ccfecc-httpd-run\") pod \"e4e598de-a07c-4442-95be-4e4581ccfecc\" (UID: \"e4e598de-a07c-4442-95be-4e4581ccfecc\") " Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.558471 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4e598de-a07c-4442-95be-4e4581ccfecc-scripts\") pod \"e4e598de-a07c-4442-95be-4e4581ccfecc\" (UID: \"e4e598de-a07c-4442-95be-4e4581ccfecc\") " Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.558577 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4e598de-a07c-4442-95be-4e4581ccfecc-combined-ca-bundle\") pod \"e4e598de-a07c-4442-95be-4e4581ccfecc\" (UID: \"e4e598de-a07c-4442-95be-4e4581ccfecc\") " Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.558601 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e4e598de-a07c-4442-95be-4e4581ccfecc-ceph\") pod \"e4e598de-a07c-4442-95be-4e4581ccfecc\" (UID: \"e4e598de-a07c-4442-95be-4e4581ccfecc\") " Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.558626 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4e598de-a07c-4442-95be-4e4581ccfecc-config-data\") pod \"e4e598de-a07c-4442-95be-4e4581ccfecc\" (UID: \"e4e598de-a07c-4442-95be-4e4581ccfecc\") " Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.558676 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4e598de-a07c-4442-95be-4e4581ccfecc-logs\") pod \"e4e598de-a07c-4442-95be-4e4581ccfecc\" (UID: \"e4e598de-a07c-4442-95be-4e4581ccfecc\") " Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.559406 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4e598de-a07c-4442-95be-4e4581ccfecc-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e4e598de-a07c-4442-95be-4e4581ccfecc" (UID: "e4e598de-a07c-4442-95be-4e4581ccfecc"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.560281 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4e598de-a07c-4442-95be-4e4581ccfecc-logs" (OuterVolumeSpecName: "logs") pod "e4e598de-a07c-4442-95be-4e4581ccfecc" (UID: "e4e598de-a07c-4442-95be-4e4581ccfecc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.565068 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4e598de-a07c-4442-95be-4e4581ccfecc-ceph" (OuterVolumeSpecName: "ceph") pod "e4e598de-a07c-4442-95be-4e4581ccfecc" (UID: "e4e598de-a07c-4442-95be-4e4581ccfecc"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.565727 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4e598de-a07c-4442-95be-4e4581ccfecc-kube-api-access-h6t9h" (OuterVolumeSpecName: "kube-api-access-h6t9h") pod "e4e598de-a07c-4442-95be-4e4581ccfecc" (UID: "e4e598de-a07c-4442-95be-4e4581ccfecc"). InnerVolumeSpecName "kube-api-access-h6t9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.567958 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4e598de-a07c-4442-95be-4e4581ccfecc-scripts" (OuterVolumeSpecName: "scripts") pod "e4e598de-a07c-4442-95be-4e4581ccfecc" (UID: "e4e598de-a07c-4442-95be-4e4581ccfecc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.597965 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4e598de-a07c-4442-95be-4e4581ccfecc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e4e598de-a07c-4442-95be-4e4581ccfecc" (UID: "e4e598de-a07c-4442-95be-4e4581ccfecc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.633496 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4e598de-a07c-4442-95be-4e4581ccfecc-config-data" (OuterVolumeSpecName: "config-data") pod "e4e598de-a07c-4442-95be-4e4581ccfecc" (UID: "e4e598de-a07c-4442-95be-4e4581ccfecc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.660961 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4e598de-a07c-4442-95be-4e4581ccfecc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.660998 4790 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e4e598de-a07c-4442-95be-4e4581ccfecc-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.661009 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4e598de-a07c-4442-95be-4e4581ccfecc-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.661018 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4e598de-a07c-4442-95be-4e4581ccfecc-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.661027 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6t9h\" (UniqueName: \"kubernetes.io/projected/e4e598de-a07c-4442-95be-4e4581ccfecc-kube-api-access-h6t9h\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.661037 4790 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e4e598de-a07c-4442-95be-4e4581ccfecc-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.661046 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4e598de-a07c-4442-95be-4e4581ccfecc-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.836413 4790 generic.go:334] "Generic (PLEG): container finished" podID="e4e598de-a07c-4442-95be-4e4581ccfecc" containerID="f1c86dc59c7e15be123331c76847778676fb8cb7b331a1654d6b1ccd5514004e" exitCode=0 Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.837108 4790 generic.go:334] "Generic (PLEG): container finished" podID="e4e598de-a07c-4442-95be-4e4581ccfecc" containerID="9b0a53cf24994feff22fa9e596463277a5fed73b33d48ac34110c61327797409" exitCode=143 Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.836544 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.836489 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e4e598de-a07c-4442-95be-4e4581ccfecc","Type":"ContainerDied","Data":"f1c86dc59c7e15be123331c76847778676fb8cb7b331a1654d6b1ccd5514004e"} Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.837298 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e4e598de-a07c-4442-95be-4e4581ccfecc","Type":"ContainerDied","Data":"9b0a53cf24994feff22fa9e596463277a5fed73b33d48ac34110c61327797409"} Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.837339 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e4e598de-a07c-4442-95be-4e4581ccfecc","Type":"ContainerDied","Data":"0663fd424e343355438a85cd6cba93f84222d7065e6511f3c1c8dae912275ded"} Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.837386 4790 scope.go:117] "RemoveContainer" containerID="f1c86dc59c7e15be123331c76847778676fb8cb7b331a1654d6b1ccd5514004e" Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.879120 4790 scope.go:117] "RemoveContainer" containerID="9b0a53cf24994feff22fa9e596463277a5fed73b33d48ac34110c61327797409" Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.889333 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.896173 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.915027 4790 scope.go:117] "RemoveContainer" containerID="f1c86dc59c7e15be123331c76847778676fb8cb7b331a1654d6b1ccd5514004e" Nov 24 14:43:19 crc kubenswrapper[4790]: E1124 14:43:19.915758 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1c86dc59c7e15be123331c76847778676fb8cb7b331a1654d6b1ccd5514004e\": container with ID starting with f1c86dc59c7e15be123331c76847778676fb8cb7b331a1654d6b1ccd5514004e not found: ID does not exist" containerID="f1c86dc59c7e15be123331c76847778676fb8cb7b331a1654d6b1ccd5514004e" Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.915914 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1c86dc59c7e15be123331c76847778676fb8cb7b331a1654d6b1ccd5514004e"} err="failed to get container status \"f1c86dc59c7e15be123331c76847778676fb8cb7b331a1654d6b1ccd5514004e\": rpc error: code = NotFound desc = could not find container \"f1c86dc59c7e15be123331c76847778676fb8cb7b331a1654d6b1ccd5514004e\": container with ID starting with f1c86dc59c7e15be123331c76847778676fb8cb7b331a1654d6b1ccd5514004e not found: ID does not exist" Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.916043 4790 scope.go:117] "RemoveContainer" containerID="9b0a53cf24994feff22fa9e596463277a5fed73b33d48ac34110c61327797409" Nov 24 14:43:19 crc kubenswrapper[4790]: E1124 14:43:19.916371 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b0a53cf24994feff22fa9e596463277a5fed73b33d48ac34110c61327797409\": container with ID starting with 9b0a53cf24994feff22fa9e596463277a5fed73b33d48ac34110c61327797409 not found: ID does not exist" containerID="9b0a53cf24994feff22fa9e596463277a5fed73b33d48ac34110c61327797409" Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.916472 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b0a53cf24994feff22fa9e596463277a5fed73b33d48ac34110c61327797409"} err="failed to get container status \"9b0a53cf24994feff22fa9e596463277a5fed73b33d48ac34110c61327797409\": rpc error: code = NotFound desc = could not find container \"9b0a53cf24994feff22fa9e596463277a5fed73b33d48ac34110c61327797409\": container with ID starting with 9b0a53cf24994feff22fa9e596463277a5fed73b33d48ac34110c61327797409 not found: ID does not exist" Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.916581 4790 scope.go:117] "RemoveContainer" containerID="f1c86dc59c7e15be123331c76847778676fb8cb7b331a1654d6b1ccd5514004e" Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.916985 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1c86dc59c7e15be123331c76847778676fb8cb7b331a1654d6b1ccd5514004e"} err="failed to get container status \"f1c86dc59c7e15be123331c76847778676fb8cb7b331a1654d6b1ccd5514004e\": rpc error: code = NotFound desc = could not find container \"f1c86dc59c7e15be123331c76847778676fb8cb7b331a1654d6b1ccd5514004e\": container with ID starting with f1c86dc59c7e15be123331c76847778676fb8cb7b331a1654d6b1ccd5514004e not found: ID does not exist" Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.917097 4790 scope.go:117] "RemoveContainer" containerID="9b0a53cf24994feff22fa9e596463277a5fed73b33d48ac34110c61327797409" Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.917546 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b0a53cf24994feff22fa9e596463277a5fed73b33d48ac34110c61327797409"} err="failed to get container status \"9b0a53cf24994feff22fa9e596463277a5fed73b33d48ac34110c61327797409\": rpc error: code = NotFound desc = could not find container \"9b0a53cf24994feff22fa9e596463277a5fed73b33d48ac34110c61327797409\": container with ID starting with 9b0a53cf24994feff22fa9e596463277a5fed73b33d48ac34110c61327797409 not found: ID does not exist" Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.930264 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:43:19 crc kubenswrapper[4790]: E1124 14:43:19.930723 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4e598de-a07c-4442-95be-4e4581ccfecc" containerName="glance-httpd" Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.930745 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4e598de-a07c-4442-95be-4e4581ccfecc" containerName="glance-httpd" Nov 24 14:43:19 crc kubenswrapper[4790]: E1124 14:43:19.930779 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4e598de-a07c-4442-95be-4e4581ccfecc" containerName="glance-log" Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.930789 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4e598de-a07c-4442-95be-4e4581ccfecc" containerName="glance-log" Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.931024 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4e598de-a07c-4442-95be-4e4581ccfecc" containerName="glance-httpd" Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.931054 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4e598de-a07c-4442-95be-4e4581ccfecc" containerName="glance-log" Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.932322 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.935070 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 14:43:19 crc kubenswrapper[4790]: I1124 14:43:19.946387 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:43:20 crc kubenswrapper[4790]: I1124 14:43:20.079202 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b8cad57-cddc-4273-a1be-498c1172ee79-scripts\") pod \"glance-default-external-api-0\" (UID: \"4b8cad57-cddc-4273-a1be-498c1172ee79\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:20 crc kubenswrapper[4790]: I1124 14:43:20.079308 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b8cad57-cddc-4273-a1be-498c1172ee79-config-data\") pod \"glance-default-external-api-0\" (UID: \"4b8cad57-cddc-4273-a1be-498c1172ee79\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:20 crc kubenswrapper[4790]: I1124 14:43:20.079358 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6wpp\" (UniqueName: \"kubernetes.io/projected/4b8cad57-cddc-4273-a1be-498c1172ee79-kube-api-access-p6wpp\") pod \"glance-default-external-api-0\" (UID: \"4b8cad57-cddc-4273-a1be-498c1172ee79\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:20 crc kubenswrapper[4790]: I1124 14:43:20.079462 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4b8cad57-cddc-4273-a1be-498c1172ee79-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4b8cad57-cddc-4273-a1be-498c1172ee79\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:20 crc kubenswrapper[4790]: I1124 14:43:20.079512 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b8cad57-cddc-4273-a1be-498c1172ee79-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4b8cad57-cddc-4273-a1be-498c1172ee79\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:20 crc kubenswrapper[4790]: I1124 14:43:20.079534 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b8cad57-cddc-4273-a1be-498c1172ee79-logs\") pod \"glance-default-external-api-0\" (UID: \"4b8cad57-cddc-4273-a1be-498c1172ee79\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:20 crc kubenswrapper[4790]: I1124 14:43:20.079574 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4b8cad57-cddc-4273-a1be-498c1172ee79-ceph\") pod \"glance-default-external-api-0\" (UID: \"4b8cad57-cddc-4273-a1be-498c1172ee79\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:20 crc kubenswrapper[4790]: I1124 14:43:20.181520 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4b8cad57-cddc-4273-a1be-498c1172ee79-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4b8cad57-cddc-4273-a1be-498c1172ee79\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:20 crc kubenswrapper[4790]: I1124 14:43:20.181652 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b8cad57-cddc-4273-a1be-498c1172ee79-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4b8cad57-cddc-4273-a1be-498c1172ee79\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:20 crc kubenswrapper[4790]: I1124 14:43:20.181691 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b8cad57-cddc-4273-a1be-498c1172ee79-logs\") pod \"glance-default-external-api-0\" (UID: \"4b8cad57-cddc-4273-a1be-498c1172ee79\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:20 crc kubenswrapper[4790]: I1124 14:43:20.181756 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4b8cad57-cddc-4273-a1be-498c1172ee79-ceph\") pod \"glance-default-external-api-0\" (UID: \"4b8cad57-cddc-4273-a1be-498c1172ee79\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:20 crc kubenswrapper[4790]: I1124 14:43:20.181836 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b8cad57-cddc-4273-a1be-498c1172ee79-scripts\") pod \"glance-default-external-api-0\" (UID: \"4b8cad57-cddc-4273-a1be-498c1172ee79\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:20 crc kubenswrapper[4790]: I1124 14:43:20.181968 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b8cad57-cddc-4273-a1be-498c1172ee79-config-data\") pod \"glance-default-external-api-0\" (UID: \"4b8cad57-cddc-4273-a1be-498c1172ee79\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:20 crc kubenswrapper[4790]: I1124 14:43:20.182028 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6wpp\" (UniqueName: \"kubernetes.io/projected/4b8cad57-cddc-4273-a1be-498c1172ee79-kube-api-access-p6wpp\") pod \"glance-default-external-api-0\" (UID: \"4b8cad57-cddc-4273-a1be-498c1172ee79\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:20 crc kubenswrapper[4790]: I1124 14:43:20.182413 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b8cad57-cddc-4273-a1be-498c1172ee79-logs\") pod \"glance-default-external-api-0\" (UID: \"4b8cad57-cddc-4273-a1be-498c1172ee79\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:20 crc kubenswrapper[4790]: I1124 14:43:20.182561 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4b8cad57-cddc-4273-a1be-498c1172ee79-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4b8cad57-cddc-4273-a1be-498c1172ee79\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:20 crc kubenswrapper[4790]: I1124 14:43:20.187264 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b8cad57-cddc-4273-a1be-498c1172ee79-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4b8cad57-cddc-4273-a1be-498c1172ee79\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:20 crc kubenswrapper[4790]: I1124 14:43:20.187629 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4b8cad57-cddc-4273-a1be-498c1172ee79-ceph\") pod \"glance-default-external-api-0\" (UID: \"4b8cad57-cddc-4273-a1be-498c1172ee79\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:20 crc kubenswrapper[4790]: I1124 14:43:20.188493 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b8cad57-cddc-4273-a1be-498c1172ee79-scripts\") pod \"glance-default-external-api-0\" (UID: \"4b8cad57-cddc-4273-a1be-498c1172ee79\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:20 crc kubenswrapper[4790]: I1124 14:43:20.190041 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b8cad57-cddc-4273-a1be-498c1172ee79-config-data\") pod \"glance-default-external-api-0\" (UID: \"4b8cad57-cddc-4273-a1be-498c1172ee79\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:20 crc kubenswrapper[4790]: I1124 14:43:20.203138 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6wpp\" (UniqueName: \"kubernetes.io/projected/4b8cad57-cddc-4273-a1be-498c1172ee79-kube-api-access-p6wpp\") pod \"glance-default-external-api-0\" (UID: \"4b8cad57-cddc-4273-a1be-498c1172ee79\") " pod="openstack/glance-default-external-api-0" Nov 24 14:43:20 crc kubenswrapper[4790]: I1124 14:43:20.247830 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:43:20 crc kubenswrapper[4790]: I1124 14:43:20.330734 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4e598de-a07c-4442-95be-4e4581ccfecc" path="/var/lib/kubelet/pods/e4e598de-a07c-4442-95be-4e4581ccfecc/volumes" Nov 24 14:43:20 crc kubenswrapper[4790]: I1124 14:43:20.858026 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="876e67dc-9b8e-4be5-8540-1016be2b320b" containerName="glance-log" containerID="cri-o://f849df65aaf49ab550a0e1b52f4272e5e72e4c702435be5e9bd86899280eb996" gracePeriod=30 Nov 24 14:43:20 crc kubenswrapper[4790]: I1124 14:43:20.858114 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="876e67dc-9b8e-4be5-8540-1016be2b320b" containerName="glance-httpd" containerID="cri-o://7a4eaa8754c042c3e140e113fb7d8e11fea794f34887e23cfdf8227f4339e622" gracePeriod=30 Nov 24 14:43:20 crc kubenswrapper[4790]: I1124 14:43:20.891759 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:43:20 crc kubenswrapper[4790]: W1124 14:43:20.899745 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b8cad57_cddc_4273_a1be_498c1172ee79.slice/crio-a881c971ddc1a135586fc2b7988aabfaa7c968d7125d6ed5dff19f23c3166fa5 WatchSource:0}: Error finding container a881c971ddc1a135586fc2b7988aabfaa7c968d7125d6ed5dff19f23c3166fa5: Status 404 returned error can't find the container with id a881c971ddc1a135586fc2b7988aabfaa7c968d7125d6ed5dff19f23c3166fa5 Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.570261 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.710892 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/876e67dc-9b8e-4be5-8540-1016be2b320b-httpd-run\") pod \"876e67dc-9b8e-4be5-8540-1016be2b320b\" (UID: \"876e67dc-9b8e-4be5-8540-1016be2b320b\") " Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.711184 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/876e67dc-9b8e-4be5-8540-1016be2b320b-config-data\") pod \"876e67dc-9b8e-4be5-8540-1016be2b320b\" (UID: \"876e67dc-9b8e-4be5-8540-1016be2b320b\") " Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.711323 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/876e67dc-9b8e-4be5-8540-1016be2b320b-scripts\") pod \"876e67dc-9b8e-4be5-8540-1016be2b320b\" (UID: \"876e67dc-9b8e-4be5-8540-1016be2b320b\") " Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.711362 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/876e67dc-9b8e-4be5-8540-1016be2b320b-combined-ca-bundle\") pod \"876e67dc-9b8e-4be5-8540-1016be2b320b\" (UID: \"876e67dc-9b8e-4be5-8540-1016be2b320b\") " Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.711418 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/876e67dc-9b8e-4be5-8540-1016be2b320b-ceph\") pod \"876e67dc-9b8e-4be5-8540-1016be2b320b\" (UID: \"876e67dc-9b8e-4be5-8540-1016be2b320b\") " Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.711454 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/876e67dc-9b8e-4be5-8540-1016be2b320b-logs\") pod \"876e67dc-9b8e-4be5-8540-1016be2b320b\" (UID: \"876e67dc-9b8e-4be5-8540-1016be2b320b\") " Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.711481 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfkrk\" (UniqueName: \"kubernetes.io/projected/876e67dc-9b8e-4be5-8540-1016be2b320b-kube-api-access-sfkrk\") pod \"876e67dc-9b8e-4be5-8540-1016be2b320b\" (UID: \"876e67dc-9b8e-4be5-8540-1016be2b320b\") " Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.711694 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/876e67dc-9b8e-4be5-8540-1016be2b320b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "876e67dc-9b8e-4be5-8540-1016be2b320b" (UID: "876e67dc-9b8e-4be5-8540-1016be2b320b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.711935 4790 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/876e67dc-9b8e-4be5-8540-1016be2b320b-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.715967 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/876e67dc-9b8e-4be5-8540-1016be2b320b-scripts" (OuterVolumeSpecName: "scripts") pod "876e67dc-9b8e-4be5-8540-1016be2b320b" (UID: "876e67dc-9b8e-4be5-8540-1016be2b320b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.716320 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/876e67dc-9b8e-4be5-8540-1016be2b320b-kube-api-access-sfkrk" (OuterVolumeSpecName: "kube-api-access-sfkrk") pod "876e67dc-9b8e-4be5-8540-1016be2b320b" (UID: "876e67dc-9b8e-4be5-8540-1016be2b320b"). InnerVolumeSpecName "kube-api-access-sfkrk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.716569 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/876e67dc-9b8e-4be5-8540-1016be2b320b-logs" (OuterVolumeSpecName: "logs") pod "876e67dc-9b8e-4be5-8540-1016be2b320b" (UID: "876e67dc-9b8e-4be5-8540-1016be2b320b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.719435 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/876e67dc-9b8e-4be5-8540-1016be2b320b-ceph" (OuterVolumeSpecName: "ceph") pod "876e67dc-9b8e-4be5-8540-1016be2b320b" (UID: "876e67dc-9b8e-4be5-8540-1016be2b320b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.745019 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/876e67dc-9b8e-4be5-8540-1016be2b320b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "876e67dc-9b8e-4be5-8540-1016be2b320b" (UID: "876e67dc-9b8e-4be5-8540-1016be2b320b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.777078 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/876e67dc-9b8e-4be5-8540-1016be2b320b-config-data" (OuterVolumeSpecName: "config-data") pod "876e67dc-9b8e-4be5-8540-1016be2b320b" (UID: "876e67dc-9b8e-4be5-8540-1016be2b320b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.813930 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/876e67dc-9b8e-4be5-8540-1016be2b320b-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.813966 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/876e67dc-9b8e-4be5-8540-1016be2b320b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.813977 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/876e67dc-9b8e-4be5-8540-1016be2b320b-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.813986 4790 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/876e67dc-9b8e-4be5-8540-1016be2b320b-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.813996 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/876e67dc-9b8e-4be5-8540-1016be2b320b-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.814005 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfkrk\" (UniqueName: \"kubernetes.io/projected/876e67dc-9b8e-4be5-8540-1016be2b320b-kube-api-access-sfkrk\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.869928 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4b8cad57-cddc-4273-a1be-498c1172ee79","Type":"ContainerStarted","Data":"6b3329faf3f91547a252762a0890d5dba6b36153a7b56704b3a84abd4e8aeeb3"} Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.869983 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4b8cad57-cddc-4273-a1be-498c1172ee79","Type":"ContainerStarted","Data":"a881c971ddc1a135586fc2b7988aabfaa7c968d7125d6ed5dff19f23c3166fa5"} Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.874769 4790 generic.go:334] "Generic (PLEG): container finished" podID="876e67dc-9b8e-4be5-8540-1016be2b320b" containerID="7a4eaa8754c042c3e140e113fb7d8e11fea794f34887e23cfdf8227f4339e622" exitCode=0 Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.874809 4790 generic.go:334] "Generic (PLEG): container finished" podID="876e67dc-9b8e-4be5-8540-1016be2b320b" containerID="f849df65aaf49ab550a0e1b52f4272e5e72e4c702435be5e9bd86899280eb996" exitCode=143 Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.874834 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"876e67dc-9b8e-4be5-8540-1016be2b320b","Type":"ContainerDied","Data":"7a4eaa8754c042c3e140e113fb7d8e11fea794f34887e23cfdf8227f4339e622"} Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.874863 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.874913 4790 scope.go:117] "RemoveContainer" containerID="7a4eaa8754c042c3e140e113fb7d8e11fea794f34887e23cfdf8227f4339e622" Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.874869 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"876e67dc-9b8e-4be5-8540-1016be2b320b","Type":"ContainerDied","Data":"f849df65aaf49ab550a0e1b52f4272e5e72e4c702435be5e9bd86899280eb996"} Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.875050 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"876e67dc-9b8e-4be5-8540-1016be2b320b","Type":"ContainerDied","Data":"c4b693566d300070b0e59273b796db3a31ba1688e4c61bc7db1264db1b99b3f4"} Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.909644 4790 scope.go:117] "RemoveContainer" containerID="f849df65aaf49ab550a0e1b52f4272e5e72e4c702435be5e9bd86899280eb996" Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.919022 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.925171 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.933375 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:43:21 crc kubenswrapper[4790]: E1124 14:43:21.933707 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="876e67dc-9b8e-4be5-8540-1016be2b320b" containerName="glance-httpd" Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.933727 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="876e67dc-9b8e-4be5-8540-1016be2b320b" containerName="glance-httpd" Nov 24 14:43:21 crc kubenswrapper[4790]: E1124 14:43:21.933742 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="876e67dc-9b8e-4be5-8540-1016be2b320b" containerName="glance-log" Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.933750 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="876e67dc-9b8e-4be5-8540-1016be2b320b" containerName="glance-log" Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.933986 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="876e67dc-9b8e-4be5-8540-1016be2b320b" containerName="glance-httpd" Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.934010 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="876e67dc-9b8e-4be5-8540-1016be2b320b" containerName="glance-log" Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.936322 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.938872 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.942437 4790 scope.go:117] "RemoveContainer" containerID="7a4eaa8754c042c3e140e113fb7d8e11fea794f34887e23cfdf8227f4339e622" Nov 24 14:43:21 crc kubenswrapper[4790]: E1124 14:43:21.942934 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a4eaa8754c042c3e140e113fb7d8e11fea794f34887e23cfdf8227f4339e622\": container with ID starting with 7a4eaa8754c042c3e140e113fb7d8e11fea794f34887e23cfdf8227f4339e622 not found: ID does not exist" containerID="7a4eaa8754c042c3e140e113fb7d8e11fea794f34887e23cfdf8227f4339e622" Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.942964 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a4eaa8754c042c3e140e113fb7d8e11fea794f34887e23cfdf8227f4339e622"} err="failed to get container status \"7a4eaa8754c042c3e140e113fb7d8e11fea794f34887e23cfdf8227f4339e622\": rpc error: code = NotFound desc = could not find container \"7a4eaa8754c042c3e140e113fb7d8e11fea794f34887e23cfdf8227f4339e622\": container with ID starting with 7a4eaa8754c042c3e140e113fb7d8e11fea794f34887e23cfdf8227f4339e622 not found: ID does not exist" Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.942987 4790 scope.go:117] "RemoveContainer" containerID="f849df65aaf49ab550a0e1b52f4272e5e72e4c702435be5e9bd86899280eb996" Nov 24 14:43:21 crc kubenswrapper[4790]: E1124 14:43:21.943408 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f849df65aaf49ab550a0e1b52f4272e5e72e4c702435be5e9bd86899280eb996\": container with ID starting with f849df65aaf49ab550a0e1b52f4272e5e72e4c702435be5e9bd86899280eb996 not found: ID does not exist" containerID="f849df65aaf49ab550a0e1b52f4272e5e72e4c702435be5e9bd86899280eb996" Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.943429 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f849df65aaf49ab550a0e1b52f4272e5e72e4c702435be5e9bd86899280eb996"} err="failed to get container status \"f849df65aaf49ab550a0e1b52f4272e5e72e4c702435be5e9bd86899280eb996\": rpc error: code = NotFound desc = could not find container \"f849df65aaf49ab550a0e1b52f4272e5e72e4c702435be5e9bd86899280eb996\": container with ID starting with f849df65aaf49ab550a0e1b52f4272e5e72e4c702435be5e9bd86899280eb996 not found: ID does not exist" Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.943445 4790 scope.go:117] "RemoveContainer" containerID="7a4eaa8754c042c3e140e113fb7d8e11fea794f34887e23cfdf8227f4339e622" Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.943637 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a4eaa8754c042c3e140e113fb7d8e11fea794f34887e23cfdf8227f4339e622"} err="failed to get container status \"7a4eaa8754c042c3e140e113fb7d8e11fea794f34887e23cfdf8227f4339e622\": rpc error: code = NotFound desc = could not find container \"7a4eaa8754c042c3e140e113fb7d8e11fea794f34887e23cfdf8227f4339e622\": container with ID starting with 7a4eaa8754c042c3e140e113fb7d8e11fea794f34887e23cfdf8227f4339e622 not found: ID does not exist" Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.943650 4790 scope.go:117] "RemoveContainer" containerID="f849df65aaf49ab550a0e1b52f4272e5e72e4c702435be5e9bd86899280eb996" Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.944146 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f849df65aaf49ab550a0e1b52f4272e5e72e4c702435be5e9bd86899280eb996"} err="failed to get container status \"f849df65aaf49ab550a0e1b52f4272e5e72e4c702435be5e9bd86899280eb996\": rpc error: code = NotFound desc = could not find container \"f849df65aaf49ab550a0e1b52f4272e5e72e4c702435be5e9bd86899280eb996\": container with ID starting with f849df65aaf49ab550a0e1b52f4272e5e72e4c702435be5e9bd86899280eb996 not found: ID does not exist" Nov 24 14:43:21 crc kubenswrapper[4790]: I1124 14:43:21.948001 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:43:22 crc kubenswrapper[4790]: I1124 14:43:22.018672 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcrjz\" (UniqueName: \"kubernetes.io/projected/82be1eb5-9ab6-471e-baf5-400b4adc92a2-kube-api-access-bcrjz\") pod \"glance-default-internal-api-0\" (UID: \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:22 crc kubenswrapper[4790]: I1124 14:43:22.018715 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/82be1eb5-9ab6-471e-baf5-400b4adc92a2-ceph\") pod \"glance-default-internal-api-0\" (UID: \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:22 crc kubenswrapper[4790]: I1124 14:43:22.018743 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82be1eb5-9ab6-471e-baf5-400b4adc92a2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:22 crc kubenswrapper[4790]: I1124 14:43:22.018980 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82be1eb5-9ab6-471e-baf5-400b4adc92a2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:22 crc kubenswrapper[4790]: I1124 14:43:22.019136 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82be1eb5-9ab6-471e-baf5-400b4adc92a2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:22 crc kubenswrapper[4790]: I1124 14:43:22.019185 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82be1eb5-9ab6-471e-baf5-400b4adc92a2-logs\") pod \"glance-default-internal-api-0\" (UID: \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:22 crc kubenswrapper[4790]: I1124 14:43:22.019213 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/82be1eb5-9ab6-471e-baf5-400b4adc92a2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:22 crc kubenswrapper[4790]: I1124 14:43:22.122960 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82be1eb5-9ab6-471e-baf5-400b4adc92a2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:22 crc kubenswrapper[4790]: I1124 14:43:22.123098 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82be1eb5-9ab6-471e-baf5-400b4adc92a2-logs\") pod \"glance-default-internal-api-0\" (UID: \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:22 crc kubenswrapper[4790]: I1124 14:43:22.123135 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/82be1eb5-9ab6-471e-baf5-400b4adc92a2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:22 crc kubenswrapper[4790]: I1124 14:43:22.123212 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcrjz\" (UniqueName: \"kubernetes.io/projected/82be1eb5-9ab6-471e-baf5-400b4adc92a2-kube-api-access-bcrjz\") pod \"glance-default-internal-api-0\" (UID: \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:22 crc kubenswrapper[4790]: I1124 14:43:22.123237 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/82be1eb5-9ab6-471e-baf5-400b4adc92a2-ceph\") pod \"glance-default-internal-api-0\" (UID: \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:22 crc kubenswrapper[4790]: I1124 14:43:22.123308 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82be1eb5-9ab6-471e-baf5-400b4adc92a2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:22 crc kubenswrapper[4790]: I1124 14:43:22.123385 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82be1eb5-9ab6-471e-baf5-400b4adc92a2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:22 crc kubenswrapper[4790]: I1124 14:43:22.123738 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82be1eb5-9ab6-471e-baf5-400b4adc92a2-logs\") pod \"glance-default-internal-api-0\" (UID: \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:22 crc kubenswrapper[4790]: I1124 14:43:22.123847 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/82be1eb5-9ab6-471e-baf5-400b4adc92a2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:22 crc kubenswrapper[4790]: I1124 14:43:22.128587 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82be1eb5-9ab6-471e-baf5-400b4adc92a2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:22 crc kubenswrapper[4790]: I1124 14:43:22.129242 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82be1eb5-9ab6-471e-baf5-400b4adc92a2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:22 crc kubenswrapper[4790]: I1124 14:43:22.132135 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/82be1eb5-9ab6-471e-baf5-400b4adc92a2-ceph\") pod \"glance-default-internal-api-0\" (UID: \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:22 crc kubenswrapper[4790]: I1124 14:43:22.132557 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82be1eb5-9ab6-471e-baf5-400b4adc92a2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:22 crc kubenswrapper[4790]: I1124 14:43:22.148146 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcrjz\" (UniqueName: \"kubernetes.io/projected/82be1eb5-9ab6-471e-baf5-400b4adc92a2-kube-api-access-bcrjz\") pod \"glance-default-internal-api-0\" (UID: \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:43:22 crc kubenswrapper[4790]: I1124 14:43:22.269129 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:43:22 crc kubenswrapper[4790]: I1124 14:43:22.327313 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="876e67dc-9b8e-4be5-8540-1016be2b320b" path="/var/lib/kubelet/pods/876e67dc-9b8e-4be5-8540-1016be2b320b/volumes" Nov 24 14:43:22 crc kubenswrapper[4790]: I1124 14:43:22.842958 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:43:22 crc kubenswrapper[4790]: W1124 14:43:22.848418 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82be1eb5_9ab6_471e_baf5_400b4adc92a2.slice/crio-a06e368eb0a87c206fdbc844fb69d40ab93bd1924c3e8fd68e58ba1531731205 WatchSource:0}: Error finding container a06e368eb0a87c206fdbc844fb69d40ab93bd1924c3e8fd68e58ba1531731205: Status 404 returned error can't find the container with id a06e368eb0a87c206fdbc844fb69d40ab93bd1924c3e8fd68e58ba1531731205 Nov 24 14:43:22 crc kubenswrapper[4790]: I1124 14:43:22.886510 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"82be1eb5-9ab6-471e-baf5-400b4adc92a2","Type":"ContainerStarted","Data":"a06e368eb0a87c206fdbc844fb69d40ab93bd1924c3e8fd68e58ba1531731205"} Nov 24 14:43:22 crc kubenswrapper[4790]: I1124 14:43:22.889766 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4b8cad57-cddc-4273-a1be-498c1172ee79","Type":"ContainerStarted","Data":"c65baf0b9f2f5600591225f558e76d78f8d8dd850c6290154a8a5b152f89bf66"} Nov 24 14:43:22 crc kubenswrapper[4790]: I1124 14:43:22.917088 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.9170644489999997 podStartE2EDuration="3.917064449s" podCreationTimestamp="2025-11-24 14:43:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:43:22.911128229 +0000 UTC m=+5451.291021901" watchObservedRunningTime="2025-11-24 14:43:22.917064449 +0000 UTC m=+5451.296958111" Nov 24 14:43:23 crc kubenswrapper[4790]: I1124 14:43:23.910510 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"82be1eb5-9ab6-471e-baf5-400b4adc92a2","Type":"ContainerStarted","Data":"c559ac37f2e2e1a61c0e69a6aa52909d5eebcf31936385ceb56e5d85c035f07e"} Nov 24 14:43:24 crc kubenswrapper[4790]: I1124 14:43:24.924515 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"82be1eb5-9ab6-471e-baf5-400b4adc92a2","Type":"ContainerStarted","Data":"99f6d9838ea1b4bced54c8d65874120944a9a70b875a07da856f63661fe54f0f"} Nov 24 14:43:24 crc kubenswrapper[4790]: I1124 14:43:24.959985 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.959940864 podStartE2EDuration="3.959940864s" podCreationTimestamp="2025-11-24 14:43:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:43:24.957201285 +0000 UTC m=+5453.337094987" watchObservedRunningTime="2025-11-24 14:43:24.959940864 +0000 UTC m=+5453.339834566" Nov 24 14:43:26 crc kubenswrapper[4790]: I1124 14:43:26.556206 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5dd6c96d9c-4mllx" Nov 24 14:43:26 crc kubenswrapper[4790]: I1124 14:43:26.639623 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5688cb67b5-sfbn9"] Nov 24 14:43:26 crc kubenswrapper[4790]: I1124 14:43:26.639952 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5688cb67b5-sfbn9" podUID="f65fadf5-f57c-44a8-9f38-0b8298add279" containerName="dnsmasq-dns" containerID="cri-o://117f92bb49cd70bb69470788b5d891afeb4610c25b55e9215dc309366b270d59" gracePeriod=10 Nov 24 14:43:26 crc kubenswrapper[4790]: I1124 14:43:26.952683 4790 generic.go:334] "Generic (PLEG): container finished" podID="f65fadf5-f57c-44a8-9f38-0b8298add279" containerID="117f92bb49cd70bb69470788b5d891afeb4610c25b55e9215dc309366b270d59" exitCode=0 Nov 24 14:43:26 crc kubenswrapper[4790]: I1124 14:43:26.953078 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5688cb67b5-sfbn9" event={"ID":"f65fadf5-f57c-44a8-9f38-0b8298add279","Type":"ContainerDied","Data":"117f92bb49cd70bb69470788b5d891afeb4610c25b55e9215dc309366b270d59"} Nov 24 14:43:27 crc kubenswrapper[4790]: I1124 14:43:27.085369 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5688cb67b5-sfbn9" Nov 24 14:43:27 crc kubenswrapper[4790]: I1124 14:43:27.227002 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f65fadf5-f57c-44a8-9f38-0b8298add279-dns-svc\") pod \"f65fadf5-f57c-44a8-9f38-0b8298add279\" (UID: \"f65fadf5-f57c-44a8-9f38-0b8298add279\") " Nov 24 14:43:27 crc kubenswrapper[4790]: I1124 14:43:27.227083 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f65fadf5-f57c-44a8-9f38-0b8298add279-config\") pod \"f65fadf5-f57c-44a8-9f38-0b8298add279\" (UID: \"f65fadf5-f57c-44a8-9f38-0b8298add279\") " Nov 24 14:43:27 crc kubenswrapper[4790]: I1124 14:43:27.227168 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f65fadf5-f57c-44a8-9f38-0b8298add279-ovsdbserver-nb\") pod \"f65fadf5-f57c-44a8-9f38-0b8298add279\" (UID: \"f65fadf5-f57c-44a8-9f38-0b8298add279\") " Nov 24 14:43:27 crc kubenswrapper[4790]: I1124 14:43:27.227307 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hm5x\" (UniqueName: \"kubernetes.io/projected/f65fadf5-f57c-44a8-9f38-0b8298add279-kube-api-access-4hm5x\") pod \"f65fadf5-f57c-44a8-9f38-0b8298add279\" (UID: \"f65fadf5-f57c-44a8-9f38-0b8298add279\") " Nov 24 14:43:27 crc kubenswrapper[4790]: I1124 14:43:27.227388 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f65fadf5-f57c-44a8-9f38-0b8298add279-ovsdbserver-sb\") pod \"f65fadf5-f57c-44a8-9f38-0b8298add279\" (UID: \"f65fadf5-f57c-44a8-9f38-0b8298add279\") " Nov 24 14:43:27 crc kubenswrapper[4790]: I1124 14:43:27.234861 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f65fadf5-f57c-44a8-9f38-0b8298add279-kube-api-access-4hm5x" (OuterVolumeSpecName: "kube-api-access-4hm5x") pod "f65fadf5-f57c-44a8-9f38-0b8298add279" (UID: "f65fadf5-f57c-44a8-9f38-0b8298add279"). InnerVolumeSpecName "kube-api-access-4hm5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:43:27 crc kubenswrapper[4790]: I1124 14:43:27.277494 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f65fadf5-f57c-44a8-9f38-0b8298add279-config" (OuterVolumeSpecName: "config") pod "f65fadf5-f57c-44a8-9f38-0b8298add279" (UID: "f65fadf5-f57c-44a8-9f38-0b8298add279"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:43:27 crc kubenswrapper[4790]: I1124 14:43:27.282573 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f65fadf5-f57c-44a8-9f38-0b8298add279-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f65fadf5-f57c-44a8-9f38-0b8298add279" (UID: "f65fadf5-f57c-44a8-9f38-0b8298add279"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:43:27 crc kubenswrapper[4790]: I1124 14:43:27.288270 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f65fadf5-f57c-44a8-9f38-0b8298add279-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f65fadf5-f57c-44a8-9f38-0b8298add279" (UID: "f65fadf5-f57c-44a8-9f38-0b8298add279"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:43:27 crc kubenswrapper[4790]: I1124 14:43:27.290304 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f65fadf5-f57c-44a8-9f38-0b8298add279-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f65fadf5-f57c-44a8-9f38-0b8298add279" (UID: "f65fadf5-f57c-44a8-9f38-0b8298add279"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:43:27 crc kubenswrapper[4790]: I1124 14:43:27.314522 4790 scope.go:117] "RemoveContainer" containerID="92c086e44ccb28a62777d612e499209f7483422ca70b460ce6ec73c416e8e8b8" Nov 24 14:43:27 crc kubenswrapper[4790]: E1124 14:43:27.315218 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:43:27 crc kubenswrapper[4790]: I1124 14:43:27.329999 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hm5x\" (UniqueName: \"kubernetes.io/projected/f65fadf5-f57c-44a8-9f38-0b8298add279-kube-api-access-4hm5x\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:27 crc kubenswrapper[4790]: I1124 14:43:27.330121 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f65fadf5-f57c-44a8-9f38-0b8298add279-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:27 crc kubenswrapper[4790]: I1124 14:43:27.330139 4790 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f65fadf5-f57c-44a8-9f38-0b8298add279-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:27 crc kubenswrapper[4790]: I1124 14:43:27.330153 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f65fadf5-f57c-44a8-9f38-0b8298add279-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:27 crc kubenswrapper[4790]: I1124 14:43:27.330167 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f65fadf5-f57c-44a8-9f38-0b8298add279-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:27 crc kubenswrapper[4790]: I1124 14:43:27.969484 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5688cb67b5-sfbn9" event={"ID":"f65fadf5-f57c-44a8-9f38-0b8298add279","Type":"ContainerDied","Data":"331648eb92825368130e3d0017bb220fe8a6c407596dcb7e40e52e3ae07419cd"} Nov 24 14:43:27 crc kubenswrapper[4790]: I1124 14:43:27.969950 4790 scope.go:117] "RemoveContainer" containerID="117f92bb49cd70bb69470788b5d891afeb4610c25b55e9215dc309366b270d59" Nov 24 14:43:27 crc kubenswrapper[4790]: I1124 14:43:27.969621 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5688cb67b5-sfbn9" Nov 24 14:43:28 crc kubenswrapper[4790]: I1124 14:43:28.010237 4790 scope.go:117] "RemoveContainer" containerID="c1a9a9423008222f6a4bdbe517e15a3db74df3b99a3146347a86bc07adb28778" Nov 24 14:43:28 crc kubenswrapper[4790]: I1124 14:43:28.041063 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5688cb67b5-sfbn9"] Nov 24 14:43:28 crc kubenswrapper[4790]: I1124 14:43:28.052262 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5688cb67b5-sfbn9"] Nov 24 14:43:28 crc kubenswrapper[4790]: I1124 14:43:28.330313 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f65fadf5-f57c-44a8-9f38-0b8298add279" path="/var/lib/kubelet/pods/f65fadf5-f57c-44a8-9f38-0b8298add279/volumes" Nov 24 14:43:30 crc kubenswrapper[4790]: I1124 14:43:30.248927 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 14:43:30 crc kubenswrapper[4790]: I1124 14:43:30.249308 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 14:43:30 crc kubenswrapper[4790]: I1124 14:43:30.300378 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 14:43:30 crc kubenswrapper[4790]: I1124 14:43:30.338254 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 14:43:31 crc kubenswrapper[4790]: I1124 14:43:31.012002 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 14:43:31 crc kubenswrapper[4790]: I1124 14:43:31.012085 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 14:43:32 crc kubenswrapper[4790]: I1124 14:43:32.269904 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 14:43:32 crc kubenswrapper[4790]: I1124 14:43:32.271081 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 14:43:32 crc kubenswrapper[4790]: I1124 14:43:32.329710 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 14:43:32 crc kubenswrapper[4790]: I1124 14:43:32.330469 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 14:43:32 crc kubenswrapper[4790]: I1124 14:43:32.997407 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 14:43:32 crc kubenswrapper[4790]: I1124 14:43:32.998249 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 14:43:33 crc kubenswrapper[4790]: I1124 14:43:33.040075 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 14:43:33 crc kubenswrapper[4790]: I1124 14:43:33.040478 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 14:43:34 crc kubenswrapper[4790]: I1124 14:43:34.843582 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 14:43:34 crc kubenswrapper[4790]: I1124 14:43:34.921482 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 14:43:38 crc kubenswrapper[4790]: I1124 14:43:38.314867 4790 scope.go:117] "RemoveContainer" containerID="92c086e44ccb28a62777d612e499209f7483422ca70b460ce6ec73c416e8e8b8" Nov 24 14:43:38 crc kubenswrapper[4790]: E1124 14:43:38.316325 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:43:40 crc kubenswrapper[4790]: I1124 14:43:40.956339 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-2qmwz"] Nov 24 14:43:40 crc kubenswrapper[4790]: E1124 14:43:40.957106 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f65fadf5-f57c-44a8-9f38-0b8298add279" containerName="dnsmasq-dns" Nov 24 14:43:40 crc kubenswrapper[4790]: I1124 14:43:40.957126 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="f65fadf5-f57c-44a8-9f38-0b8298add279" containerName="dnsmasq-dns" Nov 24 14:43:40 crc kubenswrapper[4790]: E1124 14:43:40.957145 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f65fadf5-f57c-44a8-9f38-0b8298add279" containerName="init" Nov 24 14:43:40 crc kubenswrapper[4790]: I1124 14:43:40.957152 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="f65fadf5-f57c-44a8-9f38-0b8298add279" containerName="init" Nov 24 14:43:40 crc kubenswrapper[4790]: I1124 14:43:40.957394 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="f65fadf5-f57c-44a8-9f38-0b8298add279" containerName="dnsmasq-dns" Nov 24 14:43:40 crc kubenswrapper[4790]: I1124 14:43:40.958223 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-2qmwz" Nov 24 14:43:40 crc kubenswrapper[4790]: I1124 14:43:40.966188 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6251-account-create-9fg9t"] Nov 24 14:43:40 crc kubenswrapper[4790]: I1124 14:43:40.967800 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6251-account-create-9fg9t" Nov 24 14:43:40 crc kubenswrapper[4790]: I1124 14:43:40.969552 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 24 14:43:40 crc kubenswrapper[4790]: I1124 14:43:40.976420 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-2qmwz"] Nov 24 14:43:40 crc kubenswrapper[4790]: I1124 14:43:40.985575 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6251-account-create-9fg9t"] Nov 24 14:43:41 crc kubenswrapper[4790]: I1124 14:43:41.025859 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8rbb\" (UniqueName: \"kubernetes.io/projected/c460e843-5c12-44dd-8245-e58ef2ea3546-kube-api-access-w8rbb\") pod \"placement-6251-account-create-9fg9t\" (UID: \"c460e843-5c12-44dd-8245-e58ef2ea3546\") " pod="openstack/placement-6251-account-create-9fg9t" Nov 24 14:43:41 crc kubenswrapper[4790]: I1124 14:43:41.026091 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80e59caa-53d1-4697-b987-e93cba1bf143-operator-scripts\") pod \"placement-db-create-2qmwz\" (UID: \"80e59caa-53d1-4697-b987-e93cba1bf143\") " pod="openstack/placement-db-create-2qmwz" Nov 24 14:43:41 crc kubenswrapper[4790]: I1124 14:43:41.026175 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c460e843-5c12-44dd-8245-e58ef2ea3546-operator-scripts\") pod \"placement-6251-account-create-9fg9t\" (UID: \"c460e843-5c12-44dd-8245-e58ef2ea3546\") " pod="openstack/placement-6251-account-create-9fg9t" Nov 24 14:43:41 crc kubenswrapper[4790]: I1124 14:43:41.026328 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6b6k\" (UniqueName: \"kubernetes.io/projected/80e59caa-53d1-4697-b987-e93cba1bf143-kube-api-access-k6b6k\") pod \"placement-db-create-2qmwz\" (UID: \"80e59caa-53d1-4697-b987-e93cba1bf143\") " pod="openstack/placement-db-create-2qmwz" Nov 24 14:43:41 crc kubenswrapper[4790]: I1124 14:43:41.128525 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6b6k\" (UniqueName: \"kubernetes.io/projected/80e59caa-53d1-4697-b987-e93cba1bf143-kube-api-access-k6b6k\") pod \"placement-db-create-2qmwz\" (UID: \"80e59caa-53d1-4697-b987-e93cba1bf143\") " pod="openstack/placement-db-create-2qmwz" Nov 24 14:43:41 crc kubenswrapper[4790]: I1124 14:43:41.128654 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8rbb\" (UniqueName: \"kubernetes.io/projected/c460e843-5c12-44dd-8245-e58ef2ea3546-kube-api-access-w8rbb\") pod \"placement-6251-account-create-9fg9t\" (UID: \"c460e843-5c12-44dd-8245-e58ef2ea3546\") " pod="openstack/placement-6251-account-create-9fg9t" Nov 24 14:43:41 crc kubenswrapper[4790]: I1124 14:43:41.128725 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80e59caa-53d1-4697-b987-e93cba1bf143-operator-scripts\") pod \"placement-db-create-2qmwz\" (UID: \"80e59caa-53d1-4697-b987-e93cba1bf143\") " pod="openstack/placement-db-create-2qmwz" Nov 24 14:43:41 crc kubenswrapper[4790]: I1124 14:43:41.128762 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c460e843-5c12-44dd-8245-e58ef2ea3546-operator-scripts\") pod \"placement-6251-account-create-9fg9t\" (UID: \"c460e843-5c12-44dd-8245-e58ef2ea3546\") " pod="openstack/placement-6251-account-create-9fg9t" Nov 24 14:43:41 crc kubenswrapper[4790]: I1124 14:43:41.129639 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c460e843-5c12-44dd-8245-e58ef2ea3546-operator-scripts\") pod \"placement-6251-account-create-9fg9t\" (UID: \"c460e843-5c12-44dd-8245-e58ef2ea3546\") " pod="openstack/placement-6251-account-create-9fg9t" Nov 24 14:43:41 crc kubenswrapper[4790]: I1124 14:43:41.130385 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80e59caa-53d1-4697-b987-e93cba1bf143-operator-scripts\") pod \"placement-db-create-2qmwz\" (UID: \"80e59caa-53d1-4697-b987-e93cba1bf143\") " pod="openstack/placement-db-create-2qmwz" Nov 24 14:43:41 crc kubenswrapper[4790]: I1124 14:43:41.155426 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8rbb\" (UniqueName: \"kubernetes.io/projected/c460e843-5c12-44dd-8245-e58ef2ea3546-kube-api-access-w8rbb\") pod \"placement-6251-account-create-9fg9t\" (UID: \"c460e843-5c12-44dd-8245-e58ef2ea3546\") " pod="openstack/placement-6251-account-create-9fg9t" Nov 24 14:43:41 crc kubenswrapper[4790]: I1124 14:43:41.159718 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6b6k\" (UniqueName: \"kubernetes.io/projected/80e59caa-53d1-4697-b987-e93cba1bf143-kube-api-access-k6b6k\") pod \"placement-db-create-2qmwz\" (UID: \"80e59caa-53d1-4697-b987-e93cba1bf143\") " pod="openstack/placement-db-create-2qmwz" Nov 24 14:43:41 crc kubenswrapper[4790]: I1124 14:43:41.284687 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-2qmwz" Nov 24 14:43:41 crc kubenswrapper[4790]: I1124 14:43:41.295334 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6251-account-create-9fg9t" Nov 24 14:43:41 crc kubenswrapper[4790]: I1124 14:43:41.643015 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-2qmwz"] Nov 24 14:43:41 crc kubenswrapper[4790]: I1124 14:43:41.773965 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6251-account-create-9fg9t"] Nov 24 14:43:41 crc kubenswrapper[4790]: W1124 14:43:41.778665 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc460e843_5c12_44dd_8245_e58ef2ea3546.slice/crio-215b9aa6a63f7da899f5afea72bb81b95e6544c4732dfb468ddf887bf6688345 WatchSource:0}: Error finding container 215b9aa6a63f7da899f5afea72bb81b95e6544c4732dfb468ddf887bf6688345: Status 404 returned error can't find the container with id 215b9aa6a63f7da899f5afea72bb81b95e6544c4732dfb468ddf887bf6688345 Nov 24 14:43:42 crc kubenswrapper[4790]: I1124 14:43:42.132532 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-2qmwz" event={"ID":"80e59caa-53d1-4697-b987-e93cba1bf143","Type":"ContainerStarted","Data":"b816cb66ae39bd0fb01fd1d5efb8914832138243781b2438acf9996a2c387fce"} Nov 24 14:43:42 crc kubenswrapper[4790]: I1124 14:43:42.132600 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-2qmwz" event={"ID":"80e59caa-53d1-4697-b987-e93cba1bf143","Type":"ContainerStarted","Data":"e66b844398737b17c6fcb25aac4eda2a554cabe1f2cd981a596c1405de066cec"} Nov 24 14:43:42 crc kubenswrapper[4790]: I1124 14:43:42.133629 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6251-account-create-9fg9t" event={"ID":"c460e843-5c12-44dd-8245-e58ef2ea3546","Type":"ContainerStarted","Data":"1113d3dcf1845850aa7b42b43e945aac7f247747af2f5dc144c9265ca98d2bd4"} Nov 24 14:43:42 crc kubenswrapper[4790]: I1124 14:43:42.133675 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6251-account-create-9fg9t" event={"ID":"c460e843-5c12-44dd-8245-e58ef2ea3546","Type":"ContainerStarted","Data":"215b9aa6a63f7da899f5afea72bb81b95e6544c4732dfb468ddf887bf6688345"} Nov 24 14:43:43 crc kubenswrapper[4790]: I1124 14:43:43.191589 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-2qmwz" podStartSLOduration=3.191566779 podStartE2EDuration="3.191566779s" podCreationTimestamp="2025-11-24 14:43:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:43:43.186271607 +0000 UTC m=+5471.566165279" watchObservedRunningTime="2025-11-24 14:43:43.191566779 +0000 UTC m=+5471.571460441" Nov 24 14:43:43 crc kubenswrapper[4790]: I1124 14:43:43.195546 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6251-account-create-9fg9t" podStartSLOduration=3.195537343 podStartE2EDuration="3.195537343s" podCreationTimestamp="2025-11-24 14:43:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:43:43.175103457 +0000 UTC m=+5471.554997149" watchObservedRunningTime="2025-11-24 14:43:43.195537343 +0000 UTC m=+5471.575431005" Nov 24 14:43:44 crc kubenswrapper[4790]: I1124 14:43:44.187549 4790 generic.go:334] "Generic (PLEG): container finished" podID="c460e843-5c12-44dd-8245-e58ef2ea3546" containerID="1113d3dcf1845850aa7b42b43e945aac7f247747af2f5dc144c9265ca98d2bd4" exitCode=0 Nov 24 14:43:44 crc kubenswrapper[4790]: I1124 14:43:44.187644 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6251-account-create-9fg9t" event={"ID":"c460e843-5c12-44dd-8245-e58ef2ea3546","Type":"ContainerDied","Data":"1113d3dcf1845850aa7b42b43e945aac7f247747af2f5dc144c9265ca98d2bd4"} Nov 24 14:43:44 crc kubenswrapper[4790]: I1124 14:43:44.192400 4790 generic.go:334] "Generic (PLEG): container finished" podID="80e59caa-53d1-4697-b987-e93cba1bf143" containerID="b816cb66ae39bd0fb01fd1d5efb8914832138243781b2438acf9996a2c387fce" exitCode=0 Nov 24 14:43:44 crc kubenswrapper[4790]: I1124 14:43:44.192493 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-2qmwz" event={"ID":"80e59caa-53d1-4697-b987-e93cba1bf143","Type":"ContainerDied","Data":"b816cb66ae39bd0fb01fd1d5efb8914832138243781b2438acf9996a2c387fce"} Nov 24 14:43:45 crc kubenswrapper[4790]: I1124 14:43:45.529930 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-2qmwz" Nov 24 14:43:45 crc kubenswrapper[4790]: I1124 14:43:45.611944 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6251-account-create-9fg9t" Nov 24 14:43:45 crc kubenswrapper[4790]: I1124 14:43:45.715983 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6b6k\" (UniqueName: \"kubernetes.io/projected/80e59caa-53d1-4697-b987-e93cba1bf143-kube-api-access-k6b6k\") pod \"80e59caa-53d1-4697-b987-e93cba1bf143\" (UID: \"80e59caa-53d1-4697-b987-e93cba1bf143\") " Nov 24 14:43:45 crc kubenswrapper[4790]: I1124 14:43:45.716066 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8rbb\" (UniqueName: \"kubernetes.io/projected/c460e843-5c12-44dd-8245-e58ef2ea3546-kube-api-access-w8rbb\") pod \"c460e843-5c12-44dd-8245-e58ef2ea3546\" (UID: \"c460e843-5c12-44dd-8245-e58ef2ea3546\") " Nov 24 14:43:45 crc kubenswrapper[4790]: I1124 14:43:45.716203 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c460e843-5c12-44dd-8245-e58ef2ea3546-operator-scripts\") pod \"c460e843-5c12-44dd-8245-e58ef2ea3546\" (UID: \"c460e843-5c12-44dd-8245-e58ef2ea3546\") " Nov 24 14:43:45 crc kubenswrapper[4790]: I1124 14:43:45.716402 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80e59caa-53d1-4697-b987-e93cba1bf143-operator-scripts\") pod \"80e59caa-53d1-4697-b987-e93cba1bf143\" (UID: \"80e59caa-53d1-4697-b987-e93cba1bf143\") " Nov 24 14:43:45 crc kubenswrapper[4790]: I1124 14:43:45.717245 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80e59caa-53d1-4697-b987-e93cba1bf143-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "80e59caa-53d1-4697-b987-e93cba1bf143" (UID: "80e59caa-53d1-4697-b987-e93cba1bf143"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:43:45 crc kubenswrapper[4790]: I1124 14:43:45.717444 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c460e843-5c12-44dd-8245-e58ef2ea3546-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c460e843-5c12-44dd-8245-e58ef2ea3546" (UID: "c460e843-5c12-44dd-8245-e58ef2ea3546"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:43:45 crc kubenswrapper[4790]: I1124 14:43:45.722127 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c460e843-5c12-44dd-8245-e58ef2ea3546-kube-api-access-w8rbb" (OuterVolumeSpecName: "kube-api-access-w8rbb") pod "c460e843-5c12-44dd-8245-e58ef2ea3546" (UID: "c460e843-5c12-44dd-8245-e58ef2ea3546"). InnerVolumeSpecName "kube-api-access-w8rbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:43:45 crc kubenswrapper[4790]: I1124 14:43:45.722203 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80e59caa-53d1-4697-b987-e93cba1bf143-kube-api-access-k6b6k" (OuterVolumeSpecName: "kube-api-access-k6b6k") pod "80e59caa-53d1-4697-b987-e93cba1bf143" (UID: "80e59caa-53d1-4697-b987-e93cba1bf143"). InnerVolumeSpecName "kube-api-access-k6b6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:43:45 crc kubenswrapper[4790]: I1124 14:43:45.818673 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8rbb\" (UniqueName: \"kubernetes.io/projected/c460e843-5c12-44dd-8245-e58ef2ea3546-kube-api-access-w8rbb\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:45 crc kubenswrapper[4790]: I1124 14:43:45.818751 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c460e843-5c12-44dd-8245-e58ef2ea3546-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:45 crc kubenswrapper[4790]: I1124 14:43:45.818768 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80e59caa-53d1-4697-b987-e93cba1bf143-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:45 crc kubenswrapper[4790]: I1124 14:43:45.818781 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6b6k\" (UniqueName: \"kubernetes.io/projected/80e59caa-53d1-4697-b987-e93cba1bf143-kube-api-access-k6b6k\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:46 crc kubenswrapper[4790]: I1124 14:43:46.216211 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6251-account-create-9fg9t" event={"ID":"c460e843-5c12-44dd-8245-e58ef2ea3546","Type":"ContainerDied","Data":"215b9aa6a63f7da899f5afea72bb81b95e6544c4732dfb468ddf887bf6688345"} Nov 24 14:43:46 crc kubenswrapper[4790]: I1124 14:43:46.216566 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="215b9aa6a63f7da899f5afea72bb81b95e6544c4732dfb468ddf887bf6688345" Nov 24 14:43:46 crc kubenswrapper[4790]: I1124 14:43:46.216275 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6251-account-create-9fg9t" Nov 24 14:43:46 crc kubenswrapper[4790]: I1124 14:43:46.218518 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-2qmwz" event={"ID":"80e59caa-53d1-4697-b987-e93cba1bf143","Type":"ContainerDied","Data":"e66b844398737b17c6fcb25aac4eda2a554cabe1f2cd981a596c1405de066cec"} Nov 24 14:43:46 crc kubenswrapper[4790]: I1124 14:43:46.218587 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e66b844398737b17c6fcb25aac4eda2a554cabe1f2cd981a596c1405de066cec" Nov 24 14:43:46 crc kubenswrapper[4790]: I1124 14:43:46.218669 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-2qmwz" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.427137 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-x66l4"] Nov 24 14:43:51 crc kubenswrapper[4790]: E1124 14:43:51.428928 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80e59caa-53d1-4697-b987-e93cba1bf143" containerName="mariadb-database-create" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.428952 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="80e59caa-53d1-4697-b987-e93cba1bf143" containerName="mariadb-database-create" Nov 24 14:43:51 crc kubenswrapper[4790]: E1124 14:43:51.428970 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c460e843-5c12-44dd-8245-e58ef2ea3546" containerName="mariadb-account-create" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.428980 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="c460e843-5c12-44dd-8245-e58ef2ea3546" containerName="mariadb-account-create" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.429265 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="c460e843-5c12-44dd-8245-e58ef2ea3546" containerName="mariadb-account-create" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.429289 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="80e59caa-53d1-4697-b987-e93cba1bf143" containerName="mariadb-database-create" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.430231 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-x66l4" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.433101 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.440674 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-7tfsd" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.440941 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.444507 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-76d49bc685-2dflh"] Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.463623 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76d49bc685-2dflh" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.479756 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76d49bc685-2dflh"] Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.492611 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-x66l4"] Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.534528 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbvtw\" (UniqueName: \"kubernetes.io/projected/e31ed6a6-8482-4dc5-9da6-46dd87a73053-kube-api-access-qbvtw\") pod \"placement-db-sync-x66l4\" (UID: \"e31ed6a6-8482-4dc5-9da6-46dd87a73053\") " pod="openstack/placement-db-sync-x66l4" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.534628 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e31ed6a6-8482-4dc5-9da6-46dd87a73053-config-data\") pod \"placement-db-sync-x66l4\" (UID: \"e31ed6a6-8482-4dc5-9da6-46dd87a73053\") " pod="openstack/placement-db-sync-x66l4" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.534655 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e31ed6a6-8482-4dc5-9da6-46dd87a73053-logs\") pod \"placement-db-sync-x66l4\" (UID: \"e31ed6a6-8482-4dc5-9da6-46dd87a73053\") " pod="openstack/placement-db-sync-x66l4" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.534733 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e31ed6a6-8482-4dc5-9da6-46dd87a73053-combined-ca-bundle\") pod \"placement-db-sync-x66l4\" (UID: \"e31ed6a6-8482-4dc5-9da6-46dd87a73053\") " pod="openstack/placement-db-sync-x66l4" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.534769 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e31ed6a6-8482-4dc5-9da6-46dd87a73053-scripts\") pod \"placement-db-sync-x66l4\" (UID: \"e31ed6a6-8482-4dc5-9da6-46dd87a73053\") " pod="openstack/placement-db-sync-x66l4" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.636133 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbvtw\" (UniqueName: \"kubernetes.io/projected/e31ed6a6-8482-4dc5-9da6-46dd87a73053-kube-api-access-qbvtw\") pod \"placement-db-sync-x66l4\" (UID: \"e31ed6a6-8482-4dc5-9da6-46dd87a73053\") " pod="openstack/placement-db-sync-x66l4" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.636219 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e31ed6a6-8482-4dc5-9da6-46dd87a73053-config-data\") pod \"placement-db-sync-x66l4\" (UID: \"e31ed6a6-8482-4dc5-9da6-46dd87a73053\") " pod="openstack/placement-db-sync-x66l4" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.636251 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa76fa12-375b-4c64-926f-fc0815a7d4a6-ovsdbserver-nb\") pod \"dnsmasq-dns-76d49bc685-2dflh\" (UID: \"aa76fa12-375b-4c64-926f-fc0815a7d4a6\") " pod="openstack/dnsmasq-dns-76d49bc685-2dflh" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.636269 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa76fa12-375b-4c64-926f-fc0815a7d4a6-ovsdbserver-sb\") pod \"dnsmasq-dns-76d49bc685-2dflh\" (UID: \"aa76fa12-375b-4c64-926f-fc0815a7d4a6\") " pod="openstack/dnsmasq-dns-76d49bc685-2dflh" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.636291 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e31ed6a6-8482-4dc5-9da6-46dd87a73053-logs\") pod \"placement-db-sync-x66l4\" (UID: \"e31ed6a6-8482-4dc5-9da6-46dd87a73053\") " pod="openstack/placement-db-sync-x66l4" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.636442 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jprl7\" (UniqueName: \"kubernetes.io/projected/aa76fa12-375b-4c64-926f-fc0815a7d4a6-kube-api-access-jprl7\") pod \"dnsmasq-dns-76d49bc685-2dflh\" (UID: \"aa76fa12-375b-4c64-926f-fc0815a7d4a6\") " pod="openstack/dnsmasq-dns-76d49bc685-2dflh" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.636562 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e31ed6a6-8482-4dc5-9da6-46dd87a73053-combined-ca-bundle\") pod \"placement-db-sync-x66l4\" (UID: \"e31ed6a6-8482-4dc5-9da6-46dd87a73053\") " pod="openstack/placement-db-sync-x66l4" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.636609 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa76fa12-375b-4c64-926f-fc0815a7d4a6-config\") pod \"dnsmasq-dns-76d49bc685-2dflh\" (UID: \"aa76fa12-375b-4c64-926f-fc0815a7d4a6\") " pod="openstack/dnsmasq-dns-76d49bc685-2dflh" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.636693 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e31ed6a6-8482-4dc5-9da6-46dd87a73053-scripts\") pod \"placement-db-sync-x66l4\" (UID: \"e31ed6a6-8482-4dc5-9da6-46dd87a73053\") " pod="openstack/placement-db-sync-x66l4" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.636796 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e31ed6a6-8482-4dc5-9da6-46dd87a73053-logs\") pod \"placement-db-sync-x66l4\" (UID: \"e31ed6a6-8482-4dc5-9da6-46dd87a73053\") " pod="openstack/placement-db-sync-x66l4" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.636909 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa76fa12-375b-4c64-926f-fc0815a7d4a6-dns-svc\") pod \"dnsmasq-dns-76d49bc685-2dflh\" (UID: \"aa76fa12-375b-4c64-926f-fc0815a7d4a6\") " pod="openstack/dnsmasq-dns-76d49bc685-2dflh" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.642287 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e31ed6a6-8482-4dc5-9da6-46dd87a73053-config-data\") pod \"placement-db-sync-x66l4\" (UID: \"e31ed6a6-8482-4dc5-9da6-46dd87a73053\") " pod="openstack/placement-db-sync-x66l4" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.642430 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e31ed6a6-8482-4dc5-9da6-46dd87a73053-scripts\") pod \"placement-db-sync-x66l4\" (UID: \"e31ed6a6-8482-4dc5-9da6-46dd87a73053\") " pod="openstack/placement-db-sync-x66l4" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.645605 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e31ed6a6-8482-4dc5-9da6-46dd87a73053-combined-ca-bundle\") pod \"placement-db-sync-x66l4\" (UID: \"e31ed6a6-8482-4dc5-9da6-46dd87a73053\") " pod="openstack/placement-db-sync-x66l4" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.671340 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbvtw\" (UniqueName: \"kubernetes.io/projected/e31ed6a6-8482-4dc5-9da6-46dd87a73053-kube-api-access-qbvtw\") pod \"placement-db-sync-x66l4\" (UID: \"e31ed6a6-8482-4dc5-9da6-46dd87a73053\") " pod="openstack/placement-db-sync-x66l4" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.738798 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa76fa12-375b-4c64-926f-fc0815a7d4a6-ovsdbserver-nb\") pod \"dnsmasq-dns-76d49bc685-2dflh\" (UID: \"aa76fa12-375b-4c64-926f-fc0815a7d4a6\") " pod="openstack/dnsmasq-dns-76d49bc685-2dflh" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.738853 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa76fa12-375b-4c64-926f-fc0815a7d4a6-ovsdbserver-sb\") pod \"dnsmasq-dns-76d49bc685-2dflh\" (UID: \"aa76fa12-375b-4c64-926f-fc0815a7d4a6\") " pod="openstack/dnsmasq-dns-76d49bc685-2dflh" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.738944 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jprl7\" (UniqueName: \"kubernetes.io/projected/aa76fa12-375b-4c64-926f-fc0815a7d4a6-kube-api-access-jprl7\") pod \"dnsmasq-dns-76d49bc685-2dflh\" (UID: \"aa76fa12-375b-4c64-926f-fc0815a7d4a6\") " pod="openstack/dnsmasq-dns-76d49bc685-2dflh" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.738977 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa76fa12-375b-4c64-926f-fc0815a7d4a6-config\") pod \"dnsmasq-dns-76d49bc685-2dflh\" (UID: \"aa76fa12-375b-4c64-926f-fc0815a7d4a6\") " pod="openstack/dnsmasq-dns-76d49bc685-2dflh" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.739015 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa76fa12-375b-4c64-926f-fc0815a7d4a6-dns-svc\") pod \"dnsmasq-dns-76d49bc685-2dflh\" (UID: \"aa76fa12-375b-4c64-926f-fc0815a7d4a6\") " pod="openstack/dnsmasq-dns-76d49bc685-2dflh" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.739909 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa76fa12-375b-4c64-926f-fc0815a7d4a6-dns-svc\") pod \"dnsmasq-dns-76d49bc685-2dflh\" (UID: \"aa76fa12-375b-4c64-926f-fc0815a7d4a6\") " pod="openstack/dnsmasq-dns-76d49bc685-2dflh" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.740446 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa76fa12-375b-4c64-926f-fc0815a7d4a6-ovsdbserver-nb\") pod \"dnsmasq-dns-76d49bc685-2dflh\" (UID: \"aa76fa12-375b-4c64-926f-fc0815a7d4a6\") " pod="openstack/dnsmasq-dns-76d49bc685-2dflh" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.741046 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa76fa12-375b-4c64-926f-fc0815a7d4a6-ovsdbserver-sb\") pod \"dnsmasq-dns-76d49bc685-2dflh\" (UID: \"aa76fa12-375b-4c64-926f-fc0815a7d4a6\") " pod="openstack/dnsmasq-dns-76d49bc685-2dflh" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.741818 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa76fa12-375b-4c64-926f-fc0815a7d4a6-config\") pod \"dnsmasq-dns-76d49bc685-2dflh\" (UID: \"aa76fa12-375b-4c64-926f-fc0815a7d4a6\") " pod="openstack/dnsmasq-dns-76d49bc685-2dflh" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.769849 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jprl7\" (UniqueName: \"kubernetes.io/projected/aa76fa12-375b-4c64-926f-fc0815a7d4a6-kube-api-access-jprl7\") pod \"dnsmasq-dns-76d49bc685-2dflh\" (UID: \"aa76fa12-375b-4c64-926f-fc0815a7d4a6\") " pod="openstack/dnsmasq-dns-76d49bc685-2dflh" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.777865 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-x66l4" Nov 24 14:43:51 crc kubenswrapper[4790]: I1124 14:43:51.798490 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76d49bc685-2dflh" Nov 24 14:43:52 crc kubenswrapper[4790]: I1124 14:43:52.300919 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-x66l4"] Nov 24 14:43:52 crc kubenswrapper[4790]: I1124 14:43:52.325894 4790 scope.go:117] "RemoveContainer" containerID="92c086e44ccb28a62777d612e499209f7483422ca70b460ce6ec73c416e8e8b8" Nov 24 14:43:52 crc kubenswrapper[4790]: E1124 14:43:52.326218 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:43:52 crc kubenswrapper[4790]: I1124 14:43:52.357169 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76d49bc685-2dflh"] Nov 24 14:43:52 crc kubenswrapper[4790]: W1124 14:43:52.367451 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa76fa12_375b_4c64_926f_fc0815a7d4a6.slice/crio-7bd64d7b68d1f9e4ee0cf070dff4c5e9b5cd9ff9748ad304d41c4c0bebeea4a9 WatchSource:0}: Error finding container 7bd64d7b68d1f9e4ee0cf070dff4c5e9b5cd9ff9748ad304d41c4c0bebeea4a9: Status 404 returned error can't find the container with id 7bd64d7b68d1f9e4ee0cf070dff4c5e9b5cd9ff9748ad304d41c4c0bebeea4a9 Nov 24 14:43:53 crc kubenswrapper[4790]: I1124 14:43:53.319566 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-x66l4" event={"ID":"e31ed6a6-8482-4dc5-9da6-46dd87a73053","Type":"ContainerStarted","Data":"d3d13bedb07d91e6537015645cdcf203b6eb953e219100e13cd04d174716439b"} Nov 24 14:43:53 crc kubenswrapper[4790]: I1124 14:43:53.320111 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-x66l4" event={"ID":"e31ed6a6-8482-4dc5-9da6-46dd87a73053","Type":"ContainerStarted","Data":"a5e8de0b594354b97e9c075b87dfa3bde41f659446d05f857ed80016639b2ef8"} Nov 24 14:43:53 crc kubenswrapper[4790]: I1124 14:43:53.323478 4790 generic.go:334] "Generic (PLEG): container finished" podID="aa76fa12-375b-4c64-926f-fc0815a7d4a6" containerID="894257bd0519ae9d3abb6489e66869e13da246857e81314bec7f41e7f82df3f9" exitCode=0 Nov 24 14:43:53 crc kubenswrapper[4790]: I1124 14:43:53.323533 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76d49bc685-2dflh" event={"ID":"aa76fa12-375b-4c64-926f-fc0815a7d4a6","Type":"ContainerDied","Data":"894257bd0519ae9d3abb6489e66869e13da246857e81314bec7f41e7f82df3f9"} Nov 24 14:43:53 crc kubenswrapper[4790]: I1124 14:43:53.323559 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76d49bc685-2dflh" event={"ID":"aa76fa12-375b-4c64-926f-fc0815a7d4a6","Type":"ContainerStarted","Data":"7bd64d7b68d1f9e4ee0cf070dff4c5e9b5cd9ff9748ad304d41c4c0bebeea4a9"} Nov 24 14:43:53 crc kubenswrapper[4790]: I1124 14:43:53.348325 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-x66l4" podStartSLOduration=2.348308119 podStartE2EDuration="2.348308119s" podCreationTimestamp="2025-11-24 14:43:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:43:53.34556424 +0000 UTC m=+5481.725457912" watchObservedRunningTime="2025-11-24 14:43:53.348308119 +0000 UTC m=+5481.728201781" Nov 24 14:43:54 crc kubenswrapper[4790]: I1124 14:43:54.334980 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76d49bc685-2dflh" event={"ID":"aa76fa12-375b-4c64-926f-fc0815a7d4a6","Type":"ContainerStarted","Data":"ba0dd109b7f2187b0ccd1a652b87c9d4f6ec196a3ca0cdadadd9ff76ad6de458"} Nov 24 14:43:54 crc kubenswrapper[4790]: I1124 14:43:54.335365 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-76d49bc685-2dflh" Nov 24 14:43:54 crc kubenswrapper[4790]: I1124 14:43:54.337021 4790 generic.go:334] "Generic (PLEG): container finished" podID="e31ed6a6-8482-4dc5-9da6-46dd87a73053" containerID="d3d13bedb07d91e6537015645cdcf203b6eb953e219100e13cd04d174716439b" exitCode=0 Nov 24 14:43:54 crc kubenswrapper[4790]: I1124 14:43:54.337087 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-x66l4" event={"ID":"e31ed6a6-8482-4dc5-9da6-46dd87a73053","Type":"ContainerDied","Data":"d3d13bedb07d91e6537015645cdcf203b6eb953e219100e13cd04d174716439b"} Nov 24 14:43:54 crc kubenswrapper[4790]: I1124 14:43:54.358660 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-76d49bc685-2dflh" podStartSLOduration=3.358631507 podStartE2EDuration="3.358631507s" podCreationTimestamp="2025-11-24 14:43:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:43:54.353654324 +0000 UTC m=+5482.733547996" watchObservedRunningTime="2025-11-24 14:43:54.358631507 +0000 UTC m=+5482.738525199" Nov 24 14:43:55 crc kubenswrapper[4790]: I1124 14:43:55.774379 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-x66l4" Nov 24 14:43:55 crc kubenswrapper[4790]: I1124 14:43:55.928930 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e31ed6a6-8482-4dc5-9da6-46dd87a73053-logs\") pod \"e31ed6a6-8482-4dc5-9da6-46dd87a73053\" (UID: \"e31ed6a6-8482-4dc5-9da6-46dd87a73053\") " Nov 24 14:43:55 crc kubenswrapper[4790]: I1124 14:43:55.929014 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbvtw\" (UniqueName: \"kubernetes.io/projected/e31ed6a6-8482-4dc5-9da6-46dd87a73053-kube-api-access-qbvtw\") pod \"e31ed6a6-8482-4dc5-9da6-46dd87a73053\" (UID: \"e31ed6a6-8482-4dc5-9da6-46dd87a73053\") " Nov 24 14:43:55 crc kubenswrapper[4790]: I1124 14:43:55.929064 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e31ed6a6-8482-4dc5-9da6-46dd87a73053-scripts\") pod \"e31ed6a6-8482-4dc5-9da6-46dd87a73053\" (UID: \"e31ed6a6-8482-4dc5-9da6-46dd87a73053\") " Nov 24 14:43:55 crc kubenswrapper[4790]: I1124 14:43:55.929167 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e31ed6a6-8482-4dc5-9da6-46dd87a73053-combined-ca-bundle\") pod \"e31ed6a6-8482-4dc5-9da6-46dd87a73053\" (UID: \"e31ed6a6-8482-4dc5-9da6-46dd87a73053\") " Nov 24 14:43:55 crc kubenswrapper[4790]: I1124 14:43:55.929232 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e31ed6a6-8482-4dc5-9da6-46dd87a73053-config-data\") pod \"e31ed6a6-8482-4dc5-9da6-46dd87a73053\" (UID: \"e31ed6a6-8482-4dc5-9da6-46dd87a73053\") " Nov 24 14:43:55 crc kubenswrapper[4790]: I1124 14:43:55.929381 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e31ed6a6-8482-4dc5-9da6-46dd87a73053-logs" (OuterVolumeSpecName: "logs") pod "e31ed6a6-8482-4dc5-9da6-46dd87a73053" (UID: "e31ed6a6-8482-4dc5-9da6-46dd87a73053"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:43:55 crc kubenswrapper[4790]: I1124 14:43:55.929767 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e31ed6a6-8482-4dc5-9da6-46dd87a73053-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:55 crc kubenswrapper[4790]: I1124 14:43:55.935356 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e31ed6a6-8482-4dc5-9da6-46dd87a73053-scripts" (OuterVolumeSpecName: "scripts") pod "e31ed6a6-8482-4dc5-9da6-46dd87a73053" (UID: "e31ed6a6-8482-4dc5-9da6-46dd87a73053"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:43:55 crc kubenswrapper[4790]: I1124 14:43:55.935985 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e31ed6a6-8482-4dc5-9da6-46dd87a73053-kube-api-access-qbvtw" (OuterVolumeSpecName: "kube-api-access-qbvtw") pod "e31ed6a6-8482-4dc5-9da6-46dd87a73053" (UID: "e31ed6a6-8482-4dc5-9da6-46dd87a73053"). InnerVolumeSpecName "kube-api-access-qbvtw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:43:55 crc kubenswrapper[4790]: I1124 14:43:55.957664 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e31ed6a6-8482-4dc5-9da6-46dd87a73053-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e31ed6a6-8482-4dc5-9da6-46dd87a73053" (UID: "e31ed6a6-8482-4dc5-9da6-46dd87a73053"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:43:55 crc kubenswrapper[4790]: I1124 14:43:55.959396 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e31ed6a6-8482-4dc5-9da6-46dd87a73053-config-data" (OuterVolumeSpecName: "config-data") pod "e31ed6a6-8482-4dc5-9da6-46dd87a73053" (UID: "e31ed6a6-8482-4dc5-9da6-46dd87a73053"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:43:56 crc kubenswrapper[4790]: I1124 14:43:56.032956 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbvtw\" (UniqueName: \"kubernetes.io/projected/e31ed6a6-8482-4dc5-9da6-46dd87a73053-kube-api-access-qbvtw\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:56 crc kubenswrapper[4790]: I1124 14:43:56.033004 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e31ed6a6-8482-4dc5-9da6-46dd87a73053-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:56 crc kubenswrapper[4790]: I1124 14:43:56.033021 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e31ed6a6-8482-4dc5-9da6-46dd87a73053-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:56 crc kubenswrapper[4790]: I1124 14:43:56.033033 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e31ed6a6-8482-4dc5-9da6-46dd87a73053-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:43:56 crc kubenswrapper[4790]: I1124 14:43:56.362539 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-x66l4" event={"ID":"e31ed6a6-8482-4dc5-9da6-46dd87a73053","Type":"ContainerDied","Data":"a5e8de0b594354b97e9c075b87dfa3bde41f659446d05f857ed80016639b2ef8"} Nov 24 14:43:56 crc kubenswrapper[4790]: I1124 14:43:56.362588 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5e8de0b594354b97e9c075b87dfa3bde41f659446d05f857ed80016639b2ef8" Nov 24 14:43:56 crc kubenswrapper[4790]: I1124 14:43:56.362613 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-x66l4" Nov 24 14:43:56 crc kubenswrapper[4790]: I1124 14:43:56.434743 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-59d7f799d8-grzhn"] Nov 24 14:43:56 crc kubenswrapper[4790]: E1124 14:43:56.435598 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e31ed6a6-8482-4dc5-9da6-46dd87a73053" containerName="placement-db-sync" Nov 24 14:43:56 crc kubenswrapper[4790]: I1124 14:43:56.435626 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="e31ed6a6-8482-4dc5-9da6-46dd87a73053" containerName="placement-db-sync" Nov 24 14:43:56 crc kubenswrapper[4790]: I1124 14:43:56.435919 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="e31ed6a6-8482-4dc5-9da6-46dd87a73053" containerName="placement-db-sync" Nov 24 14:43:56 crc kubenswrapper[4790]: I1124 14:43:56.437143 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-59d7f799d8-grzhn" Nov 24 14:43:56 crc kubenswrapper[4790]: I1124 14:43:56.440213 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-7tfsd" Nov 24 14:43:56 crc kubenswrapper[4790]: I1124 14:43:56.443914 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 24 14:43:56 crc kubenswrapper[4790]: I1124 14:43:56.444899 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 24 14:43:56 crc kubenswrapper[4790]: I1124 14:43:56.459765 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-59d7f799d8-grzhn"] Nov 24 14:43:56 crc kubenswrapper[4790]: I1124 14:43:56.543217 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca49278e-82e5-42f8-9b83-a7f44e4e21e6-scripts\") pod \"placement-59d7f799d8-grzhn\" (UID: \"ca49278e-82e5-42f8-9b83-a7f44e4e21e6\") " pod="openstack/placement-59d7f799d8-grzhn" Nov 24 14:43:56 crc kubenswrapper[4790]: I1124 14:43:56.543268 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca49278e-82e5-42f8-9b83-a7f44e4e21e6-logs\") pod \"placement-59d7f799d8-grzhn\" (UID: \"ca49278e-82e5-42f8-9b83-a7f44e4e21e6\") " pod="openstack/placement-59d7f799d8-grzhn" Nov 24 14:43:56 crc kubenswrapper[4790]: I1124 14:43:56.543344 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca49278e-82e5-42f8-9b83-a7f44e4e21e6-config-data\") pod \"placement-59d7f799d8-grzhn\" (UID: \"ca49278e-82e5-42f8-9b83-a7f44e4e21e6\") " pod="openstack/placement-59d7f799d8-grzhn" Nov 24 14:43:56 crc kubenswrapper[4790]: I1124 14:43:56.543370 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca49278e-82e5-42f8-9b83-a7f44e4e21e6-combined-ca-bundle\") pod \"placement-59d7f799d8-grzhn\" (UID: \"ca49278e-82e5-42f8-9b83-a7f44e4e21e6\") " pod="openstack/placement-59d7f799d8-grzhn" Nov 24 14:43:56 crc kubenswrapper[4790]: I1124 14:43:56.543791 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkg79\" (UniqueName: \"kubernetes.io/projected/ca49278e-82e5-42f8-9b83-a7f44e4e21e6-kube-api-access-mkg79\") pod \"placement-59d7f799d8-grzhn\" (UID: \"ca49278e-82e5-42f8-9b83-a7f44e4e21e6\") " pod="openstack/placement-59d7f799d8-grzhn" Nov 24 14:43:56 crc kubenswrapper[4790]: I1124 14:43:56.646624 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca49278e-82e5-42f8-9b83-a7f44e4e21e6-config-data\") pod \"placement-59d7f799d8-grzhn\" (UID: \"ca49278e-82e5-42f8-9b83-a7f44e4e21e6\") " pod="openstack/placement-59d7f799d8-grzhn" Nov 24 14:43:56 crc kubenswrapper[4790]: I1124 14:43:56.646725 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca49278e-82e5-42f8-9b83-a7f44e4e21e6-combined-ca-bundle\") pod \"placement-59d7f799d8-grzhn\" (UID: \"ca49278e-82e5-42f8-9b83-a7f44e4e21e6\") " pod="openstack/placement-59d7f799d8-grzhn" Nov 24 14:43:56 crc kubenswrapper[4790]: I1124 14:43:56.646876 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkg79\" (UniqueName: \"kubernetes.io/projected/ca49278e-82e5-42f8-9b83-a7f44e4e21e6-kube-api-access-mkg79\") pod \"placement-59d7f799d8-grzhn\" (UID: \"ca49278e-82e5-42f8-9b83-a7f44e4e21e6\") " pod="openstack/placement-59d7f799d8-grzhn" Nov 24 14:43:56 crc kubenswrapper[4790]: I1124 14:43:56.646971 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca49278e-82e5-42f8-9b83-a7f44e4e21e6-scripts\") pod \"placement-59d7f799d8-grzhn\" (UID: \"ca49278e-82e5-42f8-9b83-a7f44e4e21e6\") " pod="openstack/placement-59d7f799d8-grzhn" Nov 24 14:43:56 crc kubenswrapper[4790]: I1124 14:43:56.647009 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca49278e-82e5-42f8-9b83-a7f44e4e21e6-logs\") pod \"placement-59d7f799d8-grzhn\" (UID: \"ca49278e-82e5-42f8-9b83-a7f44e4e21e6\") " pod="openstack/placement-59d7f799d8-grzhn" Nov 24 14:43:56 crc kubenswrapper[4790]: I1124 14:43:56.648020 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca49278e-82e5-42f8-9b83-a7f44e4e21e6-logs\") pod \"placement-59d7f799d8-grzhn\" (UID: \"ca49278e-82e5-42f8-9b83-a7f44e4e21e6\") " pod="openstack/placement-59d7f799d8-grzhn" Nov 24 14:43:56 crc kubenswrapper[4790]: I1124 14:43:56.651846 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca49278e-82e5-42f8-9b83-a7f44e4e21e6-combined-ca-bundle\") pod \"placement-59d7f799d8-grzhn\" (UID: \"ca49278e-82e5-42f8-9b83-a7f44e4e21e6\") " pod="openstack/placement-59d7f799d8-grzhn" Nov 24 14:43:56 crc kubenswrapper[4790]: I1124 14:43:56.653590 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca49278e-82e5-42f8-9b83-a7f44e4e21e6-scripts\") pod \"placement-59d7f799d8-grzhn\" (UID: \"ca49278e-82e5-42f8-9b83-a7f44e4e21e6\") " pod="openstack/placement-59d7f799d8-grzhn" Nov 24 14:43:56 crc kubenswrapper[4790]: I1124 14:43:56.655711 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca49278e-82e5-42f8-9b83-a7f44e4e21e6-config-data\") pod \"placement-59d7f799d8-grzhn\" (UID: \"ca49278e-82e5-42f8-9b83-a7f44e4e21e6\") " pod="openstack/placement-59d7f799d8-grzhn" Nov 24 14:43:56 crc kubenswrapper[4790]: I1124 14:43:56.677208 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkg79\" (UniqueName: \"kubernetes.io/projected/ca49278e-82e5-42f8-9b83-a7f44e4e21e6-kube-api-access-mkg79\") pod \"placement-59d7f799d8-grzhn\" (UID: \"ca49278e-82e5-42f8-9b83-a7f44e4e21e6\") " pod="openstack/placement-59d7f799d8-grzhn" Nov 24 14:43:56 crc kubenswrapper[4790]: I1124 14:43:56.767069 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-59d7f799d8-grzhn" Nov 24 14:43:57 crc kubenswrapper[4790]: I1124 14:43:57.205383 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-59d7f799d8-grzhn"] Nov 24 14:43:57 crc kubenswrapper[4790]: W1124 14:43:57.209938 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca49278e_82e5_42f8_9b83_a7f44e4e21e6.slice/crio-872f44ab45328cf3f803b0abdbc2aa396f353260e4696b0b7446ed920a0a98f7 WatchSource:0}: Error finding container 872f44ab45328cf3f803b0abdbc2aa396f353260e4696b0b7446ed920a0a98f7: Status 404 returned error can't find the container with id 872f44ab45328cf3f803b0abdbc2aa396f353260e4696b0b7446ed920a0a98f7 Nov 24 14:43:57 crc kubenswrapper[4790]: I1124 14:43:57.377028 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-59d7f799d8-grzhn" event={"ID":"ca49278e-82e5-42f8-9b83-a7f44e4e21e6","Type":"ContainerStarted","Data":"872f44ab45328cf3f803b0abdbc2aa396f353260e4696b0b7446ed920a0a98f7"} Nov 24 14:43:58 crc kubenswrapper[4790]: I1124 14:43:58.390726 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-59d7f799d8-grzhn" event={"ID":"ca49278e-82e5-42f8-9b83-a7f44e4e21e6","Type":"ContainerStarted","Data":"5b6455dfd2079a917163d2198f1485893d8770012bcc1f3a1252be195114130e"} Nov 24 14:43:58 crc kubenswrapper[4790]: I1124 14:43:58.391779 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-59d7f799d8-grzhn" Nov 24 14:43:58 crc kubenswrapper[4790]: I1124 14:43:58.391794 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-59d7f799d8-grzhn" event={"ID":"ca49278e-82e5-42f8-9b83-a7f44e4e21e6","Type":"ContainerStarted","Data":"ca90dd2337d4b0990f2da4524bbe8de6cc02ebdad9140bb52f493eef6ea9b16b"} Nov 24 14:43:58 crc kubenswrapper[4790]: I1124 14:43:58.429286 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-59d7f799d8-grzhn" podStartSLOduration=2.429243202 podStartE2EDuration="2.429243202s" podCreationTimestamp="2025-11-24 14:43:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:43:58.414368245 +0000 UTC m=+5486.794261907" watchObservedRunningTime="2025-11-24 14:43:58.429243202 +0000 UTC m=+5486.809136914" Nov 24 14:43:59 crc kubenswrapper[4790]: I1124 14:43:59.401727 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-59d7f799d8-grzhn" Nov 24 14:44:01 crc kubenswrapper[4790]: I1124 14:44:01.800950 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-76d49bc685-2dflh" Nov 24 14:44:01 crc kubenswrapper[4790]: I1124 14:44:01.888697 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5dd6c96d9c-4mllx"] Nov 24 14:44:01 crc kubenswrapper[4790]: I1124 14:44:01.889087 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5dd6c96d9c-4mllx" podUID="a903782f-4ab2-4439-92b0-bd0ddef96176" containerName="dnsmasq-dns" containerID="cri-o://1f0586302e08eb2f9b2f20f43edd13990a2e9aa8f223fe3d1d3b60518be779ed" gracePeriod=10 Nov 24 14:44:02 crc kubenswrapper[4790]: I1124 14:44:02.414855 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dd6c96d9c-4mllx" Nov 24 14:44:02 crc kubenswrapper[4790]: I1124 14:44:02.437827 4790 generic.go:334] "Generic (PLEG): container finished" podID="a903782f-4ab2-4439-92b0-bd0ddef96176" containerID="1f0586302e08eb2f9b2f20f43edd13990a2e9aa8f223fe3d1d3b60518be779ed" exitCode=0 Nov 24 14:44:02 crc kubenswrapper[4790]: I1124 14:44:02.438040 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dd6c96d9c-4mllx" event={"ID":"a903782f-4ab2-4439-92b0-bd0ddef96176","Type":"ContainerDied","Data":"1f0586302e08eb2f9b2f20f43edd13990a2e9aa8f223fe3d1d3b60518be779ed"} Nov 24 14:44:02 crc kubenswrapper[4790]: I1124 14:44:02.438209 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dd6c96d9c-4mllx" event={"ID":"a903782f-4ab2-4439-92b0-bd0ddef96176","Type":"ContainerDied","Data":"0ea4cc6c19f5b453ffb420a71677208b805a48818f1d31a3aa184267bc6f1ff2"} Nov 24 14:44:02 crc kubenswrapper[4790]: I1124 14:44:02.438334 4790 scope.go:117] "RemoveContainer" containerID="1f0586302e08eb2f9b2f20f43edd13990a2e9aa8f223fe3d1d3b60518be779ed" Nov 24 14:44:02 crc kubenswrapper[4790]: I1124 14:44:02.438157 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dd6c96d9c-4mllx" Nov 24 14:44:02 crc kubenswrapper[4790]: I1124 14:44:02.464869 4790 scope.go:117] "RemoveContainer" containerID="b9218f5e28274a6d669b38671f94950773306acf16e25bbde80f20207a697021" Nov 24 14:44:02 crc kubenswrapper[4790]: I1124 14:44:02.490079 4790 scope.go:117] "RemoveContainer" containerID="1f0586302e08eb2f9b2f20f43edd13990a2e9aa8f223fe3d1d3b60518be779ed" Nov 24 14:44:02 crc kubenswrapper[4790]: E1124 14:44:02.490523 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f0586302e08eb2f9b2f20f43edd13990a2e9aa8f223fe3d1d3b60518be779ed\": container with ID starting with 1f0586302e08eb2f9b2f20f43edd13990a2e9aa8f223fe3d1d3b60518be779ed not found: ID does not exist" containerID="1f0586302e08eb2f9b2f20f43edd13990a2e9aa8f223fe3d1d3b60518be779ed" Nov 24 14:44:02 crc kubenswrapper[4790]: I1124 14:44:02.490556 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f0586302e08eb2f9b2f20f43edd13990a2e9aa8f223fe3d1d3b60518be779ed"} err="failed to get container status \"1f0586302e08eb2f9b2f20f43edd13990a2e9aa8f223fe3d1d3b60518be779ed\": rpc error: code = NotFound desc = could not find container \"1f0586302e08eb2f9b2f20f43edd13990a2e9aa8f223fe3d1d3b60518be779ed\": container with ID starting with 1f0586302e08eb2f9b2f20f43edd13990a2e9aa8f223fe3d1d3b60518be779ed not found: ID does not exist" Nov 24 14:44:02 crc kubenswrapper[4790]: I1124 14:44:02.490579 4790 scope.go:117] "RemoveContainer" containerID="b9218f5e28274a6d669b38671f94950773306acf16e25bbde80f20207a697021" Nov 24 14:44:02 crc kubenswrapper[4790]: E1124 14:44:02.490913 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9218f5e28274a6d669b38671f94950773306acf16e25bbde80f20207a697021\": container with ID starting with b9218f5e28274a6d669b38671f94950773306acf16e25bbde80f20207a697021 not found: ID does not exist" containerID="b9218f5e28274a6d669b38671f94950773306acf16e25bbde80f20207a697021" Nov 24 14:44:02 crc kubenswrapper[4790]: I1124 14:44:02.491042 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9218f5e28274a6d669b38671f94950773306acf16e25bbde80f20207a697021"} err="failed to get container status \"b9218f5e28274a6d669b38671f94950773306acf16e25bbde80f20207a697021\": rpc error: code = NotFound desc = could not find container \"b9218f5e28274a6d669b38671f94950773306acf16e25bbde80f20207a697021\": container with ID starting with b9218f5e28274a6d669b38671f94950773306acf16e25bbde80f20207a697021 not found: ID does not exist" Nov 24 14:44:02 crc kubenswrapper[4790]: I1124 14:44:02.588826 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a903782f-4ab2-4439-92b0-bd0ddef96176-config\") pod \"a903782f-4ab2-4439-92b0-bd0ddef96176\" (UID: \"a903782f-4ab2-4439-92b0-bd0ddef96176\") " Nov 24 14:44:02 crc kubenswrapper[4790]: I1124 14:44:02.588911 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a903782f-4ab2-4439-92b0-bd0ddef96176-ovsdbserver-sb\") pod \"a903782f-4ab2-4439-92b0-bd0ddef96176\" (UID: \"a903782f-4ab2-4439-92b0-bd0ddef96176\") " Nov 24 14:44:02 crc kubenswrapper[4790]: I1124 14:44:02.588960 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a903782f-4ab2-4439-92b0-bd0ddef96176-ovsdbserver-nb\") pod \"a903782f-4ab2-4439-92b0-bd0ddef96176\" (UID: \"a903782f-4ab2-4439-92b0-bd0ddef96176\") " Nov 24 14:44:02 crc kubenswrapper[4790]: I1124 14:44:02.589090 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a903782f-4ab2-4439-92b0-bd0ddef96176-dns-svc\") pod \"a903782f-4ab2-4439-92b0-bd0ddef96176\" (UID: \"a903782f-4ab2-4439-92b0-bd0ddef96176\") " Nov 24 14:44:02 crc kubenswrapper[4790]: I1124 14:44:02.589144 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fwlc\" (UniqueName: \"kubernetes.io/projected/a903782f-4ab2-4439-92b0-bd0ddef96176-kube-api-access-7fwlc\") pod \"a903782f-4ab2-4439-92b0-bd0ddef96176\" (UID: \"a903782f-4ab2-4439-92b0-bd0ddef96176\") " Nov 24 14:44:02 crc kubenswrapper[4790]: I1124 14:44:02.595827 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a903782f-4ab2-4439-92b0-bd0ddef96176-kube-api-access-7fwlc" (OuterVolumeSpecName: "kube-api-access-7fwlc") pod "a903782f-4ab2-4439-92b0-bd0ddef96176" (UID: "a903782f-4ab2-4439-92b0-bd0ddef96176"). InnerVolumeSpecName "kube-api-access-7fwlc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:44:02 crc kubenswrapper[4790]: I1124 14:44:02.638624 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a903782f-4ab2-4439-92b0-bd0ddef96176-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a903782f-4ab2-4439-92b0-bd0ddef96176" (UID: "a903782f-4ab2-4439-92b0-bd0ddef96176"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:44:02 crc kubenswrapper[4790]: I1124 14:44:02.652612 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a903782f-4ab2-4439-92b0-bd0ddef96176-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a903782f-4ab2-4439-92b0-bd0ddef96176" (UID: "a903782f-4ab2-4439-92b0-bd0ddef96176"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:44:02 crc kubenswrapper[4790]: I1124 14:44:02.655533 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a903782f-4ab2-4439-92b0-bd0ddef96176-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a903782f-4ab2-4439-92b0-bd0ddef96176" (UID: "a903782f-4ab2-4439-92b0-bd0ddef96176"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:44:02 crc kubenswrapper[4790]: I1124 14:44:02.656363 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a903782f-4ab2-4439-92b0-bd0ddef96176-config" (OuterVolumeSpecName: "config") pod "a903782f-4ab2-4439-92b0-bd0ddef96176" (UID: "a903782f-4ab2-4439-92b0-bd0ddef96176"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:44:02 crc kubenswrapper[4790]: I1124 14:44:02.691313 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a903782f-4ab2-4439-92b0-bd0ddef96176-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:44:02 crc kubenswrapper[4790]: I1124 14:44:02.691355 4790 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a903782f-4ab2-4439-92b0-bd0ddef96176-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:44:02 crc kubenswrapper[4790]: I1124 14:44:02.691369 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fwlc\" (UniqueName: \"kubernetes.io/projected/a903782f-4ab2-4439-92b0-bd0ddef96176-kube-api-access-7fwlc\") on node \"crc\" DevicePath \"\"" Nov 24 14:44:02 crc kubenswrapper[4790]: I1124 14:44:02.691385 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a903782f-4ab2-4439-92b0-bd0ddef96176-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:44:02 crc kubenswrapper[4790]: I1124 14:44:02.691396 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a903782f-4ab2-4439-92b0-bd0ddef96176-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:44:02 crc kubenswrapper[4790]: I1124 14:44:02.781762 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5dd6c96d9c-4mllx"] Nov 24 14:44:02 crc kubenswrapper[4790]: I1124 14:44:02.790502 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5dd6c96d9c-4mllx"] Nov 24 14:44:04 crc kubenswrapper[4790]: I1124 14:44:04.315712 4790 scope.go:117] "RemoveContainer" containerID="92c086e44ccb28a62777d612e499209f7483422ca70b460ce6ec73c416e8e8b8" Nov 24 14:44:04 crc kubenswrapper[4790]: E1124 14:44:04.316229 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:44:04 crc kubenswrapper[4790]: I1124 14:44:04.335304 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a903782f-4ab2-4439-92b0-bd0ddef96176" path="/var/lib/kubelet/pods/a903782f-4ab2-4439-92b0-bd0ddef96176/volumes" Nov 24 14:44:18 crc kubenswrapper[4790]: I1124 14:44:18.316148 4790 scope.go:117] "RemoveContainer" containerID="92c086e44ccb28a62777d612e499209f7483422ca70b460ce6ec73c416e8e8b8" Nov 24 14:44:18 crc kubenswrapper[4790]: E1124 14:44:18.318034 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:44:27 crc kubenswrapper[4790]: I1124 14:44:27.771722 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-59d7f799d8-grzhn" Nov 24 14:44:28 crc kubenswrapper[4790]: I1124 14:44:28.836392 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-59d7f799d8-grzhn" Nov 24 14:44:33 crc kubenswrapper[4790]: I1124 14:44:33.315393 4790 scope.go:117] "RemoveContainer" containerID="92c086e44ccb28a62777d612e499209f7483422ca70b460ce6ec73c416e8e8b8" Nov 24 14:44:33 crc kubenswrapper[4790]: E1124 14:44:33.317009 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:44:45 crc kubenswrapper[4790]: I1124 14:44:45.315504 4790 scope.go:117] "RemoveContainer" containerID="92c086e44ccb28a62777d612e499209f7483422ca70b460ce6ec73c416e8e8b8" Nov 24 14:44:45 crc kubenswrapper[4790]: E1124 14:44:45.316705 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:44:47 crc kubenswrapper[4790]: I1124 14:44:47.772534 4790 scope.go:117] "RemoveContainer" containerID="a24444ba4dbbc1ab8e1990bd03a669237261e7dd92debd2bc53b97198fc0f7e7" Nov 24 14:44:47 crc kubenswrapper[4790]: I1124 14:44:47.805090 4790 scope.go:117] "RemoveContainer" containerID="0f8c45aaa6207a956a3591838b8194dfe4de88d9247e7a138640cd4a140fb06c" Nov 24 14:44:47 crc kubenswrapper[4790]: I1124 14:44:47.862103 4790 scope.go:117] "RemoveContainer" containerID="52d7add285f4fde8c621a513f658634507c8af9b6670f4357fba03a56395eb3e" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.246688 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-rg74q"] Nov 24 14:44:49 crc kubenswrapper[4790]: E1124 14:44:49.248974 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a903782f-4ab2-4439-92b0-bd0ddef96176" containerName="dnsmasq-dns" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.249096 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="a903782f-4ab2-4439-92b0-bd0ddef96176" containerName="dnsmasq-dns" Nov 24 14:44:49 crc kubenswrapper[4790]: E1124 14:44:49.249185 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a903782f-4ab2-4439-92b0-bd0ddef96176" containerName="init" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.249248 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="a903782f-4ab2-4439-92b0-bd0ddef96176" containerName="init" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.249507 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="a903782f-4ab2-4439-92b0-bd0ddef96176" containerName="dnsmasq-dns" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.250470 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-rg74q" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.254112 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-rg74q"] Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.340823 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-j98nt"] Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.342089 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-j98nt" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.364705 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-b690-account-create-jkmsr"] Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.366367 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-b690-account-create-jkmsr" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.368027 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4574a286-1090-4721-8e45-12a7ee77e76e-operator-scripts\") pod \"nova-api-db-create-rg74q\" (UID: \"4574a286-1090-4721-8e45-12a7ee77e76e\") " pod="openstack/nova-api-db-create-rg74q" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.368176 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hz2jd\" (UniqueName: \"kubernetes.io/projected/4574a286-1090-4721-8e45-12a7ee77e76e-kube-api-access-hz2jd\") pod \"nova-api-db-create-rg74q\" (UID: \"4574a286-1090-4721-8e45-12a7ee77e76e\") " pod="openstack/nova-api-db-create-rg74q" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.369123 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.375300 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-j98nt"] Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.392052 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-b690-account-create-jkmsr"] Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.460394 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-rm2jx"] Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.461779 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-rm2jx" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.469515 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-rm2jx"] Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.470481 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64f433c4-6665-42d3-90bd-19d5648fccf6-operator-scripts\") pod \"nova-cell0-db-create-j98nt\" (UID: \"64f433c4-6665-42d3-90bd-19d5648fccf6\") " pod="openstack/nova-cell0-db-create-j98nt" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.470664 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c63e605-bedd-4a88-b557-f8581dbf5513-operator-scripts\") pod \"nova-api-b690-account-create-jkmsr\" (UID: \"5c63e605-bedd-4a88-b557-f8581dbf5513\") " pod="openstack/nova-api-b690-account-create-jkmsr" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.470696 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4bdk\" (UniqueName: \"kubernetes.io/projected/64f433c4-6665-42d3-90bd-19d5648fccf6-kube-api-access-t4bdk\") pod \"nova-cell0-db-create-j98nt\" (UID: \"64f433c4-6665-42d3-90bd-19d5648fccf6\") " pod="openstack/nova-cell0-db-create-j98nt" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.470727 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4574a286-1090-4721-8e45-12a7ee77e76e-operator-scripts\") pod \"nova-api-db-create-rg74q\" (UID: \"4574a286-1090-4721-8e45-12a7ee77e76e\") " pod="openstack/nova-api-db-create-rg74q" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.470796 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m65vv\" (UniqueName: \"kubernetes.io/projected/5c63e605-bedd-4a88-b557-f8581dbf5513-kube-api-access-m65vv\") pod \"nova-api-b690-account-create-jkmsr\" (UID: \"5c63e605-bedd-4a88-b557-f8581dbf5513\") " pod="openstack/nova-api-b690-account-create-jkmsr" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.470942 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hz2jd\" (UniqueName: \"kubernetes.io/projected/4574a286-1090-4721-8e45-12a7ee77e76e-kube-api-access-hz2jd\") pod \"nova-api-db-create-rg74q\" (UID: \"4574a286-1090-4721-8e45-12a7ee77e76e\") " pod="openstack/nova-api-db-create-rg74q" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.472844 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4574a286-1090-4721-8e45-12a7ee77e76e-operator-scripts\") pod \"nova-api-db-create-rg74q\" (UID: \"4574a286-1090-4721-8e45-12a7ee77e76e\") " pod="openstack/nova-api-db-create-rg74q" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.492651 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hz2jd\" (UniqueName: \"kubernetes.io/projected/4574a286-1090-4721-8e45-12a7ee77e76e-kube-api-access-hz2jd\") pod \"nova-api-db-create-rg74q\" (UID: \"4574a286-1090-4721-8e45-12a7ee77e76e\") " pod="openstack/nova-api-db-create-rg74q" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.549069 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-3cfe-account-create-7wftp"] Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.550527 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-3cfe-account-create-7wftp" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.556776 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.566877 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-3cfe-account-create-7wftp"] Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.573708 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64f433c4-6665-42d3-90bd-19d5648fccf6-operator-scripts\") pod \"nova-cell0-db-create-j98nt\" (UID: \"64f433c4-6665-42d3-90bd-19d5648fccf6\") " pod="openstack/nova-cell0-db-create-j98nt" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.574219 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c63e605-bedd-4a88-b557-f8581dbf5513-operator-scripts\") pod \"nova-api-b690-account-create-jkmsr\" (UID: \"5c63e605-bedd-4a88-b557-f8581dbf5513\") " pod="openstack/nova-api-b690-account-create-jkmsr" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.574320 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4bdk\" (UniqueName: \"kubernetes.io/projected/64f433c4-6665-42d3-90bd-19d5648fccf6-kube-api-access-t4bdk\") pod \"nova-cell0-db-create-j98nt\" (UID: \"64f433c4-6665-42d3-90bd-19d5648fccf6\") " pod="openstack/nova-cell0-db-create-j98nt" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.574434 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4f5n\" (UniqueName: \"kubernetes.io/projected/4bfcc365-af75-4e73-9980-bce20e886b88-kube-api-access-p4f5n\") pod \"nova-cell1-db-create-rm2jx\" (UID: \"4bfcc365-af75-4e73-9980-bce20e886b88\") " pod="openstack/nova-cell1-db-create-rm2jx" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.574560 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m65vv\" (UniqueName: \"kubernetes.io/projected/5c63e605-bedd-4a88-b557-f8581dbf5513-kube-api-access-m65vv\") pod \"nova-api-b690-account-create-jkmsr\" (UID: \"5c63e605-bedd-4a88-b557-f8581dbf5513\") " pod="openstack/nova-api-b690-account-create-jkmsr" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.574661 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4bfcc365-af75-4e73-9980-bce20e886b88-operator-scripts\") pod \"nova-cell1-db-create-rm2jx\" (UID: \"4bfcc365-af75-4e73-9980-bce20e886b88\") " pod="openstack/nova-cell1-db-create-rm2jx" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.575851 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64f433c4-6665-42d3-90bd-19d5648fccf6-operator-scripts\") pod \"nova-cell0-db-create-j98nt\" (UID: \"64f433c4-6665-42d3-90bd-19d5648fccf6\") " pod="openstack/nova-cell0-db-create-j98nt" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.576646 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c63e605-bedd-4a88-b557-f8581dbf5513-operator-scripts\") pod \"nova-api-b690-account-create-jkmsr\" (UID: \"5c63e605-bedd-4a88-b557-f8581dbf5513\") " pod="openstack/nova-api-b690-account-create-jkmsr" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.578418 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-rg74q" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.596170 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m65vv\" (UniqueName: \"kubernetes.io/projected/5c63e605-bedd-4a88-b557-f8581dbf5513-kube-api-access-m65vv\") pod \"nova-api-b690-account-create-jkmsr\" (UID: \"5c63e605-bedd-4a88-b557-f8581dbf5513\") " pod="openstack/nova-api-b690-account-create-jkmsr" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.596439 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4bdk\" (UniqueName: \"kubernetes.io/projected/64f433c4-6665-42d3-90bd-19d5648fccf6-kube-api-access-t4bdk\") pod \"nova-cell0-db-create-j98nt\" (UID: \"64f433c4-6665-42d3-90bd-19d5648fccf6\") " pod="openstack/nova-cell0-db-create-j98nt" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.667414 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-j98nt" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.677025 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/054b0bf7-15a9-4475-9c53-f5e5baf24518-operator-scripts\") pod \"nova-cell0-3cfe-account-create-7wftp\" (UID: \"054b0bf7-15a9-4475-9c53-f5e5baf24518\") " pod="openstack/nova-cell0-3cfe-account-create-7wftp" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.677122 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4f5n\" (UniqueName: \"kubernetes.io/projected/4bfcc365-af75-4e73-9980-bce20e886b88-kube-api-access-p4f5n\") pod \"nova-cell1-db-create-rm2jx\" (UID: \"4bfcc365-af75-4e73-9980-bce20e886b88\") " pod="openstack/nova-cell1-db-create-rm2jx" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.677171 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4bfcc365-af75-4e73-9980-bce20e886b88-operator-scripts\") pod \"nova-cell1-db-create-rm2jx\" (UID: \"4bfcc365-af75-4e73-9980-bce20e886b88\") " pod="openstack/nova-cell1-db-create-rm2jx" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.677265 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5gwv\" (UniqueName: \"kubernetes.io/projected/054b0bf7-15a9-4475-9c53-f5e5baf24518-kube-api-access-c5gwv\") pod \"nova-cell0-3cfe-account-create-7wftp\" (UID: \"054b0bf7-15a9-4475-9c53-f5e5baf24518\") " pod="openstack/nova-cell0-3cfe-account-create-7wftp" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.678520 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4bfcc365-af75-4e73-9980-bce20e886b88-operator-scripts\") pod \"nova-cell1-db-create-rm2jx\" (UID: \"4bfcc365-af75-4e73-9980-bce20e886b88\") " pod="openstack/nova-cell1-db-create-rm2jx" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.700945 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4f5n\" (UniqueName: \"kubernetes.io/projected/4bfcc365-af75-4e73-9980-bce20e886b88-kube-api-access-p4f5n\") pod \"nova-cell1-db-create-rm2jx\" (UID: \"4bfcc365-af75-4e73-9980-bce20e886b88\") " pod="openstack/nova-cell1-db-create-rm2jx" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.704360 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-b690-account-create-jkmsr" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.763155 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-c1a6-account-create-lwdnx"] Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.764522 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-c1a6-account-create-lwdnx" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.767551 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.773144 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-c1a6-account-create-lwdnx"] Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.782131 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-rm2jx" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.782961 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5gwv\" (UniqueName: \"kubernetes.io/projected/054b0bf7-15a9-4475-9c53-f5e5baf24518-kube-api-access-c5gwv\") pod \"nova-cell0-3cfe-account-create-7wftp\" (UID: \"054b0bf7-15a9-4475-9c53-f5e5baf24518\") " pod="openstack/nova-cell0-3cfe-account-create-7wftp" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.783012 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/054b0bf7-15a9-4475-9c53-f5e5baf24518-operator-scripts\") pod \"nova-cell0-3cfe-account-create-7wftp\" (UID: \"054b0bf7-15a9-4475-9c53-f5e5baf24518\") " pod="openstack/nova-cell0-3cfe-account-create-7wftp" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.783668 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/054b0bf7-15a9-4475-9c53-f5e5baf24518-operator-scripts\") pod \"nova-cell0-3cfe-account-create-7wftp\" (UID: \"054b0bf7-15a9-4475-9c53-f5e5baf24518\") " pod="openstack/nova-cell0-3cfe-account-create-7wftp" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.806785 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5gwv\" (UniqueName: \"kubernetes.io/projected/054b0bf7-15a9-4475-9c53-f5e5baf24518-kube-api-access-c5gwv\") pod \"nova-cell0-3cfe-account-create-7wftp\" (UID: \"054b0bf7-15a9-4475-9c53-f5e5baf24518\") " pod="openstack/nova-cell0-3cfe-account-create-7wftp" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.881903 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-3cfe-account-create-7wftp" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.886854 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99ad8afe-4fb4-4823-a699-c99e3efb75fb-operator-scripts\") pod \"nova-cell1-c1a6-account-create-lwdnx\" (UID: \"99ad8afe-4fb4-4823-a699-c99e3efb75fb\") " pod="openstack/nova-cell1-c1a6-account-create-lwdnx" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.886949 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdwmh\" (UniqueName: \"kubernetes.io/projected/99ad8afe-4fb4-4823-a699-c99e3efb75fb-kube-api-access-mdwmh\") pod \"nova-cell1-c1a6-account-create-lwdnx\" (UID: \"99ad8afe-4fb4-4823-a699-c99e3efb75fb\") " pod="openstack/nova-cell1-c1a6-account-create-lwdnx" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.989248 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99ad8afe-4fb4-4823-a699-c99e3efb75fb-operator-scripts\") pod \"nova-cell1-c1a6-account-create-lwdnx\" (UID: \"99ad8afe-4fb4-4823-a699-c99e3efb75fb\") " pod="openstack/nova-cell1-c1a6-account-create-lwdnx" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.989353 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdwmh\" (UniqueName: \"kubernetes.io/projected/99ad8afe-4fb4-4823-a699-c99e3efb75fb-kube-api-access-mdwmh\") pod \"nova-cell1-c1a6-account-create-lwdnx\" (UID: \"99ad8afe-4fb4-4823-a699-c99e3efb75fb\") " pod="openstack/nova-cell1-c1a6-account-create-lwdnx" Nov 24 14:44:49 crc kubenswrapper[4790]: I1124 14:44:49.990850 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99ad8afe-4fb4-4823-a699-c99e3efb75fb-operator-scripts\") pod \"nova-cell1-c1a6-account-create-lwdnx\" (UID: \"99ad8afe-4fb4-4823-a699-c99e3efb75fb\") " pod="openstack/nova-cell1-c1a6-account-create-lwdnx" Nov 24 14:44:50 crc kubenswrapper[4790]: I1124 14:44:50.010316 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdwmh\" (UniqueName: \"kubernetes.io/projected/99ad8afe-4fb4-4823-a699-c99e3efb75fb-kube-api-access-mdwmh\") pod \"nova-cell1-c1a6-account-create-lwdnx\" (UID: \"99ad8afe-4fb4-4823-a699-c99e3efb75fb\") " pod="openstack/nova-cell1-c1a6-account-create-lwdnx" Nov 24 14:44:50 crc kubenswrapper[4790]: I1124 14:44:50.099857 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-c1a6-account-create-lwdnx" Nov 24 14:44:50 crc kubenswrapper[4790]: I1124 14:44:50.184337 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-j98nt"] Nov 24 14:44:50 crc kubenswrapper[4790]: I1124 14:44:50.259443 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-rg74q"] Nov 24 14:44:50 crc kubenswrapper[4790]: I1124 14:44:50.330767 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-b690-account-create-jkmsr"] Nov 24 14:44:50 crc kubenswrapper[4790]: W1124 14:44:50.346316 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5c63e605_bedd_4a88_b557_f8581dbf5513.slice/crio-e0451ba964e6be4b2dd898230b6ab2f49942487afc9b874389bb9e5161aa6a42 WatchSource:0}: Error finding container e0451ba964e6be4b2dd898230b6ab2f49942487afc9b874389bb9e5161aa6a42: Status 404 returned error can't find the container with id e0451ba964e6be4b2dd898230b6ab2f49942487afc9b874389bb9e5161aa6a42 Nov 24 14:44:50 crc kubenswrapper[4790]: I1124 14:44:50.519736 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-rm2jx"] Nov 24 14:44:50 crc kubenswrapper[4790]: I1124 14:44:50.661356 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-3cfe-account-create-7wftp"] Nov 24 14:44:50 crc kubenswrapper[4790]: W1124 14:44:50.669516 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod054b0bf7_15a9_4475_9c53_f5e5baf24518.slice/crio-2343644d83028d896d15ab0c1503f3f45e4a7aa3e5e581b323d26781205d2855 WatchSource:0}: Error finding container 2343644d83028d896d15ab0c1503f3f45e4a7aa3e5e581b323d26781205d2855: Status 404 returned error can't find the container with id 2343644d83028d896d15ab0c1503f3f45e4a7aa3e5e581b323d26781205d2855 Nov 24 14:44:50 crc kubenswrapper[4790]: I1124 14:44:50.765795 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-c1a6-account-create-lwdnx"] Nov 24 14:44:51 crc kubenswrapper[4790]: I1124 14:44:51.035733 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-c1a6-account-create-lwdnx" event={"ID":"99ad8afe-4fb4-4823-a699-c99e3efb75fb","Type":"ContainerStarted","Data":"b90dacf4b2e8751334843927b9a9f003205791d8dd957bb2e7483e4715518456"} Nov 24 14:44:51 crc kubenswrapper[4790]: I1124 14:44:51.037666 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-rm2jx" event={"ID":"4bfcc365-af75-4e73-9980-bce20e886b88","Type":"ContainerStarted","Data":"69aae9f6aa724e1fc1d9c2a392738d19f04108d96539608e37e9a69b94c393a4"} Nov 24 14:44:51 crc kubenswrapper[4790]: I1124 14:44:51.041069 4790 generic.go:334] "Generic (PLEG): container finished" podID="4574a286-1090-4721-8e45-12a7ee77e76e" containerID="20a60ed2404a9b78c22ff3b4d82a3742b2ceeda85aa4919fff5d62f3e933ade4" exitCode=0 Nov 24 14:44:51 crc kubenswrapper[4790]: I1124 14:44:51.041127 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-rg74q" event={"ID":"4574a286-1090-4721-8e45-12a7ee77e76e","Type":"ContainerDied","Data":"20a60ed2404a9b78c22ff3b4d82a3742b2ceeda85aa4919fff5d62f3e933ade4"} Nov 24 14:44:51 crc kubenswrapper[4790]: I1124 14:44:51.041145 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-rg74q" event={"ID":"4574a286-1090-4721-8e45-12a7ee77e76e","Type":"ContainerStarted","Data":"a56139b5c0b51d7273e8344794979611bcf553f2909de7d531995b667c3977c2"} Nov 24 14:44:51 crc kubenswrapper[4790]: I1124 14:44:51.044234 4790 generic.go:334] "Generic (PLEG): container finished" podID="64f433c4-6665-42d3-90bd-19d5648fccf6" containerID="d330b7dd28854cd58a09d86e5bda959ea7b9d02af5cae00a4bb708ac7457bcf0" exitCode=0 Nov 24 14:44:51 crc kubenswrapper[4790]: I1124 14:44:51.044302 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-j98nt" event={"ID":"64f433c4-6665-42d3-90bd-19d5648fccf6","Type":"ContainerDied","Data":"d330b7dd28854cd58a09d86e5bda959ea7b9d02af5cae00a4bb708ac7457bcf0"} Nov 24 14:44:51 crc kubenswrapper[4790]: I1124 14:44:51.044325 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-j98nt" event={"ID":"64f433c4-6665-42d3-90bd-19d5648fccf6","Type":"ContainerStarted","Data":"d5890b4069869501d4a02ed892f9f5ebbf9ece796efe8b31cb572c2a951be487"} Nov 24 14:44:51 crc kubenswrapper[4790]: I1124 14:44:51.046934 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-3cfe-account-create-7wftp" event={"ID":"054b0bf7-15a9-4475-9c53-f5e5baf24518","Type":"ContainerStarted","Data":"2343644d83028d896d15ab0c1503f3f45e4a7aa3e5e581b323d26781205d2855"} Nov 24 14:44:51 crc kubenswrapper[4790]: I1124 14:44:51.052218 4790 generic.go:334] "Generic (PLEG): container finished" podID="5c63e605-bedd-4a88-b557-f8581dbf5513" containerID="7d4b9f6a11c41644138576f5ccb753d918c145085f323cc28912ba50cced709b" exitCode=0 Nov 24 14:44:51 crc kubenswrapper[4790]: I1124 14:44:51.052281 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-b690-account-create-jkmsr" event={"ID":"5c63e605-bedd-4a88-b557-f8581dbf5513","Type":"ContainerDied","Data":"7d4b9f6a11c41644138576f5ccb753d918c145085f323cc28912ba50cced709b"} Nov 24 14:44:51 crc kubenswrapper[4790]: I1124 14:44:51.052322 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-b690-account-create-jkmsr" event={"ID":"5c63e605-bedd-4a88-b557-f8581dbf5513","Type":"ContainerStarted","Data":"e0451ba964e6be4b2dd898230b6ab2f49942487afc9b874389bb9e5161aa6a42"} Nov 24 14:44:52 crc kubenswrapper[4790]: I1124 14:44:52.066867 4790 generic.go:334] "Generic (PLEG): container finished" podID="99ad8afe-4fb4-4823-a699-c99e3efb75fb" containerID="ef4eb8d324481200a647ad7872c0e232d63e2484a2666846c3c212a4e8c4912e" exitCode=0 Nov 24 14:44:52 crc kubenswrapper[4790]: I1124 14:44:52.067067 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-c1a6-account-create-lwdnx" event={"ID":"99ad8afe-4fb4-4823-a699-c99e3efb75fb","Type":"ContainerDied","Data":"ef4eb8d324481200a647ad7872c0e232d63e2484a2666846c3c212a4e8c4912e"} Nov 24 14:44:52 crc kubenswrapper[4790]: I1124 14:44:52.069440 4790 generic.go:334] "Generic (PLEG): container finished" podID="4bfcc365-af75-4e73-9980-bce20e886b88" containerID="dd6f816e5859dda855280e67acfa0fd8a91290d029ab7b064470cf2a6b471786" exitCode=0 Nov 24 14:44:52 crc kubenswrapper[4790]: I1124 14:44:52.069578 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-rm2jx" event={"ID":"4bfcc365-af75-4e73-9980-bce20e886b88","Type":"ContainerDied","Data":"dd6f816e5859dda855280e67acfa0fd8a91290d029ab7b064470cf2a6b471786"} Nov 24 14:44:52 crc kubenswrapper[4790]: I1124 14:44:52.072501 4790 generic.go:334] "Generic (PLEG): container finished" podID="054b0bf7-15a9-4475-9c53-f5e5baf24518" containerID="c108fdf41c12d6c658e87825b230bc0807a4e4df5eaa6e1b7735027278a07b20" exitCode=0 Nov 24 14:44:52 crc kubenswrapper[4790]: I1124 14:44:52.072548 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-3cfe-account-create-7wftp" event={"ID":"054b0bf7-15a9-4475-9c53-f5e5baf24518","Type":"ContainerDied","Data":"c108fdf41c12d6c658e87825b230bc0807a4e4df5eaa6e1b7735027278a07b20"} Nov 24 14:44:52 crc kubenswrapper[4790]: I1124 14:44:52.502738 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-j98nt" Nov 24 14:44:52 crc kubenswrapper[4790]: I1124 14:44:52.559773 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64f433c4-6665-42d3-90bd-19d5648fccf6-operator-scripts\") pod \"64f433c4-6665-42d3-90bd-19d5648fccf6\" (UID: \"64f433c4-6665-42d3-90bd-19d5648fccf6\") " Nov 24 14:44:52 crc kubenswrapper[4790]: I1124 14:44:52.559856 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4bdk\" (UniqueName: \"kubernetes.io/projected/64f433c4-6665-42d3-90bd-19d5648fccf6-kube-api-access-t4bdk\") pod \"64f433c4-6665-42d3-90bd-19d5648fccf6\" (UID: \"64f433c4-6665-42d3-90bd-19d5648fccf6\") " Nov 24 14:44:52 crc kubenswrapper[4790]: I1124 14:44:52.561869 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64f433c4-6665-42d3-90bd-19d5648fccf6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "64f433c4-6665-42d3-90bd-19d5648fccf6" (UID: "64f433c4-6665-42d3-90bd-19d5648fccf6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:44:52 crc kubenswrapper[4790]: I1124 14:44:52.569322 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64f433c4-6665-42d3-90bd-19d5648fccf6-kube-api-access-t4bdk" (OuterVolumeSpecName: "kube-api-access-t4bdk") pod "64f433c4-6665-42d3-90bd-19d5648fccf6" (UID: "64f433c4-6665-42d3-90bd-19d5648fccf6"). InnerVolumeSpecName "kube-api-access-t4bdk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:44:52 crc kubenswrapper[4790]: I1124 14:44:52.642950 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-rg74q" Nov 24 14:44:52 crc kubenswrapper[4790]: I1124 14:44:52.651738 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-b690-account-create-jkmsr" Nov 24 14:44:52 crc kubenswrapper[4790]: I1124 14:44:52.662726 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64f433c4-6665-42d3-90bd-19d5648fccf6-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:44:52 crc kubenswrapper[4790]: I1124 14:44:52.662773 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4bdk\" (UniqueName: \"kubernetes.io/projected/64f433c4-6665-42d3-90bd-19d5648fccf6-kube-api-access-t4bdk\") on node \"crc\" DevicePath \"\"" Nov 24 14:44:52 crc kubenswrapper[4790]: I1124 14:44:52.764230 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m65vv\" (UniqueName: \"kubernetes.io/projected/5c63e605-bedd-4a88-b557-f8581dbf5513-kube-api-access-m65vv\") pod \"5c63e605-bedd-4a88-b557-f8581dbf5513\" (UID: \"5c63e605-bedd-4a88-b557-f8581dbf5513\") " Nov 24 14:44:52 crc kubenswrapper[4790]: I1124 14:44:52.764336 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hz2jd\" (UniqueName: \"kubernetes.io/projected/4574a286-1090-4721-8e45-12a7ee77e76e-kube-api-access-hz2jd\") pod \"4574a286-1090-4721-8e45-12a7ee77e76e\" (UID: \"4574a286-1090-4721-8e45-12a7ee77e76e\") " Nov 24 14:44:52 crc kubenswrapper[4790]: I1124 14:44:52.764458 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c63e605-bedd-4a88-b557-f8581dbf5513-operator-scripts\") pod \"5c63e605-bedd-4a88-b557-f8581dbf5513\" (UID: \"5c63e605-bedd-4a88-b557-f8581dbf5513\") " Nov 24 14:44:52 crc kubenswrapper[4790]: I1124 14:44:52.764501 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4574a286-1090-4721-8e45-12a7ee77e76e-operator-scripts\") pod \"4574a286-1090-4721-8e45-12a7ee77e76e\" (UID: \"4574a286-1090-4721-8e45-12a7ee77e76e\") " Nov 24 14:44:52 crc kubenswrapper[4790]: I1124 14:44:52.765563 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c63e605-bedd-4a88-b557-f8581dbf5513-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5c63e605-bedd-4a88-b557-f8581dbf5513" (UID: "5c63e605-bedd-4a88-b557-f8581dbf5513"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:44:52 crc kubenswrapper[4790]: I1124 14:44:52.765737 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4574a286-1090-4721-8e45-12a7ee77e76e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4574a286-1090-4721-8e45-12a7ee77e76e" (UID: "4574a286-1090-4721-8e45-12a7ee77e76e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:44:52 crc kubenswrapper[4790]: I1124 14:44:52.780968 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c63e605-bedd-4a88-b557-f8581dbf5513-kube-api-access-m65vv" (OuterVolumeSpecName: "kube-api-access-m65vv") pod "5c63e605-bedd-4a88-b557-f8581dbf5513" (UID: "5c63e605-bedd-4a88-b557-f8581dbf5513"). InnerVolumeSpecName "kube-api-access-m65vv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:44:52 crc kubenswrapper[4790]: I1124 14:44:52.781054 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4574a286-1090-4721-8e45-12a7ee77e76e-kube-api-access-hz2jd" (OuterVolumeSpecName: "kube-api-access-hz2jd") pod "4574a286-1090-4721-8e45-12a7ee77e76e" (UID: "4574a286-1090-4721-8e45-12a7ee77e76e"). InnerVolumeSpecName "kube-api-access-hz2jd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:44:52 crc kubenswrapper[4790]: I1124 14:44:52.867722 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m65vv\" (UniqueName: \"kubernetes.io/projected/5c63e605-bedd-4a88-b557-f8581dbf5513-kube-api-access-m65vv\") on node \"crc\" DevicePath \"\"" Nov 24 14:44:52 crc kubenswrapper[4790]: I1124 14:44:52.867784 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hz2jd\" (UniqueName: \"kubernetes.io/projected/4574a286-1090-4721-8e45-12a7ee77e76e-kube-api-access-hz2jd\") on node \"crc\" DevicePath \"\"" Nov 24 14:44:52 crc kubenswrapper[4790]: I1124 14:44:52.867805 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c63e605-bedd-4a88-b557-f8581dbf5513-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:44:52 crc kubenswrapper[4790]: I1124 14:44:52.867825 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4574a286-1090-4721-8e45-12a7ee77e76e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:44:53 crc kubenswrapper[4790]: I1124 14:44:53.088019 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-j98nt" event={"ID":"64f433c4-6665-42d3-90bd-19d5648fccf6","Type":"ContainerDied","Data":"d5890b4069869501d4a02ed892f9f5ebbf9ece796efe8b31cb572c2a951be487"} Nov 24 14:44:53 crc kubenswrapper[4790]: I1124 14:44:53.088092 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5890b4069869501d4a02ed892f9f5ebbf9ece796efe8b31cb572c2a951be487" Nov 24 14:44:53 crc kubenswrapper[4790]: I1124 14:44:53.088059 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-j98nt" Nov 24 14:44:53 crc kubenswrapper[4790]: I1124 14:44:53.090642 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-b690-account-create-jkmsr" event={"ID":"5c63e605-bedd-4a88-b557-f8581dbf5513","Type":"ContainerDied","Data":"e0451ba964e6be4b2dd898230b6ab2f49942487afc9b874389bb9e5161aa6a42"} Nov 24 14:44:53 crc kubenswrapper[4790]: I1124 14:44:53.090713 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-b690-account-create-jkmsr" Nov 24 14:44:53 crc kubenswrapper[4790]: I1124 14:44:53.090716 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0451ba964e6be4b2dd898230b6ab2f49942487afc9b874389bb9e5161aa6a42" Nov 24 14:44:53 crc kubenswrapper[4790]: I1124 14:44:53.092978 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-rg74q" event={"ID":"4574a286-1090-4721-8e45-12a7ee77e76e","Type":"ContainerDied","Data":"a56139b5c0b51d7273e8344794979611bcf553f2909de7d531995b667c3977c2"} Nov 24 14:44:53 crc kubenswrapper[4790]: I1124 14:44:53.093143 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a56139b5c0b51d7273e8344794979611bcf553f2909de7d531995b667c3977c2" Nov 24 14:44:53 crc kubenswrapper[4790]: I1124 14:44:53.093262 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-rg74q" Nov 24 14:44:53 crc kubenswrapper[4790]: I1124 14:44:53.511447 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-3cfe-account-create-7wftp" Nov 24 14:44:53 crc kubenswrapper[4790]: I1124 14:44:53.590930 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5gwv\" (UniqueName: \"kubernetes.io/projected/054b0bf7-15a9-4475-9c53-f5e5baf24518-kube-api-access-c5gwv\") pod \"054b0bf7-15a9-4475-9c53-f5e5baf24518\" (UID: \"054b0bf7-15a9-4475-9c53-f5e5baf24518\") " Nov 24 14:44:53 crc kubenswrapper[4790]: I1124 14:44:53.591138 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/054b0bf7-15a9-4475-9c53-f5e5baf24518-operator-scripts\") pod \"054b0bf7-15a9-4475-9c53-f5e5baf24518\" (UID: \"054b0bf7-15a9-4475-9c53-f5e5baf24518\") " Nov 24 14:44:53 crc kubenswrapper[4790]: I1124 14:44:53.591861 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/054b0bf7-15a9-4475-9c53-f5e5baf24518-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "054b0bf7-15a9-4475-9c53-f5e5baf24518" (UID: "054b0bf7-15a9-4475-9c53-f5e5baf24518"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:44:53 crc kubenswrapper[4790]: I1124 14:44:53.597741 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/054b0bf7-15a9-4475-9c53-f5e5baf24518-kube-api-access-c5gwv" (OuterVolumeSpecName: "kube-api-access-c5gwv") pod "054b0bf7-15a9-4475-9c53-f5e5baf24518" (UID: "054b0bf7-15a9-4475-9c53-f5e5baf24518"). InnerVolumeSpecName "kube-api-access-c5gwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:44:53 crc kubenswrapper[4790]: I1124 14:44:53.641632 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-rm2jx" Nov 24 14:44:53 crc kubenswrapper[4790]: I1124 14:44:53.649194 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-c1a6-account-create-lwdnx" Nov 24 14:44:53 crc kubenswrapper[4790]: I1124 14:44:53.693416 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5gwv\" (UniqueName: \"kubernetes.io/projected/054b0bf7-15a9-4475-9c53-f5e5baf24518-kube-api-access-c5gwv\") on node \"crc\" DevicePath \"\"" Nov 24 14:44:53 crc kubenswrapper[4790]: I1124 14:44:53.693457 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/054b0bf7-15a9-4475-9c53-f5e5baf24518-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:44:53 crc kubenswrapper[4790]: I1124 14:44:53.794769 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4bfcc365-af75-4e73-9980-bce20e886b88-operator-scripts\") pod \"4bfcc365-af75-4e73-9980-bce20e886b88\" (UID: \"4bfcc365-af75-4e73-9980-bce20e886b88\") " Nov 24 14:44:53 crc kubenswrapper[4790]: I1124 14:44:53.794923 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4f5n\" (UniqueName: \"kubernetes.io/projected/4bfcc365-af75-4e73-9980-bce20e886b88-kube-api-access-p4f5n\") pod \"4bfcc365-af75-4e73-9980-bce20e886b88\" (UID: \"4bfcc365-af75-4e73-9980-bce20e886b88\") " Nov 24 14:44:53 crc kubenswrapper[4790]: I1124 14:44:53.794964 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdwmh\" (UniqueName: \"kubernetes.io/projected/99ad8afe-4fb4-4823-a699-c99e3efb75fb-kube-api-access-mdwmh\") pod \"99ad8afe-4fb4-4823-a699-c99e3efb75fb\" (UID: \"99ad8afe-4fb4-4823-a699-c99e3efb75fb\") " Nov 24 14:44:53 crc kubenswrapper[4790]: I1124 14:44:53.795087 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99ad8afe-4fb4-4823-a699-c99e3efb75fb-operator-scripts\") pod \"99ad8afe-4fb4-4823-a699-c99e3efb75fb\" (UID: \"99ad8afe-4fb4-4823-a699-c99e3efb75fb\") " Nov 24 14:44:53 crc kubenswrapper[4790]: I1124 14:44:53.795371 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bfcc365-af75-4e73-9980-bce20e886b88-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4bfcc365-af75-4e73-9980-bce20e886b88" (UID: "4bfcc365-af75-4e73-9980-bce20e886b88"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:44:53 crc kubenswrapper[4790]: I1124 14:44:53.795629 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99ad8afe-4fb4-4823-a699-c99e3efb75fb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "99ad8afe-4fb4-4823-a699-c99e3efb75fb" (UID: "99ad8afe-4fb4-4823-a699-c99e3efb75fb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:44:53 crc kubenswrapper[4790]: I1124 14:44:53.799647 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bfcc365-af75-4e73-9980-bce20e886b88-kube-api-access-p4f5n" (OuterVolumeSpecName: "kube-api-access-p4f5n") pod "4bfcc365-af75-4e73-9980-bce20e886b88" (UID: "4bfcc365-af75-4e73-9980-bce20e886b88"). InnerVolumeSpecName "kube-api-access-p4f5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:44:53 crc kubenswrapper[4790]: I1124 14:44:53.802124 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99ad8afe-4fb4-4823-a699-c99e3efb75fb-kube-api-access-mdwmh" (OuterVolumeSpecName: "kube-api-access-mdwmh") pod "99ad8afe-4fb4-4823-a699-c99e3efb75fb" (UID: "99ad8afe-4fb4-4823-a699-c99e3efb75fb"). InnerVolumeSpecName "kube-api-access-mdwmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:44:53 crc kubenswrapper[4790]: I1124 14:44:53.897487 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4f5n\" (UniqueName: \"kubernetes.io/projected/4bfcc365-af75-4e73-9980-bce20e886b88-kube-api-access-p4f5n\") on node \"crc\" DevicePath \"\"" Nov 24 14:44:53 crc kubenswrapper[4790]: I1124 14:44:53.897957 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdwmh\" (UniqueName: \"kubernetes.io/projected/99ad8afe-4fb4-4823-a699-c99e3efb75fb-kube-api-access-mdwmh\") on node \"crc\" DevicePath \"\"" Nov 24 14:44:53 crc kubenswrapper[4790]: I1124 14:44:53.897975 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99ad8afe-4fb4-4823-a699-c99e3efb75fb-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:44:53 crc kubenswrapper[4790]: I1124 14:44:53.897992 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4bfcc365-af75-4e73-9980-bce20e886b88-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:44:54 crc kubenswrapper[4790]: I1124 14:44:54.106867 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-3cfe-account-create-7wftp" event={"ID":"054b0bf7-15a9-4475-9c53-f5e5baf24518","Type":"ContainerDied","Data":"2343644d83028d896d15ab0c1503f3f45e4a7aa3e5e581b323d26781205d2855"} Nov 24 14:44:54 crc kubenswrapper[4790]: I1124 14:44:54.106950 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2343644d83028d896d15ab0c1503f3f45e4a7aa3e5e581b323d26781205d2855" Nov 24 14:44:54 crc kubenswrapper[4790]: I1124 14:44:54.106915 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-3cfe-account-create-7wftp" Nov 24 14:44:54 crc kubenswrapper[4790]: I1124 14:44:54.109855 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-c1a6-account-create-lwdnx" event={"ID":"99ad8afe-4fb4-4823-a699-c99e3efb75fb","Type":"ContainerDied","Data":"b90dacf4b2e8751334843927b9a9f003205791d8dd957bb2e7483e4715518456"} Nov 24 14:44:54 crc kubenswrapper[4790]: I1124 14:44:54.109907 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-c1a6-account-create-lwdnx" Nov 24 14:44:54 crc kubenswrapper[4790]: I1124 14:44:54.109918 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b90dacf4b2e8751334843927b9a9f003205791d8dd957bb2e7483e4715518456" Nov 24 14:44:54 crc kubenswrapper[4790]: I1124 14:44:54.111804 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-rm2jx" event={"ID":"4bfcc365-af75-4e73-9980-bce20e886b88","Type":"ContainerDied","Data":"69aae9f6aa724e1fc1d9c2a392738d19f04108d96539608e37e9a69b94c393a4"} Nov 24 14:44:54 crc kubenswrapper[4790]: I1124 14:44:54.111826 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="69aae9f6aa724e1fc1d9c2a392738d19f04108d96539608e37e9a69b94c393a4" Nov 24 14:44:54 crc kubenswrapper[4790]: I1124 14:44:54.111895 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-rm2jx" Nov 24 14:44:54 crc kubenswrapper[4790]: I1124 14:44:54.964154 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-2w7vx"] Nov 24 14:44:54 crc kubenswrapper[4790]: E1124 14:44:54.964749 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64f433c4-6665-42d3-90bd-19d5648fccf6" containerName="mariadb-database-create" Nov 24 14:44:54 crc kubenswrapper[4790]: I1124 14:44:54.964773 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="64f433c4-6665-42d3-90bd-19d5648fccf6" containerName="mariadb-database-create" Nov 24 14:44:54 crc kubenswrapper[4790]: E1124 14:44:54.964794 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bfcc365-af75-4e73-9980-bce20e886b88" containerName="mariadb-database-create" Nov 24 14:44:54 crc kubenswrapper[4790]: I1124 14:44:54.964802 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bfcc365-af75-4e73-9980-bce20e886b88" containerName="mariadb-database-create" Nov 24 14:44:54 crc kubenswrapper[4790]: E1124 14:44:54.964819 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c63e605-bedd-4a88-b557-f8581dbf5513" containerName="mariadb-account-create" Nov 24 14:44:54 crc kubenswrapper[4790]: I1124 14:44:54.964826 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c63e605-bedd-4a88-b557-f8581dbf5513" containerName="mariadb-account-create" Nov 24 14:44:54 crc kubenswrapper[4790]: E1124 14:44:54.964845 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="054b0bf7-15a9-4475-9c53-f5e5baf24518" containerName="mariadb-account-create" Nov 24 14:44:54 crc kubenswrapper[4790]: I1124 14:44:54.964852 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="054b0bf7-15a9-4475-9c53-f5e5baf24518" containerName="mariadb-account-create" Nov 24 14:44:54 crc kubenswrapper[4790]: E1124 14:44:54.964862 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4574a286-1090-4721-8e45-12a7ee77e76e" containerName="mariadb-database-create" Nov 24 14:44:54 crc kubenswrapper[4790]: I1124 14:44:54.964868 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="4574a286-1090-4721-8e45-12a7ee77e76e" containerName="mariadb-database-create" Nov 24 14:44:54 crc kubenswrapper[4790]: E1124 14:44:54.964896 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99ad8afe-4fb4-4823-a699-c99e3efb75fb" containerName="mariadb-account-create" Nov 24 14:44:54 crc kubenswrapper[4790]: I1124 14:44:54.966987 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="99ad8afe-4fb4-4823-a699-c99e3efb75fb" containerName="mariadb-account-create" Nov 24 14:44:54 crc kubenswrapper[4790]: I1124 14:44:54.967283 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="054b0bf7-15a9-4475-9c53-f5e5baf24518" containerName="mariadb-account-create" Nov 24 14:44:54 crc kubenswrapper[4790]: I1124 14:44:54.967307 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="64f433c4-6665-42d3-90bd-19d5648fccf6" containerName="mariadb-database-create" Nov 24 14:44:54 crc kubenswrapper[4790]: I1124 14:44:54.967322 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="99ad8afe-4fb4-4823-a699-c99e3efb75fb" containerName="mariadb-account-create" Nov 24 14:44:54 crc kubenswrapper[4790]: I1124 14:44:54.967337 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c63e605-bedd-4a88-b557-f8581dbf5513" containerName="mariadb-account-create" Nov 24 14:44:54 crc kubenswrapper[4790]: I1124 14:44:54.967352 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bfcc365-af75-4e73-9980-bce20e886b88" containerName="mariadb-database-create" Nov 24 14:44:54 crc kubenswrapper[4790]: I1124 14:44:54.967368 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="4574a286-1090-4721-8e45-12a7ee77e76e" containerName="mariadb-database-create" Nov 24 14:44:54 crc kubenswrapper[4790]: I1124 14:44:54.968229 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-2w7vx" Nov 24 14:44:54 crc kubenswrapper[4790]: I1124 14:44:54.970650 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 24 14:44:54 crc kubenswrapper[4790]: I1124 14:44:54.971261 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-2brfn" Nov 24 14:44:54 crc kubenswrapper[4790]: I1124 14:44:54.971375 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 14:44:54 crc kubenswrapper[4790]: I1124 14:44:54.976111 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-2w7vx"] Nov 24 14:44:55 crc kubenswrapper[4790]: I1124 14:44:55.118842 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce20586f-aa65-426b-940e-22a8a12cd16d-config-data\") pod \"nova-cell0-conductor-db-sync-2w7vx\" (UID: \"ce20586f-aa65-426b-940e-22a8a12cd16d\") " pod="openstack/nova-cell0-conductor-db-sync-2w7vx" Nov 24 14:44:55 crc kubenswrapper[4790]: I1124 14:44:55.119090 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzfcx\" (UniqueName: \"kubernetes.io/projected/ce20586f-aa65-426b-940e-22a8a12cd16d-kube-api-access-wzfcx\") pod \"nova-cell0-conductor-db-sync-2w7vx\" (UID: \"ce20586f-aa65-426b-940e-22a8a12cd16d\") " pod="openstack/nova-cell0-conductor-db-sync-2w7vx" Nov 24 14:44:55 crc kubenswrapper[4790]: I1124 14:44:55.119122 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce20586f-aa65-426b-940e-22a8a12cd16d-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-2w7vx\" (UID: \"ce20586f-aa65-426b-940e-22a8a12cd16d\") " pod="openstack/nova-cell0-conductor-db-sync-2w7vx" Nov 24 14:44:55 crc kubenswrapper[4790]: I1124 14:44:55.119180 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce20586f-aa65-426b-940e-22a8a12cd16d-scripts\") pod \"nova-cell0-conductor-db-sync-2w7vx\" (UID: \"ce20586f-aa65-426b-940e-22a8a12cd16d\") " pod="openstack/nova-cell0-conductor-db-sync-2w7vx" Nov 24 14:44:55 crc kubenswrapper[4790]: I1124 14:44:55.220691 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzfcx\" (UniqueName: \"kubernetes.io/projected/ce20586f-aa65-426b-940e-22a8a12cd16d-kube-api-access-wzfcx\") pod \"nova-cell0-conductor-db-sync-2w7vx\" (UID: \"ce20586f-aa65-426b-940e-22a8a12cd16d\") " pod="openstack/nova-cell0-conductor-db-sync-2w7vx" Nov 24 14:44:55 crc kubenswrapper[4790]: I1124 14:44:55.220741 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce20586f-aa65-426b-940e-22a8a12cd16d-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-2w7vx\" (UID: \"ce20586f-aa65-426b-940e-22a8a12cd16d\") " pod="openstack/nova-cell0-conductor-db-sync-2w7vx" Nov 24 14:44:55 crc kubenswrapper[4790]: I1124 14:44:55.220801 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce20586f-aa65-426b-940e-22a8a12cd16d-scripts\") pod \"nova-cell0-conductor-db-sync-2w7vx\" (UID: \"ce20586f-aa65-426b-940e-22a8a12cd16d\") " pod="openstack/nova-cell0-conductor-db-sync-2w7vx" Nov 24 14:44:55 crc kubenswrapper[4790]: I1124 14:44:55.220932 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce20586f-aa65-426b-940e-22a8a12cd16d-config-data\") pod \"nova-cell0-conductor-db-sync-2w7vx\" (UID: \"ce20586f-aa65-426b-940e-22a8a12cd16d\") " pod="openstack/nova-cell0-conductor-db-sync-2w7vx" Nov 24 14:44:55 crc kubenswrapper[4790]: I1124 14:44:55.225640 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce20586f-aa65-426b-940e-22a8a12cd16d-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-2w7vx\" (UID: \"ce20586f-aa65-426b-940e-22a8a12cd16d\") " pod="openstack/nova-cell0-conductor-db-sync-2w7vx" Nov 24 14:44:55 crc kubenswrapper[4790]: I1124 14:44:55.225715 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce20586f-aa65-426b-940e-22a8a12cd16d-scripts\") pod \"nova-cell0-conductor-db-sync-2w7vx\" (UID: \"ce20586f-aa65-426b-940e-22a8a12cd16d\") " pod="openstack/nova-cell0-conductor-db-sync-2w7vx" Nov 24 14:44:55 crc kubenswrapper[4790]: I1124 14:44:55.226016 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce20586f-aa65-426b-940e-22a8a12cd16d-config-data\") pod \"nova-cell0-conductor-db-sync-2w7vx\" (UID: \"ce20586f-aa65-426b-940e-22a8a12cd16d\") " pod="openstack/nova-cell0-conductor-db-sync-2w7vx" Nov 24 14:44:55 crc kubenswrapper[4790]: I1124 14:44:55.239354 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzfcx\" (UniqueName: \"kubernetes.io/projected/ce20586f-aa65-426b-940e-22a8a12cd16d-kube-api-access-wzfcx\") pod \"nova-cell0-conductor-db-sync-2w7vx\" (UID: \"ce20586f-aa65-426b-940e-22a8a12cd16d\") " pod="openstack/nova-cell0-conductor-db-sync-2w7vx" Nov 24 14:44:55 crc kubenswrapper[4790]: I1124 14:44:55.305629 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-2w7vx" Nov 24 14:44:55 crc kubenswrapper[4790]: I1124 14:44:55.628599 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-2w7vx"] Nov 24 14:44:55 crc kubenswrapper[4790]: W1124 14:44:55.636667 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce20586f_aa65_426b_940e_22a8a12cd16d.slice/crio-d1730213d95f737cb7ed36d01a25580cc3669301640165dc3022e1aa64a8cc92 WatchSource:0}: Error finding container d1730213d95f737cb7ed36d01a25580cc3669301640165dc3022e1aa64a8cc92: Status 404 returned error can't find the container with id d1730213d95f737cb7ed36d01a25580cc3669301640165dc3022e1aa64a8cc92 Nov 24 14:44:56 crc kubenswrapper[4790]: I1124 14:44:56.136238 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-2w7vx" event={"ID":"ce20586f-aa65-426b-940e-22a8a12cd16d","Type":"ContainerStarted","Data":"cb7a331fbf30381c6e7a979dd1f51750b18f7a4083a30da259de3cc4a9286e29"} Nov 24 14:44:56 crc kubenswrapper[4790]: I1124 14:44:56.136837 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-2w7vx" event={"ID":"ce20586f-aa65-426b-940e-22a8a12cd16d","Type":"ContainerStarted","Data":"d1730213d95f737cb7ed36d01a25580cc3669301640165dc3022e1aa64a8cc92"} Nov 24 14:44:56 crc kubenswrapper[4790]: I1124 14:44:56.164959 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-2w7vx" podStartSLOduration=2.164929256 podStartE2EDuration="2.164929256s" podCreationTimestamp="2025-11-24 14:44:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:44:56.15775272 +0000 UTC m=+5544.537646422" watchObservedRunningTime="2025-11-24 14:44:56.164929256 +0000 UTC m=+5544.544822948" Nov 24 14:44:59 crc kubenswrapper[4790]: I1124 14:44:59.314205 4790 scope.go:117] "RemoveContainer" containerID="92c086e44ccb28a62777d612e499209f7483422ca70b460ce6ec73c416e8e8b8" Nov 24 14:44:59 crc kubenswrapper[4790]: E1124 14:44:59.314969 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:45:00 crc kubenswrapper[4790]: I1124 14:45:00.150035 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399925-xfp46"] Nov 24 14:45:00 crc kubenswrapper[4790]: I1124 14:45:00.152328 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-xfp46" Nov 24 14:45:00 crc kubenswrapper[4790]: I1124 14:45:00.154582 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 14:45:00 crc kubenswrapper[4790]: I1124 14:45:00.155622 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 14:45:00 crc kubenswrapper[4790]: I1124 14:45:00.165832 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399925-xfp46"] Nov 24 14:45:00 crc kubenswrapper[4790]: I1124 14:45:00.232126 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a6cd3c64-493b-4d69-8cec-7465da8136a2-secret-volume\") pod \"collect-profiles-29399925-xfp46\" (UID: \"a6cd3c64-493b-4d69-8cec-7465da8136a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-xfp46" Nov 24 14:45:00 crc kubenswrapper[4790]: I1124 14:45:00.232665 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcz4j\" (UniqueName: \"kubernetes.io/projected/a6cd3c64-493b-4d69-8cec-7465da8136a2-kube-api-access-bcz4j\") pod \"collect-profiles-29399925-xfp46\" (UID: \"a6cd3c64-493b-4d69-8cec-7465da8136a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-xfp46" Nov 24 14:45:00 crc kubenswrapper[4790]: I1124 14:45:00.232821 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a6cd3c64-493b-4d69-8cec-7465da8136a2-config-volume\") pod \"collect-profiles-29399925-xfp46\" (UID: \"a6cd3c64-493b-4d69-8cec-7465da8136a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-xfp46" Nov 24 14:45:00 crc kubenswrapper[4790]: I1124 14:45:00.335258 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcz4j\" (UniqueName: \"kubernetes.io/projected/a6cd3c64-493b-4d69-8cec-7465da8136a2-kube-api-access-bcz4j\") pod \"collect-profiles-29399925-xfp46\" (UID: \"a6cd3c64-493b-4d69-8cec-7465da8136a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-xfp46" Nov 24 14:45:00 crc kubenswrapper[4790]: I1124 14:45:00.335320 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a6cd3c64-493b-4d69-8cec-7465da8136a2-config-volume\") pod \"collect-profiles-29399925-xfp46\" (UID: \"a6cd3c64-493b-4d69-8cec-7465da8136a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-xfp46" Nov 24 14:45:00 crc kubenswrapper[4790]: I1124 14:45:00.335395 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a6cd3c64-493b-4d69-8cec-7465da8136a2-secret-volume\") pod \"collect-profiles-29399925-xfp46\" (UID: \"a6cd3c64-493b-4d69-8cec-7465da8136a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-xfp46" Nov 24 14:45:00 crc kubenswrapper[4790]: I1124 14:45:00.338011 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a6cd3c64-493b-4d69-8cec-7465da8136a2-config-volume\") pod \"collect-profiles-29399925-xfp46\" (UID: \"a6cd3c64-493b-4d69-8cec-7465da8136a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-xfp46" Nov 24 14:45:00 crc kubenswrapper[4790]: I1124 14:45:00.354910 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a6cd3c64-493b-4d69-8cec-7465da8136a2-secret-volume\") pod \"collect-profiles-29399925-xfp46\" (UID: \"a6cd3c64-493b-4d69-8cec-7465da8136a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-xfp46" Nov 24 14:45:00 crc kubenswrapper[4790]: I1124 14:45:00.364610 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcz4j\" (UniqueName: \"kubernetes.io/projected/a6cd3c64-493b-4d69-8cec-7465da8136a2-kube-api-access-bcz4j\") pod \"collect-profiles-29399925-xfp46\" (UID: \"a6cd3c64-493b-4d69-8cec-7465da8136a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-xfp46" Nov 24 14:45:00 crc kubenswrapper[4790]: I1124 14:45:00.490201 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-xfp46" Nov 24 14:45:01 crc kubenswrapper[4790]: I1124 14:45:01.005873 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399925-xfp46"] Nov 24 14:45:01 crc kubenswrapper[4790]: I1124 14:45:01.205292 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-xfp46" event={"ID":"a6cd3c64-493b-4d69-8cec-7465da8136a2","Type":"ContainerStarted","Data":"aa0b084db76668da61051be1b9de078a5ea793b509937ffc68ca303214a318f9"} Nov 24 14:45:01 crc kubenswrapper[4790]: I1124 14:45:01.205760 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-xfp46" event={"ID":"a6cd3c64-493b-4d69-8cec-7465da8136a2","Type":"ContainerStarted","Data":"ff522c418a1bd16616ba3c27700fcfea42dbcd643893c7b6557cc4c56f462d16"} Nov 24 14:45:01 crc kubenswrapper[4790]: I1124 14:45:01.222521 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-xfp46" podStartSLOduration=1.222497038 podStartE2EDuration="1.222497038s" podCreationTimestamp="2025-11-24 14:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:45:01.22153457 +0000 UTC m=+5549.601428222" watchObservedRunningTime="2025-11-24 14:45:01.222497038 +0000 UTC m=+5549.602390710" Nov 24 14:45:02 crc kubenswrapper[4790]: I1124 14:45:02.214323 4790 generic.go:334] "Generic (PLEG): container finished" podID="ce20586f-aa65-426b-940e-22a8a12cd16d" containerID="cb7a331fbf30381c6e7a979dd1f51750b18f7a4083a30da259de3cc4a9286e29" exitCode=0 Nov 24 14:45:02 crc kubenswrapper[4790]: I1124 14:45:02.214489 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-2w7vx" event={"ID":"ce20586f-aa65-426b-940e-22a8a12cd16d","Type":"ContainerDied","Data":"cb7a331fbf30381c6e7a979dd1f51750b18f7a4083a30da259de3cc4a9286e29"} Nov 24 14:45:02 crc kubenswrapper[4790]: I1124 14:45:02.216762 4790 generic.go:334] "Generic (PLEG): container finished" podID="a6cd3c64-493b-4d69-8cec-7465da8136a2" containerID="aa0b084db76668da61051be1b9de078a5ea793b509937ffc68ca303214a318f9" exitCode=0 Nov 24 14:45:02 crc kubenswrapper[4790]: I1124 14:45:02.216813 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-xfp46" event={"ID":"a6cd3c64-493b-4d69-8cec-7465da8136a2","Type":"ContainerDied","Data":"aa0b084db76668da61051be1b9de078a5ea793b509937ffc68ca303214a318f9"} Nov 24 14:45:03 crc kubenswrapper[4790]: I1124 14:45:03.614318 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-xfp46" Nov 24 14:45:03 crc kubenswrapper[4790]: I1124 14:45:03.618564 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-2w7vx" Nov 24 14:45:03 crc kubenswrapper[4790]: I1124 14:45:03.698258 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a6cd3c64-493b-4d69-8cec-7465da8136a2-secret-volume\") pod \"a6cd3c64-493b-4d69-8cec-7465da8136a2\" (UID: \"a6cd3c64-493b-4d69-8cec-7465da8136a2\") " Nov 24 14:45:03 crc kubenswrapper[4790]: I1124 14:45:03.698306 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wzfcx\" (UniqueName: \"kubernetes.io/projected/ce20586f-aa65-426b-940e-22a8a12cd16d-kube-api-access-wzfcx\") pod \"ce20586f-aa65-426b-940e-22a8a12cd16d\" (UID: \"ce20586f-aa65-426b-940e-22a8a12cd16d\") " Nov 24 14:45:03 crc kubenswrapper[4790]: I1124 14:45:03.698367 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce20586f-aa65-426b-940e-22a8a12cd16d-config-data\") pod \"ce20586f-aa65-426b-940e-22a8a12cd16d\" (UID: \"ce20586f-aa65-426b-940e-22a8a12cd16d\") " Nov 24 14:45:03 crc kubenswrapper[4790]: I1124 14:45:03.698403 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a6cd3c64-493b-4d69-8cec-7465da8136a2-config-volume\") pod \"a6cd3c64-493b-4d69-8cec-7465da8136a2\" (UID: \"a6cd3c64-493b-4d69-8cec-7465da8136a2\") " Nov 24 14:45:03 crc kubenswrapper[4790]: I1124 14:45:03.698434 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce20586f-aa65-426b-940e-22a8a12cd16d-scripts\") pod \"ce20586f-aa65-426b-940e-22a8a12cd16d\" (UID: \"ce20586f-aa65-426b-940e-22a8a12cd16d\") " Nov 24 14:45:03 crc kubenswrapper[4790]: I1124 14:45:03.698476 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce20586f-aa65-426b-940e-22a8a12cd16d-combined-ca-bundle\") pod \"ce20586f-aa65-426b-940e-22a8a12cd16d\" (UID: \"ce20586f-aa65-426b-940e-22a8a12cd16d\") " Nov 24 14:45:03 crc kubenswrapper[4790]: I1124 14:45:03.698558 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bcz4j\" (UniqueName: \"kubernetes.io/projected/a6cd3c64-493b-4d69-8cec-7465da8136a2-kube-api-access-bcz4j\") pod \"a6cd3c64-493b-4d69-8cec-7465da8136a2\" (UID: \"a6cd3c64-493b-4d69-8cec-7465da8136a2\") " Nov 24 14:45:03 crc kubenswrapper[4790]: I1124 14:45:03.701118 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6cd3c64-493b-4d69-8cec-7465da8136a2-config-volume" (OuterVolumeSpecName: "config-volume") pod "a6cd3c64-493b-4d69-8cec-7465da8136a2" (UID: "a6cd3c64-493b-4d69-8cec-7465da8136a2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:45:03 crc kubenswrapper[4790]: I1124 14:45:03.706436 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce20586f-aa65-426b-940e-22a8a12cd16d-scripts" (OuterVolumeSpecName: "scripts") pod "ce20586f-aa65-426b-940e-22a8a12cd16d" (UID: "ce20586f-aa65-426b-940e-22a8a12cd16d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:45:03 crc kubenswrapper[4790]: I1124 14:45:03.707132 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6cd3c64-493b-4d69-8cec-7465da8136a2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a6cd3c64-493b-4d69-8cec-7465da8136a2" (UID: "a6cd3c64-493b-4d69-8cec-7465da8136a2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:45:03 crc kubenswrapper[4790]: I1124 14:45:03.707308 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6cd3c64-493b-4d69-8cec-7465da8136a2-kube-api-access-bcz4j" (OuterVolumeSpecName: "kube-api-access-bcz4j") pod "a6cd3c64-493b-4d69-8cec-7465da8136a2" (UID: "a6cd3c64-493b-4d69-8cec-7465da8136a2"). InnerVolumeSpecName "kube-api-access-bcz4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:45:03 crc kubenswrapper[4790]: I1124 14:45:03.709675 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce20586f-aa65-426b-940e-22a8a12cd16d-kube-api-access-wzfcx" (OuterVolumeSpecName: "kube-api-access-wzfcx") pod "ce20586f-aa65-426b-940e-22a8a12cd16d" (UID: "ce20586f-aa65-426b-940e-22a8a12cd16d"). InnerVolumeSpecName "kube-api-access-wzfcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:45:03 crc kubenswrapper[4790]: I1124 14:45:03.730553 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce20586f-aa65-426b-940e-22a8a12cd16d-config-data" (OuterVolumeSpecName: "config-data") pod "ce20586f-aa65-426b-940e-22a8a12cd16d" (UID: "ce20586f-aa65-426b-940e-22a8a12cd16d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:45:03 crc kubenswrapper[4790]: I1124 14:45:03.745852 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce20586f-aa65-426b-940e-22a8a12cd16d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ce20586f-aa65-426b-940e-22a8a12cd16d" (UID: "ce20586f-aa65-426b-940e-22a8a12cd16d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:45:03 crc kubenswrapper[4790]: I1124 14:45:03.800420 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce20586f-aa65-426b-940e-22a8a12cd16d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:03 crc kubenswrapper[4790]: I1124 14:45:03.800454 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bcz4j\" (UniqueName: \"kubernetes.io/projected/a6cd3c64-493b-4d69-8cec-7465da8136a2-kube-api-access-bcz4j\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:03 crc kubenswrapper[4790]: I1124 14:45:03.800494 4790 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a6cd3c64-493b-4d69-8cec-7465da8136a2-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:03 crc kubenswrapper[4790]: I1124 14:45:03.800505 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wzfcx\" (UniqueName: \"kubernetes.io/projected/ce20586f-aa65-426b-940e-22a8a12cd16d-kube-api-access-wzfcx\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:03 crc kubenswrapper[4790]: I1124 14:45:03.800515 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce20586f-aa65-426b-940e-22a8a12cd16d-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:03 crc kubenswrapper[4790]: I1124 14:45:03.800523 4790 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a6cd3c64-493b-4d69-8cec-7465da8136a2-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:03 crc kubenswrapper[4790]: I1124 14:45:03.800533 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce20586f-aa65-426b-940e-22a8a12cd16d-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:04 crc kubenswrapper[4790]: I1124 14:45:04.237202 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-xfp46" Nov 24 14:45:04 crc kubenswrapper[4790]: I1124 14:45:04.237206 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-xfp46" event={"ID":"a6cd3c64-493b-4d69-8cec-7465da8136a2","Type":"ContainerDied","Data":"ff522c418a1bd16616ba3c27700fcfea42dbcd643893c7b6557cc4c56f462d16"} Nov 24 14:45:04 crc kubenswrapper[4790]: I1124 14:45:04.237334 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff522c418a1bd16616ba3c27700fcfea42dbcd643893c7b6557cc4c56f462d16" Nov 24 14:45:04 crc kubenswrapper[4790]: I1124 14:45:04.239235 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-2w7vx" event={"ID":"ce20586f-aa65-426b-940e-22a8a12cd16d","Type":"ContainerDied","Data":"d1730213d95f737cb7ed36d01a25580cc3669301640165dc3022e1aa64a8cc92"} Nov 24 14:45:04 crc kubenswrapper[4790]: I1124 14:45:04.239274 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1730213d95f737cb7ed36d01a25580cc3669301640165dc3022e1aa64a8cc92" Nov 24 14:45:04 crc kubenswrapper[4790]: I1124 14:45:04.239401 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-2w7vx" Nov 24 14:45:04 crc kubenswrapper[4790]: I1124 14:45:04.333326 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399880-6cbwh"] Nov 24 14:45:04 crc kubenswrapper[4790]: I1124 14:45:04.340674 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399880-6cbwh"] Nov 24 14:45:04 crc kubenswrapper[4790]: I1124 14:45:04.347430 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 14:45:04 crc kubenswrapper[4790]: E1124 14:45:04.348139 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6cd3c64-493b-4d69-8cec-7465da8136a2" containerName="collect-profiles" Nov 24 14:45:04 crc kubenswrapper[4790]: I1124 14:45:04.348168 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6cd3c64-493b-4d69-8cec-7465da8136a2" containerName="collect-profiles" Nov 24 14:45:04 crc kubenswrapper[4790]: E1124 14:45:04.348233 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce20586f-aa65-426b-940e-22a8a12cd16d" containerName="nova-cell0-conductor-db-sync" Nov 24 14:45:04 crc kubenswrapper[4790]: I1124 14:45:04.348247 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce20586f-aa65-426b-940e-22a8a12cd16d" containerName="nova-cell0-conductor-db-sync" Nov 24 14:45:04 crc kubenswrapper[4790]: I1124 14:45:04.348552 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6cd3c64-493b-4d69-8cec-7465da8136a2" containerName="collect-profiles" Nov 24 14:45:04 crc kubenswrapper[4790]: I1124 14:45:04.348611 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce20586f-aa65-426b-940e-22a8a12cd16d" containerName="nova-cell0-conductor-db-sync" Nov 24 14:45:04 crc kubenswrapper[4790]: I1124 14:45:04.349615 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 14:45:04 crc kubenswrapper[4790]: I1124 14:45:04.366318 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 14:45:04 crc kubenswrapper[4790]: I1124 14:45:04.366833 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-2brfn" Nov 24 14:45:04 crc kubenswrapper[4790]: I1124 14:45:04.368517 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 14:45:04 crc kubenswrapper[4790]: I1124 14:45:04.417132 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4609d59-f11c-470c-a35e-966e0873e051-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e4609d59-f11c-470c-a35e-966e0873e051\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:45:04 crc kubenswrapper[4790]: I1124 14:45:04.417266 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2w9p\" (UniqueName: \"kubernetes.io/projected/e4609d59-f11c-470c-a35e-966e0873e051-kube-api-access-b2w9p\") pod \"nova-cell0-conductor-0\" (UID: \"e4609d59-f11c-470c-a35e-966e0873e051\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:45:04 crc kubenswrapper[4790]: I1124 14:45:04.417395 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4609d59-f11c-470c-a35e-966e0873e051-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e4609d59-f11c-470c-a35e-966e0873e051\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:45:04 crc kubenswrapper[4790]: I1124 14:45:04.519592 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4609d59-f11c-470c-a35e-966e0873e051-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e4609d59-f11c-470c-a35e-966e0873e051\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:45:04 crc kubenswrapper[4790]: I1124 14:45:04.519657 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2w9p\" (UniqueName: \"kubernetes.io/projected/e4609d59-f11c-470c-a35e-966e0873e051-kube-api-access-b2w9p\") pod \"nova-cell0-conductor-0\" (UID: \"e4609d59-f11c-470c-a35e-966e0873e051\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:45:04 crc kubenswrapper[4790]: I1124 14:45:04.519713 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4609d59-f11c-470c-a35e-966e0873e051-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e4609d59-f11c-470c-a35e-966e0873e051\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:45:04 crc kubenswrapper[4790]: I1124 14:45:04.526339 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4609d59-f11c-470c-a35e-966e0873e051-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e4609d59-f11c-470c-a35e-966e0873e051\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:45:04 crc kubenswrapper[4790]: I1124 14:45:04.526393 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4609d59-f11c-470c-a35e-966e0873e051-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e4609d59-f11c-470c-a35e-966e0873e051\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:45:04 crc kubenswrapper[4790]: I1124 14:45:04.542734 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2w9p\" (UniqueName: \"kubernetes.io/projected/e4609d59-f11c-470c-a35e-966e0873e051-kube-api-access-b2w9p\") pod \"nova-cell0-conductor-0\" (UID: \"e4609d59-f11c-470c-a35e-966e0873e051\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:45:04 crc kubenswrapper[4790]: I1124 14:45:04.677320 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 14:45:05 crc kubenswrapper[4790]: I1124 14:45:05.134265 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 14:45:05 crc kubenswrapper[4790]: I1124 14:45:05.263461 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e4609d59-f11c-470c-a35e-966e0873e051","Type":"ContainerStarted","Data":"6c6e17cc58137a095331ab1ae549c6bae550cdaaecb96bfbe68ad2caaf8fa982"} Nov 24 14:45:06 crc kubenswrapper[4790]: I1124 14:45:06.273818 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e4609d59-f11c-470c-a35e-966e0873e051","Type":"ContainerStarted","Data":"3bff9a9454a0a0cf4e1d009263e54b6f1892df0dec107b632524415fcd23984a"} Nov 24 14:45:06 crc kubenswrapper[4790]: I1124 14:45:06.274301 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 24 14:45:06 crc kubenswrapper[4790]: I1124 14:45:06.306468 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.306442448 podStartE2EDuration="2.306442448s" podCreationTimestamp="2025-11-24 14:45:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:45:06.294726282 +0000 UTC m=+5554.674619954" watchObservedRunningTime="2025-11-24 14:45:06.306442448 +0000 UTC m=+5554.686336120" Nov 24 14:45:06 crc kubenswrapper[4790]: I1124 14:45:06.328581 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b64e3a97-00c5-4c24-9d42-faec03c7709a" path="/var/lib/kubelet/pods/b64e3a97-00c5-4c24-9d42-faec03c7709a/volumes" Nov 24 14:45:07 crc kubenswrapper[4790]: I1124 14:45:07.174207 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-f2g9r"] Nov 24 14:45:07 crc kubenswrapper[4790]: I1124 14:45:07.176663 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f2g9r" Nov 24 14:45:07 crc kubenswrapper[4790]: I1124 14:45:07.182624 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f2g9r"] Nov 24 14:45:07 crc kubenswrapper[4790]: I1124 14:45:07.268087 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3630d953-1c95-4c3e-bb30-4601ec9534e5-catalog-content\") pod \"redhat-marketplace-f2g9r\" (UID: \"3630d953-1c95-4c3e-bb30-4601ec9534e5\") " pod="openshift-marketplace/redhat-marketplace-f2g9r" Nov 24 14:45:07 crc kubenswrapper[4790]: I1124 14:45:07.268314 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3630d953-1c95-4c3e-bb30-4601ec9534e5-utilities\") pod \"redhat-marketplace-f2g9r\" (UID: \"3630d953-1c95-4c3e-bb30-4601ec9534e5\") " pod="openshift-marketplace/redhat-marketplace-f2g9r" Nov 24 14:45:07 crc kubenswrapper[4790]: I1124 14:45:07.268374 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2zp2\" (UniqueName: \"kubernetes.io/projected/3630d953-1c95-4c3e-bb30-4601ec9534e5-kube-api-access-q2zp2\") pod \"redhat-marketplace-f2g9r\" (UID: \"3630d953-1c95-4c3e-bb30-4601ec9534e5\") " pod="openshift-marketplace/redhat-marketplace-f2g9r" Nov 24 14:45:07 crc kubenswrapper[4790]: I1124 14:45:07.370283 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3630d953-1c95-4c3e-bb30-4601ec9534e5-utilities\") pod \"redhat-marketplace-f2g9r\" (UID: \"3630d953-1c95-4c3e-bb30-4601ec9534e5\") " pod="openshift-marketplace/redhat-marketplace-f2g9r" Nov 24 14:45:07 crc kubenswrapper[4790]: I1124 14:45:07.370379 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2zp2\" (UniqueName: \"kubernetes.io/projected/3630d953-1c95-4c3e-bb30-4601ec9534e5-kube-api-access-q2zp2\") pod \"redhat-marketplace-f2g9r\" (UID: \"3630d953-1c95-4c3e-bb30-4601ec9534e5\") " pod="openshift-marketplace/redhat-marketplace-f2g9r" Nov 24 14:45:07 crc kubenswrapper[4790]: I1124 14:45:07.370462 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3630d953-1c95-4c3e-bb30-4601ec9534e5-catalog-content\") pod \"redhat-marketplace-f2g9r\" (UID: \"3630d953-1c95-4c3e-bb30-4601ec9534e5\") " pod="openshift-marketplace/redhat-marketplace-f2g9r" Nov 24 14:45:07 crc kubenswrapper[4790]: I1124 14:45:07.371020 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3630d953-1c95-4c3e-bb30-4601ec9534e5-utilities\") pod \"redhat-marketplace-f2g9r\" (UID: \"3630d953-1c95-4c3e-bb30-4601ec9534e5\") " pod="openshift-marketplace/redhat-marketplace-f2g9r" Nov 24 14:45:07 crc kubenswrapper[4790]: I1124 14:45:07.371157 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3630d953-1c95-4c3e-bb30-4601ec9534e5-catalog-content\") pod \"redhat-marketplace-f2g9r\" (UID: \"3630d953-1c95-4c3e-bb30-4601ec9534e5\") " pod="openshift-marketplace/redhat-marketplace-f2g9r" Nov 24 14:45:07 crc kubenswrapper[4790]: I1124 14:45:07.398089 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2zp2\" (UniqueName: \"kubernetes.io/projected/3630d953-1c95-4c3e-bb30-4601ec9534e5-kube-api-access-q2zp2\") pod \"redhat-marketplace-f2g9r\" (UID: \"3630d953-1c95-4c3e-bb30-4601ec9534e5\") " pod="openshift-marketplace/redhat-marketplace-f2g9r" Nov 24 14:45:07 crc kubenswrapper[4790]: I1124 14:45:07.510824 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f2g9r" Nov 24 14:45:07 crc kubenswrapper[4790]: I1124 14:45:07.983367 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f2g9r"] Nov 24 14:45:07 crc kubenswrapper[4790]: W1124 14:45:07.984454 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3630d953_1c95_4c3e_bb30_4601ec9534e5.slice/crio-64c042a60db43b4f1039b974b37328246a5350844598fe199f9929f5503a6164 WatchSource:0}: Error finding container 64c042a60db43b4f1039b974b37328246a5350844598fe199f9929f5503a6164: Status 404 returned error can't find the container with id 64c042a60db43b4f1039b974b37328246a5350844598fe199f9929f5503a6164 Nov 24 14:45:08 crc kubenswrapper[4790]: I1124 14:45:08.297331 4790 generic.go:334] "Generic (PLEG): container finished" podID="3630d953-1c95-4c3e-bb30-4601ec9534e5" containerID="23e409afac26a0676547505b2f448f0034d69346e9bdb5686470d27822e47312" exitCode=0 Nov 24 14:45:08 crc kubenswrapper[4790]: I1124 14:45:08.297461 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f2g9r" event={"ID":"3630d953-1c95-4c3e-bb30-4601ec9534e5","Type":"ContainerDied","Data":"23e409afac26a0676547505b2f448f0034d69346e9bdb5686470d27822e47312"} Nov 24 14:45:08 crc kubenswrapper[4790]: I1124 14:45:08.297792 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f2g9r" event={"ID":"3630d953-1c95-4c3e-bb30-4601ec9534e5","Type":"ContainerStarted","Data":"64c042a60db43b4f1039b974b37328246a5350844598fe199f9929f5503a6164"} Nov 24 14:45:09 crc kubenswrapper[4790]: I1124 14:45:09.307932 4790 generic.go:334] "Generic (PLEG): container finished" podID="3630d953-1c95-4c3e-bb30-4601ec9534e5" containerID="f3b5b3021a807e072d983b8b79c270bf3938e2992499036daeb98989e7ea1e7f" exitCode=0 Nov 24 14:45:09 crc kubenswrapper[4790]: I1124 14:45:09.308056 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f2g9r" event={"ID":"3630d953-1c95-4c3e-bb30-4601ec9534e5","Type":"ContainerDied","Data":"f3b5b3021a807e072d983b8b79c270bf3938e2992499036daeb98989e7ea1e7f"} Nov 24 14:45:09 crc kubenswrapper[4790]: I1124 14:45:09.563136 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-42ld4"] Nov 24 14:45:09 crc kubenswrapper[4790]: I1124 14:45:09.565725 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-42ld4" Nov 24 14:45:09 crc kubenswrapper[4790]: I1124 14:45:09.596146 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-42ld4"] Nov 24 14:45:09 crc kubenswrapper[4790]: I1124 14:45:09.711879 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58a4a77b-8070-4da1-b716-2e9796bab87a-catalog-content\") pod \"redhat-operators-42ld4\" (UID: \"58a4a77b-8070-4da1-b716-2e9796bab87a\") " pod="openshift-marketplace/redhat-operators-42ld4" Nov 24 14:45:09 crc kubenswrapper[4790]: I1124 14:45:09.712029 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jx5d2\" (UniqueName: \"kubernetes.io/projected/58a4a77b-8070-4da1-b716-2e9796bab87a-kube-api-access-jx5d2\") pod \"redhat-operators-42ld4\" (UID: \"58a4a77b-8070-4da1-b716-2e9796bab87a\") " pod="openshift-marketplace/redhat-operators-42ld4" Nov 24 14:45:09 crc kubenswrapper[4790]: I1124 14:45:09.712064 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58a4a77b-8070-4da1-b716-2e9796bab87a-utilities\") pod \"redhat-operators-42ld4\" (UID: \"58a4a77b-8070-4da1-b716-2e9796bab87a\") " pod="openshift-marketplace/redhat-operators-42ld4" Nov 24 14:45:09 crc kubenswrapper[4790]: I1124 14:45:09.814073 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jx5d2\" (UniqueName: \"kubernetes.io/projected/58a4a77b-8070-4da1-b716-2e9796bab87a-kube-api-access-jx5d2\") pod \"redhat-operators-42ld4\" (UID: \"58a4a77b-8070-4da1-b716-2e9796bab87a\") " pod="openshift-marketplace/redhat-operators-42ld4" Nov 24 14:45:09 crc kubenswrapper[4790]: I1124 14:45:09.814121 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58a4a77b-8070-4da1-b716-2e9796bab87a-utilities\") pod \"redhat-operators-42ld4\" (UID: \"58a4a77b-8070-4da1-b716-2e9796bab87a\") " pod="openshift-marketplace/redhat-operators-42ld4" Nov 24 14:45:09 crc kubenswrapper[4790]: I1124 14:45:09.814201 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58a4a77b-8070-4da1-b716-2e9796bab87a-catalog-content\") pod \"redhat-operators-42ld4\" (UID: \"58a4a77b-8070-4da1-b716-2e9796bab87a\") " pod="openshift-marketplace/redhat-operators-42ld4" Nov 24 14:45:09 crc kubenswrapper[4790]: I1124 14:45:09.814693 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58a4a77b-8070-4da1-b716-2e9796bab87a-catalog-content\") pod \"redhat-operators-42ld4\" (UID: \"58a4a77b-8070-4da1-b716-2e9796bab87a\") " pod="openshift-marketplace/redhat-operators-42ld4" Nov 24 14:45:09 crc kubenswrapper[4790]: I1124 14:45:09.814999 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58a4a77b-8070-4da1-b716-2e9796bab87a-utilities\") pod \"redhat-operators-42ld4\" (UID: \"58a4a77b-8070-4da1-b716-2e9796bab87a\") " pod="openshift-marketplace/redhat-operators-42ld4" Nov 24 14:45:09 crc kubenswrapper[4790]: I1124 14:45:09.832236 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jx5d2\" (UniqueName: \"kubernetes.io/projected/58a4a77b-8070-4da1-b716-2e9796bab87a-kube-api-access-jx5d2\") pod \"redhat-operators-42ld4\" (UID: \"58a4a77b-8070-4da1-b716-2e9796bab87a\") " pod="openshift-marketplace/redhat-operators-42ld4" Nov 24 14:45:09 crc kubenswrapper[4790]: I1124 14:45:09.899107 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-42ld4" Nov 24 14:45:10 crc kubenswrapper[4790]: I1124 14:45:10.467080 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-42ld4"] Nov 24 14:45:11 crc kubenswrapper[4790]: I1124 14:45:11.324944 4790 generic.go:334] "Generic (PLEG): container finished" podID="58a4a77b-8070-4da1-b716-2e9796bab87a" containerID="3366f4e671bcb53b61230ffba61e286b643d010934001965ccfb6a59d8415c8d" exitCode=0 Nov 24 14:45:11 crc kubenswrapper[4790]: I1124 14:45:11.325007 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-42ld4" event={"ID":"58a4a77b-8070-4da1-b716-2e9796bab87a","Type":"ContainerDied","Data":"3366f4e671bcb53b61230ffba61e286b643d010934001965ccfb6a59d8415c8d"} Nov 24 14:45:11 crc kubenswrapper[4790]: I1124 14:45:11.325034 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-42ld4" event={"ID":"58a4a77b-8070-4da1-b716-2e9796bab87a","Type":"ContainerStarted","Data":"786a9011ea182e078a38a82760eeeaf841d73924ab13d9b4cf7ae88aed4ae13f"} Nov 24 14:45:11 crc kubenswrapper[4790]: I1124 14:45:11.327129 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f2g9r" event={"ID":"3630d953-1c95-4c3e-bb30-4601ec9534e5","Type":"ContainerStarted","Data":"dd1d2d4b80a3a85f8a97953e14147f0044617694834c503530979b3739038d0f"} Nov 24 14:45:11 crc kubenswrapper[4790]: I1124 14:45:11.392985 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-f2g9r" podStartSLOduration=2.566429318 podStartE2EDuration="4.392961652s" podCreationTimestamp="2025-11-24 14:45:07 +0000 UTC" firstStartedPulling="2025-11-24 14:45:08.299732129 +0000 UTC m=+5556.679625801" lastFinishedPulling="2025-11-24 14:45:10.126264473 +0000 UTC m=+5558.506158135" observedRunningTime="2025-11-24 14:45:11.381618917 +0000 UTC m=+5559.761512589" watchObservedRunningTime="2025-11-24 14:45:11.392961652 +0000 UTC m=+5559.772855314" Nov 24 14:45:13 crc kubenswrapper[4790]: I1124 14:45:13.314698 4790 scope.go:117] "RemoveContainer" containerID="92c086e44ccb28a62777d612e499209f7483422ca70b460ce6ec73c416e8e8b8" Nov 24 14:45:13 crc kubenswrapper[4790]: E1124 14:45:13.315829 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:45:13 crc kubenswrapper[4790]: I1124 14:45:13.364547 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-42ld4" event={"ID":"58a4a77b-8070-4da1-b716-2e9796bab87a","Type":"ContainerStarted","Data":"c9cbeb2276a393a2b9a8605885a47db2d9629fa10b83c6598627848a82f691b9"} Nov 24 14:45:14 crc kubenswrapper[4790]: I1124 14:45:14.379056 4790 generic.go:334] "Generic (PLEG): container finished" podID="58a4a77b-8070-4da1-b716-2e9796bab87a" containerID="c9cbeb2276a393a2b9a8605885a47db2d9629fa10b83c6598627848a82f691b9" exitCode=0 Nov 24 14:45:14 crc kubenswrapper[4790]: I1124 14:45:14.379127 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-42ld4" event={"ID":"58a4a77b-8070-4da1-b716-2e9796bab87a","Type":"ContainerDied","Data":"c9cbeb2276a393a2b9a8605885a47db2d9629fa10b83c6598627848a82f691b9"} Nov 24 14:45:14 crc kubenswrapper[4790]: I1124 14:45:14.717708 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.229972 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-w7f6n"] Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.231573 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-w7f6n" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.237720 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.238156 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.248063 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-w7f6n"] Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.322274 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8596f2b-213d-4a3f-940a-9cc7421e85be-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-w7f6n\" (UID: \"b8596f2b-213d-4a3f-940a-9cc7421e85be\") " pod="openstack/nova-cell0-cell-mapping-w7f6n" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.322351 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5xhg\" (UniqueName: \"kubernetes.io/projected/b8596f2b-213d-4a3f-940a-9cc7421e85be-kube-api-access-z5xhg\") pod \"nova-cell0-cell-mapping-w7f6n\" (UID: \"b8596f2b-213d-4a3f-940a-9cc7421e85be\") " pod="openstack/nova-cell0-cell-mapping-w7f6n" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.322379 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8596f2b-213d-4a3f-940a-9cc7421e85be-scripts\") pod \"nova-cell0-cell-mapping-w7f6n\" (UID: \"b8596f2b-213d-4a3f-940a-9cc7421e85be\") " pod="openstack/nova-cell0-cell-mapping-w7f6n" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.322472 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8596f2b-213d-4a3f-940a-9cc7421e85be-config-data\") pod \"nova-cell0-cell-mapping-w7f6n\" (UID: \"b8596f2b-213d-4a3f-940a-9cc7421e85be\") " pod="openstack/nova-cell0-cell-mapping-w7f6n" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.347679 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.350350 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.356093 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.370516 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.424579 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0144dbfa-e743-47eb-b9ab-8e3610fc3130-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0144dbfa-e743-47eb-b9ab-8e3610fc3130\") " pod="openstack/nova-scheduler-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.424718 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8596f2b-213d-4a3f-940a-9cc7421e85be-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-w7f6n\" (UID: \"b8596f2b-213d-4a3f-940a-9cc7421e85be\") " pod="openstack/nova-cell0-cell-mapping-w7f6n" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.424778 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5xhg\" (UniqueName: \"kubernetes.io/projected/b8596f2b-213d-4a3f-940a-9cc7421e85be-kube-api-access-z5xhg\") pod \"nova-cell0-cell-mapping-w7f6n\" (UID: \"b8596f2b-213d-4a3f-940a-9cc7421e85be\") " pod="openstack/nova-cell0-cell-mapping-w7f6n" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.424798 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8596f2b-213d-4a3f-940a-9cc7421e85be-scripts\") pod \"nova-cell0-cell-mapping-w7f6n\" (UID: \"b8596f2b-213d-4a3f-940a-9cc7421e85be\") " pod="openstack/nova-cell0-cell-mapping-w7f6n" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.424823 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0144dbfa-e743-47eb-b9ab-8e3610fc3130-config-data\") pod \"nova-scheduler-0\" (UID: \"0144dbfa-e743-47eb-b9ab-8e3610fc3130\") " pod="openstack/nova-scheduler-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.424863 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msfzt\" (UniqueName: \"kubernetes.io/projected/0144dbfa-e743-47eb-b9ab-8e3610fc3130-kube-api-access-msfzt\") pod \"nova-scheduler-0\" (UID: \"0144dbfa-e743-47eb-b9ab-8e3610fc3130\") " pod="openstack/nova-scheduler-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.424918 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8596f2b-213d-4a3f-940a-9cc7421e85be-config-data\") pod \"nova-cell0-cell-mapping-w7f6n\" (UID: \"b8596f2b-213d-4a3f-940a-9cc7421e85be\") " pod="openstack/nova-cell0-cell-mapping-w7f6n" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.434790 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8596f2b-213d-4a3f-940a-9cc7421e85be-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-w7f6n\" (UID: \"b8596f2b-213d-4a3f-940a-9cc7421e85be\") " pod="openstack/nova-cell0-cell-mapping-w7f6n" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.449467 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8596f2b-213d-4a3f-940a-9cc7421e85be-config-data\") pod \"nova-cell0-cell-mapping-w7f6n\" (UID: \"b8596f2b-213d-4a3f-940a-9cc7421e85be\") " pod="openstack/nova-cell0-cell-mapping-w7f6n" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.460171 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8596f2b-213d-4a3f-940a-9cc7421e85be-scripts\") pod \"nova-cell0-cell-mapping-w7f6n\" (UID: \"b8596f2b-213d-4a3f-940a-9cc7421e85be\") " pod="openstack/nova-cell0-cell-mapping-w7f6n" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.485620 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.588633 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0144dbfa-e743-47eb-b9ab-8e3610fc3130-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0144dbfa-e743-47eb-b9ab-8e3610fc3130\") " pod="openstack/nova-scheduler-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.589014 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0144dbfa-e743-47eb-b9ab-8e3610fc3130-config-data\") pod \"nova-scheduler-0\" (UID: \"0144dbfa-e743-47eb-b9ab-8e3610fc3130\") " pod="openstack/nova-scheduler-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.589058 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msfzt\" (UniqueName: \"kubernetes.io/projected/0144dbfa-e743-47eb-b9ab-8e3610fc3130-kube-api-access-msfzt\") pod \"nova-scheduler-0\" (UID: \"0144dbfa-e743-47eb-b9ab-8e3610fc3130\") " pod="openstack/nova-scheduler-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.589720 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5xhg\" (UniqueName: \"kubernetes.io/projected/b8596f2b-213d-4a3f-940a-9cc7421e85be-kube-api-access-z5xhg\") pod \"nova-cell0-cell-mapping-w7f6n\" (UID: \"b8596f2b-213d-4a3f-940a-9cc7421e85be\") " pod="openstack/nova-cell0-cell-mapping-w7f6n" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.593238 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.593285 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.594823 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.604103 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.608701 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.615370 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.615981 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.616833 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0144dbfa-e743-47eb-b9ab-8e3610fc3130-config-data\") pod \"nova-scheduler-0\" (UID: \"0144dbfa-e743-47eb-b9ab-8e3610fc3130\") " pod="openstack/nova-scheduler-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.620400 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-w7f6n" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.627702 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0144dbfa-e743-47eb-b9ab-8e3610fc3130-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0144dbfa-e743-47eb-b9ab-8e3610fc3130\") " pod="openstack/nova-scheduler-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.648851 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msfzt\" (UniqueName: \"kubernetes.io/projected/0144dbfa-e743-47eb-b9ab-8e3610fc3130-kube-api-access-msfzt\") pod \"nova-scheduler-0\" (UID: \"0144dbfa-e743-47eb-b9ab-8e3610fc3130\") " pod="openstack/nova-scheduler-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.690478 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.691924 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/691daa40-76bd-4926-928b-2e4e25f7a50c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"691daa40-76bd-4926-928b-2e4e25f7a50c\") " pod="openstack/nova-metadata-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.691973 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3811846-d6c0-47d7-8d40-93bae3eb1e49-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c3811846-d6c0-47d7-8d40-93bae3eb1e49\") " pod="openstack/nova-api-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.691992 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3811846-d6c0-47d7-8d40-93bae3eb1e49-config-data\") pod \"nova-api-0\" (UID: \"c3811846-d6c0-47d7-8d40-93bae3eb1e49\") " pod="openstack/nova-api-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.692010 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/691daa40-76bd-4926-928b-2e4e25f7a50c-config-data\") pod \"nova-metadata-0\" (UID: \"691daa40-76bd-4926-928b-2e4e25f7a50c\") " pod="openstack/nova-metadata-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.692039 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3811846-d6c0-47d7-8d40-93bae3eb1e49-logs\") pod \"nova-api-0\" (UID: \"c3811846-d6c0-47d7-8d40-93bae3eb1e49\") " pod="openstack/nova-api-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.692148 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/691daa40-76bd-4926-928b-2e4e25f7a50c-logs\") pod \"nova-metadata-0\" (UID: \"691daa40-76bd-4926-928b-2e4e25f7a50c\") " pod="openstack/nova-metadata-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.692166 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxvrm\" (UniqueName: \"kubernetes.io/projected/c3811846-d6c0-47d7-8d40-93bae3eb1e49-kube-api-access-jxvrm\") pod \"nova-api-0\" (UID: \"c3811846-d6c0-47d7-8d40-93bae3eb1e49\") " pod="openstack/nova-api-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.692191 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7h5h\" (UniqueName: \"kubernetes.io/projected/691daa40-76bd-4926-928b-2e4e25f7a50c-kube-api-access-b7h5h\") pod \"nova-metadata-0\" (UID: \"691daa40-76bd-4926-928b-2e4e25f7a50c\") " pod="openstack/nova-metadata-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.697117 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.698399 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.709530 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.765372 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.802316 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/691daa40-76bd-4926-928b-2e4e25f7a50c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"691daa40-76bd-4926-928b-2e4e25f7a50c\") " pod="openstack/nova-metadata-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.802632 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbsjx\" (UniqueName: \"kubernetes.io/projected/70d99ce8-fc77-4c8a-84ad-092c20c65e1d-kube-api-access-mbsjx\") pod \"nova-cell1-novncproxy-0\" (UID: \"70d99ce8-fc77-4c8a-84ad-092c20c65e1d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.802833 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3811846-d6c0-47d7-8d40-93bae3eb1e49-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c3811846-d6c0-47d7-8d40-93bae3eb1e49\") " pod="openstack/nova-api-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.802965 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3811846-d6c0-47d7-8d40-93bae3eb1e49-config-data\") pod \"nova-api-0\" (UID: \"c3811846-d6c0-47d7-8d40-93bae3eb1e49\") " pod="openstack/nova-api-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.803076 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/691daa40-76bd-4926-928b-2e4e25f7a50c-config-data\") pod \"nova-metadata-0\" (UID: \"691daa40-76bd-4926-928b-2e4e25f7a50c\") " pod="openstack/nova-metadata-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.803201 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3811846-d6c0-47d7-8d40-93bae3eb1e49-logs\") pod \"nova-api-0\" (UID: \"c3811846-d6c0-47d7-8d40-93bae3eb1e49\") " pod="openstack/nova-api-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.803393 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70d99ce8-fc77-4c8a-84ad-092c20c65e1d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"70d99ce8-fc77-4c8a-84ad-092c20c65e1d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.803576 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/691daa40-76bd-4926-928b-2e4e25f7a50c-logs\") pod \"nova-metadata-0\" (UID: \"691daa40-76bd-4926-928b-2e4e25f7a50c\") " pod="openstack/nova-metadata-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.803695 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxvrm\" (UniqueName: \"kubernetes.io/projected/c3811846-d6c0-47d7-8d40-93bae3eb1e49-kube-api-access-jxvrm\") pod \"nova-api-0\" (UID: \"c3811846-d6c0-47d7-8d40-93bae3eb1e49\") " pod="openstack/nova-api-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.803810 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7h5h\" (UniqueName: \"kubernetes.io/projected/691daa40-76bd-4926-928b-2e4e25f7a50c-kube-api-access-b7h5h\") pod \"nova-metadata-0\" (UID: \"691daa40-76bd-4926-928b-2e4e25f7a50c\") " pod="openstack/nova-metadata-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.804050 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70d99ce8-fc77-4c8a-84ad-092c20c65e1d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"70d99ce8-fc77-4c8a-84ad-092c20c65e1d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.805231 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3811846-d6c0-47d7-8d40-93bae3eb1e49-logs\") pod \"nova-api-0\" (UID: \"c3811846-d6c0-47d7-8d40-93bae3eb1e49\") " pod="openstack/nova-api-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.807942 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3811846-d6c0-47d7-8d40-93bae3eb1e49-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c3811846-d6c0-47d7-8d40-93bae3eb1e49\") " pod="openstack/nova-api-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.810305 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3811846-d6c0-47d7-8d40-93bae3eb1e49-config-data\") pod \"nova-api-0\" (UID: \"c3811846-d6c0-47d7-8d40-93bae3eb1e49\") " pod="openstack/nova-api-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.811182 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/691daa40-76bd-4926-928b-2e4e25f7a50c-logs\") pod \"nova-metadata-0\" (UID: \"691daa40-76bd-4926-928b-2e4e25f7a50c\") " pod="openstack/nova-metadata-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.814681 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/691daa40-76bd-4926-928b-2e4e25f7a50c-config-data\") pod \"nova-metadata-0\" (UID: \"691daa40-76bd-4926-928b-2e4e25f7a50c\") " pod="openstack/nova-metadata-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.814984 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/691daa40-76bd-4926-928b-2e4e25f7a50c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"691daa40-76bd-4926-928b-2e4e25f7a50c\") " pod="openstack/nova-metadata-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.827948 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bbb866757-8frpq"] Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.829422 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbb866757-8frpq" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.834227 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7h5h\" (UniqueName: \"kubernetes.io/projected/691daa40-76bd-4926-928b-2e4e25f7a50c-kube-api-access-b7h5h\") pod \"nova-metadata-0\" (UID: \"691daa40-76bd-4926-928b-2e4e25f7a50c\") " pod="openstack/nova-metadata-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.845222 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxvrm\" (UniqueName: \"kubernetes.io/projected/c3811846-d6c0-47d7-8d40-93bae3eb1e49-kube-api-access-jxvrm\") pod \"nova-api-0\" (UID: \"c3811846-d6c0-47d7-8d40-93bae3eb1e49\") " pod="openstack/nova-api-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.846521 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bbb866757-8frpq"] Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.907210 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbsjx\" (UniqueName: \"kubernetes.io/projected/70d99ce8-fc77-4c8a-84ad-092c20c65e1d-kube-api-access-mbsjx\") pod \"nova-cell1-novncproxy-0\" (UID: \"70d99ce8-fc77-4c8a-84ad-092c20c65e1d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.907315 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70d99ce8-fc77-4c8a-84ad-092c20c65e1d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"70d99ce8-fc77-4c8a-84ad-092c20c65e1d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.907376 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70d99ce8-fc77-4c8a-84ad-092c20c65e1d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"70d99ce8-fc77-4c8a-84ad-092c20c65e1d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.912069 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70d99ce8-fc77-4c8a-84ad-092c20c65e1d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"70d99ce8-fc77-4c8a-84ad-092c20c65e1d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.916014 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70d99ce8-fc77-4c8a-84ad-092c20c65e1d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"70d99ce8-fc77-4c8a-84ad-092c20c65e1d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:45:15 crc kubenswrapper[4790]: I1124 14:45:15.926136 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbsjx\" (UniqueName: \"kubernetes.io/projected/70d99ce8-fc77-4c8a-84ad-092c20c65e1d-kube-api-access-mbsjx\") pod \"nova-cell1-novncproxy-0\" (UID: \"70d99ce8-fc77-4c8a-84ad-092c20c65e1d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.008923 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11caa523-04fb-4f1d-a6e5-b9323fc00b3f-ovsdbserver-nb\") pod \"dnsmasq-dns-bbb866757-8frpq\" (UID: \"11caa523-04fb-4f1d-a6e5-b9323fc00b3f\") " pod="openstack/dnsmasq-dns-bbb866757-8frpq" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.009474 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11caa523-04fb-4f1d-a6e5-b9323fc00b3f-dns-svc\") pod \"dnsmasq-dns-bbb866757-8frpq\" (UID: \"11caa523-04fb-4f1d-a6e5-b9323fc00b3f\") " pod="openstack/dnsmasq-dns-bbb866757-8frpq" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.009635 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11caa523-04fb-4f1d-a6e5-b9323fc00b3f-ovsdbserver-sb\") pod \"dnsmasq-dns-bbb866757-8frpq\" (UID: \"11caa523-04fb-4f1d-a6e5-b9323fc00b3f\") " pod="openstack/dnsmasq-dns-bbb866757-8frpq" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.009808 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjzsk\" (UniqueName: \"kubernetes.io/projected/11caa523-04fb-4f1d-a6e5-b9323fc00b3f-kube-api-access-vjzsk\") pod \"dnsmasq-dns-bbb866757-8frpq\" (UID: \"11caa523-04fb-4f1d-a6e5-b9323fc00b3f\") " pod="openstack/dnsmasq-dns-bbb866757-8frpq" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.009933 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11caa523-04fb-4f1d-a6e5-b9323fc00b3f-config\") pod \"dnsmasq-dns-bbb866757-8frpq\" (UID: \"11caa523-04fb-4f1d-a6e5-b9323fc00b3f\") " pod="openstack/dnsmasq-dns-bbb866757-8frpq" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.013746 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.067337 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.104620 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.113745 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11caa523-04fb-4f1d-a6e5-b9323fc00b3f-ovsdbserver-nb\") pod \"dnsmasq-dns-bbb866757-8frpq\" (UID: \"11caa523-04fb-4f1d-a6e5-b9323fc00b3f\") " pod="openstack/dnsmasq-dns-bbb866757-8frpq" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.113859 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11caa523-04fb-4f1d-a6e5-b9323fc00b3f-dns-svc\") pod \"dnsmasq-dns-bbb866757-8frpq\" (UID: \"11caa523-04fb-4f1d-a6e5-b9323fc00b3f\") " pod="openstack/dnsmasq-dns-bbb866757-8frpq" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.113977 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11caa523-04fb-4f1d-a6e5-b9323fc00b3f-ovsdbserver-sb\") pod \"dnsmasq-dns-bbb866757-8frpq\" (UID: \"11caa523-04fb-4f1d-a6e5-b9323fc00b3f\") " pod="openstack/dnsmasq-dns-bbb866757-8frpq" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.114027 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjzsk\" (UniqueName: \"kubernetes.io/projected/11caa523-04fb-4f1d-a6e5-b9323fc00b3f-kube-api-access-vjzsk\") pod \"dnsmasq-dns-bbb866757-8frpq\" (UID: \"11caa523-04fb-4f1d-a6e5-b9323fc00b3f\") " pod="openstack/dnsmasq-dns-bbb866757-8frpq" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.114112 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11caa523-04fb-4f1d-a6e5-b9323fc00b3f-config\") pod \"dnsmasq-dns-bbb866757-8frpq\" (UID: \"11caa523-04fb-4f1d-a6e5-b9323fc00b3f\") " pod="openstack/dnsmasq-dns-bbb866757-8frpq" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.115748 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11caa523-04fb-4f1d-a6e5-b9323fc00b3f-config\") pod \"dnsmasq-dns-bbb866757-8frpq\" (UID: \"11caa523-04fb-4f1d-a6e5-b9323fc00b3f\") " pod="openstack/dnsmasq-dns-bbb866757-8frpq" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.116099 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11caa523-04fb-4f1d-a6e5-b9323fc00b3f-dns-svc\") pod \"dnsmasq-dns-bbb866757-8frpq\" (UID: \"11caa523-04fb-4f1d-a6e5-b9323fc00b3f\") " pod="openstack/dnsmasq-dns-bbb866757-8frpq" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.116577 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11caa523-04fb-4f1d-a6e5-b9323fc00b3f-ovsdbserver-sb\") pod \"dnsmasq-dns-bbb866757-8frpq\" (UID: \"11caa523-04fb-4f1d-a6e5-b9323fc00b3f\") " pod="openstack/dnsmasq-dns-bbb866757-8frpq" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.116669 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11caa523-04fb-4f1d-a6e5-b9323fc00b3f-ovsdbserver-nb\") pod \"dnsmasq-dns-bbb866757-8frpq\" (UID: \"11caa523-04fb-4f1d-a6e5-b9323fc00b3f\") " pod="openstack/dnsmasq-dns-bbb866757-8frpq" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.140006 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjzsk\" (UniqueName: \"kubernetes.io/projected/11caa523-04fb-4f1d-a6e5-b9323fc00b3f-kube-api-access-vjzsk\") pod \"dnsmasq-dns-bbb866757-8frpq\" (UID: \"11caa523-04fb-4f1d-a6e5-b9323fc00b3f\") " pod="openstack/dnsmasq-dns-bbb866757-8frpq" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.179094 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbb866757-8frpq" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.415365 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-42ld4" event={"ID":"58a4a77b-8070-4da1-b716-2e9796bab87a","Type":"ContainerStarted","Data":"bb87531f06dd0e96e23905f5d34428dfa751ec6106733c2c510dce7a6cab644e"} Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.435994 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zv4x8"] Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.437561 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-zv4x8" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.441814 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.447337 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.448816 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zv4x8"] Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.449842 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-42ld4" podStartSLOduration=3.627744246 podStartE2EDuration="7.449823434s" podCreationTimestamp="2025-11-24 14:45:09 +0000 UTC" firstStartedPulling="2025-11-24 14:45:11.326481114 +0000 UTC m=+5559.706374776" lastFinishedPulling="2025-11-24 14:45:15.148560292 +0000 UTC m=+5563.528453964" observedRunningTime="2025-11-24 14:45:16.443319207 +0000 UTC m=+5564.823212869" watchObservedRunningTime="2025-11-24 14:45:16.449823434 +0000 UTC m=+5564.829717086" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.625265 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txw5p\" (UniqueName: \"kubernetes.io/projected/873b8408-4a0a-42dd-bb32-5e3cba17a985-kube-api-access-txw5p\") pod \"nova-cell1-conductor-db-sync-zv4x8\" (UID: \"873b8408-4a0a-42dd-bb32-5e3cba17a985\") " pod="openstack/nova-cell1-conductor-db-sync-zv4x8" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.625676 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/873b8408-4a0a-42dd-bb32-5e3cba17a985-scripts\") pod \"nova-cell1-conductor-db-sync-zv4x8\" (UID: \"873b8408-4a0a-42dd-bb32-5e3cba17a985\") " pod="openstack/nova-cell1-conductor-db-sync-zv4x8" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.625776 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/873b8408-4a0a-42dd-bb32-5e3cba17a985-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-zv4x8\" (UID: \"873b8408-4a0a-42dd-bb32-5e3cba17a985\") " pod="openstack/nova-cell1-conductor-db-sync-zv4x8" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.625992 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/873b8408-4a0a-42dd-bb32-5e3cba17a985-config-data\") pod \"nova-cell1-conductor-db-sync-zv4x8\" (UID: \"873b8408-4a0a-42dd-bb32-5e3cba17a985\") " pod="openstack/nova-cell1-conductor-db-sync-zv4x8" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.728295 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txw5p\" (UniqueName: \"kubernetes.io/projected/873b8408-4a0a-42dd-bb32-5e3cba17a985-kube-api-access-txw5p\") pod \"nova-cell1-conductor-db-sync-zv4x8\" (UID: \"873b8408-4a0a-42dd-bb32-5e3cba17a985\") " pod="openstack/nova-cell1-conductor-db-sync-zv4x8" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.728390 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/873b8408-4a0a-42dd-bb32-5e3cba17a985-scripts\") pod \"nova-cell1-conductor-db-sync-zv4x8\" (UID: \"873b8408-4a0a-42dd-bb32-5e3cba17a985\") " pod="openstack/nova-cell1-conductor-db-sync-zv4x8" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.728422 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/873b8408-4a0a-42dd-bb32-5e3cba17a985-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-zv4x8\" (UID: \"873b8408-4a0a-42dd-bb32-5e3cba17a985\") " pod="openstack/nova-cell1-conductor-db-sync-zv4x8" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.728452 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/873b8408-4a0a-42dd-bb32-5e3cba17a985-config-data\") pod \"nova-cell1-conductor-db-sync-zv4x8\" (UID: \"873b8408-4a0a-42dd-bb32-5e3cba17a985\") " pod="openstack/nova-cell1-conductor-db-sync-zv4x8" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.743364 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/873b8408-4a0a-42dd-bb32-5e3cba17a985-scripts\") pod \"nova-cell1-conductor-db-sync-zv4x8\" (UID: \"873b8408-4a0a-42dd-bb32-5e3cba17a985\") " pod="openstack/nova-cell1-conductor-db-sync-zv4x8" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.746183 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/873b8408-4a0a-42dd-bb32-5e3cba17a985-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-zv4x8\" (UID: \"873b8408-4a0a-42dd-bb32-5e3cba17a985\") " pod="openstack/nova-cell1-conductor-db-sync-zv4x8" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.746754 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/873b8408-4a0a-42dd-bb32-5e3cba17a985-config-data\") pod \"nova-cell1-conductor-db-sync-zv4x8\" (UID: \"873b8408-4a0a-42dd-bb32-5e3cba17a985\") " pod="openstack/nova-cell1-conductor-db-sync-zv4x8" Nov 24 14:45:16 crc kubenswrapper[4790]: I1124 14:45:16.759925 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txw5p\" (UniqueName: \"kubernetes.io/projected/873b8408-4a0a-42dd-bb32-5e3cba17a985-kube-api-access-txw5p\") pod \"nova-cell1-conductor-db-sync-zv4x8\" (UID: \"873b8408-4a0a-42dd-bb32-5e3cba17a985\") " pod="openstack/nova-cell1-conductor-db-sync-zv4x8" Nov 24 14:45:17 crc kubenswrapper[4790]: W1124 14:45:17.032548 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11caa523_04fb_4f1d_a6e5_b9323fc00b3f.slice/crio-a5c06777417489fcfce2dfb1348897c317b73255126f723560249b448056f2b5 WatchSource:0}: Error finding container a5c06777417489fcfce2dfb1348897c317b73255126f723560249b448056f2b5: Status 404 returned error can't find the container with id a5c06777417489fcfce2dfb1348897c317b73255126f723560249b448056f2b5 Nov 24 14:45:17 crc kubenswrapper[4790]: I1124 14:45:17.039284 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bbb866757-8frpq"] Nov 24 14:45:17 crc kubenswrapper[4790]: W1124 14:45:17.040977 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod691daa40_76bd_4926_928b_2e4e25f7a50c.slice/crio-f634bb2ba321bf14dce1f79ae199de8a69a85bf7288babc51e51f7f0ba99336f WatchSource:0}: Error finding container f634bb2ba321bf14dce1f79ae199de8a69a85bf7288babc51e51f7f0ba99336f: Status 404 returned error can't find the container with id f634bb2ba321bf14dce1f79ae199de8a69a85bf7288babc51e51f7f0ba99336f Nov 24 14:45:17 crc kubenswrapper[4790]: I1124 14:45:17.060652 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-zv4x8" Nov 24 14:45:17 crc kubenswrapper[4790]: I1124 14:45:17.066100 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:45:17 crc kubenswrapper[4790]: W1124 14:45:17.073027 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb8596f2b_213d_4a3f_940a_9cc7421e85be.slice/crio-c3d995294475670ef0c303a1dd27b0c5f3542d19aff69c267200da06928f8ebf WatchSource:0}: Error finding container c3d995294475670ef0c303a1dd27b0c5f3542d19aff69c267200da06928f8ebf: Status 404 returned error can't find the container with id c3d995294475670ef0c303a1dd27b0c5f3542d19aff69c267200da06928f8ebf Nov 24 14:45:17 crc kubenswrapper[4790]: W1124 14:45:17.076959 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70d99ce8_fc77_4c8a_84ad_092c20c65e1d.slice/crio-3fee0eb97a8176dd0c0e500504c034593a3a627d101974fd2a6c3b2c670462cc WatchSource:0}: Error finding container 3fee0eb97a8176dd0c0e500504c034593a3a627d101974fd2a6c3b2c670462cc: Status 404 returned error can't find the container with id 3fee0eb97a8176dd0c0e500504c034593a3a627d101974fd2a6c3b2c670462cc Nov 24 14:45:17 crc kubenswrapper[4790]: I1124 14:45:17.079456 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:45:17 crc kubenswrapper[4790]: I1124 14:45:17.099351 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 14:45:17 crc kubenswrapper[4790]: I1124 14:45:17.110933 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-w7f6n"] Nov 24 14:45:17 crc kubenswrapper[4790]: I1124 14:45:17.121363 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:45:17 crc kubenswrapper[4790]: I1124 14:45:17.428033 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbb866757-8frpq" event={"ID":"11caa523-04fb-4f1d-a6e5-b9323fc00b3f","Type":"ContainerStarted","Data":"a5c06777417489fcfce2dfb1348897c317b73255126f723560249b448056f2b5"} Nov 24 14:45:17 crc kubenswrapper[4790]: I1124 14:45:17.430006 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0144dbfa-e743-47eb-b9ab-8e3610fc3130","Type":"ContainerStarted","Data":"8eabfa899ecc915f6a24fe03691b3cb2e6bb7fae3a0e1456438934a7131aed52"} Nov 24 14:45:17 crc kubenswrapper[4790]: I1124 14:45:17.431487 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"691daa40-76bd-4926-928b-2e4e25f7a50c","Type":"ContainerStarted","Data":"f634bb2ba321bf14dce1f79ae199de8a69a85bf7288babc51e51f7f0ba99336f"} Nov 24 14:45:17 crc kubenswrapper[4790]: I1124 14:45:17.432497 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"70d99ce8-fc77-4c8a-84ad-092c20c65e1d","Type":"ContainerStarted","Data":"3fee0eb97a8176dd0c0e500504c034593a3a627d101974fd2a6c3b2c670462cc"} Nov 24 14:45:17 crc kubenswrapper[4790]: I1124 14:45:17.433869 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c3811846-d6c0-47d7-8d40-93bae3eb1e49","Type":"ContainerStarted","Data":"3e80e07d5b39eacf4ecc5c65880beb8a0b12349edb19f7ff9daa636d78ed5c6f"} Nov 24 14:45:17 crc kubenswrapper[4790]: I1124 14:45:17.433941 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c3811846-d6c0-47d7-8d40-93bae3eb1e49","Type":"ContainerStarted","Data":"c6d0bbf11a146f07d50980786fb31d3463c27b10c0888c9d36c453a97585cd24"} Nov 24 14:45:17 crc kubenswrapper[4790]: I1124 14:45:17.434962 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-w7f6n" event={"ID":"b8596f2b-213d-4a3f-940a-9cc7421e85be","Type":"ContainerStarted","Data":"c3d995294475670ef0c303a1dd27b0c5f3542d19aff69c267200da06928f8ebf"} Nov 24 14:45:17 crc kubenswrapper[4790]: I1124 14:45:17.512022 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-f2g9r" Nov 24 14:45:17 crc kubenswrapper[4790]: I1124 14:45:17.512349 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-f2g9r" Nov 24 14:45:17 crc kubenswrapper[4790]: I1124 14:45:17.540563 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zv4x8"] Nov 24 14:45:17 crc kubenswrapper[4790]: W1124 14:45:17.547696 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod873b8408_4a0a_42dd_bb32_5e3cba17a985.slice/crio-04704c6b5213e9da5d30d034026bfa8ba7725c0def1611d1391eaa79ae9758a5 WatchSource:0}: Error finding container 04704c6b5213e9da5d30d034026bfa8ba7725c0def1611d1391eaa79ae9758a5: Status 404 returned error can't find the container with id 04704c6b5213e9da5d30d034026bfa8ba7725c0def1611d1391eaa79ae9758a5 Nov 24 14:45:17 crc kubenswrapper[4790]: I1124 14:45:17.646809 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-f2g9r" Nov 24 14:45:18 crc kubenswrapper[4790]: I1124 14:45:18.445343 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"70d99ce8-fc77-4c8a-84ad-092c20c65e1d","Type":"ContainerStarted","Data":"227c333fcf1503133bd1d880a42aedf4b6cdd3f4bee234545d23a2a38dbb84e1"} Nov 24 14:45:18 crc kubenswrapper[4790]: I1124 14:45:18.448130 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-zv4x8" event={"ID":"873b8408-4a0a-42dd-bb32-5e3cba17a985","Type":"ContainerStarted","Data":"c1d044acc60c098abe9db072b72ce2adfa61682892f24054ec2fd9a97c4ccef6"} Nov 24 14:45:18 crc kubenswrapper[4790]: I1124 14:45:18.448282 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-zv4x8" event={"ID":"873b8408-4a0a-42dd-bb32-5e3cba17a985","Type":"ContainerStarted","Data":"04704c6b5213e9da5d30d034026bfa8ba7725c0def1611d1391eaa79ae9758a5"} Nov 24 14:45:18 crc kubenswrapper[4790]: I1124 14:45:18.451063 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c3811846-d6c0-47d7-8d40-93bae3eb1e49","Type":"ContainerStarted","Data":"798ee33e3e264e95996e4a02983c7bc25162fcdc0f45adcf6513ae8caec908ce"} Nov 24 14:45:18 crc kubenswrapper[4790]: I1124 14:45:18.452759 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-w7f6n" event={"ID":"b8596f2b-213d-4a3f-940a-9cc7421e85be","Type":"ContainerStarted","Data":"f4132b7f7a234c5c9e8140136b0824fcdd3d69623d363a43024e4ccfa3f7c0b7"} Nov 24 14:45:18 crc kubenswrapper[4790]: I1124 14:45:18.454301 4790 generic.go:334] "Generic (PLEG): container finished" podID="11caa523-04fb-4f1d-a6e5-b9323fc00b3f" containerID="3097bfa54223828dfffefb07013b4ee068ae2b0dfca245a5c841dd79c726304c" exitCode=0 Nov 24 14:45:18 crc kubenswrapper[4790]: I1124 14:45:18.454439 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbb866757-8frpq" event={"ID":"11caa523-04fb-4f1d-a6e5-b9323fc00b3f","Type":"ContainerDied","Data":"3097bfa54223828dfffefb07013b4ee068ae2b0dfca245a5c841dd79c726304c"} Nov 24 14:45:18 crc kubenswrapper[4790]: I1124 14:45:18.462263 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.462244495 podStartE2EDuration="3.462244495s" podCreationTimestamp="2025-11-24 14:45:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:45:18.461827223 +0000 UTC m=+5566.841720905" watchObservedRunningTime="2025-11-24 14:45:18.462244495 +0000 UTC m=+5566.842138157" Nov 24 14:45:18 crc kubenswrapper[4790]: I1124 14:45:18.477673 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0144dbfa-e743-47eb-b9ab-8e3610fc3130","Type":"ContainerStarted","Data":"c0c89e471a403e94f493268f25fac09952986b15ccdb311c78f230f2f42e6b2b"} Nov 24 14:45:18 crc kubenswrapper[4790]: I1124 14:45:18.477723 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"691daa40-76bd-4926-928b-2e4e25f7a50c","Type":"ContainerStarted","Data":"33f876c67cb8be531ece33f2ccff4da7636cfb0876bcf64391415f5821dc389d"} Nov 24 14:45:18 crc kubenswrapper[4790]: I1124 14:45:18.538873 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-w7f6n" podStartSLOduration=3.538845675 podStartE2EDuration="3.538845675s" podCreationTimestamp="2025-11-24 14:45:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:45:18.503362706 +0000 UTC m=+5566.883256378" watchObservedRunningTime="2025-11-24 14:45:18.538845675 +0000 UTC m=+5566.918739337" Nov 24 14:45:18 crc kubenswrapper[4790]: I1124 14:45:18.549493 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.5494649689999997 podStartE2EDuration="3.549464969s" podCreationTimestamp="2025-11-24 14:45:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:45:18.526320015 +0000 UTC m=+5566.906213687" watchObservedRunningTime="2025-11-24 14:45:18.549464969 +0000 UTC m=+5566.929358631" Nov 24 14:45:18 crc kubenswrapper[4790]: I1124 14:45:18.555369 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-f2g9r" Nov 24 14:45:18 crc kubenswrapper[4790]: I1124 14:45:18.560308 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.56027383 podStartE2EDuration="3.56027383s" podCreationTimestamp="2025-11-24 14:45:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:45:18.545405363 +0000 UTC m=+5566.925299025" watchObservedRunningTime="2025-11-24 14:45:18.56027383 +0000 UTC m=+5566.940167492" Nov 24 14:45:19 crc kubenswrapper[4790]: I1124 14:45:19.485856 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"691daa40-76bd-4926-928b-2e4e25f7a50c","Type":"ContainerStarted","Data":"45b17714e06fc213a80f08f40c248339405436e9a2c50951cf61eebf672c6962"} Nov 24 14:45:19 crc kubenswrapper[4790]: I1124 14:45:19.489222 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbb866757-8frpq" event={"ID":"11caa523-04fb-4f1d-a6e5-b9323fc00b3f","Type":"ContainerStarted","Data":"8fc61441c83ed7c15701f95649565263bde6d8a5e7518cddc59e2934f3b1a193"} Nov 24 14:45:19 crc kubenswrapper[4790]: I1124 14:45:19.489323 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bbb866757-8frpq" Nov 24 14:45:19 crc kubenswrapper[4790]: I1124 14:45:19.507098 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=4.507067643 podStartE2EDuration="4.507067643s" podCreationTimestamp="2025-11-24 14:45:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:45:19.503699187 +0000 UTC m=+5567.883592879" watchObservedRunningTime="2025-11-24 14:45:19.507067643 +0000 UTC m=+5567.886961315" Nov 24 14:45:19 crc kubenswrapper[4790]: I1124 14:45:19.541743 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bbb866757-8frpq" podStartSLOduration=4.541717738 podStartE2EDuration="4.541717738s" podCreationTimestamp="2025-11-24 14:45:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:45:19.528741436 +0000 UTC m=+5567.908635088" watchObservedRunningTime="2025-11-24 14:45:19.541717738 +0000 UTC m=+5567.921611400" Nov 24 14:45:19 crc kubenswrapper[4790]: I1124 14:45:19.558254 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f2g9r"] Nov 24 14:45:19 crc kubenswrapper[4790]: I1124 14:45:19.562355 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-zv4x8" podStartSLOduration=3.56234343 podStartE2EDuration="3.56234343s" podCreationTimestamp="2025-11-24 14:45:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:45:19.54908385 +0000 UTC m=+5567.928977532" watchObservedRunningTime="2025-11-24 14:45:19.56234343 +0000 UTC m=+5567.942237092" Nov 24 14:45:19 crc kubenswrapper[4790]: I1124 14:45:19.899624 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-42ld4" Nov 24 14:45:19 crc kubenswrapper[4790]: I1124 14:45:19.899705 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-42ld4" Nov 24 14:45:20 crc kubenswrapper[4790]: I1124 14:45:20.691990 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 14:45:20 crc kubenswrapper[4790]: I1124 14:45:20.956232 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-42ld4" podUID="58a4a77b-8070-4da1-b716-2e9796bab87a" containerName="registry-server" probeResult="failure" output=< Nov 24 14:45:20 crc kubenswrapper[4790]: timeout: failed to connect service ":50051" within 1s Nov 24 14:45:20 crc kubenswrapper[4790]: > Nov 24 14:45:21 crc kubenswrapper[4790]: I1124 14:45:21.014787 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 14:45:21 crc kubenswrapper[4790]: I1124 14:45:21.015387 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 14:45:21 crc kubenswrapper[4790]: I1124 14:45:21.105105 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:45:21 crc kubenswrapper[4790]: I1124 14:45:21.515310 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-f2g9r" podUID="3630d953-1c95-4c3e-bb30-4601ec9534e5" containerName="registry-server" containerID="cri-o://dd1d2d4b80a3a85f8a97953e14147f0044617694834c503530979b3739038d0f" gracePeriod=2 Nov 24 14:45:22 crc kubenswrapper[4790]: I1124 14:45:22.204381 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f2g9r" Nov 24 14:45:22 crc kubenswrapper[4790]: I1124 14:45:22.360710 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3630d953-1c95-4c3e-bb30-4601ec9534e5-utilities\") pod \"3630d953-1c95-4c3e-bb30-4601ec9534e5\" (UID: \"3630d953-1c95-4c3e-bb30-4601ec9534e5\") " Nov 24 14:45:22 crc kubenswrapper[4790]: I1124 14:45:22.360942 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3630d953-1c95-4c3e-bb30-4601ec9534e5-catalog-content\") pod \"3630d953-1c95-4c3e-bb30-4601ec9534e5\" (UID: \"3630d953-1c95-4c3e-bb30-4601ec9534e5\") " Nov 24 14:45:22 crc kubenswrapper[4790]: I1124 14:45:22.361063 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2zp2\" (UniqueName: \"kubernetes.io/projected/3630d953-1c95-4c3e-bb30-4601ec9534e5-kube-api-access-q2zp2\") pod \"3630d953-1c95-4c3e-bb30-4601ec9534e5\" (UID: \"3630d953-1c95-4c3e-bb30-4601ec9534e5\") " Nov 24 14:45:22 crc kubenswrapper[4790]: I1124 14:45:22.362150 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3630d953-1c95-4c3e-bb30-4601ec9534e5-utilities" (OuterVolumeSpecName: "utilities") pod "3630d953-1c95-4c3e-bb30-4601ec9534e5" (UID: "3630d953-1c95-4c3e-bb30-4601ec9534e5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:45:22 crc kubenswrapper[4790]: I1124 14:45:22.371108 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3630d953-1c95-4c3e-bb30-4601ec9534e5-kube-api-access-q2zp2" (OuterVolumeSpecName: "kube-api-access-q2zp2") pod "3630d953-1c95-4c3e-bb30-4601ec9534e5" (UID: "3630d953-1c95-4c3e-bb30-4601ec9534e5"). InnerVolumeSpecName "kube-api-access-q2zp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:45:22 crc kubenswrapper[4790]: I1124 14:45:22.396198 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3630d953-1c95-4c3e-bb30-4601ec9534e5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3630d953-1c95-4c3e-bb30-4601ec9534e5" (UID: "3630d953-1c95-4c3e-bb30-4601ec9534e5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:45:22 crc kubenswrapper[4790]: I1124 14:45:22.464069 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3630d953-1c95-4c3e-bb30-4601ec9534e5-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:22 crc kubenswrapper[4790]: I1124 14:45:22.464110 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3630d953-1c95-4c3e-bb30-4601ec9534e5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:22 crc kubenswrapper[4790]: I1124 14:45:22.464122 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2zp2\" (UniqueName: \"kubernetes.io/projected/3630d953-1c95-4c3e-bb30-4601ec9534e5-kube-api-access-q2zp2\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:22 crc kubenswrapper[4790]: I1124 14:45:22.529660 4790 generic.go:334] "Generic (PLEG): container finished" podID="873b8408-4a0a-42dd-bb32-5e3cba17a985" containerID="c1d044acc60c098abe9db072b72ce2adfa61682892f24054ec2fd9a97c4ccef6" exitCode=0 Nov 24 14:45:22 crc kubenswrapper[4790]: I1124 14:45:22.529734 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-zv4x8" event={"ID":"873b8408-4a0a-42dd-bb32-5e3cba17a985","Type":"ContainerDied","Data":"c1d044acc60c098abe9db072b72ce2adfa61682892f24054ec2fd9a97c4ccef6"} Nov 24 14:45:22 crc kubenswrapper[4790]: I1124 14:45:22.532272 4790 generic.go:334] "Generic (PLEG): container finished" podID="b8596f2b-213d-4a3f-940a-9cc7421e85be" containerID="f4132b7f7a234c5c9e8140136b0824fcdd3d69623d363a43024e4ccfa3f7c0b7" exitCode=0 Nov 24 14:45:22 crc kubenswrapper[4790]: I1124 14:45:22.532397 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-w7f6n" event={"ID":"b8596f2b-213d-4a3f-940a-9cc7421e85be","Type":"ContainerDied","Data":"f4132b7f7a234c5c9e8140136b0824fcdd3d69623d363a43024e4ccfa3f7c0b7"} Nov 24 14:45:22 crc kubenswrapper[4790]: I1124 14:45:22.536997 4790 generic.go:334] "Generic (PLEG): container finished" podID="3630d953-1c95-4c3e-bb30-4601ec9534e5" containerID="dd1d2d4b80a3a85f8a97953e14147f0044617694834c503530979b3739038d0f" exitCode=0 Nov 24 14:45:22 crc kubenswrapper[4790]: I1124 14:45:22.537110 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f2g9r" event={"ID":"3630d953-1c95-4c3e-bb30-4601ec9534e5","Type":"ContainerDied","Data":"dd1d2d4b80a3a85f8a97953e14147f0044617694834c503530979b3739038d0f"} Nov 24 14:45:22 crc kubenswrapper[4790]: I1124 14:45:22.537252 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f2g9r" event={"ID":"3630d953-1c95-4c3e-bb30-4601ec9534e5","Type":"ContainerDied","Data":"64c042a60db43b4f1039b974b37328246a5350844598fe199f9929f5503a6164"} Nov 24 14:45:22 crc kubenswrapper[4790]: I1124 14:45:22.537143 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f2g9r" Nov 24 14:45:22 crc kubenswrapper[4790]: I1124 14:45:22.537314 4790 scope.go:117] "RemoveContainer" containerID="dd1d2d4b80a3a85f8a97953e14147f0044617694834c503530979b3739038d0f" Nov 24 14:45:22 crc kubenswrapper[4790]: I1124 14:45:22.584802 4790 scope.go:117] "RemoveContainer" containerID="f3b5b3021a807e072d983b8b79c270bf3938e2992499036daeb98989e7ea1e7f" Nov 24 14:45:22 crc kubenswrapper[4790]: I1124 14:45:22.628050 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f2g9r"] Nov 24 14:45:22 crc kubenswrapper[4790]: I1124 14:45:22.631610 4790 scope.go:117] "RemoveContainer" containerID="23e409afac26a0676547505b2f448f0034d69346e9bdb5686470d27822e47312" Nov 24 14:45:22 crc kubenswrapper[4790]: I1124 14:45:22.646382 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-f2g9r"] Nov 24 14:45:22 crc kubenswrapper[4790]: I1124 14:45:22.681228 4790 scope.go:117] "RemoveContainer" containerID="dd1d2d4b80a3a85f8a97953e14147f0044617694834c503530979b3739038d0f" Nov 24 14:45:22 crc kubenswrapper[4790]: E1124 14:45:22.682042 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd1d2d4b80a3a85f8a97953e14147f0044617694834c503530979b3739038d0f\": container with ID starting with dd1d2d4b80a3a85f8a97953e14147f0044617694834c503530979b3739038d0f not found: ID does not exist" containerID="dd1d2d4b80a3a85f8a97953e14147f0044617694834c503530979b3739038d0f" Nov 24 14:45:22 crc kubenswrapper[4790]: I1124 14:45:22.682095 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd1d2d4b80a3a85f8a97953e14147f0044617694834c503530979b3739038d0f"} err="failed to get container status \"dd1d2d4b80a3a85f8a97953e14147f0044617694834c503530979b3739038d0f\": rpc error: code = NotFound desc = could not find container \"dd1d2d4b80a3a85f8a97953e14147f0044617694834c503530979b3739038d0f\": container with ID starting with dd1d2d4b80a3a85f8a97953e14147f0044617694834c503530979b3739038d0f not found: ID does not exist" Nov 24 14:45:22 crc kubenswrapper[4790]: I1124 14:45:22.682128 4790 scope.go:117] "RemoveContainer" containerID="f3b5b3021a807e072d983b8b79c270bf3938e2992499036daeb98989e7ea1e7f" Nov 24 14:45:22 crc kubenswrapper[4790]: E1124 14:45:22.682622 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3b5b3021a807e072d983b8b79c270bf3938e2992499036daeb98989e7ea1e7f\": container with ID starting with f3b5b3021a807e072d983b8b79c270bf3938e2992499036daeb98989e7ea1e7f not found: ID does not exist" containerID="f3b5b3021a807e072d983b8b79c270bf3938e2992499036daeb98989e7ea1e7f" Nov 24 14:45:22 crc kubenswrapper[4790]: I1124 14:45:22.682675 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3b5b3021a807e072d983b8b79c270bf3938e2992499036daeb98989e7ea1e7f"} err="failed to get container status \"f3b5b3021a807e072d983b8b79c270bf3938e2992499036daeb98989e7ea1e7f\": rpc error: code = NotFound desc = could not find container \"f3b5b3021a807e072d983b8b79c270bf3938e2992499036daeb98989e7ea1e7f\": container with ID starting with f3b5b3021a807e072d983b8b79c270bf3938e2992499036daeb98989e7ea1e7f not found: ID does not exist" Nov 24 14:45:22 crc kubenswrapper[4790]: I1124 14:45:22.682708 4790 scope.go:117] "RemoveContainer" containerID="23e409afac26a0676547505b2f448f0034d69346e9bdb5686470d27822e47312" Nov 24 14:45:22 crc kubenswrapper[4790]: E1124 14:45:22.683170 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23e409afac26a0676547505b2f448f0034d69346e9bdb5686470d27822e47312\": container with ID starting with 23e409afac26a0676547505b2f448f0034d69346e9bdb5686470d27822e47312 not found: ID does not exist" containerID="23e409afac26a0676547505b2f448f0034d69346e9bdb5686470d27822e47312" Nov 24 14:45:22 crc kubenswrapper[4790]: I1124 14:45:22.683224 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23e409afac26a0676547505b2f448f0034d69346e9bdb5686470d27822e47312"} err="failed to get container status \"23e409afac26a0676547505b2f448f0034d69346e9bdb5686470d27822e47312\": rpc error: code = NotFound desc = could not find container \"23e409afac26a0676547505b2f448f0034d69346e9bdb5686470d27822e47312\": container with ID starting with 23e409afac26a0676547505b2f448f0034d69346e9bdb5686470d27822e47312 not found: ID does not exist" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.083658 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-zv4x8" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.092637 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-w7f6n" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.202026 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8596f2b-213d-4a3f-940a-9cc7421e85be-combined-ca-bundle\") pod \"b8596f2b-213d-4a3f-940a-9cc7421e85be\" (UID: \"b8596f2b-213d-4a3f-940a-9cc7421e85be\") " Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.202095 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txw5p\" (UniqueName: \"kubernetes.io/projected/873b8408-4a0a-42dd-bb32-5e3cba17a985-kube-api-access-txw5p\") pod \"873b8408-4a0a-42dd-bb32-5e3cba17a985\" (UID: \"873b8408-4a0a-42dd-bb32-5e3cba17a985\") " Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.202137 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/873b8408-4a0a-42dd-bb32-5e3cba17a985-config-data\") pod \"873b8408-4a0a-42dd-bb32-5e3cba17a985\" (UID: \"873b8408-4a0a-42dd-bb32-5e3cba17a985\") " Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.202176 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5xhg\" (UniqueName: \"kubernetes.io/projected/b8596f2b-213d-4a3f-940a-9cc7421e85be-kube-api-access-z5xhg\") pod \"b8596f2b-213d-4a3f-940a-9cc7421e85be\" (UID: \"b8596f2b-213d-4a3f-940a-9cc7421e85be\") " Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.202243 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/873b8408-4a0a-42dd-bb32-5e3cba17a985-combined-ca-bundle\") pod \"873b8408-4a0a-42dd-bb32-5e3cba17a985\" (UID: \"873b8408-4a0a-42dd-bb32-5e3cba17a985\") " Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.202316 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8596f2b-213d-4a3f-940a-9cc7421e85be-scripts\") pod \"b8596f2b-213d-4a3f-940a-9cc7421e85be\" (UID: \"b8596f2b-213d-4a3f-940a-9cc7421e85be\") " Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.202466 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8596f2b-213d-4a3f-940a-9cc7421e85be-config-data\") pod \"b8596f2b-213d-4a3f-940a-9cc7421e85be\" (UID: \"b8596f2b-213d-4a3f-940a-9cc7421e85be\") " Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.202561 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/873b8408-4a0a-42dd-bb32-5e3cba17a985-scripts\") pod \"873b8408-4a0a-42dd-bb32-5e3cba17a985\" (UID: \"873b8408-4a0a-42dd-bb32-5e3cba17a985\") " Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.209595 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8596f2b-213d-4a3f-940a-9cc7421e85be-scripts" (OuterVolumeSpecName: "scripts") pod "b8596f2b-213d-4a3f-940a-9cc7421e85be" (UID: "b8596f2b-213d-4a3f-940a-9cc7421e85be"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.209815 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/873b8408-4a0a-42dd-bb32-5e3cba17a985-scripts" (OuterVolumeSpecName: "scripts") pod "873b8408-4a0a-42dd-bb32-5e3cba17a985" (UID: "873b8408-4a0a-42dd-bb32-5e3cba17a985"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.210585 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8596f2b-213d-4a3f-940a-9cc7421e85be-kube-api-access-z5xhg" (OuterVolumeSpecName: "kube-api-access-z5xhg") pod "b8596f2b-213d-4a3f-940a-9cc7421e85be" (UID: "b8596f2b-213d-4a3f-940a-9cc7421e85be"). InnerVolumeSpecName "kube-api-access-z5xhg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.211404 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/873b8408-4a0a-42dd-bb32-5e3cba17a985-kube-api-access-txw5p" (OuterVolumeSpecName: "kube-api-access-txw5p") pod "873b8408-4a0a-42dd-bb32-5e3cba17a985" (UID: "873b8408-4a0a-42dd-bb32-5e3cba17a985"). InnerVolumeSpecName "kube-api-access-txw5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.235067 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8596f2b-213d-4a3f-940a-9cc7421e85be-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b8596f2b-213d-4a3f-940a-9cc7421e85be" (UID: "b8596f2b-213d-4a3f-940a-9cc7421e85be"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.241759 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8596f2b-213d-4a3f-940a-9cc7421e85be-config-data" (OuterVolumeSpecName: "config-data") pod "b8596f2b-213d-4a3f-940a-9cc7421e85be" (UID: "b8596f2b-213d-4a3f-940a-9cc7421e85be"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.242040 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/873b8408-4a0a-42dd-bb32-5e3cba17a985-config-data" (OuterVolumeSpecName: "config-data") pod "873b8408-4a0a-42dd-bb32-5e3cba17a985" (UID: "873b8408-4a0a-42dd-bb32-5e3cba17a985"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.251726 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/873b8408-4a0a-42dd-bb32-5e3cba17a985-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "873b8408-4a0a-42dd-bb32-5e3cba17a985" (UID: "873b8408-4a0a-42dd-bb32-5e3cba17a985"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.305613 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8596f2b-213d-4a3f-940a-9cc7421e85be-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.305674 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/873b8408-4a0a-42dd-bb32-5e3cba17a985-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.305691 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8596f2b-213d-4a3f-940a-9cc7421e85be-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.305704 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txw5p\" (UniqueName: \"kubernetes.io/projected/873b8408-4a0a-42dd-bb32-5e3cba17a985-kube-api-access-txw5p\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.305716 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/873b8408-4a0a-42dd-bb32-5e3cba17a985-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.305753 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5xhg\" (UniqueName: \"kubernetes.io/projected/b8596f2b-213d-4a3f-940a-9cc7421e85be-kube-api-access-z5xhg\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.305767 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/873b8408-4a0a-42dd-bb32-5e3cba17a985-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.305778 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8596f2b-213d-4a3f-940a-9cc7421e85be-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.333111 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3630d953-1c95-4c3e-bb30-4601ec9534e5" path="/var/lib/kubelet/pods/3630d953-1c95-4c3e-bb30-4601ec9534e5/volumes" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.573059 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-zv4x8" event={"ID":"873b8408-4a0a-42dd-bb32-5e3cba17a985","Type":"ContainerDied","Data":"04704c6b5213e9da5d30d034026bfa8ba7725c0def1611d1391eaa79ae9758a5"} Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.573115 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-zv4x8" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.573323 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04704c6b5213e9da5d30d034026bfa8ba7725c0def1611d1391eaa79ae9758a5" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.576691 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-w7f6n" event={"ID":"b8596f2b-213d-4a3f-940a-9cc7421e85be","Type":"ContainerDied","Data":"c3d995294475670ef0c303a1dd27b0c5f3542d19aff69c267200da06928f8ebf"} Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.576753 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c3d995294475670ef0c303a1dd27b0c5f3542d19aff69c267200da06928f8ebf" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.576838 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-w7f6n" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.684287 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 14:45:24 crc kubenswrapper[4790]: E1124 14:45:24.684952 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="873b8408-4a0a-42dd-bb32-5e3cba17a985" containerName="nova-cell1-conductor-db-sync" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.684977 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="873b8408-4a0a-42dd-bb32-5e3cba17a985" containerName="nova-cell1-conductor-db-sync" Nov 24 14:45:24 crc kubenswrapper[4790]: E1124 14:45:24.684998 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8596f2b-213d-4a3f-940a-9cc7421e85be" containerName="nova-manage" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.685008 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8596f2b-213d-4a3f-940a-9cc7421e85be" containerName="nova-manage" Nov 24 14:45:24 crc kubenswrapper[4790]: E1124 14:45:24.685031 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3630d953-1c95-4c3e-bb30-4601ec9534e5" containerName="registry-server" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.685040 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="3630d953-1c95-4c3e-bb30-4601ec9534e5" containerName="registry-server" Nov 24 14:45:24 crc kubenswrapper[4790]: E1124 14:45:24.685055 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3630d953-1c95-4c3e-bb30-4601ec9534e5" containerName="extract-utilities" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.685063 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="3630d953-1c95-4c3e-bb30-4601ec9534e5" containerName="extract-utilities" Nov 24 14:45:24 crc kubenswrapper[4790]: E1124 14:45:24.685087 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3630d953-1c95-4c3e-bb30-4601ec9534e5" containerName="extract-content" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.685098 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="3630d953-1c95-4c3e-bb30-4601ec9534e5" containerName="extract-content" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.685385 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8596f2b-213d-4a3f-940a-9cc7421e85be" containerName="nova-manage" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.685407 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="3630d953-1c95-4c3e-bb30-4601ec9534e5" containerName="registry-server" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.685443 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="873b8408-4a0a-42dd-bb32-5e3cba17a985" containerName="nova-cell1-conductor-db-sync" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.686512 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.691459 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.706007 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.816296 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfa26363-ed7c-4bd7-a32b-420e94c62a3d-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"bfa26363-ed7c-4bd7-a32b-420e94c62a3d\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.816364 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfa26363-ed7c-4bd7-a32b-420e94c62a3d-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"bfa26363-ed7c-4bd7-a32b-420e94c62a3d\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.816405 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcjpc\" (UniqueName: \"kubernetes.io/projected/bfa26363-ed7c-4bd7-a32b-420e94c62a3d-kube-api-access-wcjpc\") pod \"nova-cell1-conductor-0\" (UID: \"bfa26363-ed7c-4bd7-a32b-420e94c62a3d\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.851938 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.852304 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c3811846-d6c0-47d7-8d40-93bae3eb1e49" containerName="nova-api-log" containerID="cri-o://3e80e07d5b39eacf4ecc5c65880beb8a0b12349edb19f7ff9daa636d78ed5c6f" gracePeriod=30 Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.853845 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c3811846-d6c0-47d7-8d40-93bae3eb1e49" containerName="nova-api-api" containerID="cri-o://798ee33e3e264e95996e4a02983c7bc25162fcdc0f45adcf6513ae8caec908ce" gracePeriod=30 Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.865702 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.865998 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="0144dbfa-e743-47eb-b9ab-8e3610fc3130" containerName="nova-scheduler-scheduler" containerID="cri-o://c0c89e471a403e94f493268f25fac09952986b15ccdb311c78f230f2f42e6b2b" gracePeriod=30 Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.878580 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.878902 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="691daa40-76bd-4926-928b-2e4e25f7a50c" containerName="nova-metadata-log" containerID="cri-o://33f876c67cb8be531ece33f2ccff4da7636cfb0876bcf64391415f5821dc389d" gracePeriod=30 Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.879046 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="691daa40-76bd-4926-928b-2e4e25f7a50c" containerName="nova-metadata-metadata" containerID="cri-o://45b17714e06fc213a80f08f40c248339405436e9a2c50951cf61eebf672c6962" gracePeriod=30 Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.918523 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfa26363-ed7c-4bd7-a32b-420e94c62a3d-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"bfa26363-ed7c-4bd7-a32b-420e94c62a3d\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.918579 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfa26363-ed7c-4bd7-a32b-420e94c62a3d-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"bfa26363-ed7c-4bd7-a32b-420e94c62a3d\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.918608 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcjpc\" (UniqueName: \"kubernetes.io/projected/bfa26363-ed7c-4bd7-a32b-420e94c62a3d-kube-api-access-wcjpc\") pod \"nova-cell1-conductor-0\" (UID: \"bfa26363-ed7c-4bd7-a32b-420e94c62a3d\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.925097 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfa26363-ed7c-4bd7-a32b-420e94c62a3d-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"bfa26363-ed7c-4bd7-a32b-420e94c62a3d\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.926727 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfa26363-ed7c-4bd7-a32b-420e94c62a3d-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"bfa26363-ed7c-4bd7-a32b-420e94c62a3d\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:45:24 crc kubenswrapper[4790]: I1124 14:45:24.935342 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcjpc\" (UniqueName: \"kubernetes.io/projected/bfa26363-ed7c-4bd7-a32b-420e94c62a3d-kube-api-access-wcjpc\") pod \"nova-cell1-conductor-0\" (UID: \"bfa26363-ed7c-4bd7-a32b-420e94c62a3d\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:45:25 crc kubenswrapper[4790]: I1124 14:45:25.018069 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 14:45:25 crc kubenswrapper[4790]: I1124 14:45:25.508522 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 14:45:25 crc kubenswrapper[4790]: W1124 14:45:25.513660 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbfa26363_ed7c_4bd7_a32b_420e94c62a3d.slice/crio-9792d21095fe0bc230576595e575197ec3143c0aec72f52aa274affe775cee96 WatchSource:0}: Error finding container 9792d21095fe0bc230576595e575197ec3143c0aec72f52aa274affe775cee96: Status 404 returned error can't find the container with id 9792d21095fe0bc230576595e575197ec3143c0aec72f52aa274affe775cee96 Nov 24 14:45:25 crc kubenswrapper[4790]: I1124 14:45:25.603047 4790 generic.go:334] "Generic (PLEG): container finished" podID="691daa40-76bd-4926-928b-2e4e25f7a50c" containerID="45b17714e06fc213a80f08f40c248339405436e9a2c50951cf61eebf672c6962" exitCode=0 Nov 24 14:45:25 crc kubenswrapper[4790]: I1124 14:45:25.603485 4790 generic.go:334] "Generic (PLEG): container finished" podID="691daa40-76bd-4926-928b-2e4e25f7a50c" containerID="33f876c67cb8be531ece33f2ccff4da7636cfb0876bcf64391415f5821dc389d" exitCode=143 Nov 24 14:45:25 crc kubenswrapper[4790]: I1124 14:45:25.603131 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"691daa40-76bd-4926-928b-2e4e25f7a50c","Type":"ContainerDied","Data":"45b17714e06fc213a80f08f40c248339405436e9a2c50951cf61eebf672c6962"} Nov 24 14:45:25 crc kubenswrapper[4790]: I1124 14:45:25.603584 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"691daa40-76bd-4926-928b-2e4e25f7a50c","Type":"ContainerDied","Data":"33f876c67cb8be531ece33f2ccff4da7636cfb0876bcf64391415f5821dc389d"} Nov 24 14:45:25 crc kubenswrapper[4790]: I1124 14:45:25.607698 4790 generic.go:334] "Generic (PLEG): container finished" podID="c3811846-d6c0-47d7-8d40-93bae3eb1e49" containerID="798ee33e3e264e95996e4a02983c7bc25162fcdc0f45adcf6513ae8caec908ce" exitCode=0 Nov 24 14:45:25 crc kubenswrapper[4790]: I1124 14:45:25.607732 4790 generic.go:334] "Generic (PLEG): container finished" podID="c3811846-d6c0-47d7-8d40-93bae3eb1e49" containerID="3e80e07d5b39eacf4ecc5c65880beb8a0b12349edb19f7ff9daa636d78ed5c6f" exitCode=143 Nov 24 14:45:25 crc kubenswrapper[4790]: I1124 14:45:25.607790 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c3811846-d6c0-47d7-8d40-93bae3eb1e49","Type":"ContainerDied","Data":"798ee33e3e264e95996e4a02983c7bc25162fcdc0f45adcf6513ae8caec908ce"} Nov 24 14:45:25 crc kubenswrapper[4790]: I1124 14:45:25.607865 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c3811846-d6c0-47d7-8d40-93bae3eb1e49","Type":"ContainerDied","Data":"3e80e07d5b39eacf4ecc5c65880beb8a0b12349edb19f7ff9daa636d78ed5c6f"} Nov 24 14:45:25 crc kubenswrapper[4790]: I1124 14:45:25.608955 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"bfa26363-ed7c-4bd7-a32b-420e94c62a3d","Type":"ContainerStarted","Data":"9792d21095fe0bc230576595e575197ec3143c0aec72f52aa274affe775cee96"} Nov 24 14:45:25 crc kubenswrapper[4790]: I1124 14:45:25.849709 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:45:25 crc kubenswrapper[4790]: I1124 14:45:25.916976 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:45:25 crc kubenswrapper[4790]: I1124 14:45:25.938707 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3811846-d6c0-47d7-8d40-93bae3eb1e49-config-data\") pod \"c3811846-d6c0-47d7-8d40-93bae3eb1e49\" (UID: \"c3811846-d6c0-47d7-8d40-93bae3eb1e49\") " Nov 24 14:45:25 crc kubenswrapper[4790]: I1124 14:45:25.938860 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3811846-d6c0-47d7-8d40-93bae3eb1e49-logs\") pod \"c3811846-d6c0-47d7-8d40-93bae3eb1e49\" (UID: \"c3811846-d6c0-47d7-8d40-93bae3eb1e49\") " Nov 24 14:45:25 crc kubenswrapper[4790]: I1124 14:45:25.938928 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxvrm\" (UniqueName: \"kubernetes.io/projected/c3811846-d6c0-47d7-8d40-93bae3eb1e49-kube-api-access-jxvrm\") pod \"c3811846-d6c0-47d7-8d40-93bae3eb1e49\" (UID: \"c3811846-d6c0-47d7-8d40-93bae3eb1e49\") " Nov 24 14:45:25 crc kubenswrapper[4790]: I1124 14:45:25.939087 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3811846-d6c0-47d7-8d40-93bae3eb1e49-combined-ca-bundle\") pod \"c3811846-d6c0-47d7-8d40-93bae3eb1e49\" (UID: \"c3811846-d6c0-47d7-8d40-93bae3eb1e49\") " Nov 24 14:45:25 crc kubenswrapper[4790]: I1124 14:45:25.939593 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3811846-d6c0-47d7-8d40-93bae3eb1e49-logs" (OuterVolumeSpecName: "logs") pod "c3811846-d6c0-47d7-8d40-93bae3eb1e49" (UID: "c3811846-d6c0-47d7-8d40-93bae3eb1e49"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:45:25 crc kubenswrapper[4790]: I1124 14:45:25.944337 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3811846-d6c0-47d7-8d40-93bae3eb1e49-kube-api-access-jxvrm" (OuterVolumeSpecName: "kube-api-access-jxvrm") pod "c3811846-d6c0-47d7-8d40-93bae3eb1e49" (UID: "c3811846-d6c0-47d7-8d40-93bae3eb1e49"). InnerVolumeSpecName "kube-api-access-jxvrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:45:25 crc kubenswrapper[4790]: I1124 14:45:25.971404 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3811846-d6c0-47d7-8d40-93bae3eb1e49-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c3811846-d6c0-47d7-8d40-93bae3eb1e49" (UID: "c3811846-d6c0-47d7-8d40-93bae3eb1e49"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:45:25 crc kubenswrapper[4790]: I1124 14:45:25.975420 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3811846-d6c0-47d7-8d40-93bae3eb1e49-config-data" (OuterVolumeSpecName: "config-data") pod "c3811846-d6c0-47d7-8d40-93bae3eb1e49" (UID: "c3811846-d6c0-47d7-8d40-93bae3eb1e49"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.040428 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/691daa40-76bd-4926-928b-2e4e25f7a50c-combined-ca-bundle\") pod \"691daa40-76bd-4926-928b-2e4e25f7a50c\" (UID: \"691daa40-76bd-4926-928b-2e4e25f7a50c\") " Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.040492 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/691daa40-76bd-4926-928b-2e4e25f7a50c-logs\") pod \"691daa40-76bd-4926-928b-2e4e25f7a50c\" (UID: \"691daa40-76bd-4926-928b-2e4e25f7a50c\") " Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.040637 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7h5h\" (UniqueName: \"kubernetes.io/projected/691daa40-76bd-4926-928b-2e4e25f7a50c-kube-api-access-b7h5h\") pod \"691daa40-76bd-4926-928b-2e4e25f7a50c\" (UID: \"691daa40-76bd-4926-928b-2e4e25f7a50c\") " Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.040705 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/691daa40-76bd-4926-928b-2e4e25f7a50c-config-data\") pod \"691daa40-76bd-4926-928b-2e4e25f7a50c\" (UID: \"691daa40-76bd-4926-928b-2e4e25f7a50c\") " Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.041269 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/691daa40-76bd-4926-928b-2e4e25f7a50c-logs" (OuterVolumeSpecName: "logs") pod "691daa40-76bd-4926-928b-2e4e25f7a50c" (UID: "691daa40-76bd-4926-928b-2e4e25f7a50c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.041343 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3811846-d6c0-47d7-8d40-93bae3eb1e49-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.041369 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3811846-d6c0-47d7-8d40-93bae3eb1e49-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.041382 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3811846-d6c0-47d7-8d40-93bae3eb1e49-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.041398 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxvrm\" (UniqueName: \"kubernetes.io/projected/c3811846-d6c0-47d7-8d40-93bae3eb1e49-kube-api-access-jxvrm\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.044867 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/691daa40-76bd-4926-928b-2e4e25f7a50c-kube-api-access-b7h5h" (OuterVolumeSpecName: "kube-api-access-b7h5h") pod "691daa40-76bd-4926-928b-2e4e25f7a50c" (UID: "691daa40-76bd-4926-928b-2e4e25f7a50c"). InnerVolumeSpecName "kube-api-access-b7h5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.071651 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/691daa40-76bd-4926-928b-2e4e25f7a50c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "691daa40-76bd-4926-928b-2e4e25f7a50c" (UID: "691daa40-76bd-4926-928b-2e4e25f7a50c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.076357 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/691daa40-76bd-4926-928b-2e4e25f7a50c-config-data" (OuterVolumeSpecName: "config-data") pod "691daa40-76bd-4926-928b-2e4e25f7a50c" (UID: "691daa40-76bd-4926-928b-2e4e25f7a50c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.105355 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.126252 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.148051 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/691daa40-76bd-4926-928b-2e4e25f7a50c-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.148114 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/691daa40-76bd-4926-928b-2e4e25f7a50c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.148134 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/691daa40-76bd-4926-928b-2e4e25f7a50c-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.148153 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7h5h\" (UniqueName: \"kubernetes.io/projected/691daa40-76bd-4926-928b-2e4e25f7a50c-kube-api-access-b7h5h\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.182155 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bbb866757-8frpq" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.277104 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76d49bc685-2dflh"] Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.278111 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-76d49bc685-2dflh" podUID="aa76fa12-375b-4c64-926f-fc0815a7d4a6" containerName="dnsmasq-dns" containerID="cri-o://ba0dd109b7f2187b0ccd1a652b87c9d4f6ec196a3ca0cdadadd9ff76ad6de458" gracePeriod=10 Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.620438 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.620430 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"691daa40-76bd-4926-928b-2e4e25f7a50c","Type":"ContainerDied","Data":"f634bb2ba321bf14dce1f79ae199de8a69a85bf7288babc51e51f7f0ba99336f"} Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.620579 4790 scope.go:117] "RemoveContainer" containerID="45b17714e06fc213a80f08f40c248339405436e9a2c50951cf61eebf672c6962" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.624300 4790 generic.go:334] "Generic (PLEG): container finished" podID="aa76fa12-375b-4c64-926f-fc0815a7d4a6" containerID="ba0dd109b7f2187b0ccd1a652b87c9d4f6ec196a3ca0cdadadd9ff76ad6de458" exitCode=0 Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.624361 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76d49bc685-2dflh" event={"ID":"aa76fa12-375b-4c64-926f-fc0815a7d4a6","Type":"ContainerDied","Data":"ba0dd109b7f2187b0ccd1a652b87c9d4f6ec196a3ca0cdadadd9ff76ad6de458"} Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.626567 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c3811846-d6c0-47d7-8d40-93bae3eb1e49","Type":"ContainerDied","Data":"c6d0bbf11a146f07d50980786fb31d3463c27b10c0888c9d36c453a97585cd24"} Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.626629 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.628554 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"bfa26363-ed7c-4bd7-a32b-420e94c62a3d","Type":"ContainerStarted","Data":"659aae12e81e3a6ff594ff0fa8c227b2de1d0e0869ca6de31bef7cebd3fce403"} Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.641991 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.648409 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.663041 4790 scope.go:117] "RemoveContainer" containerID="33f876c67cb8be531ece33f2ccff4da7636cfb0876bcf64391415f5821dc389d" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.668845 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.677834 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:45:26 crc kubenswrapper[4790]: E1124 14:45:26.678223 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="691daa40-76bd-4926-928b-2e4e25f7a50c" containerName="nova-metadata-metadata" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.678240 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="691daa40-76bd-4926-928b-2e4e25f7a50c" containerName="nova-metadata-metadata" Nov 24 14:45:26 crc kubenswrapper[4790]: E1124 14:45:26.678256 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="691daa40-76bd-4926-928b-2e4e25f7a50c" containerName="nova-metadata-log" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.678264 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="691daa40-76bd-4926-928b-2e4e25f7a50c" containerName="nova-metadata-log" Nov 24 14:45:26 crc kubenswrapper[4790]: E1124 14:45:26.678297 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3811846-d6c0-47d7-8d40-93bae3eb1e49" containerName="nova-api-log" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.678303 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3811846-d6c0-47d7-8d40-93bae3eb1e49" containerName="nova-api-log" Nov 24 14:45:26 crc kubenswrapper[4790]: E1124 14:45:26.678322 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3811846-d6c0-47d7-8d40-93bae3eb1e49" containerName="nova-api-api" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.678330 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3811846-d6c0-47d7-8d40-93bae3eb1e49" containerName="nova-api-api" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.678501 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="691daa40-76bd-4926-928b-2e4e25f7a50c" containerName="nova-metadata-metadata" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.678527 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3811846-d6c0-47d7-8d40-93bae3eb1e49" containerName="nova-api-log" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.678537 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="691daa40-76bd-4926-928b-2e4e25f7a50c" containerName="nova-metadata-log" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.678549 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3811846-d6c0-47d7-8d40-93bae3eb1e49" containerName="nova-api-api" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.679492 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.692747 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.695009 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.737536 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.753959 4790 scope.go:117] "RemoveContainer" containerID="798ee33e3e264e95996e4a02983c7bc25162fcdc0f45adcf6513ae8caec908ce" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.765929 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76d49bc685-2dflh" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.769612 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.778694 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.778665615 podStartE2EDuration="2.778665615s" podCreationTimestamp="2025-11-24 14:45:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:45:26.69315318 +0000 UTC m=+5575.073046842" watchObservedRunningTime="2025-11-24 14:45:26.778665615 +0000 UTC m=+5575.158559277" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.783282 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08642d69-d1bf-4816-928c-876bd2a62d2c-logs\") pod \"nova-metadata-0\" (UID: \"08642d69-d1bf-4816-928c-876bd2a62d2c\") " pod="openstack/nova-metadata-0" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.783410 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfx47\" (UniqueName: \"kubernetes.io/projected/08642d69-d1bf-4816-928c-876bd2a62d2c-kube-api-access-zfx47\") pod \"nova-metadata-0\" (UID: \"08642d69-d1bf-4816-928c-876bd2a62d2c\") " pod="openstack/nova-metadata-0" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.783623 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08642d69-d1bf-4816-928c-876bd2a62d2c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"08642d69-d1bf-4816-928c-876bd2a62d2c\") " pod="openstack/nova-metadata-0" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.783693 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08642d69-d1bf-4816-928c-876bd2a62d2c-config-data\") pod \"nova-metadata-0\" (UID: \"08642d69-d1bf-4816-928c-876bd2a62d2c\") " pod="openstack/nova-metadata-0" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.816259 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 14:45:26 crc kubenswrapper[4790]: E1124 14:45:26.839159 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa76fa12-375b-4c64-926f-fc0815a7d4a6" containerName="init" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.839197 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa76fa12-375b-4c64-926f-fc0815a7d4a6" containerName="init" Nov 24 14:45:26 crc kubenswrapper[4790]: E1124 14:45:26.839288 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa76fa12-375b-4c64-926f-fc0815a7d4a6" containerName="dnsmasq-dns" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.839295 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa76fa12-375b-4c64-926f-fc0815a7d4a6" containerName="dnsmasq-dns" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.839687 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa76fa12-375b-4c64-926f-fc0815a7d4a6" containerName="dnsmasq-dns" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.841375 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.842630 4790 scope.go:117] "RemoveContainer" containerID="3e80e07d5b39eacf4ecc5c65880beb8a0b12349edb19f7ff9daa636d78ed5c6f" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.856257 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.860785 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.887671 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jprl7\" (UniqueName: \"kubernetes.io/projected/aa76fa12-375b-4c64-926f-fc0815a7d4a6-kube-api-access-jprl7\") pod \"aa76fa12-375b-4c64-926f-fc0815a7d4a6\" (UID: \"aa76fa12-375b-4c64-926f-fc0815a7d4a6\") " Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.887743 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa76fa12-375b-4c64-926f-fc0815a7d4a6-config\") pod \"aa76fa12-375b-4c64-926f-fc0815a7d4a6\" (UID: \"aa76fa12-375b-4c64-926f-fc0815a7d4a6\") " Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.887764 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa76fa12-375b-4c64-926f-fc0815a7d4a6-ovsdbserver-sb\") pod \"aa76fa12-375b-4c64-926f-fc0815a7d4a6\" (UID: \"aa76fa12-375b-4c64-926f-fc0815a7d4a6\") " Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.887807 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa76fa12-375b-4c64-926f-fc0815a7d4a6-ovsdbserver-nb\") pod \"aa76fa12-375b-4c64-926f-fc0815a7d4a6\" (UID: \"aa76fa12-375b-4c64-926f-fc0815a7d4a6\") " Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.887857 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa76fa12-375b-4c64-926f-fc0815a7d4a6-dns-svc\") pod \"aa76fa12-375b-4c64-926f-fc0815a7d4a6\" (UID: \"aa76fa12-375b-4c64-926f-fc0815a7d4a6\") " Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.888112 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfx47\" (UniqueName: \"kubernetes.io/projected/08642d69-d1bf-4816-928c-876bd2a62d2c-kube-api-access-zfx47\") pod \"nova-metadata-0\" (UID: \"08642d69-d1bf-4816-928c-876bd2a62d2c\") " pod="openstack/nova-metadata-0" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.888187 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08642d69-d1bf-4816-928c-876bd2a62d2c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"08642d69-d1bf-4816-928c-876bd2a62d2c\") " pod="openstack/nova-metadata-0" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.888213 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08642d69-d1bf-4816-928c-876bd2a62d2c-config-data\") pod \"nova-metadata-0\" (UID: \"08642d69-d1bf-4816-928c-876bd2a62d2c\") " pod="openstack/nova-metadata-0" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.888278 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08642d69-d1bf-4816-928c-876bd2a62d2c-logs\") pod \"nova-metadata-0\" (UID: \"08642d69-d1bf-4816-928c-876bd2a62d2c\") " pod="openstack/nova-metadata-0" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.888684 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08642d69-d1bf-4816-928c-876bd2a62d2c-logs\") pod \"nova-metadata-0\" (UID: \"08642d69-d1bf-4816-928c-876bd2a62d2c\") " pod="openstack/nova-metadata-0" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.927559 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfx47\" (UniqueName: \"kubernetes.io/projected/08642d69-d1bf-4816-928c-876bd2a62d2c-kube-api-access-zfx47\") pod \"nova-metadata-0\" (UID: \"08642d69-d1bf-4816-928c-876bd2a62d2c\") " pod="openstack/nova-metadata-0" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.949654 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08642d69-d1bf-4816-928c-876bd2a62d2c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"08642d69-d1bf-4816-928c-876bd2a62d2c\") " pod="openstack/nova-metadata-0" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.949804 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa76fa12-375b-4c64-926f-fc0815a7d4a6-kube-api-access-jprl7" (OuterVolumeSpecName: "kube-api-access-jprl7") pod "aa76fa12-375b-4c64-926f-fc0815a7d4a6" (UID: "aa76fa12-375b-4c64-926f-fc0815a7d4a6"). InnerVolumeSpecName "kube-api-access-jprl7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.957249 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08642d69-d1bf-4816-928c-876bd2a62d2c-config-data\") pod \"nova-metadata-0\" (UID: \"08642d69-d1bf-4816-928c-876bd2a62d2c\") " pod="openstack/nova-metadata-0" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.990275 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc1a0fe0-4faf-412f-a5d3-07725b64df09-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cc1a0fe0-4faf-412f-a5d3-07725b64df09\") " pod="openstack/nova-api-0" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.996119 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc1a0fe0-4faf-412f-a5d3-07725b64df09-logs\") pod \"nova-api-0\" (UID: \"cc1a0fe0-4faf-412f-a5d3-07725b64df09\") " pod="openstack/nova-api-0" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.996269 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2d7l\" (UniqueName: \"kubernetes.io/projected/cc1a0fe0-4faf-412f-a5d3-07725b64df09-kube-api-access-l2d7l\") pod \"nova-api-0\" (UID: \"cc1a0fe0-4faf-412f-a5d3-07725b64df09\") " pod="openstack/nova-api-0" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.996382 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc1a0fe0-4faf-412f-a5d3-07725b64df09-config-data\") pod \"nova-api-0\" (UID: \"cc1a0fe0-4faf-412f-a5d3-07725b64df09\") " pod="openstack/nova-api-0" Nov 24 14:45:26 crc kubenswrapper[4790]: I1124 14:45:26.996551 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jprl7\" (UniqueName: \"kubernetes.io/projected/aa76fa12-375b-4c64-926f-fc0815a7d4a6-kube-api-access-jprl7\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:27 crc kubenswrapper[4790]: I1124 14:45:27.049385 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:45:27 crc kubenswrapper[4790]: I1124 14:45:27.085089 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa76fa12-375b-4c64-926f-fc0815a7d4a6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "aa76fa12-375b-4c64-926f-fc0815a7d4a6" (UID: "aa76fa12-375b-4c64-926f-fc0815a7d4a6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:45:27 crc kubenswrapper[4790]: I1124 14:45:27.091568 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa76fa12-375b-4c64-926f-fc0815a7d4a6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "aa76fa12-375b-4c64-926f-fc0815a7d4a6" (UID: "aa76fa12-375b-4c64-926f-fc0815a7d4a6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:45:27 crc kubenswrapper[4790]: I1124 14:45:27.101552 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2d7l\" (UniqueName: \"kubernetes.io/projected/cc1a0fe0-4faf-412f-a5d3-07725b64df09-kube-api-access-l2d7l\") pod \"nova-api-0\" (UID: \"cc1a0fe0-4faf-412f-a5d3-07725b64df09\") " pod="openstack/nova-api-0" Nov 24 14:45:27 crc kubenswrapper[4790]: I1124 14:45:27.101651 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc1a0fe0-4faf-412f-a5d3-07725b64df09-config-data\") pod \"nova-api-0\" (UID: \"cc1a0fe0-4faf-412f-a5d3-07725b64df09\") " pod="openstack/nova-api-0" Nov 24 14:45:27 crc kubenswrapper[4790]: I1124 14:45:27.101815 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc1a0fe0-4faf-412f-a5d3-07725b64df09-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cc1a0fe0-4faf-412f-a5d3-07725b64df09\") " pod="openstack/nova-api-0" Nov 24 14:45:27 crc kubenswrapper[4790]: I1124 14:45:27.101841 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc1a0fe0-4faf-412f-a5d3-07725b64df09-logs\") pod \"nova-api-0\" (UID: \"cc1a0fe0-4faf-412f-a5d3-07725b64df09\") " pod="openstack/nova-api-0" Nov 24 14:45:27 crc kubenswrapper[4790]: I1124 14:45:27.101909 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa76fa12-375b-4c64-926f-fc0815a7d4a6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:27 crc kubenswrapper[4790]: I1124 14:45:27.101920 4790 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa76fa12-375b-4c64-926f-fc0815a7d4a6-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:27 crc kubenswrapper[4790]: I1124 14:45:27.102236 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc1a0fe0-4faf-412f-a5d3-07725b64df09-logs\") pod \"nova-api-0\" (UID: \"cc1a0fe0-4faf-412f-a5d3-07725b64df09\") " pod="openstack/nova-api-0" Nov 24 14:45:27 crc kubenswrapper[4790]: I1124 14:45:27.105418 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa76fa12-375b-4c64-926f-fc0815a7d4a6-config" (OuterVolumeSpecName: "config") pod "aa76fa12-375b-4c64-926f-fc0815a7d4a6" (UID: "aa76fa12-375b-4c64-926f-fc0815a7d4a6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:45:27 crc kubenswrapper[4790]: I1124 14:45:27.105987 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc1a0fe0-4faf-412f-a5d3-07725b64df09-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cc1a0fe0-4faf-412f-a5d3-07725b64df09\") " pod="openstack/nova-api-0" Nov 24 14:45:27 crc kubenswrapper[4790]: I1124 14:45:27.106147 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc1a0fe0-4faf-412f-a5d3-07725b64df09-config-data\") pod \"nova-api-0\" (UID: \"cc1a0fe0-4faf-412f-a5d3-07725b64df09\") " pod="openstack/nova-api-0" Nov 24 14:45:27 crc kubenswrapper[4790]: I1124 14:45:27.115707 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa76fa12-375b-4c64-926f-fc0815a7d4a6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "aa76fa12-375b-4c64-926f-fc0815a7d4a6" (UID: "aa76fa12-375b-4c64-926f-fc0815a7d4a6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:45:27 crc kubenswrapper[4790]: I1124 14:45:27.117266 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2d7l\" (UniqueName: \"kubernetes.io/projected/cc1a0fe0-4faf-412f-a5d3-07725b64df09-kube-api-access-l2d7l\") pod \"nova-api-0\" (UID: \"cc1a0fe0-4faf-412f-a5d3-07725b64df09\") " pod="openstack/nova-api-0" Nov 24 14:45:27 crc kubenswrapper[4790]: I1124 14:45:27.203873 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa76fa12-375b-4c64-926f-fc0815a7d4a6-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:27 crc kubenswrapper[4790]: I1124 14:45:27.203917 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa76fa12-375b-4c64-926f-fc0815a7d4a6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:27 crc kubenswrapper[4790]: I1124 14:45:27.221440 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:45:27 crc kubenswrapper[4790]: I1124 14:45:27.522839 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:45:27 crc kubenswrapper[4790]: W1124 14:45:27.525995 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08642d69_d1bf_4816_928c_876bd2a62d2c.slice/crio-a2aff05b3fd1f8d5f38265384a0f1cdd575c6a21a177917acb8ce3252cd9466e WatchSource:0}: Error finding container a2aff05b3fd1f8d5f38265384a0f1cdd575c6a21a177917acb8ce3252cd9466e: Status 404 returned error can't find the container with id a2aff05b3fd1f8d5f38265384a0f1cdd575c6a21a177917acb8ce3252cd9466e Nov 24 14:45:27 crc kubenswrapper[4790]: I1124 14:45:27.646464 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"08642d69-d1bf-4816-928c-876bd2a62d2c","Type":"ContainerStarted","Data":"a2aff05b3fd1f8d5f38265384a0f1cdd575c6a21a177917acb8ce3252cd9466e"} Nov 24 14:45:27 crc kubenswrapper[4790]: I1124 14:45:27.649989 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76d49bc685-2dflh" Nov 24 14:45:27 crc kubenswrapper[4790]: I1124 14:45:27.649985 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76d49bc685-2dflh" event={"ID":"aa76fa12-375b-4c64-926f-fc0815a7d4a6","Type":"ContainerDied","Data":"7bd64d7b68d1f9e4ee0cf070dff4c5e9b5cd9ff9748ad304d41c4c0bebeea4a9"} Nov 24 14:45:27 crc kubenswrapper[4790]: I1124 14:45:27.650222 4790 scope.go:117] "RemoveContainer" containerID="ba0dd109b7f2187b0ccd1a652b87c9d4f6ec196a3ca0cdadadd9ff76ad6de458" Nov 24 14:45:27 crc kubenswrapper[4790]: I1124 14:45:27.650804 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:45:27 crc kubenswrapper[4790]: I1124 14:45:27.655933 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 24 14:45:27 crc kubenswrapper[4790]: I1124 14:45:27.685321 4790 scope.go:117] "RemoveContainer" containerID="894257bd0519ae9d3abb6489e66869e13da246857e81314bec7f41e7f82df3f9" Nov 24 14:45:27 crc kubenswrapper[4790]: I1124 14:45:27.693079 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76d49bc685-2dflh"] Nov 24 14:45:27 crc kubenswrapper[4790]: I1124 14:45:27.715701 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-76d49bc685-2dflh"] Nov 24 14:45:28 crc kubenswrapper[4790]: I1124 14:45:28.314567 4790 scope.go:117] "RemoveContainer" containerID="92c086e44ccb28a62777d612e499209f7483422ca70b460ce6ec73c416e8e8b8" Nov 24 14:45:28 crc kubenswrapper[4790]: E1124 14:45:28.315030 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:45:28 crc kubenswrapper[4790]: I1124 14:45:28.326576 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="691daa40-76bd-4926-928b-2e4e25f7a50c" path="/var/lib/kubelet/pods/691daa40-76bd-4926-928b-2e4e25f7a50c/volumes" Nov 24 14:45:28 crc kubenswrapper[4790]: I1124 14:45:28.329474 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa76fa12-375b-4c64-926f-fc0815a7d4a6" path="/var/lib/kubelet/pods/aa76fa12-375b-4c64-926f-fc0815a7d4a6/volumes" Nov 24 14:45:28 crc kubenswrapper[4790]: I1124 14:45:28.330514 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3811846-d6c0-47d7-8d40-93bae3eb1e49" path="/var/lib/kubelet/pods/c3811846-d6c0-47d7-8d40-93bae3eb1e49/volumes" Nov 24 14:45:28 crc kubenswrapper[4790]: I1124 14:45:28.670327 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cc1a0fe0-4faf-412f-a5d3-07725b64df09","Type":"ContainerStarted","Data":"0f4d8c1d4bbda20530690ce310bf02b9c1520de832d550f6e6e9cb7c8562249d"} Nov 24 14:45:28 crc kubenswrapper[4790]: I1124 14:45:28.670373 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cc1a0fe0-4faf-412f-a5d3-07725b64df09","Type":"ContainerStarted","Data":"d7d3f25a0c0ea33125657769d27d12e5822cf9c22a591dfcc4cfcc7f19e7cd10"} Nov 24 14:45:28 crc kubenswrapper[4790]: I1124 14:45:28.670386 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cc1a0fe0-4faf-412f-a5d3-07725b64df09","Type":"ContainerStarted","Data":"b23d342f912188b62572113722897eb5d7d632206d6de57ad6c1c1ab04076f08"} Nov 24 14:45:28 crc kubenswrapper[4790]: I1124 14:45:28.675762 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"08642d69-d1bf-4816-928c-876bd2a62d2c","Type":"ContainerStarted","Data":"0b1fe97c04a07f9803771cb466a280da626a12321ff3d6f79680c0e4be77e10f"} Nov 24 14:45:28 crc kubenswrapper[4790]: I1124 14:45:28.675818 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"08642d69-d1bf-4816-928c-876bd2a62d2c","Type":"ContainerStarted","Data":"4687665d607da921a4d0738e0c70d3b8677b74d18cfbf5b6cc62a7ed05fd20ed"} Nov 24 14:45:28 crc kubenswrapper[4790]: I1124 14:45:28.699924 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.699869027 podStartE2EDuration="2.699869027s" podCreationTimestamp="2025-11-24 14:45:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:45:28.688751838 +0000 UTC m=+5577.068645500" watchObservedRunningTime="2025-11-24 14:45:28.699869027 +0000 UTC m=+5577.079762699" Nov 24 14:45:28 crc kubenswrapper[4790]: I1124 14:45:28.719103 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.719085659 podStartE2EDuration="2.719085659s" podCreationTimestamp="2025-11-24 14:45:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:45:28.713957212 +0000 UTC m=+5577.093850884" watchObservedRunningTime="2025-11-24 14:45:28.719085659 +0000 UTC m=+5577.098979311" Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.225041 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.345995 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msfzt\" (UniqueName: \"kubernetes.io/projected/0144dbfa-e743-47eb-b9ab-8e3610fc3130-kube-api-access-msfzt\") pod \"0144dbfa-e743-47eb-b9ab-8e3610fc3130\" (UID: \"0144dbfa-e743-47eb-b9ab-8e3610fc3130\") " Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.346177 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0144dbfa-e743-47eb-b9ab-8e3610fc3130-combined-ca-bundle\") pod \"0144dbfa-e743-47eb-b9ab-8e3610fc3130\" (UID: \"0144dbfa-e743-47eb-b9ab-8e3610fc3130\") " Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.346246 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0144dbfa-e743-47eb-b9ab-8e3610fc3130-config-data\") pod \"0144dbfa-e743-47eb-b9ab-8e3610fc3130\" (UID: \"0144dbfa-e743-47eb-b9ab-8e3610fc3130\") " Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.354323 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0144dbfa-e743-47eb-b9ab-8e3610fc3130-kube-api-access-msfzt" (OuterVolumeSpecName: "kube-api-access-msfzt") pod "0144dbfa-e743-47eb-b9ab-8e3610fc3130" (UID: "0144dbfa-e743-47eb-b9ab-8e3610fc3130"). InnerVolumeSpecName "kube-api-access-msfzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.381497 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0144dbfa-e743-47eb-b9ab-8e3610fc3130-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0144dbfa-e743-47eb-b9ab-8e3610fc3130" (UID: "0144dbfa-e743-47eb-b9ab-8e3610fc3130"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.393745 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0144dbfa-e743-47eb-b9ab-8e3610fc3130-config-data" (OuterVolumeSpecName: "config-data") pod "0144dbfa-e743-47eb-b9ab-8e3610fc3130" (UID: "0144dbfa-e743-47eb-b9ab-8e3610fc3130"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.448787 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0144dbfa-e743-47eb-b9ab-8e3610fc3130-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.449195 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0144dbfa-e743-47eb-b9ab-8e3610fc3130-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.449206 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msfzt\" (UniqueName: \"kubernetes.io/projected/0144dbfa-e743-47eb-b9ab-8e3610fc3130-kube-api-access-msfzt\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.688296 4790 generic.go:334] "Generic (PLEG): container finished" podID="0144dbfa-e743-47eb-b9ab-8e3610fc3130" containerID="c0c89e471a403e94f493268f25fac09952986b15ccdb311c78f230f2f42e6b2b" exitCode=0 Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.688374 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0144dbfa-e743-47eb-b9ab-8e3610fc3130","Type":"ContainerDied","Data":"c0c89e471a403e94f493268f25fac09952986b15ccdb311c78f230f2f42e6b2b"} Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.689212 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0144dbfa-e743-47eb-b9ab-8e3610fc3130","Type":"ContainerDied","Data":"8eabfa899ecc915f6a24fe03691b3cb2e6bb7fae3a0e1456438934a7131aed52"} Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.689275 4790 scope.go:117] "RemoveContainer" containerID="c0c89e471a403e94f493268f25fac09952986b15ccdb311c78f230f2f42e6b2b" Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.688523 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.728801 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.729956 4790 scope.go:117] "RemoveContainer" containerID="c0c89e471a403e94f493268f25fac09952986b15ccdb311c78f230f2f42e6b2b" Nov 24 14:45:29 crc kubenswrapper[4790]: E1124 14:45:29.730729 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0c89e471a403e94f493268f25fac09952986b15ccdb311c78f230f2f42e6b2b\": container with ID starting with c0c89e471a403e94f493268f25fac09952986b15ccdb311c78f230f2f42e6b2b not found: ID does not exist" containerID="c0c89e471a403e94f493268f25fac09952986b15ccdb311c78f230f2f42e6b2b" Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.730778 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0c89e471a403e94f493268f25fac09952986b15ccdb311c78f230f2f42e6b2b"} err="failed to get container status \"c0c89e471a403e94f493268f25fac09952986b15ccdb311c78f230f2f42e6b2b\": rpc error: code = NotFound desc = could not find container \"c0c89e471a403e94f493268f25fac09952986b15ccdb311c78f230f2f42e6b2b\": container with ID starting with c0c89e471a403e94f493268f25fac09952986b15ccdb311c78f230f2f42e6b2b not found: ID does not exist" Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.747128 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.765833 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:45:29 crc kubenswrapper[4790]: E1124 14:45:29.766751 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0144dbfa-e743-47eb-b9ab-8e3610fc3130" containerName="nova-scheduler-scheduler" Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.766776 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="0144dbfa-e743-47eb-b9ab-8e3610fc3130" containerName="nova-scheduler-scheduler" Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.767210 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="0144dbfa-e743-47eb-b9ab-8e3610fc3130" containerName="nova-scheduler-scheduler" Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.768256 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.771925 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.780803 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.856949 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70de12b3-1a25-476e-b0ae-1681385393a0-config-data\") pod \"nova-scheduler-0\" (UID: \"70de12b3-1a25-476e-b0ae-1681385393a0\") " pod="openstack/nova-scheduler-0" Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.857051 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70de12b3-1a25-476e-b0ae-1681385393a0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"70de12b3-1a25-476e-b0ae-1681385393a0\") " pod="openstack/nova-scheduler-0" Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.857209 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jth5m\" (UniqueName: \"kubernetes.io/projected/70de12b3-1a25-476e-b0ae-1681385393a0-kube-api-access-jth5m\") pod \"nova-scheduler-0\" (UID: \"70de12b3-1a25-476e-b0ae-1681385393a0\") " pod="openstack/nova-scheduler-0" Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.953025 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-42ld4" Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.961349 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jth5m\" (UniqueName: \"kubernetes.io/projected/70de12b3-1a25-476e-b0ae-1681385393a0-kube-api-access-jth5m\") pod \"nova-scheduler-0\" (UID: \"70de12b3-1a25-476e-b0ae-1681385393a0\") " pod="openstack/nova-scheduler-0" Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.961526 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70de12b3-1a25-476e-b0ae-1681385393a0-config-data\") pod \"nova-scheduler-0\" (UID: \"70de12b3-1a25-476e-b0ae-1681385393a0\") " pod="openstack/nova-scheduler-0" Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.961573 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70de12b3-1a25-476e-b0ae-1681385393a0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"70de12b3-1a25-476e-b0ae-1681385393a0\") " pod="openstack/nova-scheduler-0" Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.967175 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70de12b3-1a25-476e-b0ae-1681385393a0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"70de12b3-1a25-476e-b0ae-1681385393a0\") " pod="openstack/nova-scheduler-0" Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.975274 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70de12b3-1a25-476e-b0ae-1681385393a0-config-data\") pod \"nova-scheduler-0\" (UID: \"70de12b3-1a25-476e-b0ae-1681385393a0\") " pod="openstack/nova-scheduler-0" Nov 24 14:45:29 crc kubenswrapper[4790]: I1124 14:45:29.981728 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jth5m\" (UniqueName: \"kubernetes.io/projected/70de12b3-1a25-476e-b0ae-1681385393a0-kube-api-access-jth5m\") pod \"nova-scheduler-0\" (UID: \"70de12b3-1a25-476e-b0ae-1681385393a0\") " pod="openstack/nova-scheduler-0" Nov 24 14:45:30 crc kubenswrapper[4790]: I1124 14:45:30.022339 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-42ld4" Nov 24 14:45:30 crc kubenswrapper[4790]: I1124 14:45:30.056650 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 24 14:45:30 crc kubenswrapper[4790]: I1124 14:45:30.102547 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:45:30 crc kubenswrapper[4790]: I1124 14:45:30.196508 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-42ld4"] Nov 24 14:45:30 crc kubenswrapper[4790]: I1124 14:45:30.333999 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0144dbfa-e743-47eb-b9ab-8e3610fc3130" path="/var/lib/kubelet/pods/0144dbfa-e743-47eb-b9ab-8e3610fc3130/volumes" Nov 24 14:45:30 crc kubenswrapper[4790]: I1124 14:45:30.387008 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:45:30 crc kubenswrapper[4790]: I1124 14:45:30.550743 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-68pwg"] Nov 24 14:45:30 crc kubenswrapper[4790]: I1124 14:45:30.552868 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-68pwg" Nov 24 14:45:30 crc kubenswrapper[4790]: I1124 14:45:30.554505 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 24 14:45:30 crc kubenswrapper[4790]: I1124 14:45:30.554827 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 24 14:45:30 crc kubenswrapper[4790]: I1124 14:45:30.572405 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-68pwg"] Nov 24 14:45:30 crc kubenswrapper[4790]: I1124 14:45:30.678180 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7581458b-2c21-4297-af62-05fd707bb201-config-data\") pod \"nova-cell1-cell-mapping-68pwg\" (UID: \"7581458b-2c21-4297-af62-05fd707bb201\") " pod="openstack/nova-cell1-cell-mapping-68pwg" Nov 24 14:45:30 crc kubenswrapper[4790]: I1124 14:45:30.678246 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76qcz\" (UniqueName: \"kubernetes.io/projected/7581458b-2c21-4297-af62-05fd707bb201-kube-api-access-76qcz\") pod \"nova-cell1-cell-mapping-68pwg\" (UID: \"7581458b-2c21-4297-af62-05fd707bb201\") " pod="openstack/nova-cell1-cell-mapping-68pwg" Nov 24 14:45:30 crc kubenswrapper[4790]: I1124 14:45:30.678317 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7581458b-2c21-4297-af62-05fd707bb201-scripts\") pod \"nova-cell1-cell-mapping-68pwg\" (UID: \"7581458b-2c21-4297-af62-05fd707bb201\") " pod="openstack/nova-cell1-cell-mapping-68pwg" Nov 24 14:45:30 crc kubenswrapper[4790]: I1124 14:45:30.678388 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7581458b-2c21-4297-af62-05fd707bb201-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-68pwg\" (UID: \"7581458b-2c21-4297-af62-05fd707bb201\") " pod="openstack/nova-cell1-cell-mapping-68pwg" Nov 24 14:45:30 crc kubenswrapper[4790]: I1124 14:45:30.701826 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"70de12b3-1a25-476e-b0ae-1681385393a0","Type":"ContainerStarted","Data":"414a4472082056f08a9d288943aa88ce18cf717766968d2a4df5d5196c782b00"} Nov 24 14:45:30 crc kubenswrapper[4790]: I1124 14:45:30.701889 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"70de12b3-1a25-476e-b0ae-1681385393a0","Type":"ContainerStarted","Data":"deb58a818e40aa97b61f043a15f58ced4d888a22f598bd361642e888352efb15"} Nov 24 14:45:30 crc kubenswrapper[4790]: I1124 14:45:30.735502 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.7354609829999998 podStartE2EDuration="1.735460983s" podCreationTimestamp="2025-11-24 14:45:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:45:30.72214832 +0000 UTC m=+5579.102041992" watchObservedRunningTime="2025-11-24 14:45:30.735460983 +0000 UTC m=+5579.115354655" Nov 24 14:45:30 crc kubenswrapper[4790]: I1124 14:45:30.783383 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7581458b-2c21-4297-af62-05fd707bb201-scripts\") pod \"nova-cell1-cell-mapping-68pwg\" (UID: \"7581458b-2c21-4297-af62-05fd707bb201\") " pod="openstack/nova-cell1-cell-mapping-68pwg" Nov 24 14:45:30 crc kubenswrapper[4790]: I1124 14:45:30.783586 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7581458b-2c21-4297-af62-05fd707bb201-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-68pwg\" (UID: \"7581458b-2c21-4297-af62-05fd707bb201\") " pod="openstack/nova-cell1-cell-mapping-68pwg" Nov 24 14:45:30 crc kubenswrapper[4790]: I1124 14:45:30.785065 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7581458b-2c21-4297-af62-05fd707bb201-config-data\") pod \"nova-cell1-cell-mapping-68pwg\" (UID: \"7581458b-2c21-4297-af62-05fd707bb201\") " pod="openstack/nova-cell1-cell-mapping-68pwg" Nov 24 14:45:30 crc kubenswrapper[4790]: I1124 14:45:30.785126 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76qcz\" (UniqueName: \"kubernetes.io/projected/7581458b-2c21-4297-af62-05fd707bb201-kube-api-access-76qcz\") pod \"nova-cell1-cell-mapping-68pwg\" (UID: \"7581458b-2c21-4297-af62-05fd707bb201\") " pod="openstack/nova-cell1-cell-mapping-68pwg" Nov 24 14:45:30 crc kubenswrapper[4790]: I1124 14:45:30.787605 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7581458b-2c21-4297-af62-05fd707bb201-scripts\") pod \"nova-cell1-cell-mapping-68pwg\" (UID: \"7581458b-2c21-4297-af62-05fd707bb201\") " pod="openstack/nova-cell1-cell-mapping-68pwg" Nov 24 14:45:30 crc kubenswrapper[4790]: I1124 14:45:30.787850 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7581458b-2c21-4297-af62-05fd707bb201-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-68pwg\" (UID: \"7581458b-2c21-4297-af62-05fd707bb201\") " pod="openstack/nova-cell1-cell-mapping-68pwg" Nov 24 14:45:30 crc kubenswrapper[4790]: I1124 14:45:30.788854 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7581458b-2c21-4297-af62-05fd707bb201-config-data\") pod \"nova-cell1-cell-mapping-68pwg\" (UID: \"7581458b-2c21-4297-af62-05fd707bb201\") " pod="openstack/nova-cell1-cell-mapping-68pwg" Nov 24 14:45:30 crc kubenswrapper[4790]: I1124 14:45:30.805538 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76qcz\" (UniqueName: \"kubernetes.io/projected/7581458b-2c21-4297-af62-05fd707bb201-kube-api-access-76qcz\") pod \"nova-cell1-cell-mapping-68pwg\" (UID: \"7581458b-2c21-4297-af62-05fd707bb201\") " pod="openstack/nova-cell1-cell-mapping-68pwg" Nov 24 14:45:30 crc kubenswrapper[4790]: I1124 14:45:30.886382 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-68pwg" Nov 24 14:45:31 crc kubenswrapper[4790]: I1124 14:45:31.400979 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-68pwg"] Nov 24 14:45:31 crc kubenswrapper[4790]: W1124 14:45:31.408768 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7581458b_2c21_4297_af62_05fd707bb201.slice/crio-31a0a7a429f4e62f97eec2ca7070cda7e2421fba237d562360e52f4c0e4ebf9d WatchSource:0}: Error finding container 31a0a7a429f4e62f97eec2ca7070cda7e2421fba237d562360e52f4c0e4ebf9d: Status 404 returned error can't find the container with id 31a0a7a429f4e62f97eec2ca7070cda7e2421fba237d562360e52f4c0e4ebf9d Nov 24 14:45:31 crc kubenswrapper[4790]: I1124 14:45:31.720500 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-42ld4" podUID="58a4a77b-8070-4da1-b716-2e9796bab87a" containerName="registry-server" containerID="cri-o://bb87531f06dd0e96e23905f5d34428dfa751ec6106733c2c510dce7a6cab644e" gracePeriod=2 Nov 24 14:45:31 crc kubenswrapper[4790]: I1124 14:45:31.722283 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-68pwg" event={"ID":"7581458b-2c21-4297-af62-05fd707bb201","Type":"ContainerStarted","Data":"7617c44da6fa13317bdbd19520037809f78c2aa5cc3ee058ec952ff2b9802314"} Nov 24 14:45:31 crc kubenswrapper[4790]: I1124 14:45:31.722326 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-68pwg" event={"ID":"7581458b-2c21-4297-af62-05fd707bb201","Type":"ContainerStarted","Data":"31a0a7a429f4e62f97eec2ca7070cda7e2421fba237d562360e52f4c0e4ebf9d"} Nov 24 14:45:31 crc kubenswrapper[4790]: I1124 14:45:31.755752 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-68pwg" podStartSLOduration=1.755725787 podStartE2EDuration="1.755725787s" podCreationTimestamp="2025-11-24 14:45:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:45:31.754609865 +0000 UTC m=+5580.134503557" watchObservedRunningTime="2025-11-24 14:45:31.755725787 +0000 UTC m=+5580.135619469" Nov 24 14:45:32 crc kubenswrapper[4790]: I1124 14:45:32.050210 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 14:45:32 crc kubenswrapper[4790]: I1124 14:45:32.050706 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 14:45:32 crc kubenswrapper[4790]: I1124 14:45:32.311522 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-42ld4" Nov 24 14:45:32 crc kubenswrapper[4790]: I1124 14:45:32.421230 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jx5d2\" (UniqueName: \"kubernetes.io/projected/58a4a77b-8070-4da1-b716-2e9796bab87a-kube-api-access-jx5d2\") pod \"58a4a77b-8070-4da1-b716-2e9796bab87a\" (UID: \"58a4a77b-8070-4da1-b716-2e9796bab87a\") " Nov 24 14:45:32 crc kubenswrapper[4790]: I1124 14:45:32.421436 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58a4a77b-8070-4da1-b716-2e9796bab87a-catalog-content\") pod \"58a4a77b-8070-4da1-b716-2e9796bab87a\" (UID: \"58a4a77b-8070-4da1-b716-2e9796bab87a\") " Nov 24 14:45:32 crc kubenswrapper[4790]: I1124 14:45:32.421458 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58a4a77b-8070-4da1-b716-2e9796bab87a-utilities\") pod \"58a4a77b-8070-4da1-b716-2e9796bab87a\" (UID: \"58a4a77b-8070-4da1-b716-2e9796bab87a\") " Nov 24 14:45:32 crc kubenswrapper[4790]: I1124 14:45:32.422574 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58a4a77b-8070-4da1-b716-2e9796bab87a-utilities" (OuterVolumeSpecName: "utilities") pod "58a4a77b-8070-4da1-b716-2e9796bab87a" (UID: "58a4a77b-8070-4da1-b716-2e9796bab87a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:45:32 crc kubenswrapper[4790]: I1124 14:45:32.438059 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58a4a77b-8070-4da1-b716-2e9796bab87a-kube-api-access-jx5d2" (OuterVolumeSpecName: "kube-api-access-jx5d2") pod "58a4a77b-8070-4da1-b716-2e9796bab87a" (UID: "58a4a77b-8070-4da1-b716-2e9796bab87a"). InnerVolumeSpecName "kube-api-access-jx5d2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:45:32 crc kubenswrapper[4790]: I1124 14:45:32.521010 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58a4a77b-8070-4da1-b716-2e9796bab87a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "58a4a77b-8070-4da1-b716-2e9796bab87a" (UID: "58a4a77b-8070-4da1-b716-2e9796bab87a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:45:32 crc kubenswrapper[4790]: I1124 14:45:32.523694 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58a4a77b-8070-4da1-b716-2e9796bab87a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:32 crc kubenswrapper[4790]: I1124 14:45:32.523735 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58a4a77b-8070-4da1-b716-2e9796bab87a-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:32 crc kubenswrapper[4790]: I1124 14:45:32.523752 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jx5d2\" (UniqueName: \"kubernetes.io/projected/58a4a77b-8070-4da1-b716-2e9796bab87a-kube-api-access-jx5d2\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:32 crc kubenswrapper[4790]: I1124 14:45:32.738941 4790 generic.go:334] "Generic (PLEG): container finished" podID="58a4a77b-8070-4da1-b716-2e9796bab87a" containerID="bb87531f06dd0e96e23905f5d34428dfa751ec6106733c2c510dce7a6cab644e" exitCode=0 Nov 24 14:45:32 crc kubenswrapper[4790]: I1124 14:45:32.739048 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-42ld4" Nov 24 14:45:32 crc kubenswrapper[4790]: I1124 14:45:32.739073 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-42ld4" event={"ID":"58a4a77b-8070-4da1-b716-2e9796bab87a","Type":"ContainerDied","Data":"bb87531f06dd0e96e23905f5d34428dfa751ec6106733c2c510dce7a6cab644e"} Nov 24 14:45:32 crc kubenswrapper[4790]: I1124 14:45:32.741404 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-42ld4" event={"ID":"58a4a77b-8070-4da1-b716-2e9796bab87a","Type":"ContainerDied","Data":"786a9011ea182e078a38a82760eeeaf841d73924ab13d9b4cf7ae88aed4ae13f"} Nov 24 14:45:32 crc kubenswrapper[4790]: I1124 14:45:32.741458 4790 scope.go:117] "RemoveContainer" containerID="bb87531f06dd0e96e23905f5d34428dfa751ec6106733c2c510dce7a6cab644e" Nov 24 14:45:32 crc kubenswrapper[4790]: I1124 14:45:32.805217 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-42ld4"] Nov 24 14:45:32 crc kubenswrapper[4790]: I1124 14:45:32.810647 4790 scope.go:117] "RemoveContainer" containerID="c9cbeb2276a393a2b9a8605885a47db2d9629fa10b83c6598627848a82f691b9" Nov 24 14:45:32 crc kubenswrapper[4790]: I1124 14:45:32.814920 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-42ld4"] Nov 24 14:45:32 crc kubenswrapper[4790]: I1124 14:45:32.844593 4790 scope.go:117] "RemoveContainer" containerID="3366f4e671bcb53b61230ffba61e286b643d010934001965ccfb6a59d8415c8d" Nov 24 14:45:32 crc kubenswrapper[4790]: I1124 14:45:32.897547 4790 scope.go:117] "RemoveContainer" containerID="bb87531f06dd0e96e23905f5d34428dfa751ec6106733c2c510dce7a6cab644e" Nov 24 14:45:32 crc kubenswrapper[4790]: E1124 14:45:32.898189 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb87531f06dd0e96e23905f5d34428dfa751ec6106733c2c510dce7a6cab644e\": container with ID starting with bb87531f06dd0e96e23905f5d34428dfa751ec6106733c2c510dce7a6cab644e not found: ID does not exist" containerID="bb87531f06dd0e96e23905f5d34428dfa751ec6106733c2c510dce7a6cab644e" Nov 24 14:45:32 crc kubenswrapper[4790]: I1124 14:45:32.898250 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb87531f06dd0e96e23905f5d34428dfa751ec6106733c2c510dce7a6cab644e"} err="failed to get container status \"bb87531f06dd0e96e23905f5d34428dfa751ec6106733c2c510dce7a6cab644e\": rpc error: code = NotFound desc = could not find container \"bb87531f06dd0e96e23905f5d34428dfa751ec6106733c2c510dce7a6cab644e\": container with ID starting with bb87531f06dd0e96e23905f5d34428dfa751ec6106733c2c510dce7a6cab644e not found: ID does not exist" Nov 24 14:45:32 crc kubenswrapper[4790]: I1124 14:45:32.898275 4790 scope.go:117] "RemoveContainer" containerID="c9cbeb2276a393a2b9a8605885a47db2d9629fa10b83c6598627848a82f691b9" Nov 24 14:45:32 crc kubenswrapper[4790]: E1124 14:45:32.898827 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9cbeb2276a393a2b9a8605885a47db2d9629fa10b83c6598627848a82f691b9\": container with ID starting with c9cbeb2276a393a2b9a8605885a47db2d9629fa10b83c6598627848a82f691b9 not found: ID does not exist" containerID="c9cbeb2276a393a2b9a8605885a47db2d9629fa10b83c6598627848a82f691b9" Nov 24 14:45:32 crc kubenswrapper[4790]: I1124 14:45:32.898936 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9cbeb2276a393a2b9a8605885a47db2d9629fa10b83c6598627848a82f691b9"} err="failed to get container status \"c9cbeb2276a393a2b9a8605885a47db2d9629fa10b83c6598627848a82f691b9\": rpc error: code = NotFound desc = could not find container \"c9cbeb2276a393a2b9a8605885a47db2d9629fa10b83c6598627848a82f691b9\": container with ID starting with c9cbeb2276a393a2b9a8605885a47db2d9629fa10b83c6598627848a82f691b9 not found: ID does not exist" Nov 24 14:45:32 crc kubenswrapper[4790]: I1124 14:45:32.899214 4790 scope.go:117] "RemoveContainer" containerID="3366f4e671bcb53b61230ffba61e286b643d010934001965ccfb6a59d8415c8d" Nov 24 14:45:32 crc kubenswrapper[4790]: E1124 14:45:32.899715 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3366f4e671bcb53b61230ffba61e286b643d010934001965ccfb6a59d8415c8d\": container with ID starting with 3366f4e671bcb53b61230ffba61e286b643d010934001965ccfb6a59d8415c8d not found: ID does not exist" containerID="3366f4e671bcb53b61230ffba61e286b643d010934001965ccfb6a59d8415c8d" Nov 24 14:45:32 crc kubenswrapper[4790]: I1124 14:45:32.899760 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3366f4e671bcb53b61230ffba61e286b643d010934001965ccfb6a59d8415c8d"} err="failed to get container status \"3366f4e671bcb53b61230ffba61e286b643d010934001965ccfb6a59d8415c8d\": rpc error: code = NotFound desc = could not find container \"3366f4e671bcb53b61230ffba61e286b643d010934001965ccfb6a59d8415c8d\": container with ID starting with 3366f4e671bcb53b61230ffba61e286b643d010934001965ccfb6a59d8415c8d not found: ID does not exist" Nov 24 14:45:34 crc kubenswrapper[4790]: I1124 14:45:34.332666 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58a4a77b-8070-4da1-b716-2e9796bab87a" path="/var/lib/kubelet/pods/58a4a77b-8070-4da1-b716-2e9796bab87a/volumes" Nov 24 14:45:35 crc kubenswrapper[4790]: I1124 14:45:35.102899 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 14:45:36 crc kubenswrapper[4790]: I1124 14:45:36.793587 4790 generic.go:334] "Generic (PLEG): container finished" podID="7581458b-2c21-4297-af62-05fd707bb201" containerID="7617c44da6fa13317bdbd19520037809f78c2aa5cc3ee058ec952ff2b9802314" exitCode=0 Nov 24 14:45:36 crc kubenswrapper[4790]: I1124 14:45:36.793734 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-68pwg" event={"ID":"7581458b-2c21-4297-af62-05fd707bb201","Type":"ContainerDied","Data":"7617c44da6fa13317bdbd19520037809f78c2aa5cc3ee058ec952ff2b9802314"} Nov 24 14:45:37 crc kubenswrapper[4790]: I1124 14:45:37.051063 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 14:45:37 crc kubenswrapper[4790]: I1124 14:45:37.051172 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 14:45:37 crc kubenswrapper[4790]: I1124 14:45:37.222705 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 14:45:37 crc kubenswrapper[4790]: I1124 14:45:37.222767 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 14:45:38 crc kubenswrapper[4790]: I1124 14:45:38.134084 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="08642d69-d1bf-4816-928c-876bd2a62d2c" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.69:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 14:45:38 crc kubenswrapper[4790]: I1124 14:45:38.134142 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="08642d69-d1bf-4816-928c-876bd2a62d2c" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.69:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 14:45:38 crc kubenswrapper[4790]: I1124 14:45:38.210796 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-68pwg" Nov 24 14:45:38 crc kubenswrapper[4790]: I1124 14:45:38.306069 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="cc1a0fe0-4faf-412f-a5d3-07725b64df09" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.70:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 14:45:38 crc kubenswrapper[4790]: I1124 14:45:38.306069 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="cc1a0fe0-4faf-412f-a5d3-07725b64df09" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.70:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 14:45:38 crc kubenswrapper[4790]: I1124 14:45:38.352254 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7581458b-2c21-4297-af62-05fd707bb201-scripts\") pod \"7581458b-2c21-4297-af62-05fd707bb201\" (UID: \"7581458b-2c21-4297-af62-05fd707bb201\") " Nov 24 14:45:38 crc kubenswrapper[4790]: I1124 14:45:38.352433 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7581458b-2c21-4297-af62-05fd707bb201-config-data\") pod \"7581458b-2c21-4297-af62-05fd707bb201\" (UID: \"7581458b-2c21-4297-af62-05fd707bb201\") " Nov 24 14:45:38 crc kubenswrapper[4790]: I1124 14:45:38.352471 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76qcz\" (UniqueName: \"kubernetes.io/projected/7581458b-2c21-4297-af62-05fd707bb201-kube-api-access-76qcz\") pod \"7581458b-2c21-4297-af62-05fd707bb201\" (UID: \"7581458b-2c21-4297-af62-05fd707bb201\") " Nov 24 14:45:38 crc kubenswrapper[4790]: I1124 14:45:38.352518 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7581458b-2c21-4297-af62-05fd707bb201-combined-ca-bundle\") pod \"7581458b-2c21-4297-af62-05fd707bb201\" (UID: \"7581458b-2c21-4297-af62-05fd707bb201\") " Nov 24 14:45:38 crc kubenswrapper[4790]: I1124 14:45:38.358608 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7581458b-2c21-4297-af62-05fd707bb201-kube-api-access-76qcz" (OuterVolumeSpecName: "kube-api-access-76qcz") pod "7581458b-2c21-4297-af62-05fd707bb201" (UID: "7581458b-2c21-4297-af62-05fd707bb201"). InnerVolumeSpecName "kube-api-access-76qcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:45:38 crc kubenswrapper[4790]: I1124 14:45:38.367090 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7581458b-2c21-4297-af62-05fd707bb201-scripts" (OuterVolumeSpecName: "scripts") pod "7581458b-2c21-4297-af62-05fd707bb201" (UID: "7581458b-2c21-4297-af62-05fd707bb201"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:45:38 crc kubenswrapper[4790]: I1124 14:45:38.386903 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7581458b-2c21-4297-af62-05fd707bb201-config-data" (OuterVolumeSpecName: "config-data") pod "7581458b-2c21-4297-af62-05fd707bb201" (UID: "7581458b-2c21-4297-af62-05fd707bb201"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:45:38 crc kubenswrapper[4790]: I1124 14:45:38.389238 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7581458b-2c21-4297-af62-05fd707bb201-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7581458b-2c21-4297-af62-05fd707bb201" (UID: "7581458b-2c21-4297-af62-05fd707bb201"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:45:38 crc kubenswrapper[4790]: I1124 14:45:38.455222 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7581458b-2c21-4297-af62-05fd707bb201-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:38 crc kubenswrapper[4790]: I1124 14:45:38.455260 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7581458b-2c21-4297-af62-05fd707bb201-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:38 crc kubenswrapper[4790]: I1124 14:45:38.455271 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76qcz\" (UniqueName: \"kubernetes.io/projected/7581458b-2c21-4297-af62-05fd707bb201-kube-api-access-76qcz\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:38 crc kubenswrapper[4790]: I1124 14:45:38.455281 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7581458b-2c21-4297-af62-05fd707bb201-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:38 crc kubenswrapper[4790]: I1124 14:45:38.834128 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-68pwg" event={"ID":"7581458b-2c21-4297-af62-05fd707bb201","Type":"ContainerDied","Data":"31a0a7a429f4e62f97eec2ca7070cda7e2421fba237d562360e52f4c0e4ebf9d"} Nov 24 14:45:38 crc kubenswrapper[4790]: I1124 14:45:38.834444 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31a0a7a429f4e62f97eec2ca7070cda7e2421fba237d562360e52f4c0e4ebf9d" Nov 24 14:45:38 crc kubenswrapper[4790]: I1124 14:45:38.834516 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-68pwg" Nov 24 14:45:39 crc kubenswrapper[4790]: I1124 14:45:39.010582 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:45:39 crc kubenswrapper[4790]: I1124 14:45:39.010794 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="cc1a0fe0-4faf-412f-a5d3-07725b64df09" containerName="nova-api-log" containerID="cri-o://d7d3f25a0c0ea33125657769d27d12e5822cf9c22a591dfcc4cfcc7f19e7cd10" gracePeriod=30 Nov 24 14:45:39 crc kubenswrapper[4790]: I1124 14:45:39.010959 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="cc1a0fe0-4faf-412f-a5d3-07725b64df09" containerName="nova-api-api" containerID="cri-o://0f4d8c1d4bbda20530690ce310bf02b9c1520de832d550f6e6e9cb7c8562249d" gracePeriod=30 Nov 24 14:45:39 crc kubenswrapper[4790]: I1124 14:45:39.043324 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:45:39 crc kubenswrapper[4790]: I1124 14:45:39.043516 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="70de12b3-1a25-476e-b0ae-1681385393a0" containerName="nova-scheduler-scheduler" containerID="cri-o://414a4472082056f08a9d288943aa88ce18cf717766968d2a4df5d5196c782b00" gracePeriod=30 Nov 24 14:45:39 crc kubenswrapper[4790]: I1124 14:45:39.064106 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:45:39 crc kubenswrapper[4790]: I1124 14:45:39.064336 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="08642d69-d1bf-4816-928c-876bd2a62d2c" containerName="nova-metadata-log" containerID="cri-o://4687665d607da921a4d0738e0c70d3b8677b74d18cfbf5b6cc62a7ed05fd20ed" gracePeriod=30 Nov 24 14:45:39 crc kubenswrapper[4790]: I1124 14:45:39.064769 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="08642d69-d1bf-4816-928c-876bd2a62d2c" containerName="nova-metadata-metadata" containerID="cri-o://0b1fe97c04a07f9803771cb466a280da626a12321ff3d6f79680c0e4be77e10f" gracePeriod=30 Nov 24 14:45:39 crc kubenswrapper[4790]: I1124 14:45:39.845475 4790 generic.go:334] "Generic (PLEG): container finished" podID="08642d69-d1bf-4816-928c-876bd2a62d2c" containerID="4687665d607da921a4d0738e0c70d3b8677b74d18cfbf5b6cc62a7ed05fd20ed" exitCode=143 Nov 24 14:45:39 crc kubenswrapper[4790]: I1124 14:45:39.845578 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"08642d69-d1bf-4816-928c-876bd2a62d2c","Type":"ContainerDied","Data":"4687665d607da921a4d0738e0c70d3b8677b74d18cfbf5b6cc62a7ed05fd20ed"} Nov 24 14:45:39 crc kubenswrapper[4790]: I1124 14:45:39.848626 4790 generic.go:334] "Generic (PLEG): container finished" podID="cc1a0fe0-4faf-412f-a5d3-07725b64df09" containerID="d7d3f25a0c0ea33125657769d27d12e5822cf9c22a591dfcc4cfcc7f19e7cd10" exitCode=143 Nov 24 14:45:39 crc kubenswrapper[4790]: I1124 14:45:39.848700 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cc1a0fe0-4faf-412f-a5d3-07725b64df09","Type":"ContainerDied","Data":"d7d3f25a0c0ea33125657769d27d12e5822cf9c22a591dfcc4cfcc7f19e7cd10"} Nov 24 14:45:41 crc kubenswrapper[4790]: I1124 14:45:41.873764 4790 generic.go:334] "Generic (PLEG): container finished" podID="70de12b3-1a25-476e-b0ae-1681385393a0" containerID="414a4472082056f08a9d288943aa88ce18cf717766968d2a4df5d5196c782b00" exitCode=0 Nov 24 14:45:41 crc kubenswrapper[4790]: I1124 14:45:41.873936 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"70de12b3-1a25-476e-b0ae-1681385393a0","Type":"ContainerDied","Data":"414a4472082056f08a9d288943aa88ce18cf717766968d2a4df5d5196c782b00"} Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.120633 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.238840 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jth5m\" (UniqueName: \"kubernetes.io/projected/70de12b3-1a25-476e-b0ae-1681385393a0-kube-api-access-jth5m\") pod \"70de12b3-1a25-476e-b0ae-1681385393a0\" (UID: \"70de12b3-1a25-476e-b0ae-1681385393a0\") " Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.238971 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70de12b3-1a25-476e-b0ae-1681385393a0-combined-ca-bundle\") pod \"70de12b3-1a25-476e-b0ae-1681385393a0\" (UID: \"70de12b3-1a25-476e-b0ae-1681385393a0\") " Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.239059 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70de12b3-1a25-476e-b0ae-1681385393a0-config-data\") pod \"70de12b3-1a25-476e-b0ae-1681385393a0\" (UID: \"70de12b3-1a25-476e-b0ae-1681385393a0\") " Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.245129 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70de12b3-1a25-476e-b0ae-1681385393a0-kube-api-access-jth5m" (OuterVolumeSpecName: "kube-api-access-jth5m") pod "70de12b3-1a25-476e-b0ae-1681385393a0" (UID: "70de12b3-1a25-476e-b0ae-1681385393a0"). InnerVolumeSpecName "kube-api-access-jth5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.266293 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70de12b3-1a25-476e-b0ae-1681385393a0-config-data" (OuterVolumeSpecName: "config-data") pod "70de12b3-1a25-476e-b0ae-1681385393a0" (UID: "70de12b3-1a25-476e-b0ae-1681385393a0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.287548 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70de12b3-1a25-476e-b0ae-1681385393a0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "70de12b3-1a25-476e-b0ae-1681385393a0" (UID: "70de12b3-1a25-476e-b0ae-1681385393a0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.340786 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70de12b3-1a25-476e-b0ae-1681385393a0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.341124 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70de12b3-1a25-476e-b0ae-1681385393a0-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.341241 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jth5m\" (UniqueName: \"kubernetes.io/projected/70de12b3-1a25-476e-b0ae-1681385393a0-kube-api-access-jth5m\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.579115 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.746387 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfx47\" (UniqueName: \"kubernetes.io/projected/08642d69-d1bf-4816-928c-876bd2a62d2c-kube-api-access-zfx47\") pod \"08642d69-d1bf-4816-928c-876bd2a62d2c\" (UID: \"08642d69-d1bf-4816-928c-876bd2a62d2c\") " Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.746498 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08642d69-d1bf-4816-928c-876bd2a62d2c-logs\") pod \"08642d69-d1bf-4816-928c-876bd2a62d2c\" (UID: \"08642d69-d1bf-4816-928c-876bd2a62d2c\") " Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.746572 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08642d69-d1bf-4816-928c-876bd2a62d2c-combined-ca-bundle\") pod \"08642d69-d1bf-4816-928c-876bd2a62d2c\" (UID: \"08642d69-d1bf-4816-928c-876bd2a62d2c\") " Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.747067 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08642d69-d1bf-4816-928c-876bd2a62d2c-logs" (OuterVolumeSpecName: "logs") pod "08642d69-d1bf-4816-928c-876bd2a62d2c" (UID: "08642d69-d1bf-4816-928c-876bd2a62d2c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.747623 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08642d69-d1bf-4816-928c-876bd2a62d2c-config-data\") pod \"08642d69-d1bf-4816-928c-876bd2a62d2c\" (UID: \"08642d69-d1bf-4816-928c-876bd2a62d2c\") " Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.748607 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08642d69-d1bf-4816-928c-876bd2a62d2c-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.751493 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08642d69-d1bf-4816-928c-876bd2a62d2c-kube-api-access-zfx47" (OuterVolumeSpecName: "kube-api-access-zfx47") pod "08642d69-d1bf-4816-928c-876bd2a62d2c" (UID: "08642d69-d1bf-4816-928c-876bd2a62d2c"). InnerVolumeSpecName "kube-api-access-zfx47". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.775783 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08642d69-d1bf-4816-928c-876bd2a62d2c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "08642d69-d1bf-4816-928c-876bd2a62d2c" (UID: "08642d69-d1bf-4816-928c-876bd2a62d2c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.777357 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08642d69-d1bf-4816-928c-876bd2a62d2c-config-data" (OuterVolumeSpecName: "config-data") pod "08642d69-d1bf-4816-928c-876bd2a62d2c" (UID: "08642d69-d1bf-4816-928c-876bd2a62d2c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.853271 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08642d69-d1bf-4816-928c-876bd2a62d2c-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.853322 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfx47\" (UniqueName: \"kubernetes.io/projected/08642d69-d1bf-4816-928c-876bd2a62d2c-kube-api-access-zfx47\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.853337 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08642d69-d1bf-4816-928c-876bd2a62d2c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.885814 4790 generic.go:334] "Generic (PLEG): container finished" podID="08642d69-d1bf-4816-928c-876bd2a62d2c" containerID="0b1fe97c04a07f9803771cb466a280da626a12321ff3d6f79680c0e4be77e10f" exitCode=0 Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.885912 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.885930 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"08642d69-d1bf-4816-928c-876bd2a62d2c","Type":"ContainerDied","Data":"0b1fe97c04a07f9803771cb466a280da626a12321ff3d6f79680c0e4be77e10f"} Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.886030 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"08642d69-d1bf-4816-928c-876bd2a62d2c","Type":"ContainerDied","Data":"a2aff05b3fd1f8d5f38265384a0f1cdd575c6a21a177917acb8ce3252cd9466e"} Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.886063 4790 scope.go:117] "RemoveContainer" containerID="0b1fe97c04a07f9803771cb466a280da626a12321ff3d6f79680c0e4be77e10f" Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.892247 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"70de12b3-1a25-476e-b0ae-1681385393a0","Type":"ContainerDied","Data":"deb58a818e40aa97b61f043a15f58ced4d888a22f598bd361642e888352efb15"} Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.892377 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.933068 4790 scope.go:117] "RemoveContainer" containerID="4687665d607da921a4d0738e0c70d3b8677b74d18cfbf5b6cc62a7ed05fd20ed" Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.949987 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.973808 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.978084 4790 scope.go:117] "RemoveContainer" containerID="0b1fe97c04a07f9803771cb466a280da626a12321ff3d6f79680c0e4be77e10f" Nov 24 14:45:42 crc kubenswrapper[4790]: E1124 14:45:42.979383 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b1fe97c04a07f9803771cb466a280da626a12321ff3d6f79680c0e4be77e10f\": container with ID starting with 0b1fe97c04a07f9803771cb466a280da626a12321ff3d6f79680c0e4be77e10f not found: ID does not exist" containerID="0b1fe97c04a07f9803771cb466a280da626a12321ff3d6f79680c0e4be77e10f" Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.979416 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b1fe97c04a07f9803771cb466a280da626a12321ff3d6f79680c0e4be77e10f"} err="failed to get container status \"0b1fe97c04a07f9803771cb466a280da626a12321ff3d6f79680c0e4be77e10f\": rpc error: code = NotFound desc = could not find container \"0b1fe97c04a07f9803771cb466a280da626a12321ff3d6f79680c0e4be77e10f\": container with ID starting with 0b1fe97c04a07f9803771cb466a280da626a12321ff3d6f79680c0e4be77e10f not found: ID does not exist" Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.979439 4790 scope.go:117] "RemoveContainer" containerID="4687665d607da921a4d0738e0c70d3b8677b74d18cfbf5b6cc62a7ed05fd20ed" Nov 24 14:45:42 crc kubenswrapper[4790]: E1124 14:45:42.982482 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4687665d607da921a4d0738e0c70d3b8677b74d18cfbf5b6cc62a7ed05fd20ed\": container with ID starting with 4687665d607da921a4d0738e0c70d3b8677b74d18cfbf5b6cc62a7ed05fd20ed not found: ID does not exist" containerID="4687665d607da921a4d0738e0c70d3b8677b74d18cfbf5b6cc62a7ed05fd20ed" Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.982533 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4687665d607da921a4d0738e0c70d3b8677b74d18cfbf5b6cc62a7ed05fd20ed"} err="failed to get container status \"4687665d607da921a4d0738e0c70d3b8677b74d18cfbf5b6cc62a7ed05fd20ed\": rpc error: code = NotFound desc = could not find container \"4687665d607da921a4d0738e0c70d3b8677b74d18cfbf5b6cc62a7ed05fd20ed\": container with ID starting with 4687665d607da921a4d0738e0c70d3b8677b74d18cfbf5b6cc62a7ed05fd20ed not found: ID does not exist" Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.982561 4790 scope.go:117] "RemoveContainer" containerID="414a4472082056f08a9d288943aa88ce18cf717766968d2a4df5d5196c782b00" Nov 24 14:45:42 crc kubenswrapper[4790]: I1124 14:45:42.986529 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.020982 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.028905 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:45:43 crc kubenswrapper[4790]: E1124 14:45:43.029542 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7581458b-2c21-4297-af62-05fd707bb201" containerName="nova-manage" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.029560 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="7581458b-2c21-4297-af62-05fd707bb201" containerName="nova-manage" Nov 24 14:45:43 crc kubenswrapper[4790]: E1124 14:45:43.029578 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58a4a77b-8070-4da1-b716-2e9796bab87a" containerName="extract-utilities" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.029589 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="58a4a77b-8070-4da1-b716-2e9796bab87a" containerName="extract-utilities" Nov 24 14:45:43 crc kubenswrapper[4790]: E1124 14:45:43.029616 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58a4a77b-8070-4da1-b716-2e9796bab87a" containerName="registry-server" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.029625 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="58a4a77b-8070-4da1-b716-2e9796bab87a" containerName="registry-server" Nov 24 14:45:43 crc kubenswrapper[4790]: E1124 14:45:43.029646 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70de12b3-1a25-476e-b0ae-1681385393a0" containerName="nova-scheduler-scheduler" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.029655 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="70de12b3-1a25-476e-b0ae-1681385393a0" containerName="nova-scheduler-scheduler" Nov 24 14:45:43 crc kubenswrapper[4790]: E1124 14:45:43.029682 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08642d69-d1bf-4816-928c-876bd2a62d2c" containerName="nova-metadata-metadata" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.029692 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="08642d69-d1bf-4816-928c-876bd2a62d2c" containerName="nova-metadata-metadata" Nov 24 14:45:43 crc kubenswrapper[4790]: E1124 14:45:43.029712 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58a4a77b-8070-4da1-b716-2e9796bab87a" containerName="extract-content" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.029721 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="58a4a77b-8070-4da1-b716-2e9796bab87a" containerName="extract-content" Nov 24 14:45:43 crc kubenswrapper[4790]: E1124 14:45:43.029738 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08642d69-d1bf-4816-928c-876bd2a62d2c" containerName="nova-metadata-log" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.029748 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="08642d69-d1bf-4816-928c-876bd2a62d2c" containerName="nova-metadata-log" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.030020 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="7581458b-2c21-4297-af62-05fd707bb201" containerName="nova-manage" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.030037 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="08642d69-d1bf-4816-928c-876bd2a62d2c" containerName="nova-metadata-metadata" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.030049 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="08642d69-d1bf-4816-928c-876bd2a62d2c" containerName="nova-metadata-log" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.030070 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="58a4a77b-8070-4da1-b716-2e9796bab87a" containerName="registry-server" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.030087 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="70de12b3-1a25-476e-b0ae-1681385393a0" containerName="nova-scheduler-scheduler" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.030981 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.035834 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.039639 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.051174 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.053767 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.056281 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.060847 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.159816 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a167644f-5164-4943-97a3-6031e71f2f28-config-data\") pod \"nova-scheduler-0\" (UID: \"a167644f-5164-4943-97a3-6031e71f2f28\") " pod="openstack/nova-scheduler-0" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.160361 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3a43449-bf3d-4104-a89e-ef4d45191a38-config-data\") pod \"nova-metadata-0\" (UID: \"e3a43449-bf3d-4104-a89e-ef4d45191a38\") " pod="openstack/nova-metadata-0" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.160547 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6rdr\" (UniqueName: \"kubernetes.io/projected/a167644f-5164-4943-97a3-6031e71f2f28-kube-api-access-d6rdr\") pod \"nova-scheduler-0\" (UID: \"a167644f-5164-4943-97a3-6031e71f2f28\") " pod="openstack/nova-scheduler-0" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.160750 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csmpn\" (UniqueName: \"kubernetes.io/projected/e3a43449-bf3d-4104-a89e-ef4d45191a38-kube-api-access-csmpn\") pod \"nova-metadata-0\" (UID: \"e3a43449-bf3d-4104-a89e-ef4d45191a38\") " pod="openstack/nova-metadata-0" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.160855 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a167644f-5164-4943-97a3-6031e71f2f28-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a167644f-5164-4943-97a3-6031e71f2f28\") " pod="openstack/nova-scheduler-0" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.161215 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3a43449-bf3d-4104-a89e-ef4d45191a38-logs\") pod \"nova-metadata-0\" (UID: \"e3a43449-bf3d-4104-a89e-ef4d45191a38\") " pod="openstack/nova-metadata-0" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.161330 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3a43449-bf3d-4104-a89e-ef4d45191a38-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e3a43449-bf3d-4104-a89e-ef4d45191a38\") " pod="openstack/nova-metadata-0" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.262919 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3a43449-bf3d-4104-a89e-ef4d45191a38-config-data\") pod \"nova-metadata-0\" (UID: \"e3a43449-bf3d-4104-a89e-ef4d45191a38\") " pod="openstack/nova-metadata-0" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.262992 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6rdr\" (UniqueName: \"kubernetes.io/projected/a167644f-5164-4943-97a3-6031e71f2f28-kube-api-access-d6rdr\") pod \"nova-scheduler-0\" (UID: \"a167644f-5164-4943-97a3-6031e71f2f28\") " pod="openstack/nova-scheduler-0" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.263030 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csmpn\" (UniqueName: \"kubernetes.io/projected/e3a43449-bf3d-4104-a89e-ef4d45191a38-kube-api-access-csmpn\") pod \"nova-metadata-0\" (UID: \"e3a43449-bf3d-4104-a89e-ef4d45191a38\") " pod="openstack/nova-metadata-0" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.263050 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a167644f-5164-4943-97a3-6031e71f2f28-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a167644f-5164-4943-97a3-6031e71f2f28\") " pod="openstack/nova-scheduler-0" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.263097 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3a43449-bf3d-4104-a89e-ef4d45191a38-logs\") pod \"nova-metadata-0\" (UID: \"e3a43449-bf3d-4104-a89e-ef4d45191a38\") " pod="openstack/nova-metadata-0" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.263120 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3a43449-bf3d-4104-a89e-ef4d45191a38-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e3a43449-bf3d-4104-a89e-ef4d45191a38\") " pod="openstack/nova-metadata-0" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.263168 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a167644f-5164-4943-97a3-6031e71f2f28-config-data\") pod \"nova-scheduler-0\" (UID: \"a167644f-5164-4943-97a3-6031e71f2f28\") " pod="openstack/nova-scheduler-0" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.264188 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3a43449-bf3d-4104-a89e-ef4d45191a38-logs\") pod \"nova-metadata-0\" (UID: \"e3a43449-bf3d-4104-a89e-ef4d45191a38\") " pod="openstack/nova-metadata-0" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.269309 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a167644f-5164-4943-97a3-6031e71f2f28-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a167644f-5164-4943-97a3-6031e71f2f28\") " pod="openstack/nova-scheduler-0" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.269327 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3a43449-bf3d-4104-a89e-ef4d45191a38-config-data\") pod \"nova-metadata-0\" (UID: \"e3a43449-bf3d-4104-a89e-ef4d45191a38\") " pod="openstack/nova-metadata-0" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.269561 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a167644f-5164-4943-97a3-6031e71f2f28-config-data\") pod \"nova-scheduler-0\" (UID: \"a167644f-5164-4943-97a3-6031e71f2f28\") " pod="openstack/nova-scheduler-0" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.271268 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3a43449-bf3d-4104-a89e-ef4d45191a38-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e3a43449-bf3d-4104-a89e-ef4d45191a38\") " pod="openstack/nova-metadata-0" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.287397 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csmpn\" (UniqueName: \"kubernetes.io/projected/e3a43449-bf3d-4104-a89e-ef4d45191a38-kube-api-access-csmpn\") pod \"nova-metadata-0\" (UID: \"e3a43449-bf3d-4104-a89e-ef4d45191a38\") " pod="openstack/nova-metadata-0" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.290194 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6rdr\" (UniqueName: \"kubernetes.io/projected/a167644f-5164-4943-97a3-6031e71f2f28-kube-api-access-d6rdr\") pod \"nova-scheduler-0\" (UID: \"a167644f-5164-4943-97a3-6031e71f2f28\") " pod="openstack/nova-scheduler-0" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.315241 4790 scope.go:117] "RemoveContainer" containerID="92c086e44ccb28a62777d612e499209f7483422ca70b460ce6ec73c416e8e8b8" Nov 24 14:45:43 crc kubenswrapper[4790]: E1124 14:45:43.315772 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.355898 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.374279 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.802127 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:45:43 crc kubenswrapper[4790]: W1124 14:45:43.807184 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda167644f_5164_4943_97a3_6031e71f2f28.slice/crio-9db79f6dff61f24e734eee32f8574d81ccedbb3554915450e0b846e0a88d7cd2 WatchSource:0}: Error finding container 9db79f6dff61f24e734eee32f8574d81ccedbb3554915450e0b846e0a88d7cd2: Status 404 returned error can't find the container with id 9db79f6dff61f24e734eee32f8574d81ccedbb3554915450e0b846e0a88d7cd2 Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.905977 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.912951 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a167644f-5164-4943-97a3-6031e71f2f28","Type":"ContainerStarted","Data":"9db79f6dff61f24e734eee32f8574d81ccedbb3554915450e0b846e0a88d7cd2"} Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.915826 4790 generic.go:334] "Generic (PLEG): container finished" podID="cc1a0fe0-4faf-412f-a5d3-07725b64df09" containerID="0f4d8c1d4bbda20530690ce310bf02b9c1520de832d550f6e6e9cb7c8562249d" exitCode=0 Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.915855 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cc1a0fe0-4faf-412f-a5d3-07725b64df09","Type":"ContainerDied","Data":"0f4d8c1d4bbda20530690ce310bf02b9c1520de832d550f6e6e9cb7c8562249d"} Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.915872 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cc1a0fe0-4faf-412f-a5d3-07725b64df09","Type":"ContainerDied","Data":"b23d342f912188b62572113722897eb5d7d632206d6de57ad6c1c1ab04076f08"} Nov 24 14:45:43 crc kubenswrapper[4790]: I1124 14:45:43.915904 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b23d342f912188b62572113722897eb5d7d632206d6de57ad6c1c1ab04076f08" Nov 24 14:45:44 crc kubenswrapper[4790]: I1124 14:45:44.105069 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:45:44 crc kubenswrapper[4790]: I1124 14:45:44.281227 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc1a0fe0-4faf-412f-a5d3-07725b64df09-combined-ca-bundle\") pod \"cc1a0fe0-4faf-412f-a5d3-07725b64df09\" (UID: \"cc1a0fe0-4faf-412f-a5d3-07725b64df09\") " Nov 24 14:45:44 crc kubenswrapper[4790]: I1124 14:45:44.281356 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2d7l\" (UniqueName: \"kubernetes.io/projected/cc1a0fe0-4faf-412f-a5d3-07725b64df09-kube-api-access-l2d7l\") pod \"cc1a0fe0-4faf-412f-a5d3-07725b64df09\" (UID: \"cc1a0fe0-4faf-412f-a5d3-07725b64df09\") " Nov 24 14:45:44 crc kubenswrapper[4790]: I1124 14:45:44.281438 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc1a0fe0-4faf-412f-a5d3-07725b64df09-logs\") pod \"cc1a0fe0-4faf-412f-a5d3-07725b64df09\" (UID: \"cc1a0fe0-4faf-412f-a5d3-07725b64df09\") " Nov 24 14:45:44 crc kubenswrapper[4790]: I1124 14:45:44.281535 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc1a0fe0-4faf-412f-a5d3-07725b64df09-config-data\") pod \"cc1a0fe0-4faf-412f-a5d3-07725b64df09\" (UID: \"cc1a0fe0-4faf-412f-a5d3-07725b64df09\") " Nov 24 14:45:44 crc kubenswrapper[4790]: I1124 14:45:44.282912 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc1a0fe0-4faf-412f-a5d3-07725b64df09-logs" (OuterVolumeSpecName: "logs") pod "cc1a0fe0-4faf-412f-a5d3-07725b64df09" (UID: "cc1a0fe0-4faf-412f-a5d3-07725b64df09"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:45:44 crc kubenswrapper[4790]: I1124 14:45:44.285858 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc1a0fe0-4faf-412f-a5d3-07725b64df09-kube-api-access-l2d7l" (OuterVolumeSpecName: "kube-api-access-l2d7l") pod "cc1a0fe0-4faf-412f-a5d3-07725b64df09" (UID: "cc1a0fe0-4faf-412f-a5d3-07725b64df09"). InnerVolumeSpecName "kube-api-access-l2d7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:45:44 crc kubenswrapper[4790]: I1124 14:45:44.315557 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc1a0fe0-4faf-412f-a5d3-07725b64df09-config-data" (OuterVolumeSpecName: "config-data") pod "cc1a0fe0-4faf-412f-a5d3-07725b64df09" (UID: "cc1a0fe0-4faf-412f-a5d3-07725b64df09"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:45:44 crc kubenswrapper[4790]: I1124 14:45:44.327455 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08642d69-d1bf-4816-928c-876bd2a62d2c" path="/var/lib/kubelet/pods/08642d69-d1bf-4816-928c-876bd2a62d2c/volumes" Nov 24 14:45:44 crc kubenswrapper[4790]: I1124 14:45:44.328407 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70de12b3-1a25-476e-b0ae-1681385393a0" path="/var/lib/kubelet/pods/70de12b3-1a25-476e-b0ae-1681385393a0/volumes" Nov 24 14:45:44 crc kubenswrapper[4790]: I1124 14:45:44.328748 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc1a0fe0-4faf-412f-a5d3-07725b64df09-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cc1a0fe0-4faf-412f-a5d3-07725b64df09" (UID: "cc1a0fe0-4faf-412f-a5d3-07725b64df09"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:45:44 crc kubenswrapper[4790]: I1124 14:45:44.384192 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc1a0fe0-4faf-412f-a5d3-07725b64df09-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:44 crc kubenswrapper[4790]: I1124 14:45:44.384244 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc1a0fe0-4faf-412f-a5d3-07725b64df09-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:44 crc kubenswrapper[4790]: I1124 14:45:44.384263 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2d7l\" (UniqueName: \"kubernetes.io/projected/cc1a0fe0-4faf-412f-a5d3-07725b64df09-kube-api-access-l2d7l\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:44 crc kubenswrapper[4790]: I1124 14:45:44.384280 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc1a0fe0-4faf-412f-a5d3-07725b64df09-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:44 crc kubenswrapper[4790]: I1124 14:45:44.955860 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a167644f-5164-4943-97a3-6031e71f2f28","Type":"ContainerStarted","Data":"82ca39eedfff2d3a9609182b4df1638283f236abbef9e67ab174b401f08d4b21"} Nov 24 14:45:44 crc kubenswrapper[4790]: I1124 14:45:44.964419 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:45:44 crc kubenswrapper[4790]: I1124 14:45:44.964490 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e3a43449-bf3d-4104-a89e-ef4d45191a38","Type":"ContainerStarted","Data":"50fdb87f7287bb1564344a783de02660937ef2fce769a9a2a4309f9457eebfe5"} Nov 24 14:45:44 crc kubenswrapper[4790]: I1124 14:45:44.964521 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e3a43449-bf3d-4104-a89e-ef4d45191a38","Type":"ContainerStarted","Data":"a964645c523ed311dc320cc96961d73dc4521e7f4cf3e145e4099d7bd09eae4c"} Nov 24 14:45:44 crc kubenswrapper[4790]: I1124 14:45:44.964531 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e3a43449-bf3d-4104-a89e-ef4d45191a38","Type":"ContainerStarted","Data":"3501e5d87dcd28433f6686936569f742009987b35ed9500cab5dd67818983b38"} Nov 24 14:45:44 crc kubenswrapper[4790]: I1124 14:45:44.980226 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.980205787 podStartE2EDuration="2.980205787s" podCreationTimestamp="2025-11-24 14:45:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:45:44.974629947 +0000 UTC m=+5593.354523619" watchObservedRunningTime="2025-11-24 14:45:44.980205787 +0000 UTC m=+5593.360099449" Nov 24 14:45:45 crc kubenswrapper[4790]: I1124 14:45:45.000322 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.000302244 podStartE2EDuration="3.000302244s" podCreationTimestamp="2025-11-24 14:45:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:45:44.997212315 +0000 UTC m=+5593.377105997" watchObservedRunningTime="2025-11-24 14:45:45.000302244 +0000 UTC m=+5593.380195916" Nov 24 14:45:45 crc kubenswrapper[4790]: I1124 14:45:45.021559 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:45:45 crc kubenswrapper[4790]: I1124 14:45:45.036932 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:45:45 crc kubenswrapper[4790]: I1124 14:45:45.045248 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 14:45:45 crc kubenswrapper[4790]: E1124 14:45:45.045728 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc1a0fe0-4faf-412f-a5d3-07725b64df09" containerName="nova-api-api" Nov 24 14:45:45 crc kubenswrapper[4790]: I1124 14:45:45.045744 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc1a0fe0-4faf-412f-a5d3-07725b64df09" containerName="nova-api-api" Nov 24 14:45:45 crc kubenswrapper[4790]: E1124 14:45:45.045760 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc1a0fe0-4faf-412f-a5d3-07725b64df09" containerName="nova-api-log" Nov 24 14:45:45 crc kubenswrapper[4790]: I1124 14:45:45.045769 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc1a0fe0-4faf-412f-a5d3-07725b64df09" containerName="nova-api-log" Nov 24 14:45:45 crc kubenswrapper[4790]: I1124 14:45:45.046031 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc1a0fe0-4faf-412f-a5d3-07725b64df09" containerName="nova-api-log" Nov 24 14:45:45 crc kubenswrapper[4790]: I1124 14:45:45.046064 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc1a0fe0-4faf-412f-a5d3-07725b64df09" containerName="nova-api-api" Nov 24 14:45:45 crc kubenswrapper[4790]: I1124 14:45:45.047329 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:45:45 crc kubenswrapper[4790]: I1124 14:45:45.050350 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 14:45:45 crc kubenswrapper[4790]: I1124 14:45:45.056001 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:45:45 crc kubenswrapper[4790]: I1124 14:45:45.200675 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/efc434d3-6cab-448c-a4aa-c441cefaafe2-logs\") pod \"nova-api-0\" (UID: \"efc434d3-6cab-448c-a4aa-c441cefaafe2\") " pod="openstack/nova-api-0" Nov 24 14:45:45 crc kubenswrapper[4790]: I1124 14:45:45.200744 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efc434d3-6cab-448c-a4aa-c441cefaafe2-config-data\") pod \"nova-api-0\" (UID: \"efc434d3-6cab-448c-a4aa-c441cefaafe2\") " pod="openstack/nova-api-0" Nov 24 14:45:45 crc kubenswrapper[4790]: I1124 14:45:45.200988 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efc434d3-6cab-448c-a4aa-c441cefaafe2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"efc434d3-6cab-448c-a4aa-c441cefaafe2\") " pod="openstack/nova-api-0" Nov 24 14:45:45 crc kubenswrapper[4790]: I1124 14:45:45.201174 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tlwn\" (UniqueName: \"kubernetes.io/projected/efc434d3-6cab-448c-a4aa-c441cefaafe2-kube-api-access-6tlwn\") pod \"nova-api-0\" (UID: \"efc434d3-6cab-448c-a4aa-c441cefaafe2\") " pod="openstack/nova-api-0" Nov 24 14:45:45 crc kubenswrapper[4790]: I1124 14:45:45.303283 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/efc434d3-6cab-448c-a4aa-c441cefaafe2-logs\") pod \"nova-api-0\" (UID: \"efc434d3-6cab-448c-a4aa-c441cefaafe2\") " pod="openstack/nova-api-0" Nov 24 14:45:45 crc kubenswrapper[4790]: I1124 14:45:45.303357 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efc434d3-6cab-448c-a4aa-c441cefaafe2-config-data\") pod \"nova-api-0\" (UID: \"efc434d3-6cab-448c-a4aa-c441cefaafe2\") " pod="openstack/nova-api-0" Nov 24 14:45:45 crc kubenswrapper[4790]: I1124 14:45:45.303389 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efc434d3-6cab-448c-a4aa-c441cefaafe2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"efc434d3-6cab-448c-a4aa-c441cefaafe2\") " pod="openstack/nova-api-0" Nov 24 14:45:45 crc kubenswrapper[4790]: I1124 14:45:45.303434 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tlwn\" (UniqueName: \"kubernetes.io/projected/efc434d3-6cab-448c-a4aa-c441cefaafe2-kube-api-access-6tlwn\") pod \"nova-api-0\" (UID: \"efc434d3-6cab-448c-a4aa-c441cefaafe2\") " pod="openstack/nova-api-0" Nov 24 14:45:45 crc kubenswrapper[4790]: I1124 14:45:45.303839 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/efc434d3-6cab-448c-a4aa-c441cefaafe2-logs\") pod \"nova-api-0\" (UID: \"efc434d3-6cab-448c-a4aa-c441cefaafe2\") " pod="openstack/nova-api-0" Nov 24 14:45:45 crc kubenswrapper[4790]: I1124 14:45:45.308431 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efc434d3-6cab-448c-a4aa-c441cefaafe2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"efc434d3-6cab-448c-a4aa-c441cefaafe2\") " pod="openstack/nova-api-0" Nov 24 14:45:45 crc kubenswrapper[4790]: I1124 14:45:45.309287 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efc434d3-6cab-448c-a4aa-c441cefaafe2-config-data\") pod \"nova-api-0\" (UID: \"efc434d3-6cab-448c-a4aa-c441cefaafe2\") " pod="openstack/nova-api-0" Nov 24 14:45:45 crc kubenswrapper[4790]: I1124 14:45:45.323271 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tlwn\" (UniqueName: \"kubernetes.io/projected/efc434d3-6cab-448c-a4aa-c441cefaafe2-kube-api-access-6tlwn\") pod \"nova-api-0\" (UID: \"efc434d3-6cab-448c-a4aa-c441cefaafe2\") " pod="openstack/nova-api-0" Nov 24 14:45:45 crc kubenswrapper[4790]: I1124 14:45:45.379377 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:45:45 crc kubenswrapper[4790]: I1124 14:45:45.900787 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:45:45 crc kubenswrapper[4790]: W1124 14:45:45.903049 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefc434d3_6cab_448c_a4aa_c441cefaafe2.slice/crio-e9444683156dfb9b328b0ed05a3b41559491b4fec29ace89ae8d890c434806e2 WatchSource:0}: Error finding container e9444683156dfb9b328b0ed05a3b41559491b4fec29ace89ae8d890c434806e2: Status 404 returned error can't find the container with id e9444683156dfb9b328b0ed05a3b41559491b4fec29ace89ae8d890c434806e2 Nov 24 14:45:45 crc kubenswrapper[4790]: I1124 14:45:45.978275 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"efc434d3-6cab-448c-a4aa-c441cefaafe2","Type":"ContainerStarted","Data":"e9444683156dfb9b328b0ed05a3b41559491b4fec29ace89ae8d890c434806e2"} Nov 24 14:45:46 crc kubenswrapper[4790]: I1124 14:45:46.325550 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc1a0fe0-4faf-412f-a5d3-07725b64df09" path="/var/lib/kubelet/pods/cc1a0fe0-4faf-412f-a5d3-07725b64df09/volumes" Nov 24 14:45:46 crc kubenswrapper[4790]: I1124 14:45:46.992770 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"efc434d3-6cab-448c-a4aa-c441cefaafe2","Type":"ContainerStarted","Data":"3e021067604d72526d82368a35fae3df9f6210fd0b182908f5c549fbd94d2c3a"} Nov 24 14:45:46 crc kubenswrapper[4790]: I1124 14:45:46.992836 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"efc434d3-6cab-448c-a4aa-c441cefaafe2","Type":"ContainerStarted","Data":"8fc8c835dcd207c0e9672856b4f27278b2b63f832ce4e3ef10ed59cc50b82aa6"} Nov 24 14:45:47 crc kubenswrapper[4790]: I1124 14:45:47.016954 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.016932555 podStartE2EDuration="2.016932555s" podCreationTimestamp="2025-11-24 14:45:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:45:47.016544814 +0000 UTC m=+5595.396438526" watchObservedRunningTime="2025-11-24 14:45:47.016932555 +0000 UTC m=+5595.396826227" Nov 24 14:45:47 crc kubenswrapper[4790]: I1124 14:45:47.973779 4790 scope.go:117] "RemoveContainer" containerID="90eaf19c23289a7e2dcfee76d9001e56223190ed57614857ef7469ace4095ff6" Nov 24 14:45:48 crc kubenswrapper[4790]: I1124 14:45:48.021902 4790 scope.go:117] "RemoveContainer" containerID="9ec50958969b0ff5c3ef1791470136baa532899ab450dc32dcd2c2677cf16c67" Nov 24 14:45:48 crc kubenswrapper[4790]: I1124 14:45:48.356571 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 14:45:48 crc kubenswrapper[4790]: I1124 14:45:48.375343 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 14:45:48 crc kubenswrapper[4790]: I1124 14:45:48.375411 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 14:45:53 crc kubenswrapper[4790]: I1124 14:45:53.357021 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 14:45:53 crc kubenswrapper[4790]: I1124 14:45:53.375681 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 14:45:53 crc kubenswrapper[4790]: I1124 14:45:53.375752 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 14:45:53 crc kubenswrapper[4790]: I1124 14:45:53.408781 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 14:45:54 crc kubenswrapper[4790]: I1124 14:45:54.111648 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 14:45:54 crc kubenswrapper[4790]: I1124 14:45:54.459111 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e3a43449-bf3d-4104-a89e-ef4d45191a38" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 14:45:54 crc kubenswrapper[4790]: I1124 14:45:54.459111 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e3a43449-bf3d-4104-a89e-ef4d45191a38" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 14:45:55 crc kubenswrapper[4790]: I1124 14:45:55.379940 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 14:45:55 crc kubenswrapper[4790]: I1124 14:45:55.380285 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 14:45:56 crc kubenswrapper[4790]: I1124 14:45:56.314955 4790 scope.go:117] "RemoveContainer" containerID="92c086e44ccb28a62777d612e499209f7483422ca70b460ce6ec73c416e8e8b8" Nov 24 14:45:56 crc kubenswrapper[4790]: I1124 14:45:56.462108 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="efc434d3-6cab-448c-a4aa-c441cefaafe2" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.75:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 14:45:56 crc kubenswrapper[4790]: I1124 14:45:56.462132 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="efc434d3-6cab-448c-a4aa-c441cefaafe2" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.75:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 14:45:57 crc kubenswrapper[4790]: I1124 14:45:57.113139 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"cd74b939f536300843ca659100b982ff6a09937cde93ea7f086d7cd5d1e85fc5"} Nov 24 14:46:03 crc kubenswrapper[4790]: I1124 14:46:03.377173 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 14:46:03 crc kubenswrapper[4790]: I1124 14:46:03.398009 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 14:46:03 crc kubenswrapper[4790]: I1124 14:46:03.401441 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 14:46:04 crc kubenswrapper[4790]: I1124 14:46:04.226688 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 14:46:05 crc kubenswrapper[4790]: I1124 14:46:05.383390 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 14:46:05 crc kubenswrapper[4790]: I1124 14:46:05.384373 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 14:46:05 crc kubenswrapper[4790]: I1124 14:46:05.384797 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 14:46:05 crc kubenswrapper[4790]: I1124 14:46:05.384862 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 14:46:05 crc kubenswrapper[4790]: I1124 14:46:05.388247 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 14:46:05 crc kubenswrapper[4790]: I1124 14:46:05.390280 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 14:46:05 crc kubenswrapper[4790]: I1124 14:46:05.612190 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5869b86df7-lfcd7"] Nov 24 14:46:05 crc kubenswrapper[4790]: I1124 14:46:05.613589 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5869b86df7-lfcd7" Nov 24 14:46:05 crc kubenswrapper[4790]: I1124 14:46:05.625120 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5869b86df7-lfcd7"] Nov 24 14:46:05 crc kubenswrapper[4790]: I1124 14:46:05.706347 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4765f4c7-8776-46c8-9a6d-2050e0201a83-ovsdbserver-sb\") pod \"dnsmasq-dns-5869b86df7-lfcd7\" (UID: \"4765f4c7-8776-46c8-9a6d-2050e0201a83\") " pod="openstack/dnsmasq-dns-5869b86df7-lfcd7" Nov 24 14:46:05 crc kubenswrapper[4790]: I1124 14:46:05.706403 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4765f4c7-8776-46c8-9a6d-2050e0201a83-config\") pod \"dnsmasq-dns-5869b86df7-lfcd7\" (UID: \"4765f4c7-8776-46c8-9a6d-2050e0201a83\") " pod="openstack/dnsmasq-dns-5869b86df7-lfcd7" Nov 24 14:46:05 crc kubenswrapper[4790]: I1124 14:46:05.706614 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4765f4c7-8776-46c8-9a6d-2050e0201a83-dns-svc\") pod \"dnsmasq-dns-5869b86df7-lfcd7\" (UID: \"4765f4c7-8776-46c8-9a6d-2050e0201a83\") " pod="openstack/dnsmasq-dns-5869b86df7-lfcd7" Nov 24 14:46:05 crc kubenswrapper[4790]: I1124 14:46:05.706713 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4765f4c7-8776-46c8-9a6d-2050e0201a83-ovsdbserver-nb\") pod \"dnsmasq-dns-5869b86df7-lfcd7\" (UID: \"4765f4c7-8776-46c8-9a6d-2050e0201a83\") " pod="openstack/dnsmasq-dns-5869b86df7-lfcd7" Nov 24 14:46:05 crc kubenswrapper[4790]: I1124 14:46:05.706914 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvjrf\" (UniqueName: \"kubernetes.io/projected/4765f4c7-8776-46c8-9a6d-2050e0201a83-kube-api-access-kvjrf\") pod \"dnsmasq-dns-5869b86df7-lfcd7\" (UID: \"4765f4c7-8776-46c8-9a6d-2050e0201a83\") " pod="openstack/dnsmasq-dns-5869b86df7-lfcd7" Nov 24 14:46:05 crc kubenswrapper[4790]: I1124 14:46:05.808216 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvjrf\" (UniqueName: \"kubernetes.io/projected/4765f4c7-8776-46c8-9a6d-2050e0201a83-kube-api-access-kvjrf\") pod \"dnsmasq-dns-5869b86df7-lfcd7\" (UID: \"4765f4c7-8776-46c8-9a6d-2050e0201a83\") " pod="openstack/dnsmasq-dns-5869b86df7-lfcd7" Nov 24 14:46:05 crc kubenswrapper[4790]: I1124 14:46:05.808286 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4765f4c7-8776-46c8-9a6d-2050e0201a83-ovsdbserver-sb\") pod \"dnsmasq-dns-5869b86df7-lfcd7\" (UID: \"4765f4c7-8776-46c8-9a6d-2050e0201a83\") " pod="openstack/dnsmasq-dns-5869b86df7-lfcd7" Nov 24 14:46:05 crc kubenswrapper[4790]: I1124 14:46:05.808310 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4765f4c7-8776-46c8-9a6d-2050e0201a83-config\") pod \"dnsmasq-dns-5869b86df7-lfcd7\" (UID: \"4765f4c7-8776-46c8-9a6d-2050e0201a83\") " pod="openstack/dnsmasq-dns-5869b86df7-lfcd7" Nov 24 14:46:05 crc kubenswrapper[4790]: I1124 14:46:05.808371 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4765f4c7-8776-46c8-9a6d-2050e0201a83-dns-svc\") pod \"dnsmasq-dns-5869b86df7-lfcd7\" (UID: \"4765f4c7-8776-46c8-9a6d-2050e0201a83\") " pod="openstack/dnsmasq-dns-5869b86df7-lfcd7" Nov 24 14:46:05 crc kubenswrapper[4790]: I1124 14:46:05.808410 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4765f4c7-8776-46c8-9a6d-2050e0201a83-ovsdbserver-nb\") pod \"dnsmasq-dns-5869b86df7-lfcd7\" (UID: \"4765f4c7-8776-46c8-9a6d-2050e0201a83\") " pod="openstack/dnsmasq-dns-5869b86df7-lfcd7" Nov 24 14:46:05 crc kubenswrapper[4790]: I1124 14:46:05.809302 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4765f4c7-8776-46c8-9a6d-2050e0201a83-ovsdbserver-nb\") pod \"dnsmasq-dns-5869b86df7-lfcd7\" (UID: \"4765f4c7-8776-46c8-9a6d-2050e0201a83\") " pod="openstack/dnsmasq-dns-5869b86df7-lfcd7" Nov 24 14:46:05 crc kubenswrapper[4790]: I1124 14:46:05.809847 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4765f4c7-8776-46c8-9a6d-2050e0201a83-dns-svc\") pod \"dnsmasq-dns-5869b86df7-lfcd7\" (UID: \"4765f4c7-8776-46c8-9a6d-2050e0201a83\") " pod="openstack/dnsmasq-dns-5869b86df7-lfcd7" Nov 24 14:46:05 crc kubenswrapper[4790]: I1124 14:46:05.809943 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4765f4c7-8776-46c8-9a6d-2050e0201a83-ovsdbserver-sb\") pod \"dnsmasq-dns-5869b86df7-lfcd7\" (UID: \"4765f4c7-8776-46c8-9a6d-2050e0201a83\") " pod="openstack/dnsmasq-dns-5869b86df7-lfcd7" Nov 24 14:46:05 crc kubenswrapper[4790]: I1124 14:46:05.809949 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4765f4c7-8776-46c8-9a6d-2050e0201a83-config\") pod \"dnsmasq-dns-5869b86df7-lfcd7\" (UID: \"4765f4c7-8776-46c8-9a6d-2050e0201a83\") " pod="openstack/dnsmasq-dns-5869b86df7-lfcd7" Nov 24 14:46:05 crc kubenswrapper[4790]: I1124 14:46:05.831947 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvjrf\" (UniqueName: \"kubernetes.io/projected/4765f4c7-8776-46c8-9a6d-2050e0201a83-kube-api-access-kvjrf\") pod \"dnsmasq-dns-5869b86df7-lfcd7\" (UID: \"4765f4c7-8776-46c8-9a6d-2050e0201a83\") " pod="openstack/dnsmasq-dns-5869b86df7-lfcd7" Nov 24 14:46:05 crc kubenswrapper[4790]: I1124 14:46:05.943563 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5869b86df7-lfcd7" Nov 24 14:46:06 crc kubenswrapper[4790]: I1124 14:46:06.335963 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5869b86df7-lfcd7"] Nov 24 14:46:07 crc kubenswrapper[4790]: I1124 14:46:07.258620 4790 generic.go:334] "Generic (PLEG): container finished" podID="4765f4c7-8776-46c8-9a6d-2050e0201a83" containerID="3bf14ef8b7710dc673a5450025822120a2515aaa43fa74889018f5c0db0be514" exitCode=0 Nov 24 14:46:07 crc kubenswrapper[4790]: I1124 14:46:07.258693 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5869b86df7-lfcd7" event={"ID":"4765f4c7-8776-46c8-9a6d-2050e0201a83","Type":"ContainerDied","Data":"3bf14ef8b7710dc673a5450025822120a2515aaa43fa74889018f5c0db0be514"} Nov 24 14:46:07 crc kubenswrapper[4790]: I1124 14:46:07.259029 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5869b86df7-lfcd7" event={"ID":"4765f4c7-8776-46c8-9a6d-2050e0201a83","Type":"ContainerStarted","Data":"4a3d812ee3797311df1a8e6f52cb268f9e49795e6e98d7fa55398bc635071de3"} Nov 24 14:46:08 crc kubenswrapper[4790]: I1124 14:46:08.267745 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5869b86df7-lfcd7" event={"ID":"4765f4c7-8776-46c8-9a6d-2050e0201a83","Type":"ContainerStarted","Data":"e5b1a9d3b9fd1f27c8f1dadb54b561065f1e91adaf02d5112dff492510776d2f"} Nov 24 14:46:08 crc kubenswrapper[4790]: I1124 14:46:08.268163 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5869b86df7-lfcd7" Nov 24 14:46:08 crc kubenswrapper[4790]: I1124 14:46:08.288409 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5869b86df7-lfcd7" podStartSLOduration=3.28838577 podStartE2EDuration="3.28838577s" podCreationTimestamp="2025-11-24 14:46:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:46:08.284042845 +0000 UTC m=+5616.663936507" watchObservedRunningTime="2025-11-24 14:46:08.28838577 +0000 UTC m=+5616.668279432" Nov 24 14:46:15 crc kubenswrapper[4790]: I1124 14:46:15.945121 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5869b86df7-lfcd7" Nov 24 14:46:16 crc kubenswrapper[4790]: I1124 14:46:16.053616 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbb866757-8frpq"] Nov 24 14:46:16 crc kubenswrapper[4790]: I1124 14:46:16.053929 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bbb866757-8frpq" podUID="11caa523-04fb-4f1d-a6e5-b9323fc00b3f" containerName="dnsmasq-dns" containerID="cri-o://8fc61441c83ed7c15701f95649565263bde6d8a5e7518cddc59e2934f3b1a193" gracePeriod=10 Nov 24 14:46:16 crc kubenswrapper[4790]: I1124 14:46:16.181088 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-bbb866757-8frpq" podUID="11caa523-04fb-4f1d-a6e5-b9323fc00b3f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.66:5353: connect: connection refused" Nov 24 14:46:16 crc kubenswrapper[4790]: I1124 14:46:16.379445 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbb866757-8frpq" event={"ID":"11caa523-04fb-4f1d-a6e5-b9323fc00b3f","Type":"ContainerDied","Data":"8fc61441c83ed7c15701f95649565263bde6d8a5e7518cddc59e2934f3b1a193"} Nov 24 14:46:16 crc kubenswrapper[4790]: I1124 14:46:16.379486 4790 generic.go:334] "Generic (PLEG): container finished" podID="11caa523-04fb-4f1d-a6e5-b9323fc00b3f" containerID="8fc61441c83ed7c15701f95649565263bde6d8a5e7518cddc59e2934f3b1a193" exitCode=0 Nov 24 14:46:16 crc kubenswrapper[4790]: I1124 14:46:16.539359 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbb866757-8frpq" Nov 24 14:46:16 crc kubenswrapper[4790]: I1124 14:46:16.654903 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11caa523-04fb-4f1d-a6e5-b9323fc00b3f-config\") pod \"11caa523-04fb-4f1d-a6e5-b9323fc00b3f\" (UID: \"11caa523-04fb-4f1d-a6e5-b9323fc00b3f\") " Nov 24 14:46:16 crc kubenswrapper[4790]: I1124 14:46:16.654962 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11caa523-04fb-4f1d-a6e5-b9323fc00b3f-ovsdbserver-sb\") pod \"11caa523-04fb-4f1d-a6e5-b9323fc00b3f\" (UID: \"11caa523-04fb-4f1d-a6e5-b9323fc00b3f\") " Nov 24 14:46:16 crc kubenswrapper[4790]: I1124 14:46:16.655007 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11caa523-04fb-4f1d-a6e5-b9323fc00b3f-ovsdbserver-nb\") pod \"11caa523-04fb-4f1d-a6e5-b9323fc00b3f\" (UID: \"11caa523-04fb-4f1d-a6e5-b9323fc00b3f\") " Nov 24 14:46:16 crc kubenswrapper[4790]: I1124 14:46:16.655068 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjzsk\" (UniqueName: \"kubernetes.io/projected/11caa523-04fb-4f1d-a6e5-b9323fc00b3f-kube-api-access-vjzsk\") pod \"11caa523-04fb-4f1d-a6e5-b9323fc00b3f\" (UID: \"11caa523-04fb-4f1d-a6e5-b9323fc00b3f\") " Nov 24 14:46:16 crc kubenswrapper[4790]: I1124 14:46:16.655256 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11caa523-04fb-4f1d-a6e5-b9323fc00b3f-dns-svc\") pod \"11caa523-04fb-4f1d-a6e5-b9323fc00b3f\" (UID: \"11caa523-04fb-4f1d-a6e5-b9323fc00b3f\") " Nov 24 14:46:16 crc kubenswrapper[4790]: I1124 14:46:16.665034 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11caa523-04fb-4f1d-a6e5-b9323fc00b3f-kube-api-access-vjzsk" (OuterVolumeSpecName: "kube-api-access-vjzsk") pod "11caa523-04fb-4f1d-a6e5-b9323fc00b3f" (UID: "11caa523-04fb-4f1d-a6e5-b9323fc00b3f"). InnerVolumeSpecName "kube-api-access-vjzsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:46:16 crc kubenswrapper[4790]: I1124 14:46:16.708924 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11caa523-04fb-4f1d-a6e5-b9323fc00b3f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "11caa523-04fb-4f1d-a6e5-b9323fc00b3f" (UID: "11caa523-04fb-4f1d-a6e5-b9323fc00b3f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:46:16 crc kubenswrapper[4790]: I1124 14:46:16.710098 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11caa523-04fb-4f1d-a6e5-b9323fc00b3f-config" (OuterVolumeSpecName: "config") pod "11caa523-04fb-4f1d-a6e5-b9323fc00b3f" (UID: "11caa523-04fb-4f1d-a6e5-b9323fc00b3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:46:16 crc kubenswrapper[4790]: I1124 14:46:16.712688 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11caa523-04fb-4f1d-a6e5-b9323fc00b3f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "11caa523-04fb-4f1d-a6e5-b9323fc00b3f" (UID: "11caa523-04fb-4f1d-a6e5-b9323fc00b3f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:46:16 crc kubenswrapper[4790]: I1124 14:46:16.719470 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11caa523-04fb-4f1d-a6e5-b9323fc00b3f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "11caa523-04fb-4f1d-a6e5-b9323fc00b3f" (UID: "11caa523-04fb-4f1d-a6e5-b9323fc00b3f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:46:16 crc kubenswrapper[4790]: I1124 14:46:16.757335 4790 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11caa523-04fb-4f1d-a6e5-b9323fc00b3f-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:16 crc kubenswrapper[4790]: I1124 14:46:16.757360 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11caa523-04fb-4f1d-a6e5-b9323fc00b3f-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:16 crc kubenswrapper[4790]: I1124 14:46:16.757370 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11caa523-04fb-4f1d-a6e5-b9323fc00b3f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:16 crc kubenswrapper[4790]: I1124 14:46:16.757381 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11caa523-04fb-4f1d-a6e5-b9323fc00b3f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:16 crc kubenswrapper[4790]: I1124 14:46:16.757391 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjzsk\" (UniqueName: \"kubernetes.io/projected/11caa523-04fb-4f1d-a6e5-b9323fc00b3f-kube-api-access-vjzsk\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:17 crc kubenswrapper[4790]: I1124 14:46:17.398444 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbb866757-8frpq" event={"ID":"11caa523-04fb-4f1d-a6e5-b9323fc00b3f","Type":"ContainerDied","Data":"a5c06777417489fcfce2dfb1348897c317b73255126f723560249b448056f2b5"} Nov 24 14:46:17 crc kubenswrapper[4790]: I1124 14:46:17.399499 4790 scope.go:117] "RemoveContainer" containerID="8fc61441c83ed7c15701f95649565263bde6d8a5e7518cddc59e2934f3b1a193" Nov 24 14:46:17 crc kubenswrapper[4790]: I1124 14:46:17.398619 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbb866757-8frpq" Nov 24 14:46:17 crc kubenswrapper[4790]: I1124 14:46:17.426678 4790 scope.go:117] "RemoveContainer" containerID="3097bfa54223828dfffefb07013b4ee068ae2b0dfca245a5c841dd79c726304c" Nov 24 14:46:17 crc kubenswrapper[4790]: I1124 14:46:17.471173 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbb866757-8frpq"] Nov 24 14:46:17 crc kubenswrapper[4790]: I1124 14:46:17.480902 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bbb866757-8frpq"] Nov 24 14:46:18 crc kubenswrapper[4790]: I1124 14:46:18.340332 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11caa523-04fb-4f1d-a6e5-b9323fc00b3f" path="/var/lib/kubelet/pods/11caa523-04fb-4f1d-a6e5-b9323fc00b3f/volumes" Nov 24 14:46:19 crc kubenswrapper[4790]: I1124 14:46:19.063370 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-vf46c"] Nov 24 14:46:19 crc kubenswrapper[4790]: E1124 14:46:19.063809 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11caa523-04fb-4f1d-a6e5-b9323fc00b3f" containerName="init" Nov 24 14:46:19 crc kubenswrapper[4790]: I1124 14:46:19.063828 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="11caa523-04fb-4f1d-a6e5-b9323fc00b3f" containerName="init" Nov 24 14:46:19 crc kubenswrapper[4790]: E1124 14:46:19.063841 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11caa523-04fb-4f1d-a6e5-b9323fc00b3f" containerName="dnsmasq-dns" Nov 24 14:46:19 crc kubenswrapper[4790]: I1124 14:46:19.063851 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="11caa523-04fb-4f1d-a6e5-b9323fc00b3f" containerName="dnsmasq-dns" Nov 24 14:46:19 crc kubenswrapper[4790]: I1124 14:46:19.064202 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="11caa523-04fb-4f1d-a6e5-b9323fc00b3f" containerName="dnsmasq-dns" Nov 24 14:46:19 crc kubenswrapper[4790]: I1124 14:46:19.065523 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vf46c" Nov 24 14:46:19 crc kubenswrapper[4790]: I1124 14:46:19.078960 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-vf46c"] Nov 24 14:46:19 crc kubenswrapper[4790]: I1124 14:46:19.161119 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-f478-account-create-lpf9g"] Nov 24 14:46:19 crc kubenswrapper[4790]: I1124 14:46:19.162648 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f478-account-create-lpf9g" Nov 24 14:46:19 crc kubenswrapper[4790]: I1124 14:46:19.168876 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 24 14:46:19 crc kubenswrapper[4790]: I1124 14:46:19.173507 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-f478-account-create-lpf9g"] Nov 24 14:46:19 crc kubenswrapper[4790]: I1124 14:46:19.203019 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9b185743-a8f8-43bc-bdd3-c62e5275309e-operator-scripts\") pod \"cinder-db-create-vf46c\" (UID: \"9b185743-a8f8-43bc-bdd3-c62e5275309e\") " pod="openstack/cinder-db-create-vf46c" Nov 24 14:46:19 crc kubenswrapper[4790]: I1124 14:46:19.203195 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8g6t\" (UniqueName: \"kubernetes.io/projected/9b185743-a8f8-43bc-bdd3-c62e5275309e-kube-api-access-g8g6t\") pod \"cinder-db-create-vf46c\" (UID: \"9b185743-a8f8-43bc-bdd3-c62e5275309e\") " pod="openstack/cinder-db-create-vf46c" Nov 24 14:46:19 crc kubenswrapper[4790]: I1124 14:46:19.305378 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c65f65d6-1e55-46a4-a310-ae5fa73ae854-operator-scripts\") pod \"cinder-f478-account-create-lpf9g\" (UID: \"c65f65d6-1e55-46a4-a310-ae5fa73ae854\") " pod="openstack/cinder-f478-account-create-lpf9g" Nov 24 14:46:19 crc kubenswrapper[4790]: I1124 14:46:19.305541 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwmnt\" (UniqueName: \"kubernetes.io/projected/c65f65d6-1e55-46a4-a310-ae5fa73ae854-kube-api-access-jwmnt\") pod \"cinder-f478-account-create-lpf9g\" (UID: \"c65f65d6-1e55-46a4-a310-ae5fa73ae854\") " pod="openstack/cinder-f478-account-create-lpf9g" Nov 24 14:46:19 crc kubenswrapper[4790]: I1124 14:46:19.305664 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8g6t\" (UniqueName: \"kubernetes.io/projected/9b185743-a8f8-43bc-bdd3-c62e5275309e-kube-api-access-g8g6t\") pod \"cinder-db-create-vf46c\" (UID: \"9b185743-a8f8-43bc-bdd3-c62e5275309e\") " pod="openstack/cinder-db-create-vf46c" Nov 24 14:46:19 crc kubenswrapper[4790]: I1124 14:46:19.305913 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9b185743-a8f8-43bc-bdd3-c62e5275309e-operator-scripts\") pod \"cinder-db-create-vf46c\" (UID: \"9b185743-a8f8-43bc-bdd3-c62e5275309e\") " pod="openstack/cinder-db-create-vf46c" Nov 24 14:46:19 crc kubenswrapper[4790]: I1124 14:46:19.306838 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9b185743-a8f8-43bc-bdd3-c62e5275309e-operator-scripts\") pod \"cinder-db-create-vf46c\" (UID: \"9b185743-a8f8-43bc-bdd3-c62e5275309e\") " pod="openstack/cinder-db-create-vf46c" Nov 24 14:46:19 crc kubenswrapper[4790]: I1124 14:46:19.324787 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8g6t\" (UniqueName: \"kubernetes.io/projected/9b185743-a8f8-43bc-bdd3-c62e5275309e-kube-api-access-g8g6t\") pod \"cinder-db-create-vf46c\" (UID: \"9b185743-a8f8-43bc-bdd3-c62e5275309e\") " pod="openstack/cinder-db-create-vf46c" Nov 24 14:46:19 crc kubenswrapper[4790]: I1124 14:46:19.388720 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vf46c" Nov 24 14:46:19 crc kubenswrapper[4790]: I1124 14:46:19.407200 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c65f65d6-1e55-46a4-a310-ae5fa73ae854-operator-scripts\") pod \"cinder-f478-account-create-lpf9g\" (UID: \"c65f65d6-1e55-46a4-a310-ae5fa73ae854\") " pod="openstack/cinder-f478-account-create-lpf9g" Nov 24 14:46:19 crc kubenswrapper[4790]: I1124 14:46:19.407266 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwmnt\" (UniqueName: \"kubernetes.io/projected/c65f65d6-1e55-46a4-a310-ae5fa73ae854-kube-api-access-jwmnt\") pod \"cinder-f478-account-create-lpf9g\" (UID: \"c65f65d6-1e55-46a4-a310-ae5fa73ae854\") " pod="openstack/cinder-f478-account-create-lpf9g" Nov 24 14:46:19 crc kubenswrapper[4790]: I1124 14:46:19.408383 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c65f65d6-1e55-46a4-a310-ae5fa73ae854-operator-scripts\") pod \"cinder-f478-account-create-lpf9g\" (UID: \"c65f65d6-1e55-46a4-a310-ae5fa73ae854\") " pod="openstack/cinder-f478-account-create-lpf9g" Nov 24 14:46:19 crc kubenswrapper[4790]: I1124 14:46:19.432740 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwmnt\" (UniqueName: \"kubernetes.io/projected/c65f65d6-1e55-46a4-a310-ae5fa73ae854-kube-api-access-jwmnt\") pod \"cinder-f478-account-create-lpf9g\" (UID: \"c65f65d6-1e55-46a4-a310-ae5fa73ae854\") " pod="openstack/cinder-f478-account-create-lpf9g" Nov 24 14:46:19 crc kubenswrapper[4790]: I1124 14:46:19.484752 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f478-account-create-lpf9g" Nov 24 14:46:19 crc kubenswrapper[4790]: I1124 14:46:19.880161 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-vf46c"] Nov 24 14:46:19 crc kubenswrapper[4790]: I1124 14:46:19.976107 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-f478-account-create-lpf9g"] Nov 24 14:46:19 crc kubenswrapper[4790]: W1124 14:46:19.977190 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc65f65d6_1e55_46a4_a310_ae5fa73ae854.slice/crio-7219d8e780b702e4c9b1f030e77a5ba123b33a6483d53ab33cc6fa9f754ba9fe WatchSource:0}: Error finding container 7219d8e780b702e4c9b1f030e77a5ba123b33a6483d53ab33cc6fa9f754ba9fe: Status 404 returned error can't find the container with id 7219d8e780b702e4c9b1f030e77a5ba123b33a6483d53ab33cc6fa9f754ba9fe Nov 24 14:46:20 crc kubenswrapper[4790]: I1124 14:46:20.444062 4790 generic.go:334] "Generic (PLEG): container finished" podID="9b185743-a8f8-43bc-bdd3-c62e5275309e" containerID="1aca79fde163fdd085dbb216e6ef776da218068f8d70b5ad9db5568f87c7d80f" exitCode=0 Nov 24 14:46:20 crc kubenswrapper[4790]: I1124 14:46:20.444134 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vf46c" event={"ID":"9b185743-a8f8-43bc-bdd3-c62e5275309e","Type":"ContainerDied","Data":"1aca79fde163fdd085dbb216e6ef776da218068f8d70b5ad9db5568f87c7d80f"} Nov 24 14:46:20 crc kubenswrapper[4790]: I1124 14:46:20.444693 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vf46c" event={"ID":"9b185743-a8f8-43bc-bdd3-c62e5275309e","Type":"ContainerStarted","Data":"e6bcc1a4b0937e9248895033192a5af79181ea3fa7df34a4d9f18573e6e1d6ec"} Nov 24 14:46:20 crc kubenswrapper[4790]: I1124 14:46:20.446581 4790 generic.go:334] "Generic (PLEG): container finished" podID="c65f65d6-1e55-46a4-a310-ae5fa73ae854" containerID="6e0567e71c58f0b072f917c99cde349dcf07f95554b9b7029917e653f0b4cabf" exitCode=0 Nov 24 14:46:20 crc kubenswrapper[4790]: I1124 14:46:20.446615 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f478-account-create-lpf9g" event={"ID":"c65f65d6-1e55-46a4-a310-ae5fa73ae854","Type":"ContainerDied","Data":"6e0567e71c58f0b072f917c99cde349dcf07f95554b9b7029917e653f0b4cabf"} Nov 24 14:46:20 crc kubenswrapper[4790]: I1124 14:46:20.446633 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f478-account-create-lpf9g" event={"ID":"c65f65d6-1e55-46a4-a310-ae5fa73ae854","Type":"ContainerStarted","Data":"7219d8e780b702e4c9b1f030e77a5ba123b33a6483d53ab33cc6fa9f754ba9fe"} Nov 24 14:46:21 crc kubenswrapper[4790]: I1124 14:46:21.831404 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vf46c" Nov 24 14:46:21 crc kubenswrapper[4790]: I1124 14:46:21.837234 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f478-account-create-lpf9g" Nov 24 14:46:21 crc kubenswrapper[4790]: I1124 14:46:21.956935 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c65f65d6-1e55-46a4-a310-ae5fa73ae854-operator-scripts\") pod \"c65f65d6-1e55-46a4-a310-ae5fa73ae854\" (UID: \"c65f65d6-1e55-46a4-a310-ae5fa73ae854\") " Nov 24 14:46:21 crc kubenswrapper[4790]: I1124 14:46:21.957005 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8g6t\" (UniqueName: \"kubernetes.io/projected/9b185743-a8f8-43bc-bdd3-c62e5275309e-kube-api-access-g8g6t\") pod \"9b185743-a8f8-43bc-bdd3-c62e5275309e\" (UID: \"9b185743-a8f8-43bc-bdd3-c62e5275309e\") " Nov 24 14:46:21 crc kubenswrapper[4790]: I1124 14:46:21.957055 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9b185743-a8f8-43bc-bdd3-c62e5275309e-operator-scripts\") pod \"9b185743-a8f8-43bc-bdd3-c62e5275309e\" (UID: \"9b185743-a8f8-43bc-bdd3-c62e5275309e\") " Nov 24 14:46:21 crc kubenswrapper[4790]: I1124 14:46:21.957152 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwmnt\" (UniqueName: \"kubernetes.io/projected/c65f65d6-1e55-46a4-a310-ae5fa73ae854-kube-api-access-jwmnt\") pod \"c65f65d6-1e55-46a4-a310-ae5fa73ae854\" (UID: \"c65f65d6-1e55-46a4-a310-ae5fa73ae854\") " Nov 24 14:46:21 crc kubenswrapper[4790]: I1124 14:46:21.957751 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b185743-a8f8-43bc-bdd3-c62e5275309e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9b185743-a8f8-43bc-bdd3-c62e5275309e" (UID: "9b185743-a8f8-43bc-bdd3-c62e5275309e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:46:21 crc kubenswrapper[4790]: I1124 14:46:21.957776 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c65f65d6-1e55-46a4-a310-ae5fa73ae854-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c65f65d6-1e55-46a4-a310-ae5fa73ae854" (UID: "c65f65d6-1e55-46a4-a310-ae5fa73ae854"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:46:21 crc kubenswrapper[4790]: I1124 14:46:21.966616 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c65f65d6-1e55-46a4-a310-ae5fa73ae854-kube-api-access-jwmnt" (OuterVolumeSpecName: "kube-api-access-jwmnt") pod "c65f65d6-1e55-46a4-a310-ae5fa73ae854" (UID: "c65f65d6-1e55-46a4-a310-ae5fa73ae854"). InnerVolumeSpecName "kube-api-access-jwmnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:46:21 crc kubenswrapper[4790]: I1124 14:46:21.966689 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b185743-a8f8-43bc-bdd3-c62e5275309e-kube-api-access-g8g6t" (OuterVolumeSpecName: "kube-api-access-g8g6t") pod "9b185743-a8f8-43bc-bdd3-c62e5275309e" (UID: "9b185743-a8f8-43bc-bdd3-c62e5275309e"). InnerVolumeSpecName "kube-api-access-g8g6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:46:22 crc kubenswrapper[4790]: I1124 14:46:22.059276 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c65f65d6-1e55-46a4-a310-ae5fa73ae854-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:22 crc kubenswrapper[4790]: I1124 14:46:22.059323 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8g6t\" (UniqueName: \"kubernetes.io/projected/9b185743-a8f8-43bc-bdd3-c62e5275309e-kube-api-access-g8g6t\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:22 crc kubenswrapper[4790]: I1124 14:46:22.059337 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9b185743-a8f8-43bc-bdd3-c62e5275309e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:22 crc kubenswrapper[4790]: I1124 14:46:22.059349 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwmnt\" (UniqueName: \"kubernetes.io/projected/c65f65d6-1e55-46a4-a310-ae5fa73ae854-kube-api-access-jwmnt\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:22 crc kubenswrapper[4790]: I1124 14:46:22.475645 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vf46c" event={"ID":"9b185743-a8f8-43bc-bdd3-c62e5275309e","Type":"ContainerDied","Data":"e6bcc1a4b0937e9248895033192a5af79181ea3fa7df34a4d9f18573e6e1d6ec"} Nov 24 14:46:22 crc kubenswrapper[4790]: I1124 14:46:22.475695 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6bcc1a4b0937e9248895033192a5af79181ea3fa7df34a4d9f18573e6e1d6ec" Nov 24 14:46:22 crc kubenswrapper[4790]: I1124 14:46:22.475690 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vf46c" Nov 24 14:46:22 crc kubenswrapper[4790]: I1124 14:46:22.478419 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f478-account-create-lpf9g" event={"ID":"c65f65d6-1e55-46a4-a310-ae5fa73ae854","Type":"ContainerDied","Data":"7219d8e780b702e4c9b1f030e77a5ba123b33a6483d53ab33cc6fa9f754ba9fe"} Nov 24 14:46:22 crc kubenswrapper[4790]: I1124 14:46:22.478447 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7219d8e780b702e4c9b1f030e77a5ba123b33a6483d53ab33cc6fa9f754ba9fe" Nov 24 14:46:22 crc kubenswrapper[4790]: I1124 14:46:22.478502 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f478-account-create-lpf9g" Nov 24 14:46:24 crc kubenswrapper[4790]: I1124 14:46:24.455352 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-tjhvv"] Nov 24 14:46:24 crc kubenswrapper[4790]: E1124 14:46:24.456075 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c65f65d6-1e55-46a4-a310-ae5fa73ae854" containerName="mariadb-account-create" Nov 24 14:46:24 crc kubenswrapper[4790]: I1124 14:46:24.456090 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="c65f65d6-1e55-46a4-a310-ae5fa73ae854" containerName="mariadb-account-create" Nov 24 14:46:24 crc kubenswrapper[4790]: E1124 14:46:24.456110 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b185743-a8f8-43bc-bdd3-c62e5275309e" containerName="mariadb-database-create" Nov 24 14:46:24 crc kubenswrapper[4790]: I1124 14:46:24.456119 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b185743-a8f8-43bc-bdd3-c62e5275309e" containerName="mariadb-database-create" Nov 24 14:46:24 crc kubenswrapper[4790]: I1124 14:46:24.456334 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="c65f65d6-1e55-46a4-a310-ae5fa73ae854" containerName="mariadb-account-create" Nov 24 14:46:24 crc kubenswrapper[4790]: I1124 14:46:24.456442 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b185743-a8f8-43bc-bdd3-c62e5275309e" containerName="mariadb-database-create" Nov 24 14:46:24 crc kubenswrapper[4790]: I1124 14:46:24.457189 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-tjhvv" Nov 24 14:46:24 crc kubenswrapper[4790]: I1124 14:46:24.461552 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 24 14:46:24 crc kubenswrapper[4790]: I1124 14:46:24.461687 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-ltds6" Nov 24 14:46:24 crc kubenswrapper[4790]: I1124 14:46:24.462080 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 24 14:46:24 crc kubenswrapper[4790]: I1124 14:46:24.471475 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-tjhvv"] Nov 24 14:46:24 crc kubenswrapper[4790]: I1124 14:46:24.609632 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f89e3f88-652b-4467-9a73-dbcf6514e674-combined-ca-bundle\") pod \"cinder-db-sync-tjhvv\" (UID: \"f89e3f88-652b-4467-9a73-dbcf6514e674\") " pod="openstack/cinder-db-sync-tjhvv" Nov 24 14:46:24 crc kubenswrapper[4790]: I1124 14:46:24.609912 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f89e3f88-652b-4467-9a73-dbcf6514e674-etc-machine-id\") pod \"cinder-db-sync-tjhvv\" (UID: \"f89e3f88-652b-4467-9a73-dbcf6514e674\") " pod="openstack/cinder-db-sync-tjhvv" Nov 24 14:46:24 crc kubenswrapper[4790]: I1124 14:46:24.610022 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f89e3f88-652b-4467-9a73-dbcf6514e674-config-data\") pod \"cinder-db-sync-tjhvv\" (UID: \"f89e3f88-652b-4467-9a73-dbcf6514e674\") " pod="openstack/cinder-db-sync-tjhvv" Nov 24 14:46:24 crc kubenswrapper[4790]: I1124 14:46:24.610154 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f89e3f88-652b-4467-9a73-dbcf6514e674-scripts\") pod \"cinder-db-sync-tjhvv\" (UID: \"f89e3f88-652b-4467-9a73-dbcf6514e674\") " pod="openstack/cinder-db-sync-tjhvv" Nov 24 14:46:24 crc kubenswrapper[4790]: I1124 14:46:24.610258 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f89e3f88-652b-4467-9a73-dbcf6514e674-db-sync-config-data\") pod \"cinder-db-sync-tjhvv\" (UID: \"f89e3f88-652b-4467-9a73-dbcf6514e674\") " pod="openstack/cinder-db-sync-tjhvv" Nov 24 14:46:24 crc kubenswrapper[4790]: I1124 14:46:24.610373 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97wg6\" (UniqueName: \"kubernetes.io/projected/f89e3f88-652b-4467-9a73-dbcf6514e674-kube-api-access-97wg6\") pod \"cinder-db-sync-tjhvv\" (UID: \"f89e3f88-652b-4467-9a73-dbcf6514e674\") " pod="openstack/cinder-db-sync-tjhvv" Nov 24 14:46:24 crc kubenswrapper[4790]: I1124 14:46:24.712150 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97wg6\" (UniqueName: \"kubernetes.io/projected/f89e3f88-652b-4467-9a73-dbcf6514e674-kube-api-access-97wg6\") pod \"cinder-db-sync-tjhvv\" (UID: \"f89e3f88-652b-4467-9a73-dbcf6514e674\") " pod="openstack/cinder-db-sync-tjhvv" Nov 24 14:46:24 crc kubenswrapper[4790]: I1124 14:46:24.712324 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f89e3f88-652b-4467-9a73-dbcf6514e674-combined-ca-bundle\") pod \"cinder-db-sync-tjhvv\" (UID: \"f89e3f88-652b-4467-9a73-dbcf6514e674\") " pod="openstack/cinder-db-sync-tjhvv" Nov 24 14:46:24 crc kubenswrapper[4790]: I1124 14:46:24.712373 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f89e3f88-652b-4467-9a73-dbcf6514e674-etc-machine-id\") pod \"cinder-db-sync-tjhvv\" (UID: \"f89e3f88-652b-4467-9a73-dbcf6514e674\") " pod="openstack/cinder-db-sync-tjhvv" Nov 24 14:46:24 crc kubenswrapper[4790]: I1124 14:46:24.712426 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f89e3f88-652b-4467-9a73-dbcf6514e674-config-data\") pod \"cinder-db-sync-tjhvv\" (UID: \"f89e3f88-652b-4467-9a73-dbcf6514e674\") " pod="openstack/cinder-db-sync-tjhvv" Nov 24 14:46:24 crc kubenswrapper[4790]: I1124 14:46:24.712520 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f89e3f88-652b-4467-9a73-dbcf6514e674-etc-machine-id\") pod \"cinder-db-sync-tjhvv\" (UID: \"f89e3f88-652b-4467-9a73-dbcf6514e674\") " pod="openstack/cinder-db-sync-tjhvv" Nov 24 14:46:24 crc kubenswrapper[4790]: I1124 14:46:24.712531 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f89e3f88-652b-4467-9a73-dbcf6514e674-scripts\") pod \"cinder-db-sync-tjhvv\" (UID: \"f89e3f88-652b-4467-9a73-dbcf6514e674\") " pod="openstack/cinder-db-sync-tjhvv" Nov 24 14:46:24 crc kubenswrapper[4790]: I1124 14:46:24.712769 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f89e3f88-652b-4467-9a73-dbcf6514e674-db-sync-config-data\") pod \"cinder-db-sync-tjhvv\" (UID: \"f89e3f88-652b-4467-9a73-dbcf6514e674\") " pod="openstack/cinder-db-sync-tjhvv" Nov 24 14:46:24 crc kubenswrapper[4790]: I1124 14:46:24.718230 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f89e3f88-652b-4467-9a73-dbcf6514e674-scripts\") pod \"cinder-db-sync-tjhvv\" (UID: \"f89e3f88-652b-4467-9a73-dbcf6514e674\") " pod="openstack/cinder-db-sync-tjhvv" Nov 24 14:46:24 crc kubenswrapper[4790]: I1124 14:46:24.718546 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f89e3f88-652b-4467-9a73-dbcf6514e674-combined-ca-bundle\") pod \"cinder-db-sync-tjhvv\" (UID: \"f89e3f88-652b-4467-9a73-dbcf6514e674\") " pod="openstack/cinder-db-sync-tjhvv" Nov 24 14:46:24 crc kubenswrapper[4790]: I1124 14:46:24.720588 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f89e3f88-652b-4467-9a73-dbcf6514e674-config-data\") pod \"cinder-db-sync-tjhvv\" (UID: \"f89e3f88-652b-4467-9a73-dbcf6514e674\") " pod="openstack/cinder-db-sync-tjhvv" Nov 24 14:46:24 crc kubenswrapper[4790]: I1124 14:46:24.720877 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f89e3f88-652b-4467-9a73-dbcf6514e674-db-sync-config-data\") pod \"cinder-db-sync-tjhvv\" (UID: \"f89e3f88-652b-4467-9a73-dbcf6514e674\") " pod="openstack/cinder-db-sync-tjhvv" Nov 24 14:46:24 crc kubenswrapper[4790]: I1124 14:46:24.739800 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97wg6\" (UniqueName: \"kubernetes.io/projected/f89e3f88-652b-4467-9a73-dbcf6514e674-kube-api-access-97wg6\") pod \"cinder-db-sync-tjhvv\" (UID: \"f89e3f88-652b-4467-9a73-dbcf6514e674\") " pod="openstack/cinder-db-sync-tjhvv" Nov 24 14:46:24 crc kubenswrapper[4790]: I1124 14:46:24.784416 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-tjhvv" Nov 24 14:46:25 crc kubenswrapper[4790]: I1124 14:46:25.300592 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-tjhvv"] Nov 24 14:46:25 crc kubenswrapper[4790]: I1124 14:46:25.511177 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-tjhvv" event={"ID":"f89e3f88-652b-4467-9a73-dbcf6514e674","Type":"ContainerStarted","Data":"07f4edc14e96331475afa8b107c25561a3fc84ea11f801aa2976d46a88f3ae66"} Nov 24 14:46:26 crc kubenswrapper[4790]: I1124 14:46:26.533346 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-tjhvv" event={"ID":"f89e3f88-652b-4467-9a73-dbcf6514e674","Type":"ContainerStarted","Data":"0f08f5749ee72d72ca517437ac8315d8ad0d67680f0eaa3ada7413a972972e1f"} Nov 24 14:46:26 crc kubenswrapper[4790]: I1124 14:46:26.563836 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-tjhvv" podStartSLOduration=2.563809921 podStartE2EDuration="2.563809921s" podCreationTimestamp="2025-11-24 14:46:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:46:26.557691505 +0000 UTC m=+5634.937585227" watchObservedRunningTime="2025-11-24 14:46:26.563809921 +0000 UTC m=+5634.943703623" Nov 24 14:46:28 crc kubenswrapper[4790]: I1124 14:46:28.558022 4790 generic.go:334] "Generic (PLEG): container finished" podID="f89e3f88-652b-4467-9a73-dbcf6514e674" containerID="0f08f5749ee72d72ca517437ac8315d8ad0d67680f0eaa3ada7413a972972e1f" exitCode=0 Nov 24 14:46:28 crc kubenswrapper[4790]: I1124 14:46:28.558073 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-tjhvv" event={"ID":"f89e3f88-652b-4467-9a73-dbcf6514e674","Type":"ContainerDied","Data":"0f08f5749ee72d72ca517437ac8315d8ad0d67680f0eaa3ada7413a972972e1f"} Nov 24 14:46:29 crc kubenswrapper[4790]: I1124 14:46:29.978956 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-tjhvv" Nov 24 14:46:30 crc kubenswrapper[4790]: I1124 14:46:30.130369 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f89e3f88-652b-4467-9a73-dbcf6514e674-combined-ca-bundle\") pod \"f89e3f88-652b-4467-9a73-dbcf6514e674\" (UID: \"f89e3f88-652b-4467-9a73-dbcf6514e674\") " Nov 24 14:46:30 crc kubenswrapper[4790]: I1124 14:46:30.130716 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f89e3f88-652b-4467-9a73-dbcf6514e674-scripts\") pod \"f89e3f88-652b-4467-9a73-dbcf6514e674\" (UID: \"f89e3f88-652b-4467-9a73-dbcf6514e674\") " Nov 24 14:46:30 crc kubenswrapper[4790]: I1124 14:46:30.130810 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f89e3f88-652b-4467-9a73-dbcf6514e674-db-sync-config-data\") pod \"f89e3f88-652b-4467-9a73-dbcf6514e674\" (UID: \"f89e3f88-652b-4467-9a73-dbcf6514e674\") " Nov 24 14:46:30 crc kubenswrapper[4790]: I1124 14:46:30.130853 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f89e3f88-652b-4467-9a73-dbcf6514e674-etc-machine-id\") pod \"f89e3f88-652b-4467-9a73-dbcf6514e674\" (UID: \"f89e3f88-652b-4467-9a73-dbcf6514e674\") " Nov 24 14:46:30 crc kubenswrapper[4790]: I1124 14:46:30.130920 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f89e3f88-652b-4467-9a73-dbcf6514e674-config-data\") pod \"f89e3f88-652b-4467-9a73-dbcf6514e674\" (UID: \"f89e3f88-652b-4467-9a73-dbcf6514e674\") " Nov 24 14:46:30 crc kubenswrapper[4790]: I1124 14:46:30.130972 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97wg6\" (UniqueName: \"kubernetes.io/projected/f89e3f88-652b-4467-9a73-dbcf6514e674-kube-api-access-97wg6\") pod \"f89e3f88-652b-4467-9a73-dbcf6514e674\" (UID: \"f89e3f88-652b-4467-9a73-dbcf6514e674\") " Nov 24 14:46:30 crc kubenswrapper[4790]: I1124 14:46:30.131017 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f89e3f88-652b-4467-9a73-dbcf6514e674-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f89e3f88-652b-4467-9a73-dbcf6514e674" (UID: "f89e3f88-652b-4467-9a73-dbcf6514e674"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:46:30 crc kubenswrapper[4790]: I1124 14:46:30.131284 4790 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f89e3f88-652b-4467-9a73-dbcf6514e674-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:30 crc kubenswrapper[4790]: I1124 14:46:30.136355 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f89e3f88-652b-4467-9a73-dbcf6514e674-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f89e3f88-652b-4467-9a73-dbcf6514e674" (UID: "f89e3f88-652b-4467-9a73-dbcf6514e674"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:46:30 crc kubenswrapper[4790]: I1124 14:46:30.136786 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f89e3f88-652b-4467-9a73-dbcf6514e674-scripts" (OuterVolumeSpecName: "scripts") pod "f89e3f88-652b-4467-9a73-dbcf6514e674" (UID: "f89e3f88-652b-4467-9a73-dbcf6514e674"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:46:30 crc kubenswrapper[4790]: I1124 14:46:30.137146 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f89e3f88-652b-4467-9a73-dbcf6514e674-kube-api-access-97wg6" (OuterVolumeSpecName: "kube-api-access-97wg6") pod "f89e3f88-652b-4467-9a73-dbcf6514e674" (UID: "f89e3f88-652b-4467-9a73-dbcf6514e674"). InnerVolumeSpecName "kube-api-access-97wg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:46:30 crc kubenswrapper[4790]: I1124 14:46:30.183104 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f89e3f88-652b-4467-9a73-dbcf6514e674-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f89e3f88-652b-4467-9a73-dbcf6514e674" (UID: "f89e3f88-652b-4467-9a73-dbcf6514e674"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:46:30 crc kubenswrapper[4790]: I1124 14:46:30.198142 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f89e3f88-652b-4467-9a73-dbcf6514e674-config-data" (OuterVolumeSpecName: "config-data") pod "f89e3f88-652b-4467-9a73-dbcf6514e674" (UID: "f89e3f88-652b-4467-9a73-dbcf6514e674"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:46:30 crc kubenswrapper[4790]: I1124 14:46:30.239726 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f89e3f88-652b-4467-9a73-dbcf6514e674-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:30 crc kubenswrapper[4790]: I1124 14:46:30.239773 4790 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f89e3f88-652b-4467-9a73-dbcf6514e674-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:30 crc kubenswrapper[4790]: I1124 14:46:30.239790 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f89e3f88-652b-4467-9a73-dbcf6514e674-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:30 crc kubenswrapper[4790]: I1124 14:46:30.239803 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97wg6\" (UniqueName: \"kubernetes.io/projected/f89e3f88-652b-4467-9a73-dbcf6514e674-kube-api-access-97wg6\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:30 crc kubenswrapper[4790]: I1124 14:46:30.239821 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f89e3f88-652b-4467-9a73-dbcf6514e674-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:30 crc kubenswrapper[4790]: I1124 14:46:30.583952 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-tjhvv" event={"ID":"f89e3f88-652b-4467-9a73-dbcf6514e674","Type":"ContainerDied","Data":"07f4edc14e96331475afa8b107c25561a3fc84ea11f801aa2976d46a88f3ae66"} Nov 24 14:46:30 crc kubenswrapper[4790]: I1124 14:46:30.584009 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07f4edc14e96331475afa8b107c25561a3fc84ea11f801aa2976d46a88f3ae66" Nov 24 14:46:30 crc kubenswrapper[4790]: I1124 14:46:30.584058 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-tjhvv" Nov 24 14:46:30 crc kubenswrapper[4790]: I1124 14:46:30.928578 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-76c6d786bf-q6t82"] Nov 24 14:46:30 crc kubenswrapper[4790]: E1124 14:46:30.935925 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f89e3f88-652b-4467-9a73-dbcf6514e674" containerName="cinder-db-sync" Nov 24 14:46:30 crc kubenswrapper[4790]: I1124 14:46:30.935965 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="f89e3f88-652b-4467-9a73-dbcf6514e674" containerName="cinder-db-sync" Nov 24 14:46:30 crc kubenswrapper[4790]: I1124 14:46:30.936525 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="f89e3f88-652b-4467-9a73-dbcf6514e674" containerName="cinder-db-sync" Nov 24 14:46:30 crc kubenswrapper[4790]: I1124 14:46:30.937739 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76c6d786bf-q6t82" Nov 24 14:46:30 crc kubenswrapper[4790]: I1124 14:46:30.944951 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76c6d786bf-q6t82"] Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.056726 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20f13a25-af88-4161-ae40-d4a73a9a28f5-ovsdbserver-sb\") pod \"dnsmasq-dns-76c6d786bf-q6t82\" (UID: \"20f13a25-af88-4161-ae40-d4a73a9a28f5\") " pod="openstack/dnsmasq-dns-76c6d786bf-q6t82" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.056835 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20f13a25-af88-4161-ae40-d4a73a9a28f5-ovsdbserver-nb\") pod \"dnsmasq-dns-76c6d786bf-q6t82\" (UID: \"20f13a25-af88-4161-ae40-d4a73a9a28f5\") " pod="openstack/dnsmasq-dns-76c6d786bf-q6t82" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.056869 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4hws\" (UniqueName: \"kubernetes.io/projected/20f13a25-af88-4161-ae40-d4a73a9a28f5-kube-api-access-b4hws\") pod \"dnsmasq-dns-76c6d786bf-q6t82\" (UID: \"20f13a25-af88-4161-ae40-d4a73a9a28f5\") " pod="openstack/dnsmasq-dns-76c6d786bf-q6t82" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.056901 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20f13a25-af88-4161-ae40-d4a73a9a28f5-config\") pod \"dnsmasq-dns-76c6d786bf-q6t82\" (UID: \"20f13a25-af88-4161-ae40-d4a73a9a28f5\") " pod="openstack/dnsmasq-dns-76c6d786bf-q6t82" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.056969 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20f13a25-af88-4161-ae40-d4a73a9a28f5-dns-svc\") pod \"dnsmasq-dns-76c6d786bf-q6t82\" (UID: \"20f13a25-af88-4161-ae40-d4a73a9a28f5\") " pod="openstack/dnsmasq-dns-76c6d786bf-q6t82" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.068054 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.069781 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.075465 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.075674 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.075780 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-ltds6" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.075966 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.080223 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.158839 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3942dd50-c47f-4c59-95e8-5145f6fcc87c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\") " pod="openstack/cinder-api-0" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.159210 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3942dd50-c47f-4c59-95e8-5145f6fcc87c-scripts\") pod \"cinder-api-0\" (UID: \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\") " pod="openstack/cinder-api-0" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.159233 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3942dd50-c47f-4c59-95e8-5145f6fcc87c-logs\") pod \"cinder-api-0\" (UID: \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\") " pod="openstack/cinder-api-0" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.159292 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20f13a25-af88-4161-ae40-d4a73a9a28f5-ovsdbserver-nb\") pod \"dnsmasq-dns-76c6d786bf-q6t82\" (UID: \"20f13a25-af88-4161-ae40-d4a73a9a28f5\") " pod="openstack/dnsmasq-dns-76c6d786bf-q6t82" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.159311 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3942dd50-c47f-4c59-95e8-5145f6fcc87c-config-data\") pod \"cinder-api-0\" (UID: \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\") " pod="openstack/cinder-api-0" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.159329 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmb4c\" (UniqueName: \"kubernetes.io/projected/3942dd50-c47f-4c59-95e8-5145f6fcc87c-kube-api-access-vmb4c\") pod \"cinder-api-0\" (UID: \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\") " pod="openstack/cinder-api-0" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.159354 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4hws\" (UniqueName: \"kubernetes.io/projected/20f13a25-af88-4161-ae40-d4a73a9a28f5-kube-api-access-b4hws\") pod \"dnsmasq-dns-76c6d786bf-q6t82\" (UID: \"20f13a25-af88-4161-ae40-d4a73a9a28f5\") " pod="openstack/dnsmasq-dns-76c6d786bf-q6t82" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.159371 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20f13a25-af88-4161-ae40-d4a73a9a28f5-config\") pod \"dnsmasq-dns-76c6d786bf-q6t82\" (UID: \"20f13a25-af88-4161-ae40-d4a73a9a28f5\") " pod="openstack/dnsmasq-dns-76c6d786bf-q6t82" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.159413 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20f13a25-af88-4161-ae40-d4a73a9a28f5-dns-svc\") pod \"dnsmasq-dns-76c6d786bf-q6t82\" (UID: \"20f13a25-af88-4161-ae40-d4a73a9a28f5\") " pod="openstack/dnsmasq-dns-76c6d786bf-q6t82" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.159439 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3942dd50-c47f-4c59-95e8-5145f6fcc87c-config-data-custom\") pod \"cinder-api-0\" (UID: \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\") " pod="openstack/cinder-api-0" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.159463 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3942dd50-c47f-4c59-95e8-5145f6fcc87c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\") " pod="openstack/cinder-api-0" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.159488 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20f13a25-af88-4161-ae40-d4a73a9a28f5-ovsdbserver-sb\") pod \"dnsmasq-dns-76c6d786bf-q6t82\" (UID: \"20f13a25-af88-4161-ae40-d4a73a9a28f5\") " pod="openstack/dnsmasq-dns-76c6d786bf-q6t82" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.160444 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20f13a25-af88-4161-ae40-d4a73a9a28f5-ovsdbserver-sb\") pod \"dnsmasq-dns-76c6d786bf-q6t82\" (UID: \"20f13a25-af88-4161-ae40-d4a73a9a28f5\") " pod="openstack/dnsmasq-dns-76c6d786bf-q6t82" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.161081 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20f13a25-af88-4161-ae40-d4a73a9a28f5-ovsdbserver-nb\") pod \"dnsmasq-dns-76c6d786bf-q6t82\" (UID: \"20f13a25-af88-4161-ae40-d4a73a9a28f5\") " pod="openstack/dnsmasq-dns-76c6d786bf-q6t82" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.161848 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20f13a25-af88-4161-ae40-d4a73a9a28f5-config\") pod \"dnsmasq-dns-76c6d786bf-q6t82\" (UID: \"20f13a25-af88-4161-ae40-d4a73a9a28f5\") " pod="openstack/dnsmasq-dns-76c6d786bf-q6t82" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.162368 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20f13a25-af88-4161-ae40-d4a73a9a28f5-dns-svc\") pod \"dnsmasq-dns-76c6d786bf-q6t82\" (UID: \"20f13a25-af88-4161-ae40-d4a73a9a28f5\") " pod="openstack/dnsmasq-dns-76c6d786bf-q6t82" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.186070 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4hws\" (UniqueName: \"kubernetes.io/projected/20f13a25-af88-4161-ae40-d4a73a9a28f5-kube-api-access-b4hws\") pod \"dnsmasq-dns-76c6d786bf-q6t82\" (UID: \"20f13a25-af88-4161-ae40-d4a73a9a28f5\") " pod="openstack/dnsmasq-dns-76c6d786bf-q6t82" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.255433 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76c6d786bf-q6t82" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.261669 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3942dd50-c47f-4c59-95e8-5145f6fcc87c-config-data-custom\") pod \"cinder-api-0\" (UID: \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\") " pod="openstack/cinder-api-0" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.261731 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3942dd50-c47f-4c59-95e8-5145f6fcc87c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\") " pod="openstack/cinder-api-0" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.261811 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3942dd50-c47f-4c59-95e8-5145f6fcc87c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\") " pod="openstack/cinder-api-0" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.261846 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3942dd50-c47f-4c59-95e8-5145f6fcc87c-scripts\") pod \"cinder-api-0\" (UID: \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\") " pod="openstack/cinder-api-0" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.261876 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3942dd50-c47f-4c59-95e8-5145f6fcc87c-logs\") pod \"cinder-api-0\" (UID: \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\") " pod="openstack/cinder-api-0" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.261925 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3942dd50-c47f-4c59-95e8-5145f6fcc87c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\") " pod="openstack/cinder-api-0" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.261964 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3942dd50-c47f-4c59-95e8-5145f6fcc87c-config-data\") pod \"cinder-api-0\" (UID: \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\") " pod="openstack/cinder-api-0" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.261991 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmb4c\" (UniqueName: \"kubernetes.io/projected/3942dd50-c47f-4c59-95e8-5145f6fcc87c-kube-api-access-vmb4c\") pod \"cinder-api-0\" (UID: \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\") " pod="openstack/cinder-api-0" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.262484 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3942dd50-c47f-4c59-95e8-5145f6fcc87c-logs\") pod \"cinder-api-0\" (UID: \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\") " pod="openstack/cinder-api-0" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.265153 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3942dd50-c47f-4c59-95e8-5145f6fcc87c-scripts\") pod \"cinder-api-0\" (UID: \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\") " pod="openstack/cinder-api-0" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.269594 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3942dd50-c47f-4c59-95e8-5145f6fcc87c-config-data-custom\") pod \"cinder-api-0\" (UID: \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\") " pod="openstack/cinder-api-0" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.274403 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3942dd50-c47f-4c59-95e8-5145f6fcc87c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\") " pod="openstack/cinder-api-0" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.274470 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3942dd50-c47f-4c59-95e8-5145f6fcc87c-config-data\") pod \"cinder-api-0\" (UID: \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\") " pod="openstack/cinder-api-0" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.282548 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmb4c\" (UniqueName: \"kubernetes.io/projected/3942dd50-c47f-4c59-95e8-5145f6fcc87c-kube-api-access-vmb4c\") pod \"cinder-api-0\" (UID: \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\") " pod="openstack/cinder-api-0" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.389543 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.729604 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76c6d786bf-q6t82"] Nov 24 14:46:31 crc kubenswrapper[4790]: W1124 14:46:31.881038 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3942dd50_c47f_4c59_95e8_5145f6fcc87c.slice/crio-cc623835879da0b1c0d8789dd0272fcbd1889783d1b7e0e66b93bb9c253e11ec WatchSource:0}: Error finding container cc623835879da0b1c0d8789dd0272fcbd1889783d1b7e0e66b93bb9c253e11ec: Status 404 returned error can't find the container with id cc623835879da0b1c0d8789dd0272fcbd1889783d1b7e0e66b93bb9c253e11ec Nov 24 14:46:31 crc kubenswrapper[4790]: I1124 14:46:31.883325 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 14:46:32 crc kubenswrapper[4790]: I1124 14:46:32.599978 4790 generic.go:334] "Generic (PLEG): container finished" podID="20f13a25-af88-4161-ae40-d4a73a9a28f5" containerID="a5c1f2159c9241a5c830fda662bbbc4261983da3d535ef066afb4531c13910be" exitCode=0 Nov 24 14:46:32 crc kubenswrapper[4790]: I1124 14:46:32.600049 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76c6d786bf-q6t82" event={"ID":"20f13a25-af88-4161-ae40-d4a73a9a28f5","Type":"ContainerDied","Data":"a5c1f2159c9241a5c830fda662bbbc4261983da3d535ef066afb4531c13910be"} Nov 24 14:46:32 crc kubenswrapper[4790]: I1124 14:46:32.601384 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76c6d786bf-q6t82" event={"ID":"20f13a25-af88-4161-ae40-d4a73a9a28f5","Type":"ContainerStarted","Data":"77e613a81a3dc8f8c9ee402a58a37f92813cb995a5478e09f37c8aedb2d19c6a"} Nov 24 14:46:32 crc kubenswrapper[4790]: I1124 14:46:32.617094 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3942dd50-c47f-4c59-95e8-5145f6fcc87c","Type":"ContainerStarted","Data":"86d55160de7791fb15fb8c40316f8bbda418225d85d816ad257a0b100f170727"} Nov 24 14:46:32 crc kubenswrapper[4790]: I1124 14:46:32.617283 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3942dd50-c47f-4c59-95e8-5145f6fcc87c","Type":"ContainerStarted","Data":"cc623835879da0b1c0d8789dd0272fcbd1889783d1b7e0e66b93bb9c253e11ec"} Nov 24 14:46:33 crc kubenswrapper[4790]: I1124 14:46:33.632928 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76c6d786bf-q6t82" event={"ID":"20f13a25-af88-4161-ae40-d4a73a9a28f5","Type":"ContainerStarted","Data":"038cdb82e65884f32057a1a3349f4d2a6e48160457f6d2effd5f1379b132367c"} Nov 24 14:46:33 crc kubenswrapper[4790]: I1124 14:46:33.633315 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-76c6d786bf-q6t82" Nov 24 14:46:33 crc kubenswrapper[4790]: I1124 14:46:33.636086 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3942dd50-c47f-4c59-95e8-5145f6fcc87c","Type":"ContainerStarted","Data":"2a469101e678b8c8d235d59534fde9844becde0d76b4a87a5ab89e7b0dddd473"} Nov 24 14:46:33 crc kubenswrapper[4790]: I1124 14:46:33.636353 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 24 14:46:33 crc kubenswrapper[4790]: I1124 14:46:33.672015 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-76c6d786bf-q6t82" podStartSLOduration=3.671987977 podStartE2EDuration="3.671987977s" podCreationTimestamp="2025-11-24 14:46:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:46:33.659153229 +0000 UTC m=+5642.039046931" watchObservedRunningTime="2025-11-24 14:46:33.671987977 +0000 UTC m=+5642.051881679" Nov 24 14:46:33 crc kubenswrapper[4790]: I1124 14:46:33.695475 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=2.695446801 podStartE2EDuration="2.695446801s" podCreationTimestamp="2025-11-24 14:46:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:46:33.68217667 +0000 UTC m=+5642.062070372" watchObservedRunningTime="2025-11-24 14:46:33.695446801 +0000 UTC m=+5642.075340503" Nov 24 14:46:41 crc kubenswrapper[4790]: I1124 14:46:41.258351 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-76c6d786bf-q6t82" Nov 24 14:46:41 crc kubenswrapper[4790]: I1124 14:46:41.352696 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5869b86df7-lfcd7"] Nov 24 14:46:41 crc kubenswrapper[4790]: I1124 14:46:41.352987 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5869b86df7-lfcd7" podUID="4765f4c7-8776-46c8-9a6d-2050e0201a83" containerName="dnsmasq-dns" containerID="cri-o://e5b1a9d3b9fd1f27c8f1dadb54b561065f1e91adaf02d5112dff492510776d2f" gracePeriod=10 Nov 24 14:46:41 crc kubenswrapper[4790]: I1124 14:46:41.746476 4790 generic.go:334] "Generic (PLEG): container finished" podID="4765f4c7-8776-46c8-9a6d-2050e0201a83" containerID="e5b1a9d3b9fd1f27c8f1dadb54b561065f1e91adaf02d5112dff492510776d2f" exitCode=0 Nov 24 14:46:41 crc kubenswrapper[4790]: I1124 14:46:41.746675 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5869b86df7-lfcd7" event={"ID":"4765f4c7-8776-46c8-9a6d-2050e0201a83","Type":"ContainerDied","Data":"e5b1a9d3b9fd1f27c8f1dadb54b561065f1e91adaf02d5112dff492510776d2f"} Nov 24 14:46:41 crc kubenswrapper[4790]: I1124 14:46:41.880086 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5869b86df7-lfcd7" Nov 24 14:46:41 crc kubenswrapper[4790]: I1124 14:46:41.988193 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4765f4c7-8776-46c8-9a6d-2050e0201a83-dns-svc\") pod \"4765f4c7-8776-46c8-9a6d-2050e0201a83\" (UID: \"4765f4c7-8776-46c8-9a6d-2050e0201a83\") " Nov 24 14:46:41 crc kubenswrapper[4790]: I1124 14:46:41.988259 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4765f4c7-8776-46c8-9a6d-2050e0201a83-ovsdbserver-nb\") pod \"4765f4c7-8776-46c8-9a6d-2050e0201a83\" (UID: \"4765f4c7-8776-46c8-9a6d-2050e0201a83\") " Nov 24 14:46:41 crc kubenswrapper[4790]: I1124 14:46:41.988297 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4765f4c7-8776-46c8-9a6d-2050e0201a83-ovsdbserver-sb\") pod \"4765f4c7-8776-46c8-9a6d-2050e0201a83\" (UID: \"4765f4c7-8776-46c8-9a6d-2050e0201a83\") " Nov 24 14:46:41 crc kubenswrapper[4790]: I1124 14:46:41.988324 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvjrf\" (UniqueName: \"kubernetes.io/projected/4765f4c7-8776-46c8-9a6d-2050e0201a83-kube-api-access-kvjrf\") pod \"4765f4c7-8776-46c8-9a6d-2050e0201a83\" (UID: \"4765f4c7-8776-46c8-9a6d-2050e0201a83\") " Nov 24 14:46:41 crc kubenswrapper[4790]: I1124 14:46:41.988368 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4765f4c7-8776-46c8-9a6d-2050e0201a83-config\") pod \"4765f4c7-8776-46c8-9a6d-2050e0201a83\" (UID: \"4765f4c7-8776-46c8-9a6d-2050e0201a83\") " Nov 24 14:46:42 crc kubenswrapper[4790]: I1124 14:46:42.004775 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4765f4c7-8776-46c8-9a6d-2050e0201a83-kube-api-access-kvjrf" (OuterVolumeSpecName: "kube-api-access-kvjrf") pod "4765f4c7-8776-46c8-9a6d-2050e0201a83" (UID: "4765f4c7-8776-46c8-9a6d-2050e0201a83"). InnerVolumeSpecName "kube-api-access-kvjrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:46:42 crc kubenswrapper[4790]: I1124 14:46:42.067105 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4765f4c7-8776-46c8-9a6d-2050e0201a83-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4765f4c7-8776-46c8-9a6d-2050e0201a83" (UID: "4765f4c7-8776-46c8-9a6d-2050e0201a83"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:46:42 crc kubenswrapper[4790]: I1124 14:46:42.068510 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4765f4c7-8776-46c8-9a6d-2050e0201a83-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4765f4c7-8776-46c8-9a6d-2050e0201a83" (UID: "4765f4c7-8776-46c8-9a6d-2050e0201a83"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:46:42 crc kubenswrapper[4790]: I1124 14:46:42.074211 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4765f4c7-8776-46c8-9a6d-2050e0201a83-config" (OuterVolumeSpecName: "config") pod "4765f4c7-8776-46c8-9a6d-2050e0201a83" (UID: "4765f4c7-8776-46c8-9a6d-2050e0201a83"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:46:42 crc kubenswrapper[4790]: I1124 14:46:42.086338 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4765f4c7-8776-46c8-9a6d-2050e0201a83-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4765f4c7-8776-46c8-9a6d-2050e0201a83" (UID: "4765f4c7-8776-46c8-9a6d-2050e0201a83"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:46:42 crc kubenswrapper[4790]: I1124 14:46:42.090718 4790 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4765f4c7-8776-46c8-9a6d-2050e0201a83-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:42 crc kubenswrapper[4790]: I1124 14:46:42.090766 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4765f4c7-8776-46c8-9a6d-2050e0201a83-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:42 crc kubenswrapper[4790]: I1124 14:46:42.090776 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4765f4c7-8776-46c8-9a6d-2050e0201a83-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:42 crc kubenswrapper[4790]: I1124 14:46:42.090787 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvjrf\" (UniqueName: \"kubernetes.io/projected/4765f4c7-8776-46c8-9a6d-2050e0201a83-kube-api-access-kvjrf\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:42 crc kubenswrapper[4790]: I1124 14:46:42.090796 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4765f4c7-8776-46c8-9a6d-2050e0201a83-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:42 crc kubenswrapper[4790]: I1124 14:46:42.536661 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:46:42 crc kubenswrapper[4790]: I1124 14:46:42.538197 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="a167644f-5164-4943-97a3-6031e71f2f28" containerName="nova-scheduler-scheduler" containerID="cri-o://82ca39eedfff2d3a9609182b4df1638283f236abbef9e67ab174b401f08d4b21" gracePeriod=30 Nov 24 14:46:42 crc kubenswrapper[4790]: I1124 14:46:42.546087 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 14:46:42 crc kubenswrapper[4790]: I1124 14:46:42.546318 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="e4609d59-f11c-470c-a35e-966e0873e051" containerName="nova-cell0-conductor-conductor" containerID="cri-o://3bff9a9454a0a0cf4e1d009263e54b6f1892df0dec107b632524415fcd23984a" gracePeriod=30 Nov 24 14:46:42 crc kubenswrapper[4790]: I1124 14:46:42.553326 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:46:42 crc kubenswrapper[4790]: I1124 14:46:42.553581 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e3a43449-bf3d-4104-a89e-ef4d45191a38" containerName="nova-metadata-log" containerID="cri-o://a964645c523ed311dc320cc96961d73dc4521e7f4cf3e145e4099d7bd09eae4c" gracePeriod=30 Nov 24 14:46:42 crc kubenswrapper[4790]: I1124 14:46:42.553773 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e3a43449-bf3d-4104-a89e-ef4d45191a38" containerName="nova-metadata-metadata" containerID="cri-o://50fdb87f7287bb1564344a783de02660937ef2fce769a9a2a4309f9457eebfe5" gracePeriod=30 Nov 24 14:46:42 crc kubenswrapper[4790]: I1124 14:46:42.703034 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:46:42 crc kubenswrapper[4790]: I1124 14:46:42.703380 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="efc434d3-6cab-448c-a4aa-c441cefaafe2" containerName="nova-api-log" containerID="cri-o://8fc8c835dcd207c0e9672856b4f27278b2b63f832ce4e3ef10ed59cc50b82aa6" gracePeriod=30 Nov 24 14:46:42 crc kubenswrapper[4790]: I1124 14:46:42.703426 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="efc434d3-6cab-448c-a4aa-c441cefaafe2" containerName="nova-api-api" containerID="cri-o://3e021067604d72526d82368a35fae3df9f6210fd0b182908f5c549fbd94d2c3a" gracePeriod=30 Nov 24 14:46:42 crc kubenswrapper[4790]: I1124 14:46:42.722032 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 14:46:42 crc kubenswrapper[4790]: I1124 14:46:42.722283 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="70d99ce8-fc77-4c8a-84ad-092c20c65e1d" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://227c333fcf1503133bd1d880a42aedf4b6cdd3f4bee234545d23a2a38dbb84e1" gracePeriod=30 Nov 24 14:46:42 crc kubenswrapper[4790]: I1124 14:46:42.806193 4790 generic.go:334] "Generic (PLEG): container finished" podID="e3a43449-bf3d-4104-a89e-ef4d45191a38" containerID="a964645c523ed311dc320cc96961d73dc4521e7f4cf3e145e4099d7bd09eae4c" exitCode=143 Nov 24 14:46:42 crc kubenswrapper[4790]: I1124 14:46:42.806262 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e3a43449-bf3d-4104-a89e-ef4d45191a38","Type":"ContainerDied","Data":"a964645c523ed311dc320cc96961d73dc4521e7f4cf3e145e4099d7bd09eae4c"} Nov 24 14:46:42 crc kubenswrapper[4790]: I1124 14:46:42.830425 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5869b86df7-lfcd7" event={"ID":"4765f4c7-8776-46c8-9a6d-2050e0201a83","Type":"ContainerDied","Data":"4a3d812ee3797311df1a8e6f52cb268f9e49795e6e98d7fa55398bc635071de3"} Nov 24 14:46:42 crc kubenswrapper[4790]: I1124 14:46:42.830474 4790 scope.go:117] "RemoveContainer" containerID="e5b1a9d3b9fd1f27c8f1dadb54b561065f1e91adaf02d5112dff492510776d2f" Nov 24 14:46:42 crc kubenswrapper[4790]: I1124 14:46:42.830593 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5869b86df7-lfcd7" Nov 24 14:46:42 crc kubenswrapper[4790]: I1124 14:46:42.874637 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5869b86df7-lfcd7"] Nov 24 14:46:42 crc kubenswrapper[4790]: I1124 14:46:42.885221 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5869b86df7-lfcd7"] Nov 24 14:46:42 crc kubenswrapper[4790]: I1124 14:46:42.887995 4790 scope.go:117] "RemoveContainer" containerID="3bf14ef8b7710dc673a5450025822120a2515aaa43fa74889018f5c0db0be514" Nov 24 14:46:43 crc kubenswrapper[4790]: E1124 14:46:43.358552 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="82ca39eedfff2d3a9609182b4df1638283f236abbef9e67ab174b401f08d4b21" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 14:46:43 crc kubenswrapper[4790]: E1124 14:46:43.360690 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="82ca39eedfff2d3a9609182b4df1638283f236abbef9e67ab174b401f08d4b21" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 14:46:43 crc kubenswrapper[4790]: E1124 14:46:43.364449 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="82ca39eedfff2d3a9609182b4df1638283f236abbef9e67ab174b401f08d4b21" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 14:46:43 crc kubenswrapper[4790]: E1124 14:46:43.364516 4790 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="a167644f-5164-4943-97a3-6031e71f2f28" containerName="nova-scheduler-scheduler" Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.379943 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.710782 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.718087 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.838600 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbsjx\" (UniqueName: \"kubernetes.io/projected/70d99ce8-fc77-4c8a-84ad-092c20c65e1d-kube-api-access-mbsjx\") pod \"70d99ce8-fc77-4c8a-84ad-092c20c65e1d\" (UID: \"70d99ce8-fc77-4c8a-84ad-092c20c65e1d\") " Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.838902 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4609d59-f11c-470c-a35e-966e0873e051-config-data\") pod \"e4609d59-f11c-470c-a35e-966e0873e051\" (UID: \"e4609d59-f11c-470c-a35e-966e0873e051\") " Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.838942 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70d99ce8-fc77-4c8a-84ad-092c20c65e1d-config-data\") pod \"70d99ce8-fc77-4c8a-84ad-092c20c65e1d\" (UID: \"70d99ce8-fc77-4c8a-84ad-092c20c65e1d\") " Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.838970 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70d99ce8-fc77-4c8a-84ad-092c20c65e1d-combined-ca-bundle\") pod \"70d99ce8-fc77-4c8a-84ad-092c20c65e1d\" (UID: \"70d99ce8-fc77-4c8a-84ad-092c20c65e1d\") " Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.839026 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2w9p\" (UniqueName: \"kubernetes.io/projected/e4609d59-f11c-470c-a35e-966e0873e051-kube-api-access-b2w9p\") pod \"e4609d59-f11c-470c-a35e-966e0873e051\" (UID: \"e4609d59-f11c-470c-a35e-966e0873e051\") " Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.839044 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4609d59-f11c-470c-a35e-966e0873e051-combined-ca-bundle\") pod \"e4609d59-f11c-470c-a35e-966e0873e051\" (UID: \"e4609d59-f11c-470c-a35e-966e0873e051\") " Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.839699 4790 generic.go:334] "Generic (PLEG): container finished" podID="efc434d3-6cab-448c-a4aa-c441cefaafe2" containerID="8fc8c835dcd207c0e9672856b4f27278b2b63f832ce4e3ef10ed59cc50b82aa6" exitCode=143 Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.839761 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"efc434d3-6cab-448c-a4aa-c441cefaafe2","Type":"ContainerDied","Data":"8fc8c835dcd207c0e9672856b4f27278b2b63f832ce4e3ef10ed59cc50b82aa6"} Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.841422 4790 generic.go:334] "Generic (PLEG): container finished" podID="e4609d59-f11c-470c-a35e-966e0873e051" containerID="3bff9a9454a0a0cf4e1d009263e54b6f1892df0dec107b632524415fcd23984a" exitCode=0 Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.841478 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.841490 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e4609d59-f11c-470c-a35e-966e0873e051","Type":"ContainerDied","Data":"3bff9a9454a0a0cf4e1d009263e54b6f1892df0dec107b632524415fcd23984a"} Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.841520 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e4609d59-f11c-470c-a35e-966e0873e051","Type":"ContainerDied","Data":"6c6e17cc58137a095331ab1ae549c6bae550cdaaecb96bfbe68ad2caaf8fa982"} Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.841538 4790 scope.go:117] "RemoveContainer" containerID="3bff9a9454a0a0cf4e1d009263e54b6f1892df0dec107b632524415fcd23984a" Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.844673 4790 generic.go:334] "Generic (PLEG): container finished" podID="70d99ce8-fc77-4c8a-84ad-092c20c65e1d" containerID="227c333fcf1503133bd1d880a42aedf4b6cdd3f4bee234545d23a2a38dbb84e1" exitCode=0 Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.844724 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"70d99ce8-fc77-4c8a-84ad-092c20c65e1d","Type":"ContainerDied","Data":"227c333fcf1503133bd1d880a42aedf4b6cdd3f4bee234545d23a2a38dbb84e1"} Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.844758 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"70d99ce8-fc77-4c8a-84ad-092c20c65e1d","Type":"ContainerDied","Data":"3fee0eb97a8176dd0c0e500504c034593a3a627d101974fd2a6c3b2c670462cc"} Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.844819 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.861269 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70d99ce8-fc77-4c8a-84ad-092c20c65e1d-kube-api-access-mbsjx" (OuterVolumeSpecName: "kube-api-access-mbsjx") pod "70d99ce8-fc77-4c8a-84ad-092c20c65e1d" (UID: "70d99ce8-fc77-4c8a-84ad-092c20c65e1d"). InnerVolumeSpecName "kube-api-access-mbsjx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.863610 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4609d59-f11c-470c-a35e-966e0873e051-kube-api-access-b2w9p" (OuterVolumeSpecName: "kube-api-access-b2w9p") pod "e4609d59-f11c-470c-a35e-966e0873e051" (UID: "e4609d59-f11c-470c-a35e-966e0873e051"). InnerVolumeSpecName "kube-api-access-b2w9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.865283 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70d99ce8-fc77-4c8a-84ad-092c20c65e1d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "70d99ce8-fc77-4c8a-84ad-092c20c65e1d" (UID: "70d99ce8-fc77-4c8a-84ad-092c20c65e1d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.866363 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4609d59-f11c-470c-a35e-966e0873e051-config-data" (OuterVolumeSpecName: "config-data") pod "e4609d59-f11c-470c-a35e-966e0873e051" (UID: "e4609d59-f11c-470c-a35e-966e0873e051"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.872929 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70d99ce8-fc77-4c8a-84ad-092c20c65e1d-config-data" (OuterVolumeSpecName: "config-data") pod "70d99ce8-fc77-4c8a-84ad-092c20c65e1d" (UID: "70d99ce8-fc77-4c8a-84ad-092c20c65e1d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.881049 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4609d59-f11c-470c-a35e-966e0873e051-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e4609d59-f11c-470c-a35e-966e0873e051" (UID: "e4609d59-f11c-470c-a35e-966e0873e051"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.929671 4790 scope.go:117] "RemoveContainer" containerID="3bff9a9454a0a0cf4e1d009263e54b6f1892df0dec107b632524415fcd23984a" Nov 24 14:46:43 crc kubenswrapper[4790]: E1124 14:46:43.933301 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bff9a9454a0a0cf4e1d009263e54b6f1892df0dec107b632524415fcd23984a\": container with ID starting with 3bff9a9454a0a0cf4e1d009263e54b6f1892df0dec107b632524415fcd23984a not found: ID does not exist" containerID="3bff9a9454a0a0cf4e1d009263e54b6f1892df0dec107b632524415fcd23984a" Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.933336 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bff9a9454a0a0cf4e1d009263e54b6f1892df0dec107b632524415fcd23984a"} err="failed to get container status \"3bff9a9454a0a0cf4e1d009263e54b6f1892df0dec107b632524415fcd23984a\": rpc error: code = NotFound desc = could not find container \"3bff9a9454a0a0cf4e1d009263e54b6f1892df0dec107b632524415fcd23984a\": container with ID starting with 3bff9a9454a0a0cf4e1d009263e54b6f1892df0dec107b632524415fcd23984a not found: ID does not exist" Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.933356 4790 scope.go:117] "RemoveContainer" containerID="227c333fcf1503133bd1d880a42aedf4b6cdd3f4bee234545d23a2a38dbb84e1" Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.941405 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbsjx\" (UniqueName: \"kubernetes.io/projected/70d99ce8-fc77-4c8a-84ad-092c20c65e1d-kube-api-access-mbsjx\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.941432 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4609d59-f11c-470c-a35e-966e0873e051-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.941442 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70d99ce8-fc77-4c8a-84ad-092c20c65e1d-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.941450 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70d99ce8-fc77-4c8a-84ad-092c20c65e1d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.941462 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2w9p\" (UniqueName: \"kubernetes.io/projected/e4609d59-f11c-470c-a35e-966e0873e051-kube-api-access-b2w9p\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.941473 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4609d59-f11c-470c-a35e-966e0873e051-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.957641 4790 scope.go:117] "RemoveContainer" containerID="227c333fcf1503133bd1d880a42aedf4b6cdd3f4bee234545d23a2a38dbb84e1" Nov 24 14:46:43 crc kubenswrapper[4790]: E1124 14:46:43.958605 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"227c333fcf1503133bd1d880a42aedf4b6cdd3f4bee234545d23a2a38dbb84e1\": container with ID starting with 227c333fcf1503133bd1d880a42aedf4b6cdd3f4bee234545d23a2a38dbb84e1 not found: ID does not exist" containerID="227c333fcf1503133bd1d880a42aedf4b6cdd3f4bee234545d23a2a38dbb84e1" Nov 24 14:46:43 crc kubenswrapper[4790]: I1124 14:46:43.958640 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"227c333fcf1503133bd1d880a42aedf4b6cdd3f4bee234545d23a2a38dbb84e1"} err="failed to get container status \"227c333fcf1503133bd1d880a42aedf4b6cdd3f4bee234545d23a2a38dbb84e1\": rpc error: code = NotFound desc = could not find container \"227c333fcf1503133bd1d880a42aedf4b6cdd3f4bee234545d23a2a38dbb84e1\": container with ID starting with 227c333fcf1503133bd1d880a42aedf4b6cdd3f4bee234545d23a2a38dbb84e1 not found: ID does not exist" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.188519 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.236914 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.257665 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.266163 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.274604 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 14:46:44 crc kubenswrapper[4790]: E1124 14:46:44.275746 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4765f4c7-8776-46c8-9a6d-2050e0201a83" containerName="init" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.275912 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="4765f4c7-8776-46c8-9a6d-2050e0201a83" containerName="init" Nov 24 14:46:44 crc kubenswrapper[4790]: E1124 14:46:44.276000 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4765f4c7-8776-46c8-9a6d-2050e0201a83" containerName="dnsmasq-dns" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.276065 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="4765f4c7-8776-46c8-9a6d-2050e0201a83" containerName="dnsmasq-dns" Nov 24 14:46:44 crc kubenswrapper[4790]: E1124 14:46:44.276154 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70d99ce8-fc77-4c8a-84ad-092c20c65e1d" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.276231 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="70d99ce8-fc77-4c8a-84ad-092c20c65e1d" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 14:46:44 crc kubenswrapper[4790]: E1124 14:46:44.276318 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4609d59-f11c-470c-a35e-966e0873e051" containerName="nova-cell0-conductor-conductor" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.276384 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4609d59-f11c-470c-a35e-966e0873e051" containerName="nova-cell0-conductor-conductor" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.276675 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="70d99ce8-fc77-4c8a-84ad-092c20c65e1d" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.276772 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4609d59-f11c-470c-a35e-966e0873e051" containerName="nova-cell0-conductor-conductor" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.276851 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="4765f4c7-8776-46c8-9a6d-2050e0201a83" containerName="dnsmasq-dns" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.277860 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.280790 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.283978 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.290745 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.291857 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.297406 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.299297 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.325838 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4765f4c7-8776-46c8-9a6d-2050e0201a83" path="/var/lib/kubelet/pods/4765f4c7-8776-46c8-9a6d-2050e0201a83/volumes" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.326582 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70d99ce8-fc77-4c8a-84ad-092c20c65e1d" path="/var/lib/kubelet/pods/70d99ce8-fc77-4c8a-84ad-092c20c65e1d/volumes" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.327190 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4609d59-f11c-470c-a35e-966e0873e051" path="/var/lib/kubelet/pods/e4609d59-f11c-470c-a35e-966e0873e051/volumes" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.450494 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06c8ec77-c053-44cc-b394-e3170e011b3d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"06c8ec77-c053-44cc-b394-e3170e011b3d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.450551 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06c8ec77-c053-44cc-b394-e3170e011b3d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"06c8ec77-c053-44cc-b394-e3170e011b3d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.450676 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mlpq\" (UniqueName: \"kubernetes.io/projected/cda1d09b-2432-4a64-8097-69246eb196a4-kube-api-access-2mlpq\") pod \"nova-cell0-conductor-0\" (UID: \"cda1d09b-2432-4a64-8097-69246eb196a4\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.451381 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kc5dh\" (UniqueName: \"kubernetes.io/projected/06c8ec77-c053-44cc-b394-e3170e011b3d-kube-api-access-kc5dh\") pod \"nova-cell1-novncproxy-0\" (UID: \"06c8ec77-c053-44cc-b394-e3170e011b3d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.451445 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cda1d09b-2432-4a64-8097-69246eb196a4-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"cda1d09b-2432-4a64-8097-69246eb196a4\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.451499 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cda1d09b-2432-4a64-8097-69246eb196a4-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"cda1d09b-2432-4a64-8097-69246eb196a4\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.553350 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kc5dh\" (UniqueName: \"kubernetes.io/projected/06c8ec77-c053-44cc-b394-e3170e011b3d-kube-api-access-kc5dh\") pod \"nova-cell1-novncproxy-0\" (UID: \"06c8ec77-c053-44cc-b394-e3170e011b3d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.553411 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cda1d09b-2432-4a64-8097-69246eb196a4-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"cda1d09b-2432-4a64-8097-69246eb196a4\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.553447 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cda1d09b-2432-4a64-8097-69246eb196a4-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"cda1d09b-2432-4a64-8097-69246eb196a4\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.553595 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06c8ec77-c053-44cc-b394-e3170e011b3d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"06c8ec77-c053-44cc-b394-e3170e011b3d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.553626 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06c8ec77-c053-44cc-b394-e3170e011b3d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"06c8ec77-c053-44cc-b394-e3170e011b3d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.553650 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mlpq\" (UniqueName: \"kubernetes.io/projected/cda1d09b-2432-4a64-8097-69246eb196a4-kube-api-access-2mlpq\") pod \"nova-cell0-conductor-0\" (UID: \"cda1d09b-2432-4a64-8097-69246eb196a4\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.557692 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06c8ec77-c053-44cc-b394-e3170e011b3d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"06c8ec77-c053-44cc-b394-e3170e011b3d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.557821 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cda1d09b-2432-4a64-8097-69246eb196a4-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"cda1d09b-2432-4a64-8097-69246eb196a4\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.557996 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cda1d09b-2432-4a64-8097-69246eb196a4-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"cda1d09b-2432-4a64-8097-69246eb196a4\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.558791 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06c8ec77-c053-44cc-b394-e3170e011b3d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"06c8ec77-c053-44cc-b394-e3170e011b3d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.570355 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kc5dh\" (UniqueName: \"kubernetes.io/projected/06c8ec77-c053-44cc-b394-e3170e011b3d-kube-api-access-kc5dh\") pod \"nova-cell1-novncproxy-0\" (UID: \"06c8ec77-c053-44cc-b394-e3170e011b3d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.570752 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mlpq\" (UniqueName: \"kubernetes.io/projected/cda1d09b-2432-4a64-8097-69246eb196a4-kube-api-access-2mlpq\") pod \"nova-cell0-conductor-0\" (UID: \"cda1d09b-2432-4a64-8097-69246eb196a4\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.602001 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 14:46:44 crc kubenswrapper[4790]: I1124 14:46:44.619828 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:46:45 crc kubenswrapper[4790]: I1124 14:46:45.053499 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 14:46:45 crc kubenswrapper[4790]: W1124 14:46:45.062679 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcda1d09b_2432_4a64_8097_69246eb196a4.slice/crio-1f6ae3c97f494a3db8e448cd3592e357186cd2813c61bf0619565836d892037e WatchSource:0}: Error finding container 1f6ae3c97f494a3db8e448cd3592e357186cd2813c61bf0619565836d892037e: Status 404 returned error can't find the container with id 1f6ae3c97f494a3db8e448cd3592e357186cd2813c61bf0619565836d892037e Nov 24 14:46:45 crc kubenswrapper[4790]: I1124 14:46:45.138794 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zqtws"] Nov 24 14:46:45 crc kubenswrapper[4790]: I1124 14:46:45.141448 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zqtws" Nov 24 14:46:45 crc kubenswrapper[4790]: W1124 14:46:45.142777 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06c8ec77_c053_44cc_b394_e3170e011b3d.slice/crio-74500036a6f8f9f3095af445ee89a6161015c77e61f8d212fbcc35a99891b4ef WatchSource:0}: Error finding container 74500036a6f8f9f3095af445ee89a6161015c77e61f8d212fbcc35a99891b4ef: Status 404 returned error can't find the container with id 74500036a6f8f9f3095af445ee89a6161015c77e61f8d212fbcc35a99891b4ef Nov 24 14:46:45 crc kubenswrapper[4790]: I1124 14:46:45.151519 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 14:46:45 crc kubenswrapper[4790]: I1124 14:46:45.164027 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zqtws"] Nov 24 14:46:45 crc kubenswrapper[4790]: I1124 14:46:45.264873 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fff5214-ae3c-4948-aaea-d5806cfae4b4-utilities\") pod \"community-operators-zqtws\" (UID: \"7fff5214-ae3c-4948-aaea-d5806cfae4b4\") " pod="openshift-marketplace/community-operators-zqtws" Nov 24 14:46:45 crc kubenswrapper[4790]: I1124 14:46:45.265175 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fff5214-ae3c-4948-aaea-d5806cfae4b4-catalog-content\") pod \"community-operators-zqtws\" (UID: \"7fff5214-ae3c-4948-aaea-d5806cfae4b4\") " pod="openshift-marketplace/community-operators-zqtws" Nov 24 14:46:45 crc kubenswrapper[4790]: I1124 14:46:45.265333 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gvhf\" (UniqueName: \"kubernetes.io/projected/7fff5214-ae3c-4948-aaea-d5806cfae4b4-kube-api-access-2gvhf\") pod \"community-operators-zqtws\" (UID: \"7fff5214-ae3c-4948-aaea-d5806cfae4b4\") " pod="openshift-marketplace/community-operators-zqtws" Nov 24 14:46:45 crc kubenswrapper[4790]: I1124 14:46:45.366535 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gvhf\" (UniqueName: \"kubernetes.io/projected/7fff5214-ae3c-4948-aaea-d5806cfae4b4-kube-api-access-2gvhf\") pod \"community-operators-zqtws\" (UID: \"7fff5214-ae3c-4948-aaea-d5806cfae4b4\") " pod="openshift-marketplace/community-operators-zqtws" Nov 24 14:46:45 crc kubenswrapper[4790]: I1124 14:46:45.366669 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fff5214-ae3c-4948-aaea-d5806cfae4b4-utilities\") pod \"community-operators-zqtws\" (UID: \"7fff5214-ae3c-4948-aaea-d5806cfae4b4\") " pod="openshift-marketplace/community-operators-zqtws" Nov 24 14:46:45 crc kubenswrapper[4790]: I1124 14:46:45.366723 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fff5214-ae3c-4948-aaea-d5806cfae4b4-catalog-content\") pod \"community-operators-zqtws\" (UID: \"7fff5214-ae3c-4948-aaea-d5806cfae4b4\") " pod="openshift-marketplace/community-operators-zqtws" Nov 24 14:46:45 crc kubenswrapper[4790]: I1124 14:46:45.367167 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fff5214-ae3c-4948-aaea-d5806cfae4b4-catalog-content\") pod \"community-operators-zqtws\" (UID: \"7fff5214-ae3c-4948-aaea-d5806cfae4b4\") " pod="openshift-marketplace/community-operators-zqtws" Nov 24 14:46:45 crc kubenswrapper[4790]: I1124 14:46:45.367477 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fff5214-ae3c-4948-aaea-d5806cfae4b4-utilities\") pod \"community-operators-zqtws\" (UID: \"7fff5214-ae3c-4948-aaea-d5806cfae4b4\") " pod="openshift-marketplace/community-operators-zqtws" Nov 24 14:46:45 crc kubenswrapper[4790]: I1124 14:46:45.382878 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gvhf\" (UniqueName: \"kubernetes.io/projected/7fff5214-ae3c-4948-aaea-d5806cfae4b4-kube-api-access-2gvhf\") pod \"community-operators-zqtws\" (UID: \"7fff5214-ae3c-4948-aaea-d5806cfae4b4\") " pod="openshift-marketplace/community-operators-zqtws" Nov 24 14:46:45 crc kubenswrapper[4790]: I1124 14:46:45.492951 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zqtws" Nov 24 14:46:45 crc kubenswrapper[4790]: I1124 14:46:45.752388 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="e3a43449-bf3d-4104-a89e-ef4d45191a38" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": read tcp 10.217.0.2:53044->10.217.1.74:8775: read: connection reset by peer" Nov 24 14:46:45 crc kubenswrapper[4790]: I1124 14:46:45.752437 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="e3a43449-bf3d-4104-a89e-ef4d45191a38" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": read tcp 10.217.0.2:53040->10.217.1.74:8775: read: connection reset by peer" Nov 24 14:46:45 crc kubenswrapper[4790]: I1124 14:46:45.870402 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"cda1d09b-2432-4a64-8097-69246eb196a4","Type":"ContainerStarted","Data":"abec8bd8546de72cc3c06177bd23ea100d8a5e09334932e57afe57cd227f16b7"} Nov 24 14:46:45 crc kubenswrapper[4790]: I1124 14:46:45.870695 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"cda1d09b-2432-4a64-8097-69246eb196a4","Type":"ContainerStarted","Data":"1f6ae3c97f494a3db8e448cd3592e357186cd2813c61bf0619565836d892037e"} Nov 24 14:46:45 crc kubenswrapper[4790]: I1124 14:46:45.870743 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 24 14:46:45 crc kubenswrapper[4790]: I1124 14:46:45.872018 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"06c8ec77-c053-44cc-b394-e3170e011b3d","Type":"ContainerStarted","Data":"1266752bf03e0ddf5dd37c1637abb06966f87894637d74d30f660761b3afc93b"} Nov 24 14:46:45 crc kubenswrapper[4790]: I1124 14:46:45.872042 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"06c8ec77-c053-44cc-b394-e3170e011b3d","Type":"ContainerStarted","Data":"74500036a6f8f9f3095af445ee89a6161015c77e61f8d212fbcc35a99891b4ef"} Nov 24 14:46:45 crc kubenswrapper[4790]: I1124 14:46:45.874198 4790 generic.go:334] "Generic (PLEG): container finished" podID="e3a43449-bf3d-4104-a89e-ef4d45191a38" containerID="50fdb87f7287bb1564344a783de02660937ef2fce769a9a2a4309f9457eebfe5" exitCode=0 Nov 24 14:46:45 crc kubenswrapper[4790]: I1124 14:46:45.874230 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e3a43449-bf3d-4104-a89e-ef4d45191a38","Type":"ContainerDied","Data":"50fdb87f7287bb1564344a783de02660937ef2fce769a9a2a4309f9457eebfe5"} Nov 24 14:46:45 crc kubenswrapper[4790]: I1124 14:46:45.885565 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=1.885548956 podStartE2EDuration="1.885548956s" podCreationTimestamp="2025-11-24 14:46:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:46:45.882803157 +0000 UTC m=+5654.262696839" watchObservedRunningTime="2025-11-24 14:46:45.885548956 +0000 UTC m=+5654.265442618" Nov 24 14:46:45 crc kubenswrapper[4790]: I1124 14:46:45.905059 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=1.9050350759999999 podStartE2EDuration="1.905035076s" podCreationTimestamp="2025-11-24 14:46:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:46:45.898753265 +0000 UTC m=+5654.278646927" watchObservedRunningTime="2025-11-24 14:46:45.905035076 +0000 UTC m=+5654.284928738" Nov 24 14:46:45 crc kubenswrapper[4790]: I1124 14:46:45.989826 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-api-0" podUID="efc434d3-6cab-448c-a4aa-c441cefaafe2" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.75:8774/\": read tcp 10.217.0.2:51544->10.217.1.75:8774: read: connection reset by peer" Nov 24 14:46:45 crc kubenswrapper[4790]: I1124 14:46:45.991264 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-api-0" podUID="efc434d3-6cab-448c-a4aa-c441cefaafe2" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.75:8774/\": read tcp 10.217.0.2:51546->10.217.1.75:8774: read: connection reset by peer" Nov 24 14:46:45 crc kubenswrapper[4790]: I1124 14:46:45.997257 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zqtws"] Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.042034 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.042237 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="bfa26363-ed7c-4bd7-a32b-420e94c62a3d" containerName="nova-cell1-conductor-conductor" containerID="cri-o://659aae12e81e3a6ff594ff0fa8c227b2de1d0e0869ca6de31bef7cebd3fce403" gracePeriod=30 Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.106857 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.283311 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3a43449-bf3d-4104-a89e-ef4d45191a38-logs\") pod \"e3a43449-bf3d-4104-a89e-ef4d45191a38\" (UID: \"e3a43449-bf3d-4104-a89e-ef4d45191a38\") " Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.283662 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3a43449-bf3d-4104-a89e-ef4d45191a38-config-data\") pod \"e3a43449-bf3d-4104-a89e-ef4d45191a38\" (UID: \"e3a43449-bf3d-4104-a89e-ef4d45191a38\") " Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.283724 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csmpn\" (UniqueName: \"kubernetes.io/projected/e3a43449-bf3d-4104-a89e-ef4d45191a38-kube-api-access-csmpn\") pod \"e3a43449-bf3d-4104-a89e-ef4d45191a38\" (UID: \"e3a43449-bf3d-4104-a89e-ef4d45191a38\") " Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.283750 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3a43449-bf3d-4104-a89e-ef4d45191a38-combined-ca-bundle\") pod \"e3a43449-bf3d-4104-a89e-ef4d45191a38\" (UID: \"e3a43449-bf3d-4104-a89e-ef4d45191a38\") " Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.285284 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3a43449-bf3d-4104-a89e-ef4d45191a38-logs" (OuterVolumeSpecName: "logs") pod "e3a43449-bf3d-4104-a89e-ef4d45191a38" (UID: "e3a43449-bf3d-4104-a89e-ef4d45191a38"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.289013 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3a43449-bf3d-4104-a89e-ef4d45191a38-kube-api-access-csmpn" (OuterVolumeSpecName: "kube-api-access-csmpn") pod "e3a43449-bf3d-4104-a89e-ef4d45191a38" (UID: "e3a43449-bf3d-4104-a89e-ef4d45191a38"). InnerVolumeSpecName "kube-api-access-csmpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.330076 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3a43449-bf3d-4104-a89e-ef4d45191a38-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e3a43449-bf3d-4104-a89e-ef4d45191a38" (UID: "e3a43449-bf3d-4104-a89e-ef4d45191a38"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.336984 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3a43449-bf3d-4104-a89e-ef4d45191a38-config-data" (OuterVolumeSpecName: "config-data") pod "e3a43449-bf3d-4104-a89e-ef4d45191a38" (UID: "e3a43449-bf3d-4104-a89e-ef4d45191a38"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.388605 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3a43449-bf3d-4104-a89e-ef4d45191a38-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.388637 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csmpn\" (UniqueName: \"kubernetes.io/projected/e3a43449-bf3d-4104-a89e-ef4d45191a38-kube-api-access-csmpn\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.388649 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3a43449-bf3d-4104-a89e-ef4d45191a38-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.388657 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3a43449-bf3d-4104-a89e-ef4d45191a38-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.394545 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.489325 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efc434d3-6cab-448c-a4aa-c441cefaafe2-combined-ca-bundle\") pod \"efc434d3-6cab-448c-a4aa-c441cefaafe2\" (UID: \"efc434d3-6cab-448c-a4aa-c441cefaafe2\") " Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.489385 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6tlwn\" (UniqueName: \"kubernetes.io/projected/efc434d3-6cab-448c-a4aa-c441cefaafe2-kube-api-access-6tlwn\") pod \"efc434d3-6cab-448c-a4aa-c441cefaafe2\" (UID: \"efc434d3-6cab-448c-a4aa-c441cefaafe2\") " Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.489441 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efc434d3-6cab-448c-a4aa-c441cefaafe2-config-data\") pod \"efc434d3-6cab-448c-a4aa-c441cefaafe2\" (UID: \"efc434d3-6cab-448c-a4aa-c441cefaafe2\") " Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.489471 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/efc434d3-6cab-448c-a4aa-c441cefaafe2-logs\") pod \"efc434d3-6cab-448c-a4aa-c441cefaafe2\" (UID: \"efc434d3-6cab-448c-a4aa-c441cefaafe2\") " Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.490274 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efc434d3-6cab-448c-a4aa-c441cefaafe2-logs" (OuterVolumeSpecName: "logs") pod "efc434d3-6cab-448c-a4aa-c441cefaafe2" (UID: "efc434d3-6cab-448c-a4aa-c441cefaafe2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.495321 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efc434d3-6cab-448c-a4aa-c441cefaafe2-kube-api-access-6tlwn" (OuterVolumeSpecName: "kube-api-access-6tlwn") pod "efc434d3-6cab-448c-a4aa-c441cefaafe2" (UID: "efc434d3-6cab-448c-a4aa-c441cefaafe2"). InnerVolumeSpecName "kube-api-access-6tlwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.523031 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efc434d3-6cab-448c-a4aa-c441cefaafe2-config-data" (OuterVolumeSpecName: "config-data") pod "efc434d3-6cab-448c-a4aa-c441cefaafe2" (UID: "efc434d3-6cab-448c-a4aa-c441cefaafe2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.547490 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efc434d3-6cab-448c-a4aa-c441cefaafe2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "efc434d3-6cab-448c-a4aa-c441cefaafe2" (UID: "efc434d3-6cab-448c-a4aa-c441cefaafe2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.595428 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efc434d3-6cab-448c-a4aa-c441cefaafe2-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.595477 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/efc434d3-6cab-448c-a4aa-c441cefaafe2-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.595502 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efc434d3-6cab-448c-a4aa-c441cefaafe2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.595517 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6tlwn\" (UniqueName: \"kubernetes.io/projected/efc434d3-6cab-448c-a4aa-c441cefaafe2-kube-api-access-6tlwn\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.892319 4790 generic.go:334] "Generic (PLEG): container finished" podID="7fff5214-ae3c-4948-aaea-d5806cfae4b4" containerID="11a9761a2b421c4f2f0bb9699504f9e9d6d52d8d1ccecb1661e32368dbfd4689" exitCode=0 Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.892384 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zqtws" event={"ID":"7fff5214-ae3c-4948-aaea-d5806cfae4b4","Type":"ContainerDied","Data":"11a9761a2b421c4f2f0bb9699504f9e9d6d52d8d1ccecb1661e32368dbfd4689"} Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.892410 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zqtws" event={"ID":"7fff5214-ae3c-4948-aaea-d5806cfae4b4","Type":"ContainerStarted","Data":"464a68671754697c1bc78dd6b6b481e776668904c8c45fc9f6d3482dcd8c6d5d"} Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.899359 4790 generic.go:334] "Generic (PLEG): container finished" podID="efc434d3-6cab-448c-a4aa-c441cefaafe2" containerID="3e021067604d72526d82368a35fae3df9f6210fd0b182908f5c549fbd94d2c3a" exitCode=0 Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.899548 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.899655 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"efc434d3-6cab-448c-a4aa-c441cefaafe2","Type":"ContainerDied","Data":"3e021067604d72526d82368a35fae3df9f6210fd0b182908f5c549fbd94d2c3a"} Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.899695 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"efc434d3-6cab-448c-a4aa-c441cefaafe2","Type":"ContainerDied","Data":"e9444683156dfb9b328b0ed05a3b41559491b4fec29ace89ae8d890c434806e2"} Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.899715 4790 scope.go:117] "RemoveContainer" containerID="3e021067604d72526d82368a35fae3df9f6210fd0b182908f5c549fbd94d2c3a" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.902917 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e3a43449-bf3d-4104-a89e-ef4d45191a38","Type":"ContainerDied","Data":"3501e5d87dcd28433f6686936569f742009987b35ed9500cab5dd67818983b38"} Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.903144 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.942650 4790 scope.go:117] "RemoveContainer" containerID="8fc8c835dcd207c0e9672856b4f27278b2b63f832ce4e3ef10ed59cc50b82aa6" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.954308 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.965765 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.974560 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:46:46 crc kubenswrapper[4790]: E1124 14:46:46.974945 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efc434d3-6cab-448c-a4aa-c441cefaafe2" containerName="nova-api-api" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.974957 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="efc434d3-6cab-448c-a4aa-c441cefaafe2" containerName="nova-api-api" Nov 24 14:46:46 crc kubenswrapper[4790]: E1124 14:46:46.974994 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3a43449-bf3d-4104-a89e-ef4d45191a38" containerName="nova-metadata-log" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.975000 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3a43449-bf3d-4104-a89e-ef4d45191a38" containerName="nova-metadata-log" Nov 24 14:46:46 crc kubenswrapper[4790]: E1124 14:46:46.975012 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efc434d3-6cab-448c-a4aa-c441cefaafe2" containerName="nova-api-log" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.975019 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="efc434d3-6cab-448c-a4aa-c441cefaafe2" containerName="nova-api-log" Nov 24 14:46:46 crc kubenswrapper[4790]: E1124 14:46:46.975030 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3a43449-bf3d-4104-a89e-ef4d45191a38" containerName="nova-metadata-metadata" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.975035 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3a43449-bf3d-4104-a89e-ef4d45191a38" containerName="nova-metadata-metadata" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.975189 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="efc434d3-6cab-448c-a4aa-c441cefaafe2" containerName="nova-api-api" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.975206 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3a43449-bf3d-4104-a89e-ef4d45191a38" containerName="nova-metadata-metadata" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.975220 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="efc434d3-6cab-448c-a4aa-c441cefaafe2" containerName="nova-api-log" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.975229 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3a43449-bf3d-4104-a89e-ef4d45191a38" containerName="nova-metadata-log" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.976142 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.978029 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.988195 4790 scope.go:117] "RemoveContainer" containerID="3e021067604d72526d82368a35fae3df9f6210fd0b182908f5c549fbd94d2c3a" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.988349 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:46:46 crc kubenswrapper[4790]: E1124 14:46:46.994871 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e021067604d72526d82368a35fae3df9f6210fd0b182908f5c549fbd94d2c3a\": container with ID starting with 3e021067604d72526d82368a35fae3df9f6210fd0b182908f5c549fbd94d2c3a not found: ID does not exist" containerID="3e021067604d72526d82368a35fae3df9f6210fd0b182908f5c549fbd94d2c3a" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.995016 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e021067604d72526d82368a35fae3df9f6210fd0b182908f5c549fbd94d2c3a"} err="failed to get container status \"3e021067604d72526d82368a35fae3df9f6210fd0b182908f5c549fbd94d2c3a\": rpc error: code = NotFound desc = could not find container \"3e021067604d72526d82368a35fae3df9f6210fd0b182908f5c549fbd94d2c3a\": container with ID starting with 3e021067604d72526d82368a35fae3df9f6210fd0b182908f5c549fbd94d2c3a not found: ID does not exist" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.995048 4790 scope.go:117] "RemoveContainer" containerID="8fc8c835dcd207c0e9672856b4f27278b2b63f832ce4e3ef10ed59cc50b82aa6" Nov 24 14:46:46 crc kubenswrapper[4790]: E1124 14:46:46.995641 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fc8c835dcd207c0e9672856b4f27278b2b63f832ce4e3ef10ed59cc50b82aa6\": container with ID starting with 8fc8c835dcd207c0e9672856b4f27278b2b63f832ce4e3ef10ed59cc50b82aa6 not found: ID does not exist" containerID="8fc8c835dcd207c0e9672856b4f27278b2b63f832ce4e3ef10ed59cc50b82aa6" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.995668 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fc8c835dcd207c0e9672856b4f27278b2b63f832ce4e3ef10ed59cc50b82aa6"} err="failed to get container status \"8fc8c835dcd207c0e9672856b4f27278b2b63f832ce4e3ef10ed59cc50b82aa6\": rpc error: code = NotFound desc = could not find container \"8fc8c835dcd207c0e9672856b4f27278b2b63f832ce4e3ef10ed59cc50b82aa6\": container with ID starting with 8fc8c835dcd207c0e9672856b4f27278b2b63f832ce4e3ef10ed59cc50b82aa6 not found: ID does not exist" Nov 24 14:46:46 crc kubenswrapper[4790]: I1124 14:46:46.995686 4790 scope.go:117] "RemoveContainer" containerID="50fdb87f7287bb1564344a783de02660937ef2fce769a9a2a4309f9457eebfe5" Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.001104 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.013797 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.022810 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.024924 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.028349 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.034835 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.043588 4790 scope.go:117] "RemoveContainer" containerID="a964645c523ed311dc320cc96961d73dc4521e7f4cf3e145e4099d7bd09eae4c" Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.110382 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/924b2c00-a948-46d2-b103-25f99bc9477e-logs\") pod \"nova-metadata-0\" (UID: \"924b2c00-a948-46d2-b103-25f99bc9477e\") " pod="openstack/nova-metadata-0" Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.110465 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/039efc51-838d-4aac-b0e8-7ba1c9b376ce-config-data\") pod \"nova-api-0\" (UID: \"039efc51-838d-4aac-b0e8-7ba1c9b376ce\") " pod="openstack/nova-api-0" Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.110507 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/924b2c00-a948-46d2-b103-25f99bc9477e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"924b2c00-a948-46d2-b103-25f99bc9477e\") " pod="openstack/nova-metadata-0" Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.110539 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4z2t\" (UniqueName: \"kubernetes.io/projected/924b2c00-a948-46d2-b103-25f99bc9477e-kube-api-access-b4z2t\") pod \"nova-metadata-0\" (UID: \"924b2c00-a948-46d2-b103-25f99bc9477e\") " pod="openstack/nova-metadata-0" Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.110565 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/924b2c00-a948-46d2-b103-25f99bc9477e-config-data\") pod \"nova-metadata-0\" (UID: \"924b2c00-a948-46d2-b103-25f99bc9477e\") " pod="openstack/nova-metadata-0" Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.110604 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/039efc51-838d-4aac-b0e8-7ba1c9b376ce-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"039efc51-838d-4aac-b0e8-7ba1c9b376ce\") " pod="openstack/nova-api-0" Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.110620 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/039efc51-838d-4aac-b0e8-7ba1c9b376ce-logs\") pod \"nova-api-0\" (UID: \"039efc51-838d-4aac-b0e8-7ba1c9b376ce\") " pod="openstack/nova-api-0" Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.110663 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4lpw\" (UniqueName: \"kubernetes.io/projected/039efc51-838d-4aac-b0e8-7ba1c9b376ce-kube-api-access-t4lpw\") pod \"nova-api-0\" (UID: \"039efc51-838d-4aac-b0e8-7ba1c9b376ce\") " pod="openstack/nova-api-0" Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.211774 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/924b2c00-a948-46d2-b103-25f99bc9477e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"924b2c00-a948-46d2-b103-25f99bc9477e\") " pod="openstack/nova-metadata-0" Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.211830 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4z2t\" (UniqueName: \"kubernetes.io/projected/924b2c00-a948-46d2-b103-25f99bc9477e-kube-api-access-b4z2t\") pod \"nova-metadata-0\" (UID: \"924b2c00-a948-46d2-b103-25f99bc9477e\") " pod="openstack/nova-metadata-0" Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.211852 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/924b2c00-a948-46d2-b103-25f99bc9477e-config-data\") pod \"nova-metadata-0\" (UID: \"924b2c00-a948-46d2-b103-25f99bc9477e\") " pod="openstack/nova-metadata-0" Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.211918 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/039efc51-838d-4aac-b0e8-7ba1c9b376ce-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"039efc51-838d-4aac-b0e8-7ba1c9b376ce\") " pod="openstack/nova-api-0" Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.211937 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/039efc51-838d-4aac-b0e8-7ba1c9b376ce-logs\") pod \"nova-api-0\" (UID: \"039efc51-838d-4aac-b0e8-7ba1c9b376ce\") " pod="openstack/nova-api-0" Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.211982 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4lpw\" (UniqueName: \"kubernetes.io/projected/039efc51-838d-4aac-b0e8-7ba1c9b376ce-kube-api-access-t4lpw\") pod \"nova-api-0\" (UID: \"039efc51-838d-4aac-b0e8-7ba1c9b376ce\") " pod="openstack/nova-api-0" Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.212012 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/924b2c00-a948-46d2-b103-25f99bc9477e-logs\") pod \"nova-metadata-0\" (UID: \"924b2c00-a948-46d2-b103-25f99bc9477e\") " pod="openstack/nova-metadata-0" Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.212088 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/039efc51-838d-4aac-b0e8-7ba1c9b376ce-config-data\") pod \"nova-api-0\" (UID: \"039efc51-838d-4aac-b0e8-7ba1c9b376ce\") " pod="openstack/nova-api-0" Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.213513 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/039efc51-838d-4aac-b0e8-7ba1c9b376ce-logs\") pod \"nova-api-0\" (UID: \"039efc51-838d-4aac-b0e8-7ba1c9b376ce\") " pod="openstack/nova-api-0" Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.215132 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/924b2c00-a948-46d2-b103-25f99bc9477e-logs\") pod \"nova-metadata-0\" (UID: \"924b2c00-a948-46d2-b103-25f99bc9477e\") " pod="openstack/nova-metadata-0" Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.223548 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/039efc51-838d-4aac-b0e8-7ba1c9b376ce-config-data\") pod \"nova-api-0\" (UID: \"039efc51-838d-4aac-b0e8-7ba1c9b376ce\") " pod="openstack/nova-api-0" Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.223830 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/924b2c00-a948-46d2-b103-25f99bc9477e-config-data\") pod \"nova-metadata-0\" (UID: \"924b2c00-a948-46d2-b103-25f99bc9477e\") " pod="openstack/nova-metadata-0" Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.225145 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/924b2c00-a948-46d2-b103-25f99bc9477e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"924b2c00-a948-46d2-b103-25f99bc9477e\") " pod="openstack/nova-metadata-0" Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.228576 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4lpw\" (UniqueName: \"kubernetes.io/projected/039efc51-838d-4aac-b0e8-7ba1c9b376ce-kube-api-access-t4lpw\") pod \"nova-api-0\" (UID: \"039efc51-838d-4aac-b0e8-7ba1c9b376ce\") " pod="openstack/nova-api-0" Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.229802 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4z2t\" (UniqueName: \"kubernetes.io/projected/924b2c00-a948-46d2-b103-25f99bc9477e-kube-api-access-b4z2t\") pod \"nova-metadata-0\" (UID: \"924b2c00-a948-46d2-b103-25f99bc9477e\") " pod="openstack/nova-metadata-0" Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.239023 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/039efc51-838d-4aac-b0e8-7ba1c9b376ce-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"039efc51-838d-4aac-b0e8-7ba1c9b376ce\") " pod="openstack/nova-api-0" Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.322022 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.345521 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.859488 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.908836 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.932096 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"924b2c00-a948-46d2-b103-25f99bc9477e","Type":"ContainerStarted","Data":"a70f11282a58121dc1bb2e63baa459ff35f4c79eb897248fb2f4d504419180ee"} Nov 24 14:46:47 crc kubenswrapper[4790]: I1124 14:46:47.934534 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zqtws" event={"ID":"7fff5214-ae3c-4948-aaea-d5806cfae4b4","Type":"ContainerStarted","Data":"01b373924743e503f96bf047b537e2940059cd9ccf0985481b0800cd0e3934b7"} Nov 24 14:46:48 crc kubenswrapper[4790]: I1124 14:46:48.328856 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3a43449-bf3d-4104-a89e-ef4d45191a38" path="/var/lib/kubelet/pods/e3a43449-bf3d-4104-a89e-ef4d45191a38/volumes" Nov 24 14:46:48 crc kubenswrapper[4790]: I1124 14:46:48.331080 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efc434d3-6cab-448c-a4aa-c441cefaafe2" path="/var/lib/kubelet/pods/efc434d3-6cab-448c-a4aa-c441cefaafe2/volumes" Nov 24 14:46:48 crc kubenswrapper[4790]: E1124 14:46:48.359453 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="82ca39eedfff2d3a9609182b4df1638283f236abbef9e67ab174b401f08d4b21" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 14:46:48 crc kubenswrapper[4790]: E1124 14:46:48.362623 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="82ca39eedfff2d3a9609182b4df1638283f236abbef9e67ab174b401f08d4b21" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 14:46:48 crc kubenswrapper[4790]: E1124 14:46:48.366297 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="82ca39eedfff2d3a9609182b4df1638283f236abbef9e67ab174b401f08d4b21" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 14:46:48 crc kubenswrapper[4790]: E1124 14:46:48.366332 4790 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="a167644f-5164-4943-97a3-6031e71f2f28" containerName="nova-scheduler-scheduler" Nov 24 14:46:48 crc kubenswrapper[4790]: I1124 14:46:48.957224 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"039efc51-838d-4aac-b0e8-7ba1c9b376ce","Type":"ContainerStarted","Data":"58706d8229eea72d6668f75cb9016630db40c6067ad81aa31661875ab841589a"} Nov 24 14:46:48 crc kubenswrapper[4790]: I1124 14:46:48.958660 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"039efc51-838d-4aac-b0e8-7ba1c9b376ce","Type":"ContainerStarted","Data":"4a29ff14976dcb69b6759470411ff7beb1d298155f005c3981206b31876014bd"} Nov 24 14:46:48 crc kubenswrapper[4790]: I1124 14:46:48.958733 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"039efc51-838d-4aac-b0e8-7ba1c9b376ce","Type":"ContainerStarted","Data":"41bb5a692a4693b864f8feb5f54d078f84409c52facf3089106c328c4c61a7ed"} Nov 24 14:46:48 crc kubenswrapper[4790]: I1124 14:46:48.961810 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"924b2c00-a948-46d2-b103-25f99bc9477e","Type":"ContainerStarted","Data":"4b805b27d5fe3f30506bbea66532569a1055c70574cd5abc2c7a8a50b9e863b1"} Nov 24 14:46:48 crc kubenswrapper[4790]: I1124 14:46:48.961874 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"924b2c00-a948-46d2-b103-25f99bc9477e","Type":"ContainerStarted","Data":"5864a737edda91cad9e80b03ac060cd6a2fe7b521eb61e4e209e5c0ea954c4c1"} Nov 24 14:46:48 crc kubenswrapper[4790]: I1124 14:46:48.965112 4790 generic.go:334] "Generic (PLEG): container finished" podID="7fff5214-ae3c-4948-aaea-d5806cfae4b4" containerID="01b373924743e503f96bf047b537e2940059cd9ccf0985481b0800cd0e3934b7" exitCode=0 Nov 24 14:46:48 crc kubenswrapper[4790]: I1124 14:46:48.965222 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zqtws" event={"ID":"7fff5214-ae3c-4948-aaea-d5806cfae4b4","Type":"ContainerDied","Data":"01b373924743e503f96bf047b537e2940059cd9ccf0985481b0800cd0e3934b7"} Nov 24 14:46:48 crc kubenswrapper[4790]: I1124 14:46:48.998831 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.99878892 podStartE2EDuration="2.99878892s" podCreationTimestamp="2025-11-24 14:46:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:46:48.991040467 +0000 UTC m=+5657.370934229" watchObservedRunningTime="2025-11-24 14:46:48.99878892 +0000 UTC m=+5657.378682632" Nov 24 14:46:49 crc kubenswrapper[4790]: I1124 14:46:49.041876 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.041853877 podStartE2EDuration="3.041853877s" podCreationTimestamp="2025-11-24 14:46:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:46:49.03953107 +0000 UTC m=+5657.419424752" watchObservedRunningTime="2025-11-24 14:46:49.041853877 +0000 UTC m=+5657.421747549" Nov 24 14:46:49 crc kubenswrapper[4790]: E1124 14:46:49.559809 4790 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.69:41846->38.129.56.69:36625: write tcp 38.129.56.69:41846->38.129.56.69:36625: write: broken pipe Nov 24 14:46:49 crc kubenswrapper[4790]: I1124 14:46:49.620516 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:46:49 crc kubenswrapper[4790]: I1124 14:46:49.981667 4790 generic.go:334] "Generic (PLEG): container finished" podID="bfa26363-ed7c-4bd7-a32b-420e94c62a3d" containerID="659aae12e81e3a6ff594ff0fa8c227b2de1d0e0869ca6de31bef7cebd3fce403" exitCode=0 Nov 24 14:46:49 crc kubenswrapper[4790]: I1124 14:46:49.981798 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"bfa26363-ed7c-4bd7-a32b-420e94c62a3d","Type":"ContainerDied","Data":"659aae12e81e3a6ff594ff0fa8c227b2de1d0e0869ca6de31bef7cebd3fce403"} Nov 24 14:46:49 crc kubenswrapper[4790]: I1124 14:46:49.985055 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zqtws" event={"ID":"7fff5214-ae3c-4948-aaea-d5806cfae4b4","Type":"ContainerStarted","Data":"c2da76e6da777ecc49dd8088faa1cb2e185cbbdb1e6355e71657ae28e1d9d9e0"} Nov 24 14:46:50 crc kubenswrapper[4790]: E1124 14:46:50.021212 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 659aae12e81e3a6ff594ff0fa8c227b2de1d0e0869ca6de31bef7cebd3fce403 is running failed: container process not found" containerID="659aae12e81e3a6ff594ff0fa8c227b2de1d0e0869ca6de31bef7cebd3fce403" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 24 14:46:50 crc kubenswrapper[4790]: E1124 14:46:50.021654 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 659aae12e81e3a6ff594ff0fa8c227b2de1d0e0869ca6de31bef7cebd3fce403 is running failed: container process not found" containerID="659aae12e81e3a6ff594ff0fa8c227b2de1d0e0869ca6de31bef7cebd3fce403" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 24 14:46:50 crc kubenswrapper[4790]: E1124 14:46:50.022113 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 659aae12e81e3a6ff594ff0fa8c227b2de1d0e0869ca6de31bef7cebd3fce403 is running failed: container process not found" containerID="659aae12e81e3a6ff594ff0fa8c227b2de1d0e0869ca6de31bef7cebd3fce403" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 24 14:46:50 crc kubenswrapper[4790]: E1124 14:46:50.022201 4790 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 659aae12e81e3a6ff594ff0fa8c227b2de1d0e0869ca6de31bef7cebd3fce403 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="bfa26363-ed7c-4bd7-a32b-420e94c62a3d" containerName="nova-cell1-conductor-conductor" Nov 24 14:46:50 crc kubenswrapper[4790]: I1124 14:46:50.026566 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zqtws" podStartSLOduration=2.246111382 podStartE2EDuration="5.026548625s" podCreationTimestamp="2025-11-24 14:46:45 +0000 UTC" firstStartedPulling="2025-11-24 14:46:46.894729347 +0000 UTC m=+5655.274623019" lastFinishedPulling="2025-11-24 14:46:49.67516658 +0000 UTC m=+5658.055060262" observedRunningTime="2025-11-24 14:46:50.021282303 +0000 UTC m=+5658.401176015" watchObservedRunningTime="2025-11-24 14:46:50.026548625 +0000 UTC m=+5658.406442297" Nov 24 14:46:50 crc kubenswrapper[4790]: I1124 14:46:50.370670 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 14:46:50 crc kubenswrapper[4790]: I1124 14:46:50.484207 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfa26363-ed7c-4bd7-a32b-420e94c62a3d-combined-ca-bundle\") pod \"bfa26363-ed7c-4bd7-a32b-420e94c62a3d\" (UID: \"bfa26363-ed7c-4bd7-a32b-420e94c62a3d\") " Nov 24 14:46:50 crc kubenswrapper[4790]: I1124 14:46:50.484444 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfa26363-ed7c-4bd7-a32b-420e94c62a3d-config-data\") pod \"bfa26363-ed7c-4bd7-a32b-420e94c62a3d\" (UID: \"bfa26363-ed7c-4bd7-a32b-420e94c62a3d\") " Nov 24 14:46:50 crc kubenswrapper[4790]: I1124 14:46:50.484472 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcjpc\" (UniqueName: \"kubernetes.io/projected/bfa26363-ed7c-4bd7-a32b-420e94c62a3d-kube-api-access-wcjpc\") pod \"bfa26363-ed7c-4bd7-a32b-420e94c62a3d\" (UID: \"bfa26363-ed7c-4bd7-a32b-420e94c62a3d\") " Nov 24 14:46:50 crc kubenswrapper[4790]: I1124 14:46:50.490073 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfa26363-ed7c-4bd7-a32b-420e94c62a3d-kube-api-access-wcjpc" (OuterVolumeSpecName: "kube-api-access-wcjpc") pod "bfa26363-ed7c-4bd7-a32b-420e94c62a3d" (UID: "bfa26363-ed7c-4bd7-a32b-420e94c62a3d"). InnerVolumeSpecName "kube-api-access-wcjpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:46:50 crc kubenswrapper[4790]: I1124 14:46:50.517088 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfa26363-ed7c-4bd7-a32b-420e94c62a3d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bfa26363-ed7c-4bd7-a32b-420e94c62a3d" (UID: "bfa26363-ed7c-4bd7-a32b-420e94c62a3d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:46:50 crc kubenswrapper[4790]: I1124 14:46:50.537526 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfa26363-ed7c-4bd7-a32b-420e94c62a3d-config-data" (OuterVolumeSpecName: "config-data") pod "bfa26363-ed7c-4bd7-a32b-420e94c62a3d" (UID: "bfa26363-ed7c-4bd7-a32b-420e94c62a3d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:46:50 crc kubenswrapper[4790]: I1124 14:46:50.589157 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfa26363-ed7c-4bd7-a32b-420e94c62a3d-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:50 crc kubenswrapper[4790]: I1124 14:46:50.589205 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcjpc\" (UniqueName: \"kubernetes.io/projected/bfa26363-ed7c-4bd7-a32b-420e94c62a3d-kube-api-access-wcjpc\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:50 crc kubenswrapper[4790]: I1124 14:46:50.589220 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfa26363-ed7c-4bd7-a32b-420e94c62a3d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:50 crc kubenswrapper[4790]: I1124 14:46:50.995751 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 14:46:50 crc kubenswrapper[4790]: I1124 14:46:50.995950 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"bfa26363-ed7c-4bd7-a32b-420e94c62a3d","Type":"ContainerDied","Data":"9792d21095fe0bc230576595e575197ec3143c0aec72f52aa274affe775cee96"} Nov 24 14:46:50 crc kubenswrapper[4790]: I1124 14:46:50.996836 4790 scope.go:117] "RemoveContainer" containerID="659aae12e81e3a6ff594ff0fa8c227b2de1d0e0869ca6de31bef7cebd3fce403" Nov 24 14:46:51 crc kubenswrapper[4790]: I1124 14:46:51.043065 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 14:46:51 crc kubenswrapper[4790]: I1124 14:46:51.059751 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 14:46:51 crc kubenswrapper[4790]: I1124 14:46:51.079417 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 14:46:51 crc kubenswrapper[4790]: E1124 14:46:51.079969 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfa26363-ed7c-4bd7-a32b-420e94c62a3d" containerName="nova-cell1-conductor-conductor" Nov 24 14:46:51 crc kubenswrapper[4790]: I1124 14:46:51.079986 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfa26363-ed7c-4bd7-a32b-420e94c62a3d" containerName="nova-cell1-conductor-conductor" Nov 24 14:46:51 crc kubenswrapper[4790]: I1124 14:46:51.080255 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfa26363-ed7c-4bd7-a32b-420e94c62a3d" containerName="nova-cell1-conductor-conductor" Nov 24 14:46:51 crc kubenswrapper[4790]: I1124 14:46:51.081243 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 14:46:51 crc kubenswrapper[4790]: I1124 14:46:51.085570 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 14:46:51 crc kubenswrapper[4790]: I1124 14:46:51.091481 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 14:46:51 crc kubenswrapper[4790]: I1124 14:46:51.201767 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cea27eb-31d4-4e3b-a141-d2593bc21427-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"5cea27eb-31d4-4e3b-a141-d2593bc21427\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:46:51 crc kubenswrapper[4790]: I1124 14:46:51.201827 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpt9m\" (UniqueName: \"kubernetes.io/projected/5cea27eb-31d4-4e3b-a141-d2593bc21427-kube-api-access-zpt9m\") pod \"nova-cell1-conductor-0\" (UID: \"5cea27eb-31d4-4e3b-a141-d2593bc21427\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:46:51 crc kubenswrapper[4790]: I1124 14:46:51.202239 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cea27eb-31d4-4e3b-a141-d2593bc21427-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"5cea27eb-31d4-4e3b-a141-d2593bc21427\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:46:51 crc kubenswrapper[4790]: I1124 14:46:51.304290 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cea27eb-31d4-4e3b-a141-d2593bc21427-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"5cea27eb-31d4-4e3b-a141-d2593bc21427\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:46:51 crc kubenswrapper[4790]: I1124 14:46:51.304446 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cea27eb-31d4-4e3b-a141-d2593bc21427-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"5cea27eb-31d4-4e3b-a141-d2593bc21427\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:46:51 crc kubenswrapper[4790]: I1124 14:46:51.304492 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpt9m\" (UniqueName: \"kubernetes.io/projected/5cea27eb-31d4-4e3b-a141-d2593bc21427-kube-api-access-zpt9m\") pod \"nova-cell1-conductor-0\" (UID: \"5cea27eb-31d4-4e3b-a141-d2593bc21427\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:46:51 crc kubenswrapper[4790]: I1124 14:46:51.309943 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cea27eb-31d4-4e3b-a141-d2593bc21427-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"5cea27eb-31d4-4e3b-a141-d2593bc21427\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:46:51 crc kubenswrapper[4790]: I1124 14:46:51.314499 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cea27eb-31d4-4e3b-a141-d2593bc21427-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"5cea27eb-31d4-4e3b-a141-d2593bc21427\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:46:51 crc kubenswrapper[4790]: I1124 14:46:51.332481 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpt9m\" (UniqueName: \"kubernetes.io/projected/5cea27eb-31d4-4e3b-a141-d2593bc21427-kube-api-access-zpt9m\") pod \"nova-cell1-conductor-0\" (UID: \"5cea27eb-31d4-4e3b-a141-d2593bc21427\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:46:51 crc kubenswrapper[4790]: I1124 14:46:51.400235 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 14:46:51 crc kubenswrapper[4790]: I1124 14:46:51.896080 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 14:46:52 crc kubenswrapper[4790]: I1124 14:46:52.007689 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"5cea27eb-31d4-4e3b-a141-d2593bc21427","Type":"ContainerStarted","Data":"021c10b5114e774f39d9ea1db02a0bf29efbc5c35193468f21e6e5e31f7d9e50"} Nov 24 14:46:52 crc kubenswrapper[4790]: I1124 14:46:52.382011 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfa26363-ed7c-4bd7-a32b-420e94c62a3d" path="/var/lib/kubelet/pods/bfa26363-ed7c-4bd7-a32b-420e94c62a3d/volumes" Nov 24 14:46:52 crc kubenswrapper[4790]: I1124 14:46:52.385458 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 14:46:52 crc kubenswrapper[4790]: I1124 14:46:52.385485 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 14:46:52 crc kubenswrapper[4790]: I1124 14:46:52.436721 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:46:52 crc kubenswrapper[4790]: I1124 14:46:52.553199 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a167644f-5164-4943-97a3-6031e71f2f28-config-data\") pod \"a167644f-5164-4943-97a3-6031e71f2f28\" (UID: \"a167644f-5164-4943-97a3-6031e71f2f28\") " Nov 24 14:46:52 crc kubenswrapper[4790]: I1124 14:46:52.553356 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a167644f-5164-4943-97a3-6031e71f2f28-combined-ca-bundle\") pod \"a167644f-5164-4943-97a3-6031e71f2f28\" (UID: \"a167644f-5164-4943-97a3-6031e71f2f28\") " Nov 24 14:46:52 crc kubenswrapper[4790]: I1124 14:46:52.553450 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6rdr\" (UniqueName: \"kubernetes.io/projected/a167644f-5164-4943-97a3-6031e71f2f28-kube-api-access-d6rdr\") pod \"a167644f-5164-4943-97a3-6031e71f2f28\" (UID: \"a167644f-5164-4943-97a3-6031e71f2f28\") " Nov 24 14:46:52 crc kubenswrapper[4790]: I1124 14:46:52.562317 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a167644f-5164-4943-97a3-6031e71f2f28-kube-api-access-d6rdr" (OuterVolumeSpecName: "kube-api-access-d6rdr") pod "a167644f-5164-4943-97a3-6031e71f2f28" (UID: "a167644f-5164-4943-97a3-6031e71f2f28"). InnerVolumeSpecName "kube-api-access-d6rdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:46:52 crc kubenswrapper[4790]: I1124 14:46:52.585005 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a167644f-5164-4943-97a3-6031e71f2f28-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a167644f-5164-4943-97a3-6031e71f2f28" (UID: "a167644f-5164-4943-97a3-6031e71f2f28"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:46:52 crc kubenswrapper[4790]: I1124 14:46:52.590236 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a167644f-5164-4943-97a3-6031e71f2f28-config-data" (OuterVolumeSpecName: "config-data") pod "a167644f-5164-4943-97a3-6031e71f2f28" (UID: "a167644f-5164-4943-97a3-6031e71f2f28"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:46:52 crc kubenswrapper[4790]: I1124 14:46:52.656444 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a167644f-5164-4943-97a3-6031e71f2f28-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:52 crc kubenswrapper[4790]: I1124 14:46:52.656490 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a167644f-5164-4943-97a3-6031e71f2f28-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:52 crc kubenswrapper[4790]: I1124 14:46:52.656507 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6rdr\" (UniqueName: \"kubernetes.io/projected/a167644f-5164-4943-97a3-6031e71f2f28-kube-api-access-d6rdr\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:53 crc kubenswrapper[4790]: I1124 14:46:53.021523 4790 generic.go:334] "Generic (PLEG): container finished" podID="a167644f-5164-4943-97a3-6031e71f2f28" containerID="82ca39eedfff2d3a9609182b4df1638283f236abbef9e67ab174b401f08d4b21" exitCode=0 Nov 24 14:46:53 crc kubenswrapper[4790]: I1124 14:46:53.021612 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a167644f-5164-4943-97a3-6031e71f2f28","Type":"ContainerDied","Data":"82ca39eedfff2d3a9609182b4df1638283f236abbef9e67ab174b401f08d4b21"} Nov 24 14:46:53 crc kubenswrapper[4790]: I1124 14:46:53.021650 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a167644f-5164-4943-97a3-6031e71f2f28","Type":"ContainerDied","Data":"9db79f6dff61f24e734eee32f8574d81ccedbb3554915450e0b846e0a88d7cd2"} Nov 24 14:46:53 crc kubenswrapper[4790]: I1124 14:46:53.021677 4790 scope.go:117] "RemoveContainer" containerID="82ca39eedfff2d3a9609182b4df1638283f236abbef9e67ab174b401f08d4b21" Nov 24 14:46:53 crc kubenswrapper[4790]: I1124 14:46:53.021820 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:46:53 crc kubenswrapper[4790]: I1124 14:46:53.029942 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"5cea27eb-31d4-4e3b-a141-d2593bc21427","Type":"ContainerStarted","Data":"47c107e131a16273b3940ec27fea34ff25637218453a77b80c996ad534bf1dca"} Nov 24 14:46:53 crc kubenswrapper[4790]: I1124 14:46:53.031327 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 24 14:46:53 crc kubenswrapper[4790]: I1124 14:46:53.059194 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.059173073 podStartE2EDuration="2.059173073s" podCreationTimestamp="2025-11-24 14:46:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:46:53.057931807 +0000 UTC m=+5661.437825489" watchObservedRunningTime="2025-11-24 14:46:53.059173073 +0000 UTC m=+5661.439066745" Nov 24 14:46:53 crc kubenswrapper[4790]: I1124 14:46:53.061456 4790 scope.go:117] "RemoveContainer" containerID="82ca39eedfff2d3a9609182b4df1638283f236abbef9e67ab174b401f08d4b21" Nov 24 14:46:53 crc kubenswrapper[4790]: E1124 14:46:53.061955 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82ca39eedfff2d3a9609182b4df1638283f236abbef9e67ab174b401f08d4b21\": container with ID starting with 82ca39eedfff2d3a9609182b4df1638283f236abbef9e67ab174b401f08d4b21 not found: ID does not exist" containerID="82ca39eedfff2d3a9609182b4df1638283f236abbef9e67ab174b401f08d4b21" Nov 24 14:46:53 crc kubenswrapper[4790]: I1124 14:46:53.061997 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82ca39eedfff2d3a9609182b4df1638283f236abbef9e67ab174b401f08d4b21"} err="failed to get container status \"82ca39eedfff2d3a9609182b4df1638283f236abbef9e67ab174b401f08d4b21\": rpc error: code = NotFound desc = could not find container \"82ca39eedfff2d3a9609182b4df1638283f236abbef9e67ab174b401f08d4b21\": container with ID starting with 82ca39eedfff2d3a9609182b4df1638283f236abbef9e67ab174b401f08d4b21 not found: ID does not exist" Nov 24 14:46:53 crc kubenswrapper[4790]: I1124 14:46:53.090522 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:46:53 crc kubenswrapper[4790]: I1124 14:46:53.105340 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:46:53 crc kubenswrapper[4790]: I1124 14:46:53.120999 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:46:53 crc kubenswrapper[4790]: E1124 14:46:53.121611 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a167644f-5164-4943-97a3-6031e71f2f28" containerName="nova-scheduler-scheduler" Nov 24 14:46:53 crc kubenswrapper[4790]: I1124 14:46:53.121656 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="a167644f-5164-4943-97a3-6031e71f2f28" containerName="nova-scheduler-scheduler" Nov 24 14:46:53 crc kubenswrapper[4790]: I1124 14:46:53.121966 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="a167644f-5164-4943-97a3-6031e71f2f28" containerName="nova-scheduler-scheduler" Nov 24 14:46:53 crc kubenswrapper[4790]: I1124 14:46:53.122755 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:46:53 crc kubenswrapper[4790]: I1124 14:46:53.125287 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 14:46:53 crc kubenswrapper[4790]: I1124 14:46:53.135078 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:46:53 crc kubenswrapper[4790]: I1124 14:46:53.175243 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a3a068f-92a8-43cf-83dc-8775c1d9ee84-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2a3a068f-92a8-43cf-83dc-8775c1d9ee84\") " pod="openstack/nova-scheduler-0" Nov 24 14:46:53 crc kubenswrapper[4790]: I1124 14:46:53.175729 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a3a068f-92a8-43cf-83dc-8775c1d9ee84-config-data\") pod \"nova-scheduler-0\" (UID: \"2a3a068f-92a8-43cf-83dc-8775c1d9ee84\") " pod="openstack/nova-scheduler-0" Nov 24 14:46:53 crc kubenswrapper[4790]: I1124 14:46:53.175850 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nk4ts\" (UniqueName: \"kubernetes.io/projected/2a3a068f-92a8-43cf-83dc-8775c1d9ee84-kube-api-access-nk4ts\") pod \"nova-scheduler-0\" (UID: \"2a3a068f-92a8-43cf-83dc-8775c1d9ee84\") " pod="openstack/nova-scheduler-0" Nov 24 14:46:53 crc kubenswrapper[4790]: I1124 14:46:53.277231 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a3a068f-92a8-43cf-83dc-8775c1d9ee84-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2a3a068f-92a8-43cf-83dc-8775c1d9ee84\") " pod="openstack/nova-scheduler-0" Nov 24 14:46:53 crc kubenswrapper[4790]: I1124 14:46:53.277283 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a3a068f-92a8-43cf-83dc-8775c1d9ee84-config-data\") pod \"nova-scheduler-0\" (UID: \"2a3a068f-92a8-43cf-83dc-8775c1d9ee84\") " pod="openstack/nova-scheduler-0" Nov 24 14:46:53 crc kubenswrapper[4790]: I1124 14:46:53.277351 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nk4ts\" (UniqueName: \"kubernetes.io/projected/2a3a068f-92a8-43cf-83dc-8775c1d9ee84-kube-api-access-nk4ts\") pod \"nova-scheduler-0\" (UID: \"2a3a068f-92a8-43cf-83dc-8775c1d9ee84\") " pod="openstack/nova-scheduler-0" Nov 24 14:46:53 crc kubenswrapper[4790]: I1124 14:46:53.282103 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a3a068f-92a8-43cf-83dc-8775c1d9ee84-config-data\") pod \"nova-scheduler-0\" (UID: \"2a3a068f-92a8-43cf-83dc-8775c1d9ee84\") " pod="openstack/nova-scheduler-0" Nov 24 14:46:53 crc kubenswrapper[4790]: I1124 14:46:53.282763 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a3a068f-92a8-43cf-83dc-8775c1d9ee84-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2a3a068f-92a8-43cf-83dc-8775c1d9ee84\") " pod="openstack/nova-scheduler-0" Nov 24 14:46:53 crc kubenswrapper[4790]: I1124 14:46:53.294624 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nk4ts\" (UniqueName: \"kubernetes.io/projected/2a3a068f-92a8-43cf-83dc-8775c1d9ee84-kube-api-access-nk4ts\") pod \"nova-scheduler-0\" (UID: \"2a3a068f-92a8-43cf-83dc-8775c1d9ee84\") " pod="openstack/nova-scheduler-0" Nov 24 14:46:53 crc kubenswrapper[4790]: I1124 14:46:53.454679 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:46:53 crc kubenswrapper[4790]: I1124 14:46:53.927272 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:46:54 crc kubenswrapper[4790]: I1124 14:46:54.043049 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2a3a068f-92a8-43cf-83dc-8775c1d9ee84","Type":"ContainerStarted","Data":"9adeb24326bccef100166923b4275ac81338d33224168a41977c2e353450b62c"} Nov 24 14:46:54 crc kubenswrapper[4790]: I1124 14:46:54.327421 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a167644f-5164-4943-97a3-6031e71f2f28" path="/var/lib/kubelet/pods/a167644f-5164-4943-97a3-6031e71f2f28/volumes" Nov 24 14:46:54 crc kubenswrapper[4790]: I1124 14:46:54.620252 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:46:54 crc kubenswrapper[4790]: I1124 14:46:54.643701 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:46:54 crc kubenswrapper[4790]: I1124 14:46:54.649729 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 24 14:46:55 crc kubenswrapper[4790]: I1124 14:46:55.060960 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2a3a068f-92a8-43cf-83dc-8775c1d9ee84","Type":"ContainerStarted","Data":"56b9940a224d2d309a332cfcbce228cc2bf78759b8ba90c6fba493c1a3781739"} Nov 24 14:46:55 crc kubenswrapper[4790]: I1124 14:46:55.084464 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:46:55 crc kubenswrapper[4790]: I1124 14:46:55.096052 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.096029396 podStartE2EDuration="2.096029396s" podCreationTimestamp="2025-11-24 14:46:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:46:55.080555421 +0000 UTC m=+5663.460449133" watchObservedRunningTime="2025-11-24 14:46:55.096029396 +0000 UTC m=+5663.475923078" Nov 24 14:46:55 crc kubenswrapper[4790]: I1124 14:46:55.493697 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zqtws" Nov 24 14:46:55 crc kubenswrapper[4790]: I1124 14:46:55.494364 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zqtws" Nov 24 14:46:55 crc kubenswrapper[4790]: I1124 14:46:55.582094 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zqtws" Nov 24 14:46:56 crc kubenswrapper[4790]: I1124 14:46:56.147057 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zqtws" Nov 24 14:46:56 crc kubenswrapper[4790]: I1124 14:46:56.195942 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zqtws"] Nov 24 14:46:57 crc kubenswrapper[4790]: I1124 14:46:57.323556 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 14:46:57 crc kubenswrapper[4790]: I1124 14:46:57.324169 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 14:46:57 crc kubenswrapper[4790]: I1124 14:46:57.346907 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 14:46:57 crc kubenswrapper[4790]: I1124 14:46:57.347020 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 14:46:58 crc kubenswrapper[4790]: I1124 14:46:58.091129 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zqtws" podUID="7fff5214-ae3c-4948-aaea-d5806cfae4b4" containerName="registry-server" containerID="cri-o://c2da76e6da777ecc49dd8088faa1cb2e185cbbdb1e6355e71657ae28e1d9d9e0" gracePeriod=2 Nov 24 14:46:58 crc kubenswrapper[4790]: I1124 14:46:58.364305 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="924b2c00-a948-46d2-b103-25f99bc9477e" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.85:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 14:46:58 crc kubenswrapper[4790]: I1124 14:46:58.406290 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="924b2c00-a948-46d2-b103-25f99bc9477e" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.85:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 14:46:58 crc kubenswrapper[4790]: I1124 14:46:58.456097 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 14:46:58 crc kubenswrapper[4790]: I1124 14:46:58.489157 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="039efc51-838d-4aac-b0e8-7ba1c9b376ce" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.86:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 14:46:58 crc kubenswrapper[4790]: I1124 14:46:58.489441 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="039efc51-838d-4aac-b0e8-7ba1c9b376ce" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.86:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 14:46:58 crc kubenswrapper[4790]: I1124 14:46:58.608846 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zqtws" Nov 24 14:46:58 crc kubenswrapper[4790]: I1124 14:46:58.785253 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fff5214-ae3c-4948-aaea-d5806cfae4b4-utilities\") pod \"7fff5214-ae3c-4948-aaea-d5806cfae4b4\" (UID: \"7fff5214-ae3c-4948-aaea-d5806cfae4b4\") " Nov 24 14:46:58 crc kubenswrapper[4790]: I1124 14:46:58.786407 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fff5214-ae3c-4948-aaea-d5806cfae4b4-utilities" (OuterVolumeSpecName: "utilities") pod "7fff5214-ae3c-4948-aaea-d5806cfae4b4" (UID: "7fff5214-ae3c-4948-aaea-d5806cfae4b4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:46:58 crc kubenswrapper[4790]: I1124 14:46:58.786510 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fff5214-ae3c-4948-aaea-d5806cfae4b4-catalog-content\") pod \"7fff5214-ae3c-4948-aaea-d5806cfae4b4\" (UID: \"7fff5214-ae3c-4948-aaea-d5806cfae4b4\") " Nov 24 14:46:58 crc kubenswrapper[4790]: I1124 14:46:58.800738 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gvhf\" (UniqueName: \"kubernetes.io/projected/7fff5214-ae3c-4948-aaea-d5806cfae4b4-kube-api-access-2gvhf\") pod \"7fff5214-ae3c-4948-aaea-d5806cfae4b4\" (UID: \"7fff5214-ae3c-4948-aaea-d5806cfae4b4\") " Nov 24 14:46:58 crc kubenswrapper[4790]: I1124 14:46:58.801825 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fff5214-ae3c-4948-aaea-d5806cfae4b4-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:58 crc kubenswrapper[4790]: I1124 14:46:58.822463 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fff5214-ae3c-4948-aaea-d5806cfae4b4-kube-api-access-2gvhf" (OuterVolumeSpecName: "kube-api-access-2gvhf") pod "7fff5214-ae3c-4948-aaea-d5806cfae4b4" (UID: "7fff5214-ae3c-4948-aaea-d5806cfae4b4"). InnerVolumeSpecName "kube-api-access-2gvhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:46:58 crc kubenswrapper[4790]: I1124 14:46:58.897972 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fff5214-ae3c-4948-aaea-d5806cfae4b4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7fff5214-ae3c-4948-aaea-d5806cfae4b4" (UID: "7fff5214-ae3c-4948-aaea-d5806cfae4b4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:46:58 crc kubenswrapper[4790]: I1124 14:46:58.902956 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gvhf\" (UniqueName: \"kubernetes.io/projected/7fff5214-ae3c-4948-aaea-d5806cfae4b4-kube-api-access-2gvhf\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:58 crc kubenswrapper[4790]: I1124 14:46:58.902980 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fff5214-ae3c-4948-aaea-d5806cfae4b4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:46:59 crc kubenswrapper[4790]: I1124 14:46:59.103577 4790 generic.go:334] "Generic (PLEG): container finished" podID="7fff5214-ae3c-4948-aaea-d5806cfae4b4" containerID="c2da76e6da777ecc49dd8088faa1cb2e185cbbdb1e6355e71657ae28e1d9d9e0" exitCode=0 Nov 24 14:46:59 crc kubenswrapper[4790]: I1124 14:46:59.103623 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zqtws" event={"ID":"7fff5214-ae3c-4948-aaea-d5806cfae4b4","Type":"ContainerDied","Data":"c2da76e6da777ecc49dd8088faa1cb2e185cbbdb1e6355e71657ae28e1d9d9e0"} Nov 24 14:46:59 crc kubenswrapper[4790]: I1124 14:46:59.103654 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zqtws" Nov 24 14:46:59 crc kubenswrapper[4790]: I1124 14:46:59.103674 4790 scope.go:117] "RemoveContainer" containerID="c2da76e6da777ecc49dd8088faa1cb2e185cbbdb1e6355e71657ae28e1d9d9e0" Nov 24 14:46:59 crc kubenswrapper[4790]: I1124 14:46:59.103661 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zqtws" event={"ID":"7fff5214-ae3c-4948-aaea-d5806cfae4b4","Type":"ContainerDied","Data":"464a68671754697c1bc78dd6b6b481e776668904c8c45fc9f6d3482dcd8c6d5d"} Nov 24 14:46:59 crc kubenswrapper[4790]: I1124 14:46:59.139386 4790 scope.go:117] "RemoveContainer" containerID="01b373924743e503f96bf047b537e2940059cd9ccf0985481b0800cd0e3934b7" Nov 24 14:46:59 crc kubenswrapper[4790]: I1124 14:46:59.139567 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zqtws"] Nov 24 14:46:59 crc kubenswrapper[4790]: I1124 14:46:59.150406 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zqtws"] Nov 24 14:46:59 crc kubenswrapper[4790]: I1124 14:46:59.162323 4790 scope.go:117] "RemoveContainer" containerID="11a9761a2b421c4f2f0bb9699504f9e9d6d52d8d1ccecb1661e32368dbfd4689" Nov 24 14:46:59 crc kubenswrapper[4790]: I1124 14:46:59.208654 4790 scope.go:117] "RemoveContainer" containerID="c2da76e6da777ecc49dd8088faa1cb2e185cbbdb1e6355e71657ae28e1d9d9e0" Nov 24 14:46:59 crc kubenswrapper[4790]: E1124 14:46:59.209313 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2da76e6da777ecc49dd8088faa1cb2e185cbbdb1e6355e71657ae28e1d9d9e0\": container with ID starting with c2da76e6da777ecc49dd8088faa1cb2e185cbbdb1e6355e71657ae28e1d9d9e0 not found: ID does not exist" containerID="c2da76e6da777ecc49dd8088faa1cb2e185cbbdb1e6355e71657ae28e1d9d9e0" Nov 24 14:46:59 crc kubenswrapper[4790]: I1124 14:46:59.209360 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2da76e6da777ecc49dd8088faa1cb2e185cbbdb1e6355e71657ae28e1d9d9e0"} err="failed to get container status \"c2da76e6da777ecc49dd8088faa1cb2e185cbbdb1e6355e71657ae28e1d9d9e0\": rpc error: code = NotFound desc = could not find container \"c2da76e6da777ecc49dd8088faa1cb2e185cbbdb1e6355e71657ae28e1d9d9e0\": container with ID starting with c2da76e6da777ecc49dd8088faa1cb2e185cbbdb1e6355e71657ae28e1d9d9e0 not found: ID does not exist" Nov 24 14:46:59 crc kubenswrapper[4790]: I1124 14:46:59.209394 4790 scope.go:117] "RemoveContainer" containerID="01b373924743e503f96bf047b537e2940059cd9ccf0985481b0800cd0e3934b7" Nov 24 14:46:59 crc kubenswrapper[4790]: E1124 14:46:59.210037 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01b373924743e503f96bf047b537e2940059cd9ccf0985481b0800cd0e3934b7\": container with ID starting with 01b373924743e503f96bf047b537e2940059cd9ccf0985481b0800cd0e3934b7 not found: ID does not exist" containerID="01b373924743e503f96bf047b537e2940059cd9ccf0985481b0800cd0e3934b7" Nov 24 14:46:59 crc kubenswrapper[4790]: I1124 14:46:59.210077 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01b373924743e503f96bf047b537e2940059cd9ccf0985481b0800cd0e3934b7"} err="failed to get container status \"01b373924743e503f96bf047b537e2940059cd9ccf0985481b0800cd0e3934b7\": rpc error: code = NotFound desc = could not find container \"01b373924743e503f96bf047b537e2940059cd9ccf0985481b0800cd0e3934b7\": container with ID starting with 01b373924743e503f96bf047b537e2940059cd9ccf0985481b0800cd0e3934b7 not found: ID does not exist" Nov 24 14:46:59 crc kubenswrapper[4790]: I1124 14:46:59.210106 4790 scope.go:117] "RemoveContainer" containerID="11a9761a2b421c4f2f0bb9699504f9e9d6d52d8d1ccecb1661e32368dbfd4689" Nov 24 14:46:59 crc kubenswrapper[4790]: E1124 14:46:59.210405 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11a9761a2b421c4f2f0bb9699504f9e9d6d52d8d1ccecb1661e32368dbfd4689\": container with ID starting with 11a9761a2b421c4f2f0bb9699504f9e9d6d52d8d1ccecb1661e32368dbfd4689 not found: ID does not exist" containerID="11a9761a2b421c4f2f0bb9699504f9e9d6d52d8d1ccecb1661e32368dbfd4689" Nov 24 14:46:59 crc kubenswrapper[4790]: I1124 14:46:59.210430 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11a9761a2b421c4f2f0bb9699504f9e9d6d52d8d1ccecb1661e32368dbfd4689"} err="failed to get container status \"11a9761a2b421c4f2f0bb9699504f9e9d6d52d8d1ccecb1661e32368dbfd4689\": rpc error: code = NotFound desc = could not find container \"11a9761a2b421c4f2f0bb9699504f9e9d6d52d8d1ccecb1661e32368dbfd4689\": container with ID starting with 11a9761a2b421c4f2f0bb9699504f9e9d6d52d8d1ccecb1661e32368dbfd4689 not found: ID does not exist" Nov 24 14:47:00 crc kubenswrapper[4790]: I1124 14:47:00.335492 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fff5214-ae3c-4948-aaea-d5806cfae4b4" path="/var/lib/kubelet/pods/7fff5214-ae3c-4948-aaea-d5806cfae4b4/volumes" Nov 24 14:47:00 crc kubenswrapper[4790]: I1124 14:47:00.947828 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 14:47:00 crc kubenswrapper[4790]: E1124 14:47:00.948571 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fff5214-ae3c-4948-aaea-d5806cfae4b4" containerName="registry-server" Nov 24 14:47:00 crc kubenswrapper[4790]: I1124 14:47:00.948589 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fff5214-ae3c-4948-aaea-d5806cfae4b4" containerName="registry-server" Nov 24 14:47:00 crc kubenswrapper[4790]: E1124 14:47:00.948603 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fff5214-ae3c-4948-aaea-d5806cfae4b4" containerName="extract-content" Nov 24 14:47:00 crc kubenswrapper[4790]: I1124 14:47:00.948608 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fff5214-ae3c-4948-aaea-d5806cfae4b4" containerName="extract-content" Nov 24 14:47:00 crc kubenswrapper[4790]: E1124 14:47:00.948620 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fff5214-ae3c-4948-aaea-d5806cfae4b4" containerName="extract-utilities" Nov 24 14:47:00 crc kubenswrapper[4790]: I1124 14:47:00.948626 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fff5214-ae3c-4948-aaea-d5806cfae4b4" containerName="extract-utilities" Nov 24 14:47:00 crc kubenswrapper[4790]: I1124 14:47:00.948795 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fff5214-ae3c-4948-aaea-d5806cfae4b4" containerName="registry-server" Nov 24 14:47:00 crc kubenswrapper[4790]: I1124 14:47:00.949772 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 14:47:00 crc kubenswrapper[4790]: I1124 14:47:00.952401 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 24 14:47:00 crc kubenswrapper[4790]: I1124 14:47:00.957027 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 14:47:01 crc kubenswrapper[4790]: I1124 14:47:01.046681 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/159300e3-1764-475a-aff1-1703d7420fbf-scripts\") pod \"cinder-scheduler-0\" (UID: \"159300e3-1764-475a-aff1-1703d7420fbf\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:01 crc kubenswrapper[4790]: I1124 14:47:01.046753 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/159300e3-1764-475a-aff1-1703d7420fbf-config-data\") pod \"cinder-scheduler-0\" (UID: \"159300e3-1764-475a-aff1-1703d7420fbf\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:01 crc kubenswrapper[4790]: I1124 14:47:01.046798 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/159300e3-1764-475a-aff1-1703d7420fbf-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"159300e3-1764-475a-aff1-1703d7420fbf\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:01 crc kubenswrapper[4790]: I1124 14:47:01.046874 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/159300e3-1764-475a-aff1-1703d7420fbf-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"159300e3-1764-475a-aff1-1703d7420fbf\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:01 crc kubenswrapper[4790]: I1124 14:47:01.046929 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxj62\" (UniqueName: \"kubernetes.io/projected/159300e3-1764-475a-aff1-1703d7420fbf-kube-api-access-wxj62\") pod \"cinder-scheduler-0\" (UID: \"159300e3-1764-475a-aff1-1703d7420fbf\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:01 crc kubenswrapper[4790]: I1124 14:47:01.046967 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/159300e3-1764-475a-aff1-1703d7420fbf-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"159300e3-1764-475a-aff1-1703d7420fbf\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:01 crc kubenswrapper[4790]: I1124 14:47:01.148520 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/159300e3-1764-475a-aff1-1703d7420fbf-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"159300e3-1764-475a-aff1-1703d7420fbf\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:01 crc kubenswrapper[4790]: I1124 14:47:01.148603 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/159300e3-1764-475a-aff1-1703d7420fbf-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"159300e3-1764-475a-aff1-1703d7420fbf\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:01 crc kubenswrapper[4790]: I1124 14:47:01.148627 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxj62\" (UniqueName: \"kubernetes.io/projected/159300e3-1764-475a-aff1-1703d7420fbf-kube-api-access-wxj62\") pod \"cinder-scheduler-0\" (UID: \"159300e3-1764-475a-aff1-1703d7420fbf\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:01 crc kubenswrapper[4790]: I1124 14:47:01.148655 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/159300e3-1764-475a-aff1-1703d7420fbf-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"159300e3-1764-475a-aff1-1703d7420fbf\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:01 crc kubenswrapper[4790]: I1124 14:47:01.148733 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/159300e3-1764-475a-aff1-1703d7420fbf-scripts\") pod \"cinder-scheduler-0\" (UID: \"159300e3-1764-475a-aff1-1703d7420fbf\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:01 crc kubenswrapper[4790]: I1124 14:47:01.148765 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/159300e3-1764-475a-aff1-1703d7420fbf-config-data\") pod \"cinder-scheduler-0\" (UID: \"159300e3-1764-475a-aff1-1703d7420fbf\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:01 crc kubenswrapper[4790]: I1124 14:47:01.148860 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/159300e3-1764-475a-aff1-1703d7420fbf-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"159300e3-1764-475a-aff1-1703d7420fbf\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:01 crc kubenswrapper[4790]: I1124 14:47:01.153697 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/159300e3-1764-475a-aff1-1703d7420fbf-scripts\") pod \"cinder-scheduler-0\" (UID: \"159300e3-1764-475a-aff1-1703d7420fbf\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:01 crc kubenswrapper[4790]: I1124 14:47:01.154081 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/159300e3-1764-475a-aff1-1703d7420fbf-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"159300e3-1764-475a-aff1-1703d7420fbf\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:01 crc kubenswrapper[4790]: I1124 14:47:01.154642 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/159300e3-1764-475a-aff1-1703d7420fbf-config-data\") pod \"cinder-scheduler-0\" (UID: \"159300e3-1764-475a-aff1-1703d7420fbf\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:01 crc kubenswrapper[4790]: I1124 14:47:01.155005 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/159300e3-1764-475a-aff1-1703d7420fbf-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"159300e3-1764-475a-aff1-1703d7420fbf\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:01 crc kubenswrapper[4790]: I1124 14:47:01.170034 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxj62\" (UniqueName: \"kubernetes.io/projected/159300e3-1764-475a-aff1-1703d7420fbf-kube-api-access-wxj62\") pod \"cinder-scheduler-0\" (UID: \"159300e3-1764-475a-aff1-1703d7420fbf\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:01 crc kubenswrapper[4790]: I1124 14:47:01.275115 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 14:47:01 crc kubenswrapper[4790]: I1124 14:47:01.442330 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 24 14:47:01 crc kubenswrapper[4790]: W1124 14:47:01.741982 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod159300e3_1764_475a_aff1_1703d7420fbf.slice/crio-27ff6db9268373a981c4e714b9dcf473d9b8ad58d184be07fa1a4ff2fb95de25 WatchSource:0}: Error finding container 27ff6db9268373a981c4e714b9dcf473d9b8ad58d184be07fa1a4ff2fb95de25: Status 404 returned error can't find the container with id 27ff6db9268373a981c4e714b9dcf473d9b8ad58d184be07fa1a4ff2fb95de25 Nov 24 14:47:01 crc kubenswrapper[4790]: I1124 14:47:01.743793 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 14:47:02 crc kubenswrapper[4790]: I1124 14:47:02.135833 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"159300e3-1764-475a-aff1-1703d7420fbf","Type":"ContainerStarted","Data":"27ff6db9268373a981c4e714b9dcf473d9b8ad58d184be07fa1a4ff2fb95de25"} Nov 24 14:47:02 crc kubenswrapper[4790]: I1124 14:47:02.753023 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 14:47:02 crc kubenswrapper[4790]: I1124 14:47:02.753366 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="3942dd50-c47f-4c59-95e8-5145f6fcc87c" containerName="cinder-api-log" containerID="cri-o://86d55160de7791fb15fb8c40316f8bbda418225d85d816ad257a0b100f170727" gracePeriod=30 Nov 24 14:47:02 crc kubenswrapper[4790]: I1124 14:47:02.753899 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="3942dd50-c47f-4c59-95e8-5145f6fcc87c" containerName="cinder-api" containerID="cri-o://2a469101e678b8c8d235d59534fde9844becde0d76b4a87a5ab89e7b0dddd473" gracePeriod=30 Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.153160 4790 generic.go:334] "Generic (PLEG): container finished" podID="3942dd50-c47f-4c59-95e8-5145f6fcc87c" containerID="86d55160de7791fb15fb8c40316f8bbda418225d85d816ad257a0b100f170727" exitCode=143 Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.153257 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3942dd50-c47f-4c59-95e8-5145f6fcc87c","Type":"ContainerDied","Data":"86d55160de7791fb15fb8c40316f8bbda418225d85d816ad257a0b100f170727"} Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.158197 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"159300e3-1764-475a-aff1-1703d7420fbf","Type":"ContainerStarted","Data":"b71894b4eb3a2083356ba96cd87333c125605499be6338e3ef41f26d3569ed39"} Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.158245 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"159300e3-1764-475a-aff1-1703d7420fbf","Type":"ContainerStarted","Data":"a08896726072699810185fac8556a826d3d53a911d7ea2c9a53e7a2a5f25f002"} Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.187501 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.188158 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.188134857 podStartE2EDuration="3.188134857s" podCreationTimestamp="2025-11-24 14:47:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:47:03.184764231 +0000 UTC m=+5671.564657903" watchObservedRunningTime="2025-11-24 14:47:03.188134857 +0000 UTC m=+5671.568028519" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.189605 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.196117 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.243066 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.385851 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/777b5526-f7f6-4ab3-9399-7aa596ff36a0-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.385904 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zd6sx\" (UniqueName: \"kubernetes.io/projected/777b5526-f7f6-4ab3-9399-7aa596ff36a0-kube-api-access-zd6sx\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.385960 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/777b5526-f7f6-4ab3-9399-7aa596ff36a0-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.385984 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/777b5526-f7f6-4ab3-9399-7aa596ff36a0-run\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.386004 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/777b5526-f7f6-4ab3-9399-7aa596ff36a0-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.386024 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/777b5526-f7f6-4ab3-9399-7aa596ff36a0-dev\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.386251 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/777b5526-f7f6-4ab3-9399-7aa596ff36a0-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.386344 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/777b5526-f7f6-4ab3-9399-7aa596ff36a0-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.386411 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/777b5526-f7f6-4ab3-9399-7aa596ff36a0-sys\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.386531 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/777b5526-f7f6-4ab3-9399-7aa596ff36a0-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.386575 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/777b5526-f7f6-4ab3-9399-7aa596ff36a0-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.386625 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/777b5526-f7f6-4ab3-9399-7aa596ff36a0-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.386661 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/777b5526-f7f6-4ab3-9399-7aa596ff36a0-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.386702 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/777b5526-f7f6-4ab3-9399-7aa596ff36a0-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.386751 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/777b5526-f7f6-4ab3-9399-7aa596ff36a0-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.386802 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/777b5526-f7f6-4ab3-9399-7aa596ff36a0-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.455750 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.484260 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.487921 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/777b5526-f7f6-4ab3-9399-7aa596ff36a0-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.487977 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/777b5526-f7f6-4ab3-9399-7aa596ff36a0-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.488004 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/777b5526-f7f6-4ab3-9399-7aa596ff36a0-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.488030 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/777b5526-f7f6-4ab3-9399-7aa596ff36a0-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.488056 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/777b5526-f7f6-4ab3-9399-7aa596ff36a0-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.488089 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/777b5526-f7f6-4ab3-9399-7aa596ff36a0-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.488100 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/777b5526-f7f6-4ab3-9399-7aa596ff36a0-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.488118 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/777b5526-f7f6-4ab3-9399-7aa596ff36a0-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.488139 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zd6sx\" (UniqueName: \"kubernetes.io/projected/777b5526-f7f6-4ab3-9399-7aa596ff36a0-kube-api-access-zd6sx\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.488143 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/777b5526-f7f6-4ab3-9399-7aa596ff36a0-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.488181 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/777b5526-f7f6-4ab3-9399-7aa596ff36a0-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.488204 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/777b5526-f7f6-4ab3-9399-7aa596ff36a0-run\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.488223 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/777b5526-f7f6-4ab3-9399-7aa596ff36a0-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.488250 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/777b5526-f7f6-4ab3-9399-7aa596ff36a0-dev\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.488312 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/777b5526-f7f6-4ab3-9399-7aa596ff36a0-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.488335 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/777b5526-f7f6-4ab3-9399-7aa596ff36a0-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.488353 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/777b5526-f7f6-4ab3-9399-7aa596ff36a0-sys\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.488410 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/777b5526-f7f6-4ab3-9399-7aa596ff36a0-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.488525 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/777b5526-f7f6-4ab3-9399-7aa596ff36a0-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.488586 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/777b5526-f7f6-4ab3-9399-7aa596ff36a0-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.488812 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/777b5526-f7f6-4ab3-9399-7aa596ff36a0-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.488875 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/777b5526-f7f6-4ab3-9399-7aa596ff36a0-dev\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.488921 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/777b5526-f7f6-4ab3-9399-7aa596ff36a0-run\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.489030 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/777b5526-f7f6-4ab3-9399-7aa596ff36a0-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.489266 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/777b5526-f7f6-4ab3-9399-7aa596ff36a0-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.489347 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/777b5526-f7f6-4ab3-9399-7aa596ff36a0-sys\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.497160 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/777b5526-f7f6-4ab3-9399-7aa596ff36a0-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.500952 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/777b5526-f7f6-4ab3-9399-7aa596ff36a0-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.502024 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/777b5526-f7f6-4ab3-9399-7aa596ff36a0-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.512521 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/777b5526-f7f6-4ab3-9399-7aa596ff36a0-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.513097 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/777b5526-f7f6-4ab3-9399-7aa596ff36a0-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.513297 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zd6sx\" (UniqueName: \"kubernetes.io/projected/777b5526-f7f6-4ab3-9399-7aa596ff36a0-kube-api-access-zd6sx\") pod \"cinder-volume-volume1-0\" (UID: \"777b5526-f7f6-4ab3-9399-7aa596ff36a0\") " pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.524321 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.845928 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.848134 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.852180 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.861374 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.901839 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b54f7c8e-31ff-4565-bebe-663f7de8b110-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.901920 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b54f7c8e-31ff-4565-bebe-663f7de8b110-etc-nvme\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.902051 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2x6h\" (UniqueName: \"kubernetes.io/projected/b54f7c8e-31ff-4565-bebe-663f7de8b110-kube-api-access-b2x6h\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.902100 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b54f7c8e-31ff-4565-bebe-663f7de8b110-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.902183 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b54f7c8e-31ff-4565-bebe-663f7de8b110-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.902219 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b54f7c8e-31ff-4565-bebe-663f7de8b110-scripts\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.902358 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b54f7c8e-31ff-4565-bebe-663f7de8b110-run\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.902388 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b54f7c8e-31ff-4565-bebe-663f7de8b110-config-data-custom\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.902524 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b54f7c8e-31ff-4565-bebe-663f7de8b110-ceph\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.902578 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b54f7c8e-31ff-4565-bebe-663f7de8b110-sys\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.902600 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b54f7c8e-31ff-4565-bebe-663f7de8b110-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.902627 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/b54f7c8e-31ff-4565-bebe-663f7de8b110-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.902700 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b54f7c8e-31ff-4565-bebe-663f7de8b110-config-data\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.902746 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b54f7c8e-31ff-4565-bebe-663f7de8b110-lib-modules\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.902769 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/b54f7c8e-31ff-4565-bebe-663f7de8b110-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:03 crc kubenswrapper[4790]: I1124 14:47:03.902831 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b54f7c8e-31ff-4565-bebe-663f7de8b110-dev\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.004557 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b54f7c8e-31ff-4565-bebe-663f7de8b110-run\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.004680 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b54f7c8e-31ff-4565-bebe-663f7de8b110-run\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.004922 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b54f7c8e-31ff-4565-bebe-663f7de8b110-config-data-custom\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.005038 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b54f7c8e-31ff-4565-bebe-663f7de8b110-ceph\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.005075 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b54f7c8e-31ff-4565-bebe-663f7de8b110-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.005096 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b54f7c8e-31ff-4565-bebe-663f7de8b110-sys\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.005117 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/b54f7c8e-31ff-4565-bebe-663f7de8b110-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.005164 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b54f7c8e-31ff-4565-bebe-663f7de8b110-config-data\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.005197 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b54f7c8e-31ff-4565-bebe-663f7de8b110-lib-modules\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.005198 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b54f7c8e-31ff-4565-bebe-663f7de8b110-sys\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.005230 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/b54f7c8e-31ff-4565-bebe-663f7de8b110-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.005268 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b54f7c8e-31ff-4565-bebe-663f7de8b110-dev\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.005315 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b54f7c8e-31ff-4565-bebe-663f7de8b110-lib-modules\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.005341 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/b54f7c8e-31ff-4565-bebe-663f7de8b110-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.005353 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b54f7c8e-31ff-4565-bebe-663f7de8b110-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.005313 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/b54f7c8e-31ff-4565-bebe-663f7de8b110-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.005400 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b54f7c8e-31ff-4565-bebe-663f7de8b110-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.005412 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b54f7c8e-31ff-4565-bebe-663f7de8b110-etc-nvme\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.005410 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b54f7c8e-31ff-4565-bebe-663f7de8b110-dev\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.005445 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b54f7c8e-31ff-4565-bebe-663f7de8b110-etc-nvme\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.005448 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2x6h\" (UniqueName: \"kubernetes.io/projected/b54f7c8e-31ff-4565-bebe-663f7de8b110-kube-api-access-b2x6h\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.005504 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b54f7c8e-31ff-4565-bebe-663f7de8b110-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.005548 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b54f7c8e-31ff-4565-bebe-663f7de8b110-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.005573 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b54f7c8e-31ff-4565-bebe-663f7de8b110-scripts\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.005610 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b54f7c8e-31ff-4565-bebe-663f7de8b110-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.005714 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b54f7c8e-31ff-4565-bebe-663f7de8b110-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.011865 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b54f7c8e-31ff-4565-bebe-663f7de8b110-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.012945 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b54f7c8e-31ff-4565-bebe-663f7de8b110-scripts\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.013022 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b54f7c8e-31ff-4565-bebe-663f7de8b110-config-data\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.013368 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b54f7c8e-31ff-4565-bebe-663f7de8b110-ceph\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.019301 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b54f7c8e-31ff-4565-bebe-663f7de8b110-config-data-custom\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.030126 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2x6h\" (UniqueName: \"kubernetes.io/projected/b54f7c8e-31ff-4565-bebe-663f7de8b110-kube-api-access-b2x6h\") pod \"cinder-backup-0\" (UID: \"b54f7c8e-31ff-4565-bebe-663f7de8b110\") " pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.061277 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Nov 24 14:47:04 crc kubenswrapper[4790]: W1124 14:47:04.069354 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod777b5526_f7f6_4ab3_9399_7aa596ff36a0.slice/crio-dca753746a079604e3f340d3a9e5b75339b0ecee9a7a90aa1fd3df435acf0a67 WatchSource:0}: Error finding container dca753746a079604e3f340d3a9e5b75339b0ecee9a7a90aa1fd3df435acf0a67: Status 404 returned error can't find the container with id dca753746a079604e3f340d3a9e5b75339b0ecee9a7a90aa1fd3df435acf0a67 Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.165424 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.169463 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"777b5526-f7f6-4ab3-9399-7aa596ff36a0","Type":"ContainerStarted","Data":"dca753746a079604e3f340d3a9e5b75339b0ecee9a7a90aa1fd3df435acf0a67"} Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.207818 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 14:47:04 crc kubenswrapper[4790]: I1124 14:47:04.780515 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Nov 24 14:47:04 crc kubenswrapper[4790]: W1124 14:47:04.794683 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb54f7c8e_31ff_4565_bebe_663f7de8b110.slice/crio-46404da28890e7b4563ed6ddcbbfdb5fbc84c40ee6c39dd4533f8b43b3083f42 WatchSource:0}: Error finding container 46404da28890e7b4563ed6ddcbbfdb5fbc84c40ee6c39dd4533f8b43b3083f42: Status 404 returned error can't find the container with id 46404da28890e7b4563ed6ddcbbfdb5fbc84c40ee6c39dd4533f8b43b3083f42 Nov 24 14:47:05 crc kubenswrapper[4790]: I1124 14:47:05.179743 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"b54f7c8e-31ff-4565-bebe-663f7de8b110","Type":"ContainerStarted","Data":"46404da28890e7b4563ed6ddcbbfdb5fbc84c40ee6c39dd4533f8b43b3083f42"} Nov 24 14:47:06 crc kubenswrapper[4790]: I1124 14:47:06.195703 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"777b5526-f7f6-4ab3-9399-7aa596ff36a0","Type":"ContainerStarted","Data":"6099d932aa9c2f2a305bb893c5c3548f97d4c34bf47a397d1db93bee72fc79e8"} Nov 24 14:47:06 crc kubenswrapper[4790]: I1124 14:47:06.198663 4790 generic.go:334] "Generic (PLEG): container finished" podID="3942dd50-c47f-4c59-95e8-5145f6fcc87c" containerID="2a469101e678b8c8d235d59534fde9844becde0d76b4a87a5ab89e7b0dddd473" exitCode=0 Nov 24 14:47:06 crc kubenswrapper[4790]: I1124 14:47:06.198725 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3942dd50-c47f-4c59-95e8-5145f6fcc87c","Type":"ContainerDied","Data":"2a469101e678b8c8d235d59534fde9844becde0d76b4a87a5ab89e7b0dddd473"} Nov 24 14:47:06 crc kubenswrapper[4790]: I1124 14:47:06.276937 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 24 14:47:06 crc kubenswrapper[4790]: I1124 14:47:06.346494 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 14:47:06 crc kubenswrapper[4790]: I1124 14:47:06.472325 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3942dd50-c47f-4c59-95e8-5145f6fcc87c-etc-machine-id\") pod \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\" (UID: \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\") " Nov 24 14:47:06 crc kubenswrapper[4790]: I1124 14:47:06.472684 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3942dd50-c47f-4c59-95e8-5145f6fcc87c-scripts\") pod \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\" (UID: \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\") " Nov 24 14:47:06 crc kubenswrapper[4790]: I1124 14:47:06.472707 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmb4c\" (UniqueName: \"kubernetes.io/projected/3942dd50-c47f-4c59-95e8-5145f6fcc87c-kube-api-access-vmb4c\") pod \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\" (UID: \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\") " Nov 24 14:47:06 crc kubenswrapper[4790]: I1124 14:47:06.472836 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3942dd50-c47f-4c59-95e8-5145f6fcc87c-combined-ca-bundle\") pod \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\" (UID: \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\") " Nov 24 14:47:06 crc kubenswrapper[4790]: I1124 14:47:06.472924 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3942dd50-c47f-4c59-95e8-5145f6fcc87c-config-data\") pod \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\" (UID: \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\") " Nov 24 14:47:06 crc kubenswrapper[4790]: I1124 14:47:06.472973 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3942dd50-c47f-4c59-95e8-5145f6fcc87c-config-data-custom\") pod \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\" (UID: \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\") " Nov 24 14:47:06 crc kubenswrapper[4790]: I1124 14:47:06.473018 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3942dd50-c47f-4c59-95e8-5145f6fcc87c-logs\") pod \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\" (UID: \"3942dd50-c47f-4c59-95e8-5145f6fcc87c\") " Nov 24 14:47:06 crc kubenswrapper[4790]: I1124 14:47:06.473118 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3942dd50-c47f-4c59-95e8-5145f6fcc87c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "3942dd50-c47f-4c59-95e8-5145f6fcc87c" (UID: "3942dd50-c47f-4c59-95e8-5145f6fcc87c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:47:06 crc kubenswrapper[4790]: I1124 14:47:06.473400 4790 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3942dd50-c47f-4c59-95e8-5145f6fcc87c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 14:47:06 crc kubenswrapper[4790]: I1124 14:47:06.473987 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3942dd50-c47f-4c59-95e8-5145f6fcc87c-logs" (OuterVolumeSpecName: "logs") pod "3942dd50-c47f-4c59-95e8-5145f6fcc87c" (UID: "3942dd50-c47f-4c59-95e8-5145f6fcc87c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:47:06 crc kubenswrapper[4790]: I1124 14:47:06.479097 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3942dd50-c47f-4c59-95e8-5145f6fcc87c-scripts" (OuterVolumeSpecName: "scripts") pod "3942dd50-c47f-4c59-95e8-5145f6fcc87c" (UID: "3942dd50-c47f-4c59-95e8-5145f6fcc87c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:47:06 crc kubenswrapper[4790]: I1124 14:47:06.479189 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3942dd50-c47f-4c59-95e8-5145f6fcc87c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3942dd50-c47f-4c59-95e8-5145f6fcc87c" (UID: "3942dd50-c47f-4c59-95e8-5145f6fcc87c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:47:06 crc kubenswrapper[4790]: I1124 14:47:06.479897 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3942dd50-c47f-4c59-95e8-5145f6fcc87c-kube-api-access-vmb4c" (OuterVolumeSpecName: "kube-api-access-vmb4c") pod "3942dd50-c47f-4c59-95e8-5145f6fcc87c" (UID: "3942dd50-c47f-4c59-95e8-5145f6fcc87c"). InnerVolumeSpecName "kube-api-access-vmb4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:47:06 crc kubenswrapper[4790]: I1124 14:47:06.508934 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3942dd50-c47f-4c59-95e8-5145f6fcc87c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3942dd50-c47f-4c59-95e8-5145f6fcc87c" (UID: "3942dd50-c47f-4c59-95e8-5145f6fcc87c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:47:06 crc kubenswrapper[4790]: I1124 14:47:06.529902 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3942dd50-c47f-4c59-95e8-5145f6fcc87c-config-data" (OuterVolumeSpecName: "config-data") pod "3942dd50-c47f-4c59-95e8-5145f6fcc87c" (UID: "3942dd50-c47f-4c59-95e8-5145f6fcc87c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:47:06 crc kubenswrapper[4790]: I1124 14:47:06.576807 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3942dd50-c47f-4c59-95e8-5145f6fcc87c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:47:06 crc kubenswrapper[4790]: I1124 14:47:06.576855 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3942dd50-c47f-4c59-95e8-5145f6fcc87c-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:47:06 crc kubenswrapper[4790]: I1124 14:47:06.576867 4790 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3942dd50-c47f-4c59-95e8-5145f6fcc87c-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 14:47:06 crc kubenswrapper[4790]: I1124 14:47:06.576922 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3942dd50-c47f-4c59-95e8-5145f6fcc87c-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:47:06 crc kubenswrapper[4790]: I1124 14:47:06.576935 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3942dd50-c47f-4c59-95e8-5145f6fcc87c-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:47:06 crc kubenswrapper[4790]: I1124 14:47:06.576947 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmb4c\" (UniqueName: \"kubernetes.io/projected/3942dd50-c47f-4c59-95e8-5145f6fcc87c-kube-api-access-vmb4c\") on node \"crc\" DevicePath \"\"" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.221930 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"b54f7c8e-31ff-4565-bebe-663f7de8b110","Type":"ContainerStarted","Data":"2b8cb2436190daaf3aba0f9be852b370cc8d6ba9789781e4fa7a42f22b4d1a19"} Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.222366 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"b54f7c8e-31ff-4565-bebe-663f7de8b110","Type":"ContainerStarted","Data":"ac3bdb2ae83f30068f584472ec42224cafe3de0a09f793af9dedd6e1bbcaab4a"} Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.228549 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"777b5526-f7f6-4ab3-9399-7aa596ff36a0","Type":"ContainerStarted","Data":"0f4eaec86b3fafde70c7b3ff4ad11b1fea0184b7c72d3f3cad46807b3d198400"} Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.232720 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3942dd50-c47f-4c59-95e8-5145f6fcc87c","Type":"ContainerDied","Data":"cc623835879da0b1c0d8789dd0272fcbd1889783d1b7e0e66b93bb9c253e11ec"} Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.232803 4790 scope.go:117] "RemoveContainer" containerID="2a469101e678b8c8d235d59534fde9844becde0d76b4a87a5ab89e7b0dddd473" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.232822 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.269090 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=2.955668851 podStartE2EDuration="4.269033839s" podCreationTimestamp="2025-11-24 14:47:03 +0000 UTC" firstStartedPulling="2025-11-24 14:47:04.797944613 +0000 UTC m=+5673.177838275" lastFinishedPulling="2025-11-24 14:47:06.111309591 +0000 UTC m=+5674.491203263" observedRunningTime="2025-11-24 14:47:07.262612335 +0000 UTC m=+5675.642506027" watchObservedRunningTime="2025-11-24 14:47:07.269033839 +0000 UTC m=+5675.648927531" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.275985 4790 scope.go:117] "RemoveContainer" containerID="86d55160de7791fb15fb8c40316f8bbda418225d85d816ad257a0b100f170727" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.316666 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=2.80385037 podStartE2EDuration="4.316556925s" podCreationTimestamp="2025-11-24 14:47:03 +0000 UTC" firstStartedPulling="2025-11-24 14:47:04.071551016 +0000 UTC m=+5672.451444678" lastFinishedPulling="2025-11-24 14:47:05.584257571 +0000 UTC m=+5673.964151233" observedRunningTime="2025-11-24 14:47:07.291234607 +0000 UTC m=+5675.671128289" watchObservedRunningTime="2025-11-24 14:47:07.316556925 +0000 UTC m=+5675.696450627" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.338656 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.348396 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.348491 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.348556 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.354732 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.355476 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.361396 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.362659 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.370158 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.375125 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 24 14:47:07 crc kubenswrapper[4790]: E1124 14:47:07.375802 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3942dd50-c47f-4c59-95e8-5145f6fcc87c" containerName="cinder-api-log" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.375921 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="3942dd50-c47f-4c59-95e8-5145f6fcc87c" containerName="cinder-api-log" Nov 24 14:47:07 crc kubenswrapper[4790]: E1124 14:47:07.376040 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3942dd50-c47f-4c59-95e8-5145f6fcc87c" containerName="cinder-api" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.376108 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="3942dd50-c47f-4c59-95e8-5145f6fcc87c" containerName="cinder-api" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.377155 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="3942dd50-c47f-4c59-95e8-5145f6fcc87c" containerName="cinder-api-log" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.377278 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="3942dd50-c47f-4c59-95e8-5145f6fcc87c" containerName="cinder-api" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.378641 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.382493 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.392432 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.496103 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70481c96-3eef-4652-8918-9558f9ec74ae-logs\") pod \"cinder-api-0\" (UID: \"70481c96-3eef-4652-8918-9558f9ec74ae\") " pod="openstack/cinder-api-0" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.496165 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70481c96-3eef-4652-8918-9558f9ec74ae-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"70481c96-3eef-4652-8918-9558f9ec74ae\") " pod="openstack/cinder-api-0" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.496278 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/70481c96-3eef-4652-8918-9558f9ec74ae-etc-machine-id\") pod \"cinder-api-0\" (UID: \"70481c96-3eef-4652-8918-9558f9ec74ae\") " pod="openstack/cinder-api-0" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.496473 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70481c96-3eef-4652-8918-9558f9ec74ae-config-data\") pod \"cinder-api-0\" (UID: \"70481c96-3eef-4652-8918-9558f9ec74ae\") " pod="openstack/cinder-api-0" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.496565 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9g4n\" (UniqueName: \"kubernetes.io/projected/70481c96-3eef-4652-8918-9558f9ec74ae-kube-api-access-l9g4n\") pod \"cinder-api-0\" (UID: \"70481c96-3eef-4652-8918-9558f9ec74ae\") " pod="openstack/cinder-api-0" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.497763 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70481c96-3eef-4652-8918-9558f9ec74ae-scripts\") pod \"cinder-api-0\" (UID: \"70481c96-3eef-4652-8918-9558f9ec74ae\") " pod="openstack/cinder-api-0" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.497908 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70481c96-3eef-4652-8918-9558f9ec74ae-config-data-custom\") pod \"cinder-api-0\" (UID: \"70481c96-3eef-4652-8918-9558f9ec74ae\") " pod="openstack/cinder-api-0" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.599674 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70481c96-3eef-4652-8918-9558f9ec74ae-scripts\") pod \"cinder-api-0\" (UID: \"70481c96-3eef-4652-8918-9558f9ec74ae\") " pod="openstack/cinder-api-0" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.599950 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70481c96-3eef-4652-8918-9558f9ec74ae-config-data-custom\") pod \"cinder-api-0\" (UID: \"70481c96-3eef-4652-8918-9558f9ec74ae\") " pod="openstack/cinder-api-0" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.600751 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70481c96-3eef-4652-8918-9558f9ec74ae-logs\") pod \"cinder-api-0\" (UID: \"70481c96-3eef-4652-8918-9558f9ec74ae\") " pod="openstack/cinder-api-0" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.600870 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70481c96-3eef-4652-8918-9558f9ec74ae-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"70481c96-3eef-4652-8918-9558f9ec74ae\") " pod="openstack/cinder-api-0" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.601018 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/70481c96-3eef-4652-8918-9558f9ec74ae-etc-machine-id\") pod \"cinder-api-0\" (UID: \"70481c96-3eef-4652-8918-9558f9ec74ae\") " pod="openstack/cinder-api-0" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.601047 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70481c96-3eef-4652-8918-9558f9ec74ae-logs\") pod \"cinder-api-0\" (UID: \"70481c96-3eef-4652-8918-9558f9ec74ae\") " pod="openstack/cinder-api-0" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.601074 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/70481c96-3eef-4652-8918-9558f9ec74ae-etc-machine-id\") pod \"cinder-api-0\" (UID: \"70481c96-3eef-4652-8918-9558f9ec74ae\") " pod="openstack/cinder-api-0" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.601304 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70481c96-3eef-4652-8918-9558f9ec74ae-config-data\") pod \"cinder-api-0\" (UID: \"70481c96-3eef-4652-8918-9558f9ec74ae\") " pod="openstack/cinder-api-0" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.601403 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9g4n\" (UniqueName: \"kubernetes.io/projected/70481c96-3eef-4652-8918-9558f9ec74ae-kube-api-access-l9g4n\") pod \"cinder-api-0\" (UID: \"70481c96-3eef-4652-8918-9558f9ec74ae\") " pod="openstack/cinder-api-0" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.616573 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70481c96-3eef-4652-8918-9558f9ec74ae-scripts\") pod \"cinder-api-0\" (UID: \"70481c96-3eef-4652-8918-9558f9ec74ae\") " pod="openstack/cinder-api-0" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.616628 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70481c96-3eef-4652-8918-9558f9ec74ae-config-data-custom\") pod \"cinder-api-0\" (UID: \"70481c96-3eef-4652-8918-9558f9ec74ae\") " pod="openstack/cinder-api-0" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.617331 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70481c96-3eef-4652-8918-9558f9ec74ae-config-data\") pod \"cinder-api-0\" (UID: \"70481c96-3eef-4652-8918-9558f9ec74ae\") " pod="openstack/cinder-api-0" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.619056 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70481c96-3eef-4652-8918-9558f9ec74ae-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"70481c96-3eef-4652-8918-9558f9ec74ae\") " pod="openstack/cinder-api-0" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.619069 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9g4n\" (UniqueName: \"kubernetes.io/projected/70481c96-3eef-4652-8918-9558f9ec74ae-kube-api-access-l9g4n\") pod \"cinder-api-0\" (UID: \"70481c96-3eef-4652-8918-9558f9ec74ae\") " pod="openstack/cinder-api-0" Nov 24 14:47:07 crc kubenswrapper[4790]: I1124 14:47:07.712719 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 14:47:08 crc kubenswrapper[4790]: I1124 14:47:08.195444 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 14:47:08 crc kubenswrapper[4790]: I1124 14:47:08.243726 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"70481c96-3eef-4652-8918-9558f9ec74ae","Type":"ContainerStarted","Data":"90be7979de343026129a20c37349297788f30ad45ad765c65d707d9b4c9859f1"} Nov 24 14:47:08 crc kubenswrapper[4790]: I1124 14:47:08.246515 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 14:47:08 crc kubenswrapper[4790]: I1124 14:47:08.251513 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 14:47:08 crc kubenswrapper[4790]: I1124 14:47:08.253057 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 14:47:08 crc kubenswrapper[4790]: I1124 14:47:08.328595 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3942dd50-c47f-4c59-95e8-5145f6fcc87c" path="/var/lib/kubelet/pods/3942dd50-c47f-4c59-95e8-5145f6fcc87c/volumes" Nov 24 14:47:08 crc kubenswrapper[4790]: I1124 14:47:08.524694 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:09 crc kubenswrapper[4790]: I1124 14:47:09.165782 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Nov 24 14:47:09 crc kubenswrapper[4790]: I1124 14:47:09.259705 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"70481c96-3eef-4652-8918-9558f9ec74ae","Type":"ContainerStarted","Data":"96ae8c0160fc12a5893b833fd685aebca10a760022e6140884e916d307a9623b"} Nov 24 14:47:10 crc kubenswrapper[4790]: I1124 14:47:10.273154 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"70481c96-3eef-4652-8918-9558f9ec74ae","Type":"ContainerStarted","Data":"b296a2bf728eb92223ebbf023ce22495b73a6f962e4e136567754eed1cdf5700"} Nov 24 14:47:10 crc kubenswrapper[4790]: I1124 14:47:10.273291 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 24 14:47:10 crc kubenswrapper[4790]: I1124 14:47:10.297494 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.297472227 podStartE2EDuration="3.297472227s" podCreationTimestamp="2025-11-24 14:47:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:47:10.290311022 +0000 UTC m=+5678.670204694" watchObservedRunningTime="2025-11-24 14:47:10.297472227 +0000 UTC m=+5678.677365909" Nov 24 14:47:11 crc kubenswrapper[4790]: I1124 14:47:11.532574 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 24 14:47:11 crc kubenswrapper[4790]: I1124 14:47:11.600920 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 14:47:12 crc kubenswrapper[4790]: I1124 14:47:12.292819 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="159300e3-1764-475a-aff1-1703d7420fbf" containerName="cinder-scheduler" containerID="cri-o://a08896726072699810185fac8556a826d3d53a911d7ea2c9a53e7a2a5f25f002" gracePeriod=30 Nov 24 14:47:12 crc kubenswrapper[4790]: I1124 14:47:12.292930 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="159300e3-1764-475a-aff1-1703d7420fbf" containerName="probe" containerID="cri-o://b71894b4eb3a2083356ba96cd87333c125605499be6338e3ef41f26d3569ed39" gracePeriod=30 Nov 24 14:47:13 crc kubenswrapper[4790]: I1124 14:47:13.306751 4790 generic.go:334] "Generic (PLEG): container finished" podID="159300e3-1764-475a-aff1-1703d7420fbf" containerID="b71894b4eb3a2083356ba96cd87333c125605499be6338e3ef41f26d3569ed39" exitCode=0 Nov 24 14:47:13 crc kubenswrapper[4790]: I1124 14:47:13.306858 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"159300e3-1764-475a-aff1-1703d7420fbf","Type":"ContainerDied","Data":"b71894b4eb3a2083356ba96cd87333c125605499be6338e3ef41f26d3569ed39"} Nov 24 14:47:13 crc kubenswrapper[4790]: I1124 14:47:13.763518 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Nov 24 14:47:14 crc kubenswrapper[4790]: I1124 14:47:14.404188 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.164407 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.274704 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/159300e3-1764-475a-aff1-1703d7420fbf-config-data\") pod \"159300e3-1764-475a-aff1-1703d7420fbf\" (UID: \"159300e3-1764-475a-aff1-1703d7420fbf\") " Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.274876 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/159300e3-1764-475a-aff1-1703d7420fbf-combined-ca-bundle\") pod \"159300e3-1764-475a-aff1-1703d7420fbf\" (UID: \"159300e3-1764-475a-aff1-1703d7420fbf\") " Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.274914 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxj62\" (UniqueName: \"kubernetes.io/projected/159300e3-1764-475a-aff1-1703d7420fbf-kube-api-access-wxj62\") pod \"159300e3-1764-475a-aff1-1703d7420fbf\" (UID: \"159300e3-1764-475a-aff1-1703d7420fbf\") " Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.274958 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/159300e3-1764-475a-aff1-1703d7420fbf-scripts\") pod \"159300e3-1764-475a-aff1-1703d7420fbf\" (UID: \"159300e3-1764-475a-aff1-1703d7420fbf\") " Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.274991 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/159300e3-1764-475a-aff1-1703d7420fbf-config-data-custom\") pod \"159300e3-1764-475a-aff1-1703d7420fbf\" (UID: \"159300e3-1764-475a-aff1-1703d7420fbf\") " Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.275038 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/159300e3-1764-475a-aff1-1703d7420fbf-etc-machine-id\") pod \"159300e3-1764-475a-aff1-1703d7420fbf\" (UID: \"159300e3-1764-475a-aff1-1703d7420fbf\") " Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.275484 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/159300e3-1764-475a-aff1-1703d7420fbf-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "159300e3-1764-475a-aff1-1703d7420fbf" (UID: "159300e3-1764-475a-aff1-1703d7420fbf"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.280042 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/159300e3-1764-475a-aff1-1703d7420fbf-scripts" (OuterVolumeSpecName: "scripts") pod "159300e3-1764-475a-aff1-1703d7420fbf" (UID: "159300e3-1764-475a-aff1-1703d7420fbf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.280975 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/159300e3-1764-475a-aff1-1703d7420fbf-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "159300e3-1764-475a-aff1-1703d7420fbf" (UID: "159300e3-1764-475a-aff1-1703d7420fbf"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.281106 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/159300e3-1764-475a-aff1-1703d7420fbf-kube-api-access-wxj62" (OuterVolumeSpecName: "kube-api-access-wxj62") pod "159300e3-1764-475a-aff1-1703d7420fbf" (UID: "159300e3-1764-475a-aff1-1703d7420fbf"). InnerVolumeSpecName "kube-api-access-wxj62". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.331941 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/159300e3-1764-475a-aff1-1703d7420fbf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "159300e3-1764-475a-aff1-1703d7420fbf" (UID: "159300e3-1764-475a-aff1-1703d7420fbf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.345459 4790 generic.go:334] "Generic (PLEG): container finished" podID="159300e3-1764-475a-aff1-1703d7420fbf" containerID="a08896726072699810185fac8556a826d3d53a911d7ea2c9a53e7a2a5f25f002" exitCode=0 Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.345527 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"159300e3-1764-475a-aff1-1703d7420fbf","Type":"ContainerDied","Data":"a08896726072699810185fac8556a826d3d53a911d7ea2c9a53e7a2a5f25f002"} Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.345569 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"159300e3-1764-475a-aff1-1703d7420fbf","Type":"ContainerDied","Data":"27ff6db9268373a981c4e714b9dcf473d9b8ad58d184be07fa1a4ff2fb95de25"} Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.345596 4790 scope.go:117] "RemoveContainer" containerID="b71894b4eb3a2083356ba96cd87333c125605499be6338e3ef41f26d3569ed39" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.345859 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.377756 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/159300e3-1764-475a-aff1-1703d7420fbf-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.377793 4790 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/159300e3-1764-475a-aff1-1703d7420fbf-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.377804 4790 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/159300e3-1764-475a-aff1-1703d7420fbf-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.377813 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/159300e3-1764-475a-aff1-1703d7420fbf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.377822 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxj62\" (UniqueName: \"kubernetes.io/projected/159300e3-1764-475a-aff1-1703d7420fbf-kube-api-access-wxj62\") on node \"crc\" DevicePath \"\"" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.386664 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/159300e3-1764-475a-aff1-1703d7420fbf-config-data" (OuterVolumeSpecName: "config-data") pod "159300e3-1764-475a-aff1-1703d7420fbf" (UID: "159300e3-1764-475a-aff1-1703d7420fbf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.393227 4790 scope.go:117] "RemoveContainer" containerID="a08896726072699810185fac8556a826d3d53a911d7ea2c9a53e7a2a5f25f002" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.423516 4790 scope.go:117] "RemoveContainer" containerID="b71894b4eb3a2083356ba96cd87333c125605499be6338e3ef41f26d3569ed39" Nov 24 14:47:15 crc kubenswrapper[4790]: E1124 14:47:15.424032 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b71894b4eb3a2083356ba96cd87333c125605499be6338e3ef41f26d3569ed39\": container with ID starting with b71894b4eb3a2083356ba96cd87333c125605499be6338e3ef41f26d3569ed39 not found: ID does not exist" containerID="b71894b4eb3a2083356ba96cd87333c125605499be6338e3ef41f26d3569ed39" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.424066 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b71894b4eb3a2083356ba96cd87333c125605499be6338e3ef41f26d3569ed39"} err="failed to get container status \"b71894b4eb3a2083356ba96cd87333c125605499be6338e3ef41f26d3569ed39\": rpc error: code = NotFound desc = could not find container \"b71894b4eb3a2083356ba96cd87333c125605499be6338e3ef41f26d3569ed39\": container with ID starting with b71894b4eb3a2083356ba96cd87333c125605499be6338e3ef41f26d3569ed39 not found: ID does not exist" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.424086 4790 scope.go:117] "RemoveContainer" containerID="a08896726072699810185fac8556a826d3d53a911d7ea2c9a53e7a2a5f25f002" Nov 24 14:47:15 crc kubenswrapper[4790]: E1124 14:47:15.424627 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a08896726072699810185fac8556a826d3d53a911d7ea2c9a53e7a2a5f25f002\": container with ID starting with a08896726072699810185fac8556a826d3d53a911d7ea2c9a53e7a2a5f25f002 not found: ID does not exist" containerID="a08896726072699810185fac8556a826d3d53a911d7ea2c9a53e7a2a5f25f002" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.424665 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a08896726072699810185fac8556a826d3d53a911d7ea2c9a53e7a2a5f25f002"} err="failed to get container status \"a08896726072699810185fac8556a826d3d53a911d7ea2c9a53e7a2a5f25f002\": rpc error: code = NotFound desc = could not find container \"a08896726072699810185fac8556a826d3d53a911d7ea2c9a53e7a2a5f25f002\": container with ID starting with a08896726072699810185fac8556a826d3d53a911d7ea2c9a53e7a2a5f25f002 not found: ID does not exist" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.479264 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/159300e3-1764-475a-aff1-1703d7420fbf-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.693865 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.711763 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.722035 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 14:47:15 crc kubenswrapper[4790]: E1124 14:47:15.722472 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="159300e3-1764-475a-aff1-1703d7420fbf" containerName="probe" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.722493 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="159300e3-1764-475a-aff1-1703d7420fbf" containerName="probe" Nov 24 14:47:15 crc kubenswrapper[4790]: E1124 14:47:15.722520 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="159300e3-1764-475a-aff1-1703d7420fbf" containerName="cinder-scheduler" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.722528 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="159300e3-1764-475a-aff1-1703d7420fbf" containerName="cinder-scheduler" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.722734 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="159300e3-1764-475a-aff1-1703d7420fbf" containerName="cinder-scheduler" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.722776 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="159300e3-1764-475a-aff1-1703d7420fbf" containerName="probe" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.723903 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.731125 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.739969 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.888030 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73610ce7-d271-4573-a14b-fd4563757182-config-data\") pod \"cinder-scheduler-0\" (UID: \"73610ce7-d271-4573-a14b-fd4563757182\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.888089 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73610ce7-d271-4573-a14b-fd4563757182-scripts\") pod \"cinder-scheduler-0\" (UID: \"73610ce7-d271-4573-a14b-fd4563757182\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.888148 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73610ce7-d271-4573-a14b-fd4563757182-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"73610ce7-d271-4573-a14b-fd4563757182\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.888224 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/73610ce7-d271-4573-a14b-fd4563757182-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"73610ce7-d271-4573-a14b-fd4563757182\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.888323 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/73610ce7-d271-4573-a14b-fd4563757182-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"73610ce7-d271-4573-a14b-fd4563757182\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.888356 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mj4s\" (UniqueName: \"kubernetes.io/projected/73610ce7-d271-4573-a14b-fd4563757182-kube-api-access-9mj4s\") pod \"cinder-scheduler-0\" (UID: \"73610ce7-d271-4573-a14b-fd4563757182\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.990653 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/73610ce7-d271-4573-a14b-fd4563757182-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"73610ce7-d271-4573-a14b-fd4563757182\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.990746 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mj4s\" (UniqueName: \"kubernetes.io/projected/73610ce7-d271-4573-a14b-fd4563757182-kube-api-access-9mj4s\") pod \"cinder-scheduler-0\" (UID: \"73610ce7-d271-4573-a14b-fd4563757182\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.990805 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/73610ce7-d271-4573-a14b-fd4563757182-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"73610ce7-d271-4573-a14b-fd4563757182\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.990811 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73610ce7-d271-4573-a14b-fd4563757182-config-data\") pod \"cinder-scheduler-0\" (UID: \"73610ce7-d271-4573-a14b-fd4563757182\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.990958 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73610ce7-d271-4573-a14b-fd4563757182-scripts\") pod \"cinder-scheduler-0\" (UID: \"73610ce7-d271-4573-a14b-fd4563757182\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.991020 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73610ce7-d271-4573-a14b-fd4563757182-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"73610ce7-d271-4573-a14b-fd4563757182\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.991194 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/73610ce7-d271-4573-a14b-fd4563757182-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"73610ce7-d271-4573-a14b-fd4563757182\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.995156 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73610ce7-d271-4573-a14b-fd4563757182-scripts\") pod \"cinder-scheduler-0\" (UID: \"73610ce7-d271-4573-a14b-fd4563757182\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.995243 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/73610ce7-d271-4573-a14b-fd4563757182-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"73610ce7-d271-4573-a14b-fd4563757182\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.996836 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73610ce7-d271-4573-a14b-fd4563757182-config-data\") pod \"cinder-scheduler-0\" (UID: \"73610ce7-d271-4573-a14b-fd4563757182\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:15 crc kubenswrapper[4790]: I1124 14:47:15.997257 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73610ce7-d271-4573-a14b-fd4563757182-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"73610ce7-d271-4573-a14b-fd4563757182\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:16 crc kubenswrapper[4790]: I1124 14:47:16.010422 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mj4s\" (UniqueName: \"kubernetes.io/projected/73610ce7-d271-4573-a14b-fd4563757182-kube-api-access-9mj4s\") pod \"cinder-scheduler-0\" (UID: \"73610ce7-d271-4573-a14b-fd4563757182\") " pod="openstack/cinder-scheduler-0" Nov 24 14:47:16 crc kubenswrapper[4790]: I1124 14:47:16.061228 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 14:47:16 crc kubenswrapper[4790]: I1124 14:47:16.329971 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="159300e3-1764-475a-aff1-1703d7420fbf" path="/var/lib/kubelet/pods/159300e3-1764-475a-aff1-1703d7420fbf/volumes" Nov 24 14:47:16 crc kubenswrapper[4790]: W1124 14:47:16.568656 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73610ce7_d271_4573_a14b_fd4563757182.slice/crio-0e4956ec5232fc766e423d2a4258b53aa1a3034c3b175854e6c0f6dcd408ae19 WatchSource:0}: Error finding container 0e4956ec5232fc766e423d2a4258b53aa1a3034c3b175854e6c0f6dcd408ae19: Status 404 returned error can't find the container with id 0e4956ec5232fc766e423d2a4258b53aa1a3034c3b175854e6c0f6dcd408ae19 Nov 24 14:47:16 crc kubenswrapper[4790]: I1124 14:47:16.571760 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 14:47:17 crc kubenswrapper[4790]: I1124 14:47:17.365277 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"73610ce7-d271-4573-a14b-fd4563757182","Type":"ContainerStarted","Data":"e62ddc0048e80ced3bb23ab59fe210dbbff71226f398be48a282f38804ba4a32"} Nov 24 14:47:17 crc kubenswrapper[4790]: I1124 14:47:17.365608 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"73610ce7-d271-4573-a14b-fd4563757182","Type":"ContainerStarted","Data":"0e4956ec5232fc766e423d2a4258b53aa1a3034c3b175854e6c0f6dcd408ae19"} Nov 24 14:47:18 crc kubenswrapper[4790]: I1124 14:47:18.378027 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"73610ce7-d271-4573-a14b-fd4563757182","Type":"ContainerStarted","Data":"8b5236025c15802a3d990b178b4446bcc2636aa67dbda147ff993cde8d23ee2d"} Nov 24 14:47:18 crc kubenswrapper[4790]: I1124 14:47:18.399250 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.399230117 podStartE2EDuration="3.399230117s" podCreationTimestamp="2025-11-24 14:47:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:47:18.395979994 +0000 UTC m=+5686.775873666" watchObservedRunningTime="2025-11-24 14:47:18.399230117 +0000 UTC m=+5686.779123769" Nov 24 14:47:19 crc kubenswrapper[4790]: I1124 14:47:19.442976 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 24 14:47:21 crc kubenswrapper[4790]: I1124 14:47:21.061744 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 24 14:47:26 crc kubenswrapper[4790]: I1124 14:47:26.334527 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 24 14:48:13 crc kubenswrapper[4790]: I1124 14:48:13.938642 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:48:13 crc kubenswrapper[4790]: I1124 14:48:13.939464 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:48:43 crc kubenswrapper[4790]: I1124 14:48:43.939222 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:48:43 crc kubenswrapper[4790]: I1124 14:48:43.940134 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:48:57 crc kubenswrapper[4790]: I1124 14:48:57.095458 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-grlct"] Nov 24 14:48:57 crc kubenswrapper[4790]: I1124 14:48:57.109755 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-grlct"] Nov 24 14:48:58 crc kubenswrapper[4790]: I1124 14:48:58.364234 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be69b02c-42ef-437f-8ff3-47973dc25ccb" path="/var/lib/kubelet/pods/be69b02c-42ef-437f-8ff3-47973dc25ccb/volumes" Nov 24 14:48:59 crc kubenswrapper[4790]: I1124 14:48:59.064562 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-8ec9-account-create-hzpw7"] Nov 24 14:48:59 crc kubenswrapper[4790]: I1124 14:48:59.075655 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-8ec9-account-create-hzpw7"] Nov 24 14:49:00 crc kubenswrapper[4790]: I1124 14:49:00.336093 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44" path="/var/lib/kubelet/pods/c939cf2c-81e7-4f9a-bb5b-08f04f6c1f44/volumes" Nov 24 14:49:05 crc kubenswrapper[4790]: I1124 14:49:05.034125 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-qmm47"] Nov 24 14:49:05 crc kubenswrapper[4790]: I1124 14:49:05.050690 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-qmm47"] Nov 24 14:49:06 crc kubenswrapper[4790]: I1124 14:49:06.328067 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42e1cf5d-d1e0-4a6c-8607-f37dd401f552" path="/var/lib/kubelet/pods/42e1cf5d-d1e0-4a6c-8607-f37dd401f552/volumes" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.045522 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-6p54v"] Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.047284 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6p54v" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.050510 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.050750 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-vxzbb" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.064485 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6p54v"] Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.100910 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-l46w2"] Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.106920 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-l46w2" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.141951 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-l46w2"] Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.230273 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ec4e9e28-345c-4cc4-acba-7a06cac52117-var-log-ovn\") pod \"ovn-controller-6p54v\" (UID: \"ec4e9e28-345c-4cc4-acba-7a06cac52117\") " pod="openstack/ovn-controller-6p54v" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.230358 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/d9800a51-bf87-4786-8cef-c8fcafdc72d9-etc-ovs\") pod \"ovn-controller-ovs-l46w2\" (UID: \"d9800a51-bf87-4786-8cef-c8fcafdc72d9\") " pod="openstack/ovn-controller-ovs-l46w2" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.230383 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vk64r\" (UniqueName: \"kubernetes.io/projected/ec4e9e28-345c-4cc4-acba-7a06cac52117-kube-api-access-vk64r\") pod \"ovn-controller-6p54v\" (UID: \"ec4e9e28-345c-4cc4-acba-7a06cac52117\") " pod="openstack/ovn-controller-6p54v" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.230401 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ec4e9e28-345c-4cc4-acba-7a06cac52117-var-run-ovn\") pod \"ovn-controller-6p54v\" (UID: \"ec4e9e28-345c-4cc4-acba-7a06cac52117\") " pod="openstack/ovn-controller-6p54v" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.230462 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/d9800a51-bf87-4786-8cef-c8fcafdc72d9-var-lib\") pod \"ovn-controller-ovs-l46w2\" (UID: \"d9800a51-bf87-4786-8cef-c8fcafdc72d9\") " pod="openstack/ovn-controller-ovs-l46w2" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.230500 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ec4e9e28-345c-4cc4-acba-7a06cac52117-scripts\") pod \"ovn-controller-6p54v\" (UID: \"ec4e9e28-345c-4cc4-acba-7a06cac52117\") " pod="openstack/ovn-controller-6p54v" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.230520 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tl2m8\" (UniqueName: \"kubernetes.io/projected/d9800a51-bf87-4786-8cef-c8fcafdc72d9-kube-api-access-tl2m8\") pod \"ovn-controller-ovs-l46w2\" (UID: \"d9800a51-bf87-4786-8cef-c8fcafdc72d9\") " pod="openstack/ovn-controller-ovs-l46w2" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.230541 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/d9800a51-bf87-4786-8cef-c8fcafdc72d9-var-log\") pod \"ovn-controller-ovs-l46w2\" (UID: \"d9800a51-bf87-4786-8cef-c8fcafdc72d9\") " pod="openstack/ovn-controller-ovs-l46w2" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.230620 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d9800a51-bf87-4786-8cef-c8fcafdc72d9-var-run\") pod \"ovn-controller-ovs-l46w2\" (UID: \"d9800a51-bf87-4786-8cef-c8fcafdc72d9\") " pod="openstack/ovn-controller-ovs-l46w2" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.230645 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d9800a51-bf87-4786-8cef-c8fcafdc72d9-scripts\") pod \"ovn-controller-ovs-l46w2\" (UID: \"d9800a51-bf87-4786-8cef-c8fcafdc72d9\") " pod="openstack/ovn-controller-ovs-l46w2" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.230676 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ec4e9e28-345c-4cc4-acba-7a06cac52117-var-run\") pod \"ovn-controller-6p54v\" (UID: \"ec4e9e28-345c-4cc4-acba-7a06cac52117\") " pod="openstack/ovn-controller-6p54v" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.332101 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ec4e9e28-345c-4cc4-acba-7a06cac52117-var-log-ovn\") pod \"ovn-controller-6p54v\" (UID: \"ec4e9e28-345c-4cc4-acba-7a06cac52117\") " pod="openstack/ovn-controller-6p54v" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.332241 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/d9800a51-bf87-4786-8cef-c8fcafdc72d9-etc-ovs\") pod \"ovn-controller-ovs-l46w2\" (UID: \"d9800a51-bf87-4786-8cef-c8fcafdc72d9\") " pod="openstack/ovn-controller-ovs-l46w2" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.332286 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vk64r\" (UniqueName: \"kubernetes.io/projected/ec4e9e28-345c-4cc4-acba-7a06cac52117-kube-api-access-vk64r\") pod \"ovn-controller-6p54v\" (UID: \"ec4e9e28-345c-4cc4-acba-7a06cac52117\") " pod="openstack/ovn-controller-6p54v" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.332319 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ec4e9e28-345c-4cc4-acba-7a06cac52117-var-run-ovn\") pod \"ovn-controller-6p54v\" (UID: \"ec4e9e28-345c-4cc4-acba-7a06cac52117\") " pod="openstack/ovn-controller-6p54v" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.332363 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/d9800a51-bf87-4786-8cef-c8fcafdc72d9-var-lib\") pod \"ovn-controller-ovs-l46w2\" (UID: \"d9800a51-bf87-4786-8cef-c8fcafdc72d9\") " pod="openstack/ovn-controller-ovs-l46w2" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.332392 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ec4e9e28-345c-4cc4-acba-7a06cac52117-scripts\") pod \"ovn-controller-6p54v\" (UID: \"ec4e9e28-345c-4cc4-acba-7a06cac52117\") " pod="openstack/ovn-controller-6p54v" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.332425 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tl2m8\" (UniqueName: \"kubernetes.io/projected/d9800a51-bf87-4786-8cef-c8fcafdc72d9-kube-api-access-tl2m8\") pod \"ovn-controller-ovs-l46w2\" (UID: \"d9800a51-bf87-4786-8cef-c8fcafdc72d9\") " pod="openstack/ovn-controller-ovs-l46w2" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.332460 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/d9800a51-bf87-4786-8cef-c8fcafdc72d9-var-log\") pod \"ovn-controller-ovs-l46w2\" (UID: \"d9800a51-bf87-4786-8cef-c8fcafdc72d9\") " pod="openstack/ovn-controller-ovs-l46w2" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.332576 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d9800a51-bf87-4786-8cef-c8fcafdc72d9-var-run\") pod \"ovn-controller-ovs-l46w2\" (UID: \"d9800a51-bf87-4786-8cef-c8fcafdc72d9\") " pod="openstack/ovn-controller-ovs-l46w2" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.332666 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d9800a51-bf87-4786-8cef-c8fcafdc72d9-scripts\") pod \"ovn-controller-ovs-l46w2\" (UID: \"d9800a51-bf87-4786-8cef-c8fcafdc72d9\") " pod="openstack/ovn-controller-ovs-l46w2" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.332697 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ec4e9e28-345c-4cc4-acba-7a06cac52117-var-run\") pod \"ovn-controller-6p54v\" (UID: \"ec4e9e28-345c-4cc4-acba-7a06cac52117\") " pod="openstack/ovn-controller-6p54v" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.333255 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ec4e9e28-345c-4cc4-acba-7a06cac52117-var-run\") pod \"ovn-controller-6p54v\" (UID: \"ec4e9e28-345c-4cc4-acba-7a06cac52117\") " pod="openstack/ovn-controller-6p54v" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.333382 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ec4e9e28-345c-4cc4-acba-7a06cac52117-var-log-ovn\") pod \"ovn-controller-6p54v\" (UID: \"ec4e9e28-345c-4cc4-acba-7a06cac52117\") " pod="openstack/ovn-controller-6p54v" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.333470 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/d9800a51-bf87-4786-8cef-c8fcafdc72d9-etc-ovs\") pod \"ovn-controller-ovs-l46w2\" (UID: \"d9800a51-bf87-4786-8cef-c8fcafdc72d9\") " pod="openstack/ovn-controller-ovs-l46w2" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.334007 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ec4e9e28-345c-4cc4-acba-7a06cac52117-var-run-ovn\") pod \"ovn-controller-6p54v\" (UID: \"ec4e9e28-345c-4cc4-acba-7a06cac52117\") " pod="openstack/ovn-controller-6p54v" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.334108 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/d9800a51-bf87-4786-8cef-c8fcafdc72d9-var-lib\") pod \"ovn-controller-ovs-l46w2\" (UID: \"d9800a51-bf87-4786-8cef-c8fcafdc72d9\") " pod="openstack/ovn-controller-ovs-l46w2" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.335126 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d9800a51-bf87-4786-8cef-c8fcafdc72d9-var-run\") pod \"ovn-controller-ovs-l46w2\" (UID: \"d9800a51-bf87-4786-8cef-c8fcafdc72d9\") " pod="openstack/ovn-controller-ovs-l46w2" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.335122 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/d9800a51-bf87-4786-8cef-c8fcafdc72d9-var-log\") pod \"ovn-controller-ovs-l46w2\" (UID: \"d9800a51-bf87-4786-8cef-c8fcafdc72d9\") " pod="openstack/ovn-controller-ovs-l46w2" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.337745 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d9800a51-bf87-4786-8cef-c8fcafdc72d9-scripts\") pod \"ovn-controller-ovs-l46w2\" (UID: \"d9800a51-bf87-4786-8cef-c8fcafdc72d9\") " pod="openstack/ovn-controller-ovs-l46w2" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.338026 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ec4e9e28-345c-4cc4-acba-7a06cac52117-scripts\") pod \"ovn-controller-6p54v\" (UID: \"ec4e9e28-345c-4cc4-acba-7a06cac52117\") " pod="openstack/ovn-controller-6p54v" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.354849 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vk64r\" (UniqueName: \"kubernetes.io/projected/ec4e9e28-345c-4cc4-acba-7a06cac52117-kube-api-access-vk64r\") pod \"ovn-controller-6p54v\" (UID: \"ec4e9e28-345c-4cc4-acba-7a06cac52117\") " pod="openstack/ovn-controller-6p54v" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.364526 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tl2m8\" (UniqueName: \"kubernetes.io/projected/d9800a51-bf87-4786-8cef-c8fcafdc72d9-kube-api-access-tl2m8\") pod \"ovn-controller-ovs-l46w2\" (UID: \"d9800a51-bf87-4786-8cef-c8fcafdc72d9\") " pod="openstack/ovn-controller-ovs-l46w2" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.380390 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6p54v" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.441297 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-l46w2" Nov 24 14:49:07 crc kubenswrapper[4790]: I1124 14:49:07.936187 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6p54v"] Nov 24 14:49:08 crc kubenswrapper[4790]: I1124 14:49:08.374754 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-l46w2"] Nov 24 14:49:08 crc kubenswrapper[4790]: I1124 14:49:08.574365 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-9x8wb"] Nov 24 14:49:08 crc kubenswrapper[4790]: I1124 14:49:08.576165 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-9x8wb" Nov 24 14:49:08 crc kubenswrapper[4790]: I1124 14:49:08.579070 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 24 14:49:08 crc kubenswrapper[4790]: I1124 14:49:08.584141 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-9x8wb"] Nov 24 14:49:08 crc kubenswrapper[4790]: I1124 14:49:08.681141 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6p54v" event={"ID":"ec4e9e28-345c-4cc4-acba-7a06cac52117","Type":"ContainerStarted","Data":"90591f1bfcbd41d17a8abb43bab344975bdf0d9e475c4e364c1859a896d9986a"} Nov 24 14:49:08 crc kubenswrapper[4790]: I1124 14:49:08.681188 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6p54v" event={"ID":"ec4e9e28-345c-4cc4-acba-7a06cac52117","Type":"ContainerStarted","Data":"2c1f59587959e40af44032518e3efdb0030624222ad692c34ca3537cf7508fe5"} Nov 24 14:49:08 crc kubenswrapper[4790]: I1124 14:49:08.681506 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-6p54v" Nov 24 14:49:08 crc kubenswrapper[4790]: I1124 14:49:08.683675 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-l46w2" event={"ID":"d9800a51-bf87-4786-8cef-c8fcafdc72d9","Type":"ContainerStarted","Data":"c966e69c7098821b7711e833c7a313147a0b78311498e39f3e5282913c41c8b0"} Nov 24 14:49:08 crc kubenswrapper[4790]: I1124 14:49:08.706946 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-6p54v" podStartSLOduration=1.706926747 podStartE2EDuration="1.706926747s" podCreationTimestamp="2025-11-24 14:49:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:49:08.701238483 +0000 UTC m=+5797.081132145" watchObservedRunningTime="2025-11-24 14:49:08.706926747 +0000 UTC m=+5797.086820409" Nov 24 14:49:08 crc kubenswrapper[4790]: I1124 14:49:08.767683 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f1ba98b-2cf8-471f-94c6-ce7a5699d421-config\") pod \"ovn-controller-metrics-9x8wb\" (UID: \"5f1ba98b-2cf8-471f-94c6-ce7a5699d421\") " pod="openstack/ovn-controller-metrics-9x8wb" Nov 24 14:49:08 crc kubenswrapper[4790]: I1124 14:49:08.767748 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/5f1ba98b-2cf8-471f-94c6-ce7a5699d421-ovn-rundir\") pod \"ovn-controller-metrics-9x8wb\" (UID: \"5f1ba98b-2cf8-471f-94c6-ce7a5699d421\") " pod="openstack/ovn-controller-metrics-9x8wb" Nov 24 14:49:08 crc kubenswrapper[4790]: I1124 14:49:08.767946 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5mmx\" (UniqueName: \"kubernetes.io/projected/5f1ba98b-2cf8-471f-94c6-ce7a5699d421-kube-api-access-k5mmx\") pod \"ovn-controller-metrics-9x8wb\" (UID: \"5f1ba98b-2cf8-471f-94c6-ce7a5699d421\") " pod="openstack/ovn-controller-metrics-9x8wb" Nov 24 14:49:08 crc kubenswrapper[4790]: I1124 14:49:08.768205 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/5f1ba98b-2cf8-471f-94c6-ce7a5699d421-ovs-rundir\") pod \"ovn-controller-metrics-9x8wb\" (UID: \"5f1ba98b-2cf8-471f-94c6-ce7a5699d421\") " pod="openstack/ovn-controller-metrics-9x8wb" Nov 24 14:49:08 crc kubenswrapper[4790]: I1124 14:49:08.869753 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/5f1ba98b-2cf8-471f-94c6-ce7a5699d421-ovs-rundir\") pod \"ovn-controller-metrics-9x8wb\" (UID: \"5f1ba98b-2cf8-471f-94c6-ce7a5699d421\") " pod="openstack/ovn-controller-metrics-9x8wb" Nov 24 14:49:08 crc kubenswrapper[4790]: I1124 14:49:08.869951 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f1ba98b-2cf8-471f-94c6-ce7a5699d421-config\") pod \"ovn-controller-metrics-9x8wb\" (UID: \"5f1ba98b-2cf8-471f-94c6-ce7a5699d421\") " pod="openstack/ovn-controller-metrics-9x8wb" Nov 24 14:49:08 crc kubenswrapper[4790]: I1124 14:49:08.869985 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/5f1ba98b-2cf8-471f-94c6-ce7a5699d421-ovn-rundir\") pod \"ovn-controller-metrics-9x8wb\" (UID: \"5f1ba98b-2cf8-471f-94c6-ce7a5699d421\") " pod="openstack/ovn-controller-metrics-9x8wb" Nov 24 14:49:08 crc kubenswrapper[4790]: I1124 14:49:08.870017 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5mmx\" (UniqueName: \"kubernetes.io/projected/5f1ba98b-2cf8-471f-94c6-ce7a5699d421-kube-api-access-k5mmx\") pod \"ovn-controller-metrics-9x8wb\" (UID: \"5f1ba98b-2cf8-471f-94c6-ce7a5699d421\") " pod="openstack/ovn-controller-metrics-9x8wb" Nov 24 14:49:08 crc kubenswrapper[4790]: I1124 14:49:08.870199 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/5f1ba98b-2cf8-471f-94c6-ce7a5699d421-ovs-rundir\") pod \"ovn-controller-metrics-9x8wb\" (UID: \"5f1ba98b-2cf8-471f-94c6-ce7a5699d421\") " pod="openstack/ovn-controller-metrics-9x8wb" Nov 24 14:49:08 crc kubenswrapper[4790]: I1124 14:49:08.870285 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/5f1ba98b-2cf8-471f-94c6-ce7a5699d421-ovn-rundir\") pod \"ovn-controller-metrics-9x8wb\" (UID: \"5f1ba98b-2cf8-471f-94c6-ce7a5699d421\") " pod="openstack/ovn-controller-metrics-9x8wb" Nov 24 14:49:08 crc kubenswrapper[4790]: I1124 14:49:08.870841 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f1ba98b-2cf8-471f-94c6-ce7a5699d421-config\") pod \"ovn-controller-metrics-9x8wb\" (UID: \"5f1ba98b-2cf8-471f-94c6-ce7a5699d421\") " pod="openstack/ovn-controller-metrics-9x8wb" Nov 24 14:49:08 crc kubenswrapper[4790]: I1124 14:49:08.896331 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5mmx\" (UniqueName: \"kubernetes.io/projected/5f1ba98b-2cf8-471f-94c6-ce7a5699d421-kube-api-access-k5mmx\") pod \"ovn-controller-metrics-9x8wb\" (UID: \"5f1ba98b-2cf8-471f-94c6-ce7a5699d421\") " pod="openstack/ovn-controller-metrics-9x8wb" Nov 24 14:49:08 crc kubenswrapper[4790]: I1124 14:49:08.896819 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-9x8wb" Nov 24 14:49:09 crc kubenswrapper[4790]: I1124 14:49:09.345408 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-create-bzlld"] Nov 24 14:49:09 crc kubenswrapper[4790]: I1124 14:49:09.347683 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-bzlld" Nov 24 14:49:09 crc kubenswrapper[4790]: I1124 14:49:09.356129 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-bzlld"] Nov 24 14:49:09 crc kubenswrapper[4790]: I1124 14:49:09.410076 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-9x8wb"] Nov 24 14:49:09 crc kubenswrapper[4790]: W1124 14:49:09.426358 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f1ba98b_2cf8_471f_94c6_ce7a5699d421.slice/crio-f56dc2932ef954f1fbf2cd87e1b2a2f6e938a73c5b8355f20c99aae7283dc2cd WatchSource:0}: Error finding container f56dc2932ef954f1fbf2cd87e1b2a2f6e938a73c5b8355f20c99aae7283dc2cd: Status 404 returned error can't find the container with id f56dc2932ef954f1fbf2cd87e1b2a2f6e938a73c5b8355f20c99aae7283dc2cd Nov 24 14:49:09 crc kubenswrapper[4790]: I1124 14:49:09.484395 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5t6k\" (UniqueName: \"kubernetes.io/projected/9038e28a-efcc-48e9-97ac-5c1b6417fb0f-kube-api-access-l5t6k\") pod \"octavia-db-create-bzlld\" (UID: \"9038e28a-efcc-48e9-97ac-5c1b6417fb0f\") " pod="openstack/octavia-db-create-bzlld" Nov 24 14:49:09 crc kubenswrapper[4790]: I1124 14:49:09.484733 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9038e28a-efcc-48e9-97ac-5c1b6417fb0f-operator-scripts\") pod \"octavia-db-create-bzlld\" (UID: \"9038e28a-efcc-48e9-97ac-5c1b6417fb0f\") " pod="openstack/octavia-db-create-bzlld" Nov 24 14:49:09 crc kubenswrapper[4790]: I1124 14:49:09.586631 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9038e28a-efcc-48e9-97ac-5c1b6417fb0f-operator-scripts\") pod \"octavia-db-create-bzlld\" (UID: \"9038e28a-efcc-48e9-97ac-5c1b6417fb0f\") " pod="openstack/octavia-db-create-bzlld" Nov 24 14:49:09 crc kubenswrapper[4790]: I1124 14:49:09.586760 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5t6k\" (UniqueName: \"kubernetes.io/projected/9038e28a-efcc-48e9-97ac-5c1b6417fb0f-kube-api-access-l5t6k\") pod \"octavia-db-create-bzlld\" (UID: \"9038e28a-efcc-48e9-97ac-5c1b6417fb0f\") " pod="openstack/octavia-db-create-bzlld" Nov 24 14:49:09 crc kubenswrapper[4790]: I1124 14:49:09.587789 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9038e28a-efcc-48e9-97ac-5c1b6417fb0f-operator-scripts\") pod \"octavia-db-create-bzlld\" (UID: \"9038e28a-efcc-48e9-97ac-5c1b6417fb0f\") " pod="openstack/octavia-db-create-bzlld" Nov 24 14:49:09 crc kubenswrapper[4790]: I1124 14:49:09.605218 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5t6k\" (UniqueName: \"kubernetes.io/projected/9038e28a-efcc-48e9-97ac-5c1b6417fb0f-kube-api-access-l5t6k\") pod \"octavia-db-create-bzlld\" (UID: \"9038e28a-efcc-48e9-97ac-5c1b6417fb0f\") " pod="openstack/octavia-db-create-bzlld" Nov 24 14:49:09 crc kubenswrapper[4790]: I1124 14:49:09.698825 4790 generic.go:334] "Generic (PLEG): container finished" podID="d9800a51-bf87-4786-8cef-c8fcafdc72d9" containerID="b9d97efcde441c3d0e9b3ea25493f88fdd771945105103ebf8065ece3fe27848" exitCode=0 Nov 24 14:49:09 crc kubenswrapper[4790]: I1124 14:49:09.698914 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-l46w2" event={"ID":"d9800a51-bf87-4786-8cef-c8fcafdc72d9","Type":"ContainerDied","Data":"b9d97efcde441c3d0e9b3ea25493f88fdd771945105103ebf8065ece3fe27848"} Nov 24 14:49:09 crc kubenswrapper[4790]: I1124 14:49:09.703350 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-9x8wb" event={"ID":"5f1ba98b-2cf8-471f-94c6-ce7a5699d421","Type":"ContainerStarted","Data":"cbaf60858eb27b40d4168428e87c6bfe16f6829715f64232d1257c9079b3617e"} Nov 24 14:49:09 crc kubenswrapper[4790]: I1124 14:49:09.703406 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-9x8wb" event={"ID":"5f1ba98b-2cf8-471f-94c6-ce7a5699d421","Type":"ContainerStarted","Data":"f56dc2932ef954f1fbf2cd87e1b2a2f6e938a73c5b8355f20c99aae7283dc2cd"} Nov 24 14:49:09 crc kubenswrapper[4790]: I1124 14:49:09.704661 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-bzlld" Nov 24 14:49:09 crc kubenswrapper[4790]: I1124 14:49:09.744937 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-9x8wb" podStartSLOduration=1.7449129349999999 podStartE2EDuration="1.744912935s" podCreationTimestamp="2025-11-24 14:49:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:49:09.740268572 +0000 UTC m=+5798.120162234" watchObservedRunningTime="2025-11-24 14:49:09.744912935 +0000 UTC m=+5798.124806597" Nov 24 14:49:10 crc kubenswrapper[4790]: I1124 14:49:10.270312 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-bzlld"] Nov 24 14:49:10 crc kubenswrapper[4790]: I1124 14:49:10.719030 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-l46w2" event={"ID":"d9800a51-bf87-4786-8cef-c8fcafdc72d9","Type":"ContainerStarted","Data":"10b1146e5964c9d222164e82c0de219b31174afc42ee0b924beaa0e0d450d666"} Nov 24 14:49:10 crc kubenswrapper[4790]: I1124 14:49:10.719376 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-l46w2" Nov 24 14:49:10 crc kubenswrapper[4790]: I1124 14:49:10.719393 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-l46w2" event={"ID":"d9800a51-bf87-4786-8cef-c8fcafdc72d9","Type":"ContainerStarted","Data":"6c5eca71b662618903a2e0d63e23078360122d2b93e8ffe2621b2bca9a079fa9"} Nov 24 14:49:10 crc kubenswrapper[4790]: I1124 14:49:10.720948 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-bzlld" event={"ID":"9038e28a-efcc-48e9-97ac-5c1b6417fb0f","Type":"ContainerStarted","Data":"88ef69ff632a65bbcec6feb14ea1545eba4f781f2083a2d3b071757ad3e80949"} Nov 24 14:49:10 crc kubenswrapper[4790]: I1124 14:49:10.721023 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-bzlld" event={"ID":"9038e28a-efcc-48e9-97ac-5c1b6417fb0f","Type":"ContainerStarted","Data":"9d45e991c7106222cafbbf247a2230ebd3913793a1891dbcff4c893998bea5e1"} Nov 24 14:49:10 crc kubenswrapper[4790]: I1124 14:49:10.754750 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-l46w2" podStartSLOduration=3.754720824 podStartE2EDuration="3.754720824s" podCreationTimestamp="2025-11-24 14:49:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:49:10.741238057 +0000 UTC m=+5799.121131739" watchObservedRunningTime="2025-11-24 14:49:10.754720824 +0000 UTC m=+5799.134614496" Nov 24 14:49:10 crc kubenswrapper[4790]: I1124 14:49:10.766841 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-db-create-bzlld" podStartSLOduration=1.766812402 podStartE2EDuration="1.766812402s" podCreationTimestamp="2025-11-24 14:49:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:49:10.759607645 +0000 UTC m=+5799.139501317" watchObservedRunningTime="2025-11-24 14:49:10.766812402 +0000 UTC m=+5799.146706064" Nov 24 14:49:11 crc kubenswrapper[4790]: I1124 14:49:11.045852 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-bccc-account-create-j8zg9"] Nov 24 14:49:11 crc kubenswrapper[4790]: I1124 14:49:11.047060 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-bccc-account-create-j8zg9" Nov 24 14:49:11 crc kubenswrapper[4790]: I1124 14:49:11.049812 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-db-secret" Nov 24 14:49:11 crc kubenswrapper[4790]: I1124 14:49:11.070523 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-bccc-account-create-j8zg9"] Nov 24 14:49:11 crc kubenswrapper[4790]: I1124 14:49:11.220101 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85093f1b-8746-41e3-b112-58dde9f6d54f-operator-scripts\") pod \"octavia-bccc-account-create-j8zg9\" (UID: \"85093f1b-8746-41e3-b112-58dde9f6d54f\") " pod="openstack/octavia-bccc-account-create-j8zg9" Nov 24 14:49:11 crc kubenswrapper[4790]: I1124 14:49:11.220164 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xb9kx\" (UniqueName: \"kubernetes.io/projected/85093f1b-8746-41e3-b112-58dde9f6d54f-kube-api-access-xb9kx\") pod \"octavia-bccc-account-create-j8zg9\" (UID: \"85093f1b-8746-41e3-b112-58dde9f6d54f\") " pod="openstack/octavia-bccc-account-create-j8zg9" Nov 24 14:49:11 crc kubenswrapper[4790]: I1124 14:49:11.321698 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85093f1b-8746-41e3-b112-58dde9f6d54f-operator-scripts\") pod \"octavia-bccc-account-create-j8zg9\" (UID: \"85093f1b-8746-41e3-b112-58dde9f6d54f\") " pod="openstack/octavia-bccc-account-create-j8zg9" Nov 24 14:49:11 crc kubenswrapper[4790]: I1124 14:49:11.322095 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xb9kx\" (UniqueName: \"kubernetes.io/projected/85093f1b-8746-41e3-b112-58dde9f6d54f-kube-api-access-xb9kx\") pod \"octavia-bccc-account-create-j8zg9\" (UID: \"85093f1b-8746-41e3-b112-58dde9f6d54f\") " pod="openstack/octavia-bccc-account-create-j8zg9" Nov 24 14:49:11 crc kubenswrapper[4790]: I1124 14:49:11.323528 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85093f1b-8746-41e3-b112-58dde9f6d54f-operator-scripts\") pod \"octavia-bccc-account-create-j8zg9\" (UID: \"85093f1b-8746-41e3-b112-58dde9f6d54f\") " pod="openstack/octavia-bccc-account-create-j8zg9" Nov 24 14:49:11 crc kubenswrapper[4790]: I1124 14:49:11.351047 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xb9kx\" (UniqueName: \"kubernetes.io/projected/85093f1b-8746-41e3-b112-58dde9f6d54f-kube-api-access-xb9kx\") pod \"octavia-bccc-account-create-j8zg9\" (UID: \"85093f1b-8746-41e3-b112-58dde9f6d54f\") " pod="openstack/octavia-bccc-account-create-j8zg9" Nov 24 14:49:11 crc kubenswrapper[4790]: I1124 14:49:11.411745 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-bccc-account-create-j8zg9" Nov 24 14:49:11 crc kubenswrapper[4790]: I1124 14:49:11.733828 4790 generic.go:334] "Generic (PLEG): container finished" podID="9038e28a-efcc-48e9-97ac-5c1b6417fb0f" containerID="88ef69ff632a65bbcec6feb14ea1545eba4f781f2083a2d3b071757ad3e80949" exitCode=0 Nov 24 14:49:11 crc kubenswrapper[4790]: I1124 14:49:11.733943 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-bzlld" event={"ID":"9038e28a-efcc-48e9-97ac-5c1b6417fb0f","Type":"ContainerDied","Data":"88ef69ff632a65bbcec6feb14ea1545eba4f781f2083a2d3b071757ad3e80949"} Nov 24 14:49:11 crc kubenswrapper[4790]: I1124 14:49:11.734379 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-l46w2" Nov 24 14:49:11 crc kubenswrapper[4790]: I1124 14:49:11.898017 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-bccc-account-create-j8zg9"] Nov 24 14:49:12 crc kubenswrapper[4790]: I1124 14:49:12.746816 4790 generic.go:334] "Generic (PLEG): container finished" podID="85093f1b-8746-41e3-b112-58dde9f6d54f" containerID="6ed027e62487b0ee9f3027bdeffb9da69254fa8f826f89c13563c3527f0a6634" exitCode=0 Nov 24 14:49:12 crc kubenswrapper[4790]: I1124 14:49:12.746944 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-bccc-account-create-j8zg9" event={"ID":"85093f1b-8746-41e3-b112-58dde9f6d54f","Type":"ContainerDied","Data":"6ed027e62487b0ee9f3027bdeffb9da69254fa8f826f89c13563c3527f0a6634"} Nov 24 14:49:12 crc kubenswrapper[4790]: I1124 14:49:12.747278 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-bccc-account-create-j8zg9" event={"ID":"85093f1b-8746-41e3-b112-58dde9f6d54f","Type":"ContainerStarted","Data":"44b3f133964c72541eb93dbc0ff9142dac9bbf0c391798fc42ffcdfd14c03817"} Nov 24 14:49:13 crc kubenswrapper[4790]: I1124 14:49:13.144371 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-bzlld" Nov 24 14:49:13 crc kubenswrapper[4790]: I1124 14:49:13.265237 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5t6k\" (UniqueName: \"kubernetes.io/projected/9038e28a-efcc-48e9-97ac-5c1b6417fb0f-kube-api-access-l5t6k\") pod \"9038e28a-efcc-48e9-97ac-5c1b6417fb0f\" (UID: \"9038e28a-efcc-48e9-97ac-5c1b6417fb0f\") " Nov 24 14:49:13 crc kubenswrapper[4790]: I1124 14:49:13.265513 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9038e28a-efcc-48e9-97ac-5c1b6417fb0f-operator-scripts\") pod \"9038e28a-efcc-48e9-97ac-5c1b6417fb0f\" (UID: \"9038e28a-efcc-48e9-97ac-5c1b6417fb0f\") " Nov 24 14:49:13 crc kubenswrapper[4790]: I1124 14:49:13.266389 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9038e28a-efcc-48e9-97ac-5c1b6417fb0f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9038e28a-efcc-48e9-97ac-5c1b6417fb0f" (UID: "9038e28a-efcc-48e9-97ac-5c1b6417fb0f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:49:13 crc kubenswrapper[4790]: I1124 14:49:13.272829 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9038e28a-efcc-48e9-97ac-5c1b6417fb0f-kube-api-access-l5t6k" (OuterVolumeSpecName: "kube-api-access-l5t6k") pod "9038e28a-efcc-48e9-97ac-5c1b6417fb0f" (UID: "9038e28a-efcc-48e9-97ac-5c1b6417fb0f"). InnerVolumeSpecName "kube-api-access-l5t6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:49:13 crc kubenswrapper[4790]: I1124 14:49:13.368271 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9038e28a-efcc-48e9-97ac-5c1b6417fb0f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:49:13 crc kubenswrapper[4790]: I1124 14:49:13.368315 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5t6k\" (UniqueName: \"kubernetes.io/projected/9038e28a-efcc-48e9-97ac-5c1b6417fb0f-kube-api-access-l5t6k\") on node \"crc\" DevicePath \"\"" Nov 24 14:49:13 crc kubenswrapper[4790]: I1124 14:49:13.763828 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-bzlld" event={"ID":"9038e28a-efcc-48e9-97ac-5c1b6417fb0f","Type":"ContainerDied","Data":"9d45e991c7106222cafbbf247a2230ebd3913793a1891dbcff4c893998bea5e1"} Nov 24 14:49:13 crc kubenswrapper[4790]: I1124 14:49:13.763928 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-bzlld" Nov 24 14:49:13 crc kubenswrapper[4790]: I1124 14:49:13.763972 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d45e991c7106222cafbbf247a2230ebd3913793a1891dbcff4c893998bea5e1" Nov 24 14:49:13 crc kubenswrapper[4790]: I1124 14:49:13.939207 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:49:13 crc kubenswrapper[4790]: I1124 14:49:13.939651 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:49:13 crc kubenswrapper[4790]: I1124 14:49:13.939711 4790 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 14:49:13 crc kubenswrapper[4790]: I1124 14:49:13.942265 4790 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cd74b939f536300843ca659100b982ff6a09937cde93ea7f086d7cd5d1e85fc5"} pod="openshift-machine-config-operator/machine-config-daemon-xz49t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:49:13 crc kubenswrapper[4790]: I1124 14:49:13.942369 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" containerID="cri-o://cd74b939f536300843ca659100b982ff6a09937cde93ea7f086d7cd5d1e85fc5" gracePeriod=600 Nov 24 14:49:14 crc kubenswrapper[4790]: I1124 14:49:14.220930 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-bccc-account-create-j8zg9" Nov 24 14:49:14 crc kubenswrapper[4790]: I1124 14:49:14.401714 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85093f1b-8746-41e3-b112-58dde9f6d54f-operator-scripts\") pod \"85093f1b-8746-41e3-b112-58dde9f6d54f\" (UID: \"85093f1b-8746-41e3-b112-58dde9f6d54f\") " Nov 24 14:49:14 crc kubenswrapper[4790]: I1124 14:49:14.402266 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xb9kx\" (UniqueName: \"kubernetes.io/projected/85093f1b-8746-41e3-b112-58dde9f6d54f-kube-api-access-xb9kx\") pod \"85093f1b-8746-41e3-b112-58dde9f6d54f\" (UID: \"85093f1b-8746-41e3-b112-58dde9f6d54f\") " Nov 24 14:49:14 crc kubenswrapper[4790]: I1124 14:49:14.403021 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85093f1b-8746-41e3-b112-58dde9f6d54f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "85093f1b-8746-41e3-b112-58dde9f6d54f" (UID: "85093f1b-8746-41e3-b112-58dde9f6d54f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:49:14 crc kubenswrapper[4790]: I1124 14:49:14.403861 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85093f1b-8746-41e3-b112-58dde9f6d54f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:49:14 crc kubenswrapper[4790]: I1124 14:49:14.409577 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85093f1b-8746-41e3-b112-58dde9f6d54f-kube-api-access-xb9kx" (OuterVolumeSpecName: "kube-api-access-xb9kx") pod "85093f1b-8746-41e3-b112-58dde9f6d54f" (UID: "85093f1b-8746-41e3-b112-58dde9f6d54f"). InnerVolumeSpecName "kube-api-access-xb9kx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:49:14 crc kubenswrapper[4790]: I1124 14:49:14.507133 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xb9kx\" (UniqueName: \"kubernetes.io/projected/85093f1b-8746-41e3-b112-58dde9f6d54f-kube-api-access-xb9kx\") on node \"crc\" DevicePath \"\"" Nov 24 14:49:14 crc kubenswrapper[4790]: I1124 14:49:14.773564 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-bccc-account-create-j8zg9" Nov 24 14:49:14 crc kubenswrapper[4790]: I1124 14:49:14.773556 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-bccc-account-create-j8zg9" event={"ID":"85093f1b-8746-41e3-b112-58dde9f6d54f","Type":"ContainerDied","Data":"44b3f133964c72541eb93dbc0ff9142dac9bbf0c391798fc42ffcdfd14c03817"} Nov 24 14:49:14 crc kubenswrapper[4790]: I1124 14:49:14.774051 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44b3f133964c72541eb93dbc0ff9142dac9bbf0c391798fc42ffcdfd14c03817" Nov 24 14:49:14 crc kubenswrapper[4790]: I1124 14:49:14.777103 4790 generic.go:334] "Generic (PLEG): container finished" podID="0d73b133-48f0-455f-8f6a-742e633f631a" containerID="cd74b939f536300843ca659100b982ff6a09937cde93ea7f086d7cd5d1e85fc5" exitCode=0 Nov 24 14:49:14 crc kubenswrapper[4790]: I1124 14:49:14.777146 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerDied","Data":"cd74b939f536300843ca659100b982ff6a09937cde93ea7f086d7cd5d1e85fc5"} Nov 24 14:49:14 crc kubenswrapper[4790]: I1124 14:49:14.777174 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"28308703817a1d4e0e5344ce19268127c600894f2daa6b921d52a9105f014132"} Nov 24 14:49:14 crc kubenswrapper[4790]: I1124 14:49:14.777194 4790 scope.go:117] "RemoveContainer" containerID="92c086e44ccb28a62777d612e499209f7483422ca70b460ce6ec73c416e8e8b8" Nov 24 14:49:17 crc kubenswrapper[4790]: I1124 14:49:17.392388 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-persistence-db-create-sdmzn"] Nov 24 14:49:17 crc kubenswrapper[4790]: E1124 14:49:17.393735 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9038e28a-efcc-48e9-97ac-5c1b6417fb0f" containerName="mariadb-database-create" Nov 24 14:49:17 crc kubenswrapper[4790]: I1124 14:49:17.393758 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="9038e28a-efcc-48e9-97ac-5c1b6417fb0f" containerName="mariadb-database-create" Nov 24 14:49:17 crc kubenswrapper[4790]: E1124 14:49:17.393813 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85093f1b-8746-41e3-b112-58dde9f6d54f" containerName="mariadb-account-create" Nov 24 14:49:17 crc kubenswrapper[4790]: I1124 14:49:17.393825 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="85093f1b-8746-41e3-b112-58dde9f6d54f" containerName="mariadb-account-create" Nov 24 14:49:17 crc kubenswrapper[4790]: I1124 14:49:17.394134 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="85093f1b-8746-41e3-b112-58dde9f6d54f" containerName="mariadb-account-create" Nov 24 14:49:17 crc kubenswrapper[4790]: I1124 14:49:17.394170 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="9038e28a-efcc-48e9-97ac-5c1b6417fb0f" containerName="mariadb-database-create" Nov 24 14:49:17 crc kubenswrapper[4790]: I1124 14:49:17.395122 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-sdmzn" Nov 24 14:49:17 crc kubenswrapper[4790]: I1124 14:49:17.405464 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-sdmzn"] Nov 24 14:49:17 crc kubenswrapper[4790]: I1124 14:49:17.565041 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/57b1601f-40f6-49cd-ab92-85efc8fb18e1-operator-scripts\") pod \"octavia-persistence-db-create-sdmzn\" (UID: \"57b1601f-40f6-49cd-ab92-85efc8fb18e1\") " pod="openstack/octavia-persistence-db-create-sdmzn" Nov 24 14:49:17 crc kubenswrapper[4790]: I1124 14:49:17.565181 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kclc7\" (UniqueName: \"kubernetes.io/projected/57b1601f-40f6-49cd-ab92-85efc8fb18e1-kube-api-access-kclc7\") pod \"octavia-persistence-db-create-sdmzn\" (UID: \"57b1601f-40f6-49cd-ab92-85efc8fb18e1\") " pod="openstack/octavia-persistence-db-create-sdmzn" Nov 24 14:49:17 crc kubenswrapper[4790]: I1124 14:49:17.666969 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/57b1601f-40f6-49cd-ab92-85efc8fb18e1-operator-scripts\") pod \"octavia-persistence-db-create-sdmzn\" (UID: \"57b1601f-40f6-49cd-ab92-85efc8fb18e1\") " pod="openstack/octavia-persistence-db-create-sdmzn" Nov 24 14:49:17 crc kubenswrapper[4790]: I1124 14:49:17.667103 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kclc7\" (UniqueName: \"kubernetes.io/projected/57b1601f-40f6-49cd-ab92-85efc8fb18e1-kube-api-access-kclc7\") pod \"octavia-persistence-db-create-sdmzn\" (UID: \"57b1601f-40f6-49cd-ab92-85efc8fb18e1\") " pod="openstack/octavia-persistence-db-create-sdmzn" Nov 24 14:49:17 crc kubenswrapper[4790]: I1124 14:49:17.667873 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/57b1601f-40f6-49cd-ab92-85efc8fb18e1-operator-scripts\") pod \"octavia-persistence-db-create-sdmzn\" (UID: \"57b1601f-40f6-49cd-ab92-85efc8fb18e1\") " pod="openstack/octavia-persistence-db-create-sdmzn" Nov 24 14:49:17 crc kubenswrapper[4790]: I1124 14:49:17.685368 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kclc7\" (UniqueName: \"kubernetes.io/projected/57b1601f-40f6-49cd-ab92-85efc8fb18e1-kube-api-access-kclc7\") pod \"octavia-persistence-db-create-sdmzn\" (UID: \"57b1601f-40f6-49cd-ab92-85efc8fb18e1\") " pod="openstack/octavia-persistence-db-create-sdmzn" Nov 24 14:49:17 crc kubenswrapper[4790]: I1124 14:49:17.723044 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-sdmzn" Nov 24 14:49:18 crc kubenswrapper[4790]: I1124 14:49:18.216449 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-sdmzn"] Nov 24 14:49:18 crc kubenswrapper[4790]: I1124 14:49:18.403830 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-1102-account-create-hrp2v"] Nov 24 14:49:18 crc kubenswrapper[4790]: I1124 14:49:18.407365 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-1102-account-create-hrp2v" Nov 24 14:49:18 crc kubenswrapper[4790]: I1124 14:49:18.410527 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-persistence-db-secret" Nov 24 14:49:18 crc kubenswrapper[4790]: I1124 14:49:18.423318 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-1102-account-create-hrp2v"] Nov 24 14:49:18 crc kubenswrapper[4790]: I1124 14:49:18.481233 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9f2rd\" (UniqueName: \"kubernetes.io/projected/a1e0c3ca-accf-4636-82c8-bb73ad8304e4-kube-api-access-9f2rd\") pod \"octavia-1102-account-create-hrp2v\" (UID: \"a1e0c3ca-accf-4636-82c8-bb73ad8304e4\") " pod="openstack/octavia-1102-account-create-hrp2v" Nov 24 14:49:18 crc kubenswrapper[4790]: I1124 14:49:18.481306 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1e0c3ca-accf-4636-82c8-bb73ad8304e4-operator-scripts\") pod \"octavia-1102-account-create-hrp2v\" (UID: \"a1e0c3ca-accf-4636-82c8-bb73ad8304e4\") " pod="openstack/octavia-1102-account-create-hrp2v" Nov 24 14:49:18 crc kubenswrapper[4790]: I1124 14:49:18.583401 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9f2rd\" (UniqueName: \"kubernetes.io/projected/a1e0c3ca-accf-4636-82c8-bb73ad8304e4-kube-api-access-9f2rd\") pod \"octavia-1102-account-create-hrp2v\" (UID: \"a1e0c3ca-accf-4636-82c8-bb73ad8304e4\") " pod="openstack/octavia-1102-account-create-hrp2v" Nov 24 14:49:18 crc kubenswrapper[4790]: I1124 14:49:18.583489 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1e0c3ca-accf-4636-82c8-bb73ad8304e4-operator-scripts\") pod \"octavia-1102-account-create-hrp2v\" (UID: \"a1e0c3ca-accf-4636-82c8-bb73ad8304e4\") " pod="openstack/octavia-1102-account-create-hrp2v" Nov 24 14:49:18 crc kubenswrapper[4790]: I1124 14:49:18.584341 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1e0c3ca-accf-4636-82c8-bb73ad8304e4-operator-scripts\") pod \"octavia-1102-account-create-hrp2v\" (UID: \"a1e0c3ca-accf-4636-82c8-bb73ad8304e4\") " pod="openstack/octavia-1102-account-create-hrp2v" Nov 24 14:49:18 crc kubenswrapper[4790]: I1124 14:49:18.608705 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9f2rd\" (UniqueName: \"kubernetes.io/projected/a1e0c3ca-accf-4636-82c8-bb73ad8304e4-kube-api-access-9f2rd\") pod \"octavia-1102-account-create-hrp2v\" (UID: \"a1e0c3ca-accf-4636-82c8-bb73ad8304e4\") " pod="openstack/octavia-1102-account-create-hrp2v" Nov 24 14:49:18 crc kubenswrapper[4790]: I1124 14:49:18.735394 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-1102-account-create-hrp2v" Nov 24 14:49:18 crc kubenswrapper[4790]: I1124 14:49:18.868859 4790 generic.go:334] "Generic (PLEG): container finished" podID="57b1601f-40f6-49cd-ab92-85efc8fb18e1" containerID="a49ae406ba85a7abbfc33e2a4270e26ad1729908b9d461a811cfe8f4dbf153be" exitCode=0 Nov 24 14:49:18 crc kubenswrapper[4790]: I1124 14:49:18.869231 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-sdmzn" event={"ID":"57b1601f-40f6-49cd-ab92-85efc8fb18e1","Type":"ContainerDied","Data":"a49ae406ba85a7abbfc33e2a4270e26ad1729908b9d461a811cfe8f4dbf153be"} Nov 24 14:49:18 crc kubenswrapper[4790]: I1124 14:49:18.869327 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-sdmzn" event={"ID":"57b1601f-40f6-49cd-ab92-85efc8fb18e1","Type":"ContainerStarted","Data":"c836813d9b59c42abec1cbae59f26d079c8e9176288c3ce08a0efd747f1b2c85"} Nov 24 14:49:19 crc kubenswrapper[4790]: I1124 14:49:19.040037 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-c7bhw"] Nov 24 14:49:19 crc kubenswrapper[4790]: I1124 14:49:19.047637 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-c7bhw"] Nov 24 14:49:19 crc kubenswrapper[4790]: I1124 14:49:19.196967 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-1102-account-create-hrp2v"] Nov 24 14:49:19 crc kubenswrapper[4790]: I1124 14:49:19.882096 4790 generic.go:334] "Generic (PLEG): container finished" podID="a1e0c3ca-accf-4636-82c8-bb73ad8304e4" containerID="0a4bb43787f72515fb17ba0b5afce80e2325d47ac13d50a956015c340f3ea575" exitCode=0 Nov 24 14:49:19 crc kubenswrapper[4790]: I1124 14:49:19.882147 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-1102-account-create-hrp2v" event={"ID":"a1e0c3ca-accf-4636-82c8-bb73ad8304e4","Type":"ContainerDied","Data":"0a4bb43787f72515fb17ba0b5afce80e2325d47ac13d50a956015c340f3ea575"} Nov 24 14:49:19 crc kubenswrapper[4790]: I1124 14:49:19.882569 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-1102-account-create-hrp2v" event={"ID":"a1e0c3ca-accf-4636-82c8-bb73ad8304e4","Type":"ContainerStarted","Data":"a030b8d6f6b5ebf913c19347748c0e4447c1165c1cbe75d9ba1bdb92b95dc3b8"} Nov 24 14:49:20 crc kubenswrapper[4790]: I1124 14:49:20.283849 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-sdmzn" Nov 24 14:49:20 crc kubenswrapper[4790]: I1124 14:49:20.334262 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79" path="/var/lib/kubelet/pods/c4d7cf2c-f732-4a3f-86cc-79a4c66a5c79/volumes" Nov 24 14:49:20 crc kubenswrapper[4790]: I1124 14:49:20.425705 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/57b1601f-40f6-49cd-ab92-85efc8fb18e1-operator-scripts\") pod \"57b1601f-40f6-49cd-ab92-85efc8fb18e1\" (UID: \"57b1601f-40f6-49cd-ab92-85efc8fb18e1\") " Nov 24 14:49:20 crc kubenswrapper[4790]: I1124 14:49:20.426042 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kclc7\" (UniqueName: \"kubernetes.io/projected/57b1601f-40f6-49cd-ab92-85efc8fb18e1-kube-api-access-kclc7\") pod \"57b1601f-40f6-49cd-ab92-85efc8fb18e1\" (UID: \"57b1601f-40f6-49cd-ab92-85efc8fb18e1\") " Nov 24 14:49:20 crc kubenswrapper[4790]: I1124 14:49:20.426643 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57b1601f-40f6-49cd-ab92-85efc8fb18e1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "57b1601f-40f6-49cd-ab92-85efc8fb18e1" (UID: "57b1601f-40f6-49cd-ab92-85efc8fb18e1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:49:20 crc kubenswrapper[4790]: I1124 14:49:20.427280 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/57b1601f-40f6-49cd-ab92-85efc8fb18e1-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:49:20 crc kubenswrapper[4790]: I1124 14:49:20.439151 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57b1601f-40f6-49cd-ab92-85efc8fb18e1-kube-api-access-kclc7" (OuterVolumeSpecName: "kube-api-access-kclc7") pod "57b1601f-40f6-49cd-ab92-85efc8fb18e1" (UID: "57b1601f-40f6-49cd-ab92-85efc8fb18e1"). InnerVolumeSpecName "kube-api-access-kclc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:49:20 crc kubenswrapper[4790]: I1124 14:49:20.529536 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kclc7\" (UniqueName: \"kubernetes.io/projected/57b1601f-40f6-49cd-ab92-85efc8fb18e1-kube-api-access-kclc7\") on node \"crc\" DevicePath \"\"" Nov 24 14:49:20 crc kubenswrapper[4790]: I1124 14:49:20.895782 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-sdmzn" event={"ID":"57b1601f-40f6-49cd-ab92-85efc8fb18e1","Type":"ContainerDied","Data":"c836813d9b59c42abec1cbae59f26d079c8e9176288c3ce08a0efd747f1b2c85"} Nov 24 14:49:20 crc kubenswrapper[4790]: I1124 14:49:20.895916 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c836813d9b59c42abec1cbae59f26d079c8e9176288c3ce08a0efd747f1b2c85" Nov 24 14:49:20 crc kubenswrapper[4790]: I1124 14:49:20.896032 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-sdmzn" Nov 24 14:49:21 crc kubenswrapper[4790]: I1124 14:49:21.270553 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-1102-account-create-hrp2v" Nov 24 14:49:21 crc kubenswrapper[4790]: I1124 14:49:21.346719 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1e0c3ca-accf-4636-82c8-bb73ad8304e4-operator-scripts\") pod \"a1e0c3ca-accf-4636-82c8-bb73ad8304e4\" (UID: \"a1e0c3ca-accf-4636-82c8-bb73ad8304e4\") " Nov 24 14:49:21 crc kubenswrapper[4790]: I1124 14:49:21.346969 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9f2rd\" (UniqueName: \"kubernetes.io/projected/a1e0c3ca-accf-4636-82c8-bb73ad8304e4-kube-api-access-9f2rd\") pod \"a1e0c3ca-accf-4636-82c8-bb73ad8304e4\" (UID: \"a1e0c3ca-accf-4636-82c8-bb73ad8304e4\") " Nov 24 14:49:21 crc kubenswrapper[4790]: I1124 14:49:21.347649 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1e0c3ca-accf-4636-82c8-bb73ad8304e4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a1e0c3ca-accf-4636-82c8-bb73ad8304e4" (UID: "a1e0c3ca-accf-4636-82c8-bb73ad8304e4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:49:21 crc kubenswrapper[4790]: I1124 14:49:21.372933 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1e0c3ca-accf-4636-82c8-bb73ad8304e4-kube-api-access-9f2rd" (OuterVolumeSpecName: "kube-api-access-9f2rd") pod "a1e0c3ca-accf-4636-82c8-bb73ad8304e4" (UID: "a1e0c3ca-accf-4636-82c8-bb73ad8304e4"). InnerVolumeSpecName "kube-api-access-9f2rd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:49:21 crc kubenswrapper[4790]: I1124 14:49:21.449134 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1e0c3ca-accf-4636-82c8-bb73ad8304e4-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:49:21 crc kubenswrapper[4790]: I1124 14:49:21.449190 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9f2rd\" (UniqueName: \"kubernetes.io/projected/a1e0c3ca-accf-4636-82c8-bb73ad8304e4-kube-api-access-9f2rd\") on node \"crc\" DevicePath \"\"" Nov 24 14:49:21 crc kubenswrapper[4790]: I1124 14:49:21.907472 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-1102-account-create-hrp2v" event={"ID":"a1e0c3ca-accf-4636-82c8-bb73ad8304e4","Type":"ContainerDied","Data":"a030b8d6f6b5ebf913c19347748c0e4447c1165c1cbe75d9ba1bdb92b95dc3b8"} Nov 24 14:49:21 crc kubenswrapper[4790]: I1124 14:49:21.907515 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a030b8d6f6b5ebf913c19347748c0e4447c1165c1cbe75d9ba1bdb92b95dc3b8" Nov 24 14:49:21 crc kubenswrapper[4790]: I1124 14:49:21.907603 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-1102-account-create-hrp2v" Nov 24 14:49:24 crc kubenswrapper[4790]: I1124 14:49:24.047684 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-api-5b99f4ff4b-5rkfz"] Nov 24 14:49:24 crc kubenswrapper[4790]: E1124 14:49:24.049651 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1e0c3ca-accf-4636-82c8-bb73ad8304e4" containerName="mariadb-account-create" Nov 24 14:49:24 crc kubenswrapper[4790]: I1124 14:49:24.049741 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1e0c3ca-accf-4636-82c8-bb73ad8304e4" containerName="mariadb-account-create" Nov 24 14:49:24 crc kubenswrapper[4790]: E1124 14:49:24.049835 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57b1601f-40f6-49cd-ab92-85efc8fb18e1" containerName="mariadb-database-create" Nov 24 14:49:24 crc kubenswrapper[4790]: I1124 14:49:24.049925 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="57b1601f-40f6-49cd-ab92-85efc8fb18e1" containerName="mariadb-database-create" Nov 24 14:49:24 crc kubenswrapper[4790]: I1124 14:49:24.050221 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="57b1601f-40f6-49cd-ab92-85efc8fb18e1" containerName="mariadb-database-create" Nov 24 14:49:24 crc kubenswrapper[4790]: I1124 14:49:24.050365 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1e0c3ca-accf-4636-82c8-bb73ad8304e4" containerName="mariadb-account-create" Nov 24 14:49:24 crc kubenswrapper[4790]: I1124 14:49:24.052077 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-5b99f4ff4b-5rkfz" Nov 24 14:49:24 crc kubenswrapper[4790]: I1124 14:49:24.054873 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-scripts" Nov 24 14:49:24 crc kubenswrapper[4790]: I1124 14:49:24.055290 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-config-data" Nov 24 14:49:24 crc kubenswrapper[4790]: I1124 14:49:24.055441 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-octavia-dockercfg-87n5q" Nov 24 14:49:24 crc kubenswrapper[4790]: I1124 14:49:24.058793 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-5b99f4ff4b-5rkfz"] Nov 24 14:49:24 crc kubenswrapper[4790]: I1124 14:49:24.114511 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/008c0092-bd88-4696-91a3-3f124d181264-octavia-run\") pod \"octavia-api-5b99f4ff4b-5rkfz\" (UID: \"008c0092-bd88-4696-91a3-3f124d181264\") " pod="openstack/octavia-api-5b99f4ff4b-5rkfz" Nov 24 14:49:24 crc kubenswrapper[4790]: I1124 14:49:24.114629 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/008c0092-bd88-4696-91a3-3f124d181264-scripts\") pod \"octavia-api-5b99f4ff4b-5rkfz\" (UID: \"008c0092-bd88-4696-91a3-3f124d181264\") " pod="openstack/octavia-api-5b99f4ff4b-5rkfz" Nov 24 14:49:24 crc kubenswrapper[4790]: I1124 14:49:24.114744 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/008c0092-bd88-4696-91a3-3f124d181264-config-data\") pod \"octavia-api-5b99f4ff4b-5rkfz\" (UID: \"008c0092-bd88-4696-91a3-3f124d181264\") " pod="openstack/octavia-api-5b99f4ff4b-5rkfz" Nov 24 14:49:24 crc kubenswrapper[4790]: I1124 14:49:24.114783 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/008c0092-bd88-4696-91a3-3f124d181264-config-data-merged\") pod \"octavia-api-5b99f4ff4b-5rkfz\" (UID: \"008c0092-bd88-4696-91a3-3f124d181264\") " pod="openstack/octavia-api-5b99f4ff4b-5rkfz" Nov 24 14:49:24 crc kubenswrapper[4790]: I1124 14:49:24.114843 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/008c0092-bd88-4696-91a3-3f124d181264-combined-ca-bundle\") pod \"octavia-api-5b99f4ff4b-5rkfz\" (UID: \"008c0092-bd88-4696-91a3-3f124d181264\") " pod="openstack/octavia-api-5b99f4ff4b-5rkfz" Nov 24 14:49:24 crc kubenswrapper[4790]: I1124 14:49:24.216336 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/008c0092-bd88-4696-91a3-3f124d181264-scripts\") pod \"octavia-api-5b99f4ff4b-5rkfz\" (UID: \"008c0092-bd88-4696-91a3-3f124d181264\") " pod="openstack/octavia-api-5b99f4ff4b-5rkfz" Nov 24 14:49:24 crc kubenswrapper[4790]: I1124 14:49:24.216541 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/008c0092-bd88-4696-91a3-3f124d181264-config-data\") pod \"octavia-api-5b99f4ff4b-5rkfz\" (UID: \"008c0092-bd88-4696-91a3-3f124d181264\") " pod="openstack/octavia-api-5b99f4ff4b-5rkfz" Nov 24 14:49:24 crc kubenswrapper[4790]: I1124 14:49:24.216573 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/008c0092-bd88-4696-91a3-3f124d181264-config-data-merged\") pod \"octavia-api-5b99f4ff4b-5rkfz\" (UID: \"008c0092-bd88-4696-91a3-3f124d181264\") " pod="openstack/octavia-api-5b99f4ff4b-5rkfz" Nov 24 14:49:24 crc kubenswrapper[4790]: I1124 14:49:24.216636 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/008c0092-bd88-4696-91a3-3f124d181264-combined-ca-bundle\") pod \"octavia-api-5b99f4ff4b-5rkfz\" (UID: \"008c0092-bd88-4696-91a3-3f124d181264\") " pod="openstack/octavia-api-5b99f4ff4b-5rkfz" Nov 24 14:49:24 crc kubenswrapper[4790]: I1124 14:49:24.216661 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/008c0092-bd88-4696-91a3-3f124d181264-octavia-run\") pod \"octavia-api-5b99f4ff4b-5rkfz\" (UID: \"008c0092-bd88-4696-91a3-3f124d181264\") " pod="openstack/octavia-api-5b99f4ff4b-5rkfz" Nov 24 14:49:24 crc kubenswrapper[4790]: I1124 14:49:24.217241 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/008c0092-bd88-4696-91a3-3f124d181264-octavia-run\") pod \"octavia-api-5b99f4ff4b-5rkfz\" (UID: \"008c0092-bd88-4696-91a3-3f124d181264\") " pod="openstack/octavia-api-5b99f4ff4b-5rkfz" Nov 24 14:49:24 crc kubenswrapper[4790]: I1124 14:49:24.217353 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/008c0092-bd88-4696-91a3-3f124d181264-config-data-merged\") pod \"octavia-api-5b99f4ff4b-5rkfz\" (UID: \"008c0092-bd88-4696-91a3-3f124d181264\") " pod="openstack/octavia-api-5b99f4ff4b-5rkfz" Nov 24 14:49:24 crc kubenswrapper[4790]: I1124 14:49:24.222457 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/008c0092-bd88-4696-91a3-3f124d181264-combined-ca-bundle\") pod \"octavia-api-5b99f4ff4b-5rkfz\" (UID: \"008c0092-bd88-4696-91a3-3f124d181264\") " pod="openstack/octavia-api-5b99f4ff4b-5rkfz" Nov 24 14:49:24 crc kubenswrapper[4790]: I1124 14:49:24.222566 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/008c0092-bd88-4696-91a3-3f124d181264-config-data\") pod \"octavia-api-5b99f4ff4b-5rkfz\" (UID: \"008c0092-bd88-4696-91a3-3f124d181264\") " pod="openstack/octavia-api-5b99f4ff4b-5rkfz" Nov 24 14:49:24 crc kubenswrapper[4790]: I1124 14:49:24.222719 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/008c0092-bd88-4696-91a3-3f124d181264-scripts\") pod \"octavia-api-5b99f4ff4b-5rkfz\" (UID: \"008c0092-bd88-4696-91a3-3f124d181264\") " pod="openstack/octavia-api-5b99f4ff4b-5rkfz" Nov 24 14:49:24 crc kubenswrapper[4790]: I1124 14:49:24.387036 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-5b99f4ff4b-5rkfz" Nov 24 14:49:24 crc kubenswrapper[4790]: W1124 14:49:24.877402 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod008c0092_bd88_4696_91a3_3f124d181264.slice/crio-5ba693fa8f8df7aab4262fbe8968ae89a42561f519117a9e2a5fd976b1e5cb4d WatchSource:0}: Error finding container 5ba693fa8f8df7aab4262fbe8968ae89a42561f519117a9e2a5fd976b1e5cb4d: Status 404 returned error can't find the container with id 5ba693fa8f8df7aab4262fbe8968ae89a42561f519117a9e2a5fd976b1e5cb4d Nov 24 14:49:24 crc kubenswrapper[4790]: I1124 14:49:24.880046 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-5b99f4ff4b-5rkfz"] Nov 24 14:49:24 crc kubenswrapper[4790]: I1124 14:49:24.880148 4790 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 14:49:24 crc kubenswrapper[4790]: I1124 14:49:24.937586 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-5b99f4ff4b-5rkfz" event={"ID":"008c0092-bd88-4696-91a3-3f124d181264","Type":"ContainerStarted","Data":"5ba693fa8f8df7aab4262fbe8968ae89a42561f519117a9e2a5fd976b1e5cb4d"} Nov 24 14:49:35 crc kubenswrapper[4790]: I1124 14:49:35.053091 4790 generic.go:334] "Generic (PLEG): container finished" podID="008c0092-bd88-4696-91a3-3f124d181264" containerID="4c1544a4cf62909054c82043fa22769da3947532d8d16cc8fca32907e827277b" exitCode=0 Nov 24 14:49:35 crc kubenswrapper[4790]: I1124 14:49:35.053250 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-5b99f4ff4b-5rkfz" event={"ID":"008c0092-bd88-4696-91a3-3f124d181264","Type":"ContainerDied","Data":"4c1544a4cf62909054c82043fa22769da3947532d8d16cc8fca32907e827277b"} Nov 24 14:49:36 crc kubenswrapper[4790]: I1124 14:49:36.066171 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-5b99f4ff4b-5rkfz" event={"ID":"008c0092-bd88-4696-91a3-3f124d181264","Type":"ContainerStarted","Data":"45aa84ec1977dec6fa599087185b845ae15e5cf52368b90be1320d0d34482275"} Nov 24 14:49:36 crc kubenswrapper[4790]: I1124 14:49:36.066543 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-5b99f4ff4b-5rkfz" event={"ID":"008c0092-bd88-4696-91a3-3f124d181264","Type":"ContainerStarted","Data":"f9ef46b03fe3cf4e39364afb1681eb60adc80eecb2fbdd88b09b4b185a0cee25"} Nov 24 14:49:36 crc kubenswrapper[4790]: I1124 14:49:36.066574 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-5b99f4ff4b-5rkfz" Nov 24 14:49:36 crc kubenswrapper[4790]: I1124 14:49:36.066594 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-5b99f4ff4b-5rkfz" Nov 24 14:49:42 crc kubenswrapper[4790]: I1124 14:49:42.428256 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-6p54v" podUID="ec4e9e28-345c-4cc4-acba-7a06cac52117" containerName="ovn-controller" probeResult="failure" output=< Nov 24 14:49:42 crc kubenswrapper[4790]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 24 14:49:42 crc kubenswrapper[4790]: > Nov 24 14:49:42 crc kubenswrapper[4790]: I1124 14:49:42.505967 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-l46w2" Nov 24 14:49:42 crc kubenswrapper[4790]: I1124 14:49:42.512760 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-l46w2" Nov 24 14:49:42 crc kubenswrapper[4790]: I1124 14:49:42.534164 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-api-5b99f4ff4b-5rkfz" podStartSLOduration=9.31207854 podStartE2EDuration="18.534140342s" podCreationTimestamp="2025-11-24 14:49:24 +0000 UTC" firstStartedPulling="2025-11-24 14:49:24.87995937 +0000 UTC m=+5813.259853032" lastFinishedPulling="2025-11-24 14:49:34.102021172 +0000 UTC m=+5822.481914834" observedRunningTime="2025-11-24 14:49:36.089485317 +0000 UTC m=+5824.469378999" watchObservedRunningTime="2025-11-24 14:49:42.534140342 +0000 UTC m=+5830.914034044" Nov 24 14:49:42 crc kubenswrapper[4790]: I1124 14:49:42.669141 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-6p54v-config-n66jn"] Nov 24 14:49:42 crc kubenswrapper[4790]: I1124 14:49:42.671048 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6p54v-config-n66jn" Nov 24 14:49:42 crc kubenswrapper[4790]: I1124 14:49:42.673761 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 24 14:49:42 crc kubenswrapper[4790]: I1124 14:49:42.679615 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6p54v-config-n66jn"] Nov 24 14:49:42 crc kubenswrapper[4790]: I1124 14:49:42.730811 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-var-run-ovn\") pod \"ovn-controller-6p54v-config-n66jn\" (UID: \"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d\") " pod="openstack/ovn-controller-6p54v-config-n66jn" Nov 24 14:49:42 crc kubenswrapper[4790]: I1124 14:49:42.731010 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-var-log-ovn\") pod \"ovn-controller-6p54v-config-n66jn\" (UID: \"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d\") " pod="openstack/ovn-controller-6p54v-config-n66jn" Nov 24 14:49:42 crc kubenswrapper[4790]: I1124 14:49:42.731059 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-scripts\") pod \"ovn-controller-6p54v-config-n66jn\" (UID: \"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d\") " pod="openstack/ovn-controller-6p54v-config-n66jn" Nov 24 14:49:42 crc kubenswrapper[4790]: I1124 14:49:42.731156 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjl2p\" (UniqueName: \"kubernetes.io/projected/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-kube-api-access-cjl2p\") pod \"ovn-controller-6p54v-config-n66jn\" (UID: \"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d\") " pod="openstack/ovn-controller-6p54v-config-n66jn" Nov 24 14:49:42 crc kubenswrapper[4790]: I1124 14:49:42.731488 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-additional-scripts\") pod \"ovn-controller-6p54v-config-n66jn\" (UID: \"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d\") " pod="openstack/ovn-controller-6p54v-config-n66jn" Nov 24 14:49:42 crc kubenswrapper[4790]: I1124 14:49:42.731734 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-var-run\") pod \"ovn-controller-6p54v-config-n66jn\" (UID: \"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d\") " pod="openstack/ovn-controller-6p54v-config-n66jn" Nov 24 14:49:42 crc kubenswrapper[4790]: I1124 14:49:42.833453 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-var-run\") pod \"ovn-controller-6p54v-config-n66jn\" (UID: \"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d\") " pod="openstack/ovn-controller-6p54v-config-n66jn" Nov 24 14:49:42 crc kubenswrapper[4790]: I1124 14:49:42.833613 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-var-run-ovn\") pod \"ovn-controller-6p54v-config-n66jn\" (UID: \"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d\") " pod="openstack/ovn-controller-6p54v-config-n66jn" Nov 24 14:49:42 crc kubenswrapper[4790]: I1124 14:49:42.833647 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-var-log-ovn\") pod \"ovn-controller-6p54v-config-n66jn\" (UID: \"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d\") " pod="openstack/ovn-controller-6p54v-config-n66jn" Nov 24 14:49:42 crc kubenswrapper[4790]: I1124 14:49:42.833670 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-scripts\") pod \"ovn-controller-6p54v-config-n66jn\" (UID: \"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d\") " pod="openstack/ovn-controller-6p54v-config-n66jn" Nov 24 14:49:42 crc kubenswrapper[4790]: I1124 14:49:42.833710 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjl2p\" (UniqueName: \"kubernetes.io/projected/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-kube-api-access-cjl2p\") pod \"ovn-controller-6p54v-config-n66jn\" (UID: \"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d\") " pod="openstack/ovn-controller-6p54v-config-n66jn" Nov 24 14:49:42 crc kubenswrapper[4790]: I1124 14:49:42.833818 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-additional-scripts\") pod \"ovn-controller-6p54v-config-n66jn\" (UID: \"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d\") " pod="openstack/ovn-controller-6p54v-config-n66jn" Nov 24 14:49:42 crc kubenswrapper[4790]: I1124 14:49:42.833956 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-var-run-ovn\") pod \"ovn-controller-6p54v-config-n66jn\" (UID: \"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d\") " pod="openstack/ovn-controller-6p54v-config-n66jn" Nov 24 14:49:42 crc kubenswrapper[4790]: I1124 14:49:42.833983 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-var-run\") pod \"ovn-controller-6p54v-config-n66jn\" (UID: \"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d\") " pod="openstack/ovn-controller-6p54v-config-n66jn" Nov 24 14:49:42 crc kubenswrapper[4790]: I1124 14:49:42.834132 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-var-log-ovn\") pod \"ovn-controller-6p54v-config-n66jn\" (UID: \"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d\") " pod="openstack/ovn-controller-6p54v-config-n66jn" Nov 24 14:49:42 crc kubenswrapper[4790]: I1124 14:49:42.834944 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-additional-scripts\") pod \"ovn-controller-6p54v-config-n66jn\" (UID: \"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d\") " pod="openstack/ovn-controller-6p54v-config-n66jn" Nov 24 14:49:42 crc kubenswrapper[4790]: I1124 14:49:42.836429 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-scripts\") pod \"ovn-controller-6p54v-config-n66jn\" (UID: \"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d\") " pod="openstack/ovn-controller-6p54v-config-n66jn" Nov 24 14:49:42 crc kubenswrapper[4790]: I1124 14:49:42.855936 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjl2p\" (UniqueName: \"kubernetes.io/projected/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-kube-api-access-cjl2p\") pod \"ovn-controller-6p54v-config-n66jn\" (UID: \"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d\") " pod="openstack/ovn-controller-6p54v-config-n66jn" Nov 24 14:49:42 crc kubenswrapper[4790]: I1124 14:49:42.997082 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6p54v-config-n66jn" Nov 24 14:49:43 crc kubenswrapper[4790]: I1124 14:49:43.552578 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6p54v-config-n66jn"] Nov 24 14:49:43 crc kubenswrapper[4790]: W1124 14:49:43.564059 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5be5a30f_961d_4b1f_a5b2_0aaad3b9fc4d.slice/crio-c9b23e7511c4beff79fb815a09d2f57763fc72e6b83599831d7ea3083e0f4553 WatchSource:0}: Error finding container c9b23e7511c4beff79fb815a09d2f57763fc72e6b83599831d7ea3083e0f4553: Status 404 returned error can't find the container with id c9b23e7511c4beff79fb815a09d2f57763fc72e6b83599831d7ea3083e0f4553 Nov 24 14:49:44 crc kubenswrapper[4790]: I1124 14:49:44.161278 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6p54v-config-n66jn" event={"ID":"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d","Type":"ContainerStarted","Data":"fcebf0df908b2a9cd9189fe5ede67e477dff4ce9334dc82d06f074beb6bcf297"} Nov 24 14:49:44 crc kubenswrapper[4790]: I1124 14:49:44.161765 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6p54v-config-n66jn" event={"ID":"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d","Type":"ContainerStarted","Data":"c9b23e7511c4beff79fb815a09d2f57763fc72e6b83599831d7ea3083e0f4553"} Nov 24 14:49:44 crc kubenswrapper[4790]: I1124 14:49:44.193734 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-6p54v-config-n66jn" podStartSLOduration=2.193710046 podStartE2EDuration="2.193710046s" podCreationTimestamp="2025-11-24 14:49:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:49:44.186141008 +0000 UTC m=+5832.566034670" watchObservedRunningTime="2025-11-24 14:49:44.193710046 +0000 UTC m=+5832.573603718" Nov 24 14:49:45 crc kubenswrapper[4790]: I1124 14:49:45.179729 4790 generic.go:334] "Generic (PLEG): container finished" podID="5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d" containerID="fcebf0df908b2a9cd9189fe5ede67e477dff4ce9334dc82d06f074beb6bcf297" exitCode=0 Nov 24 14:49:45 crc kubenswrapper[4790]: I1124 14:49:45.179833 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6p54v-config-n66jn" event={"ID":"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d","Type":"ContainerDied","Data":"fcebf0df908b2a9cd9189fe5ede67e477dff4ce9334dc82d06f074beb6bcf297"} Nov 24 14:49:46 crc kubenswrapper[4790]: I1124 14:49:46.616064 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6p54v-config-n66jn" Nov 24 14:49:46 crc kubenswrapper[4790]: I1124 14:49:46.718799 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjl2p\" (UniqueName: \"kubernetes.io/projected/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-kube-api-access-cjl2p\") pod \"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d\" (UID: \"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d\") " Nov 24 14:49:46 crc kubenswrapper[4790]: I1124 14:49:46.718977 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-var-run-ovn\") pod \"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d\" (UID: \"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d\") " Nov 24 14:49:46 crc kubenswrapper[4790]: I1124 14:49:46.719020 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-additional-scripts\") pod \"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d\" (UID: \"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d\") " Nov 24 14:49:46 crc kubenswrapper[4790]: I1124 14:49:46.719041 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-var-run\") pod \"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d\" (UID: \"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d\") " Nov 24 14:49:46 crc kubenswrapper[4790]: I1124 14:49:46.719084 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-var-log-ovn\") pod \"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d\" (UID: \"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d\") " Nov 24 14:49:46 crc kubenswrapper[4790]: I1124 14:49:46.719163 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-scripts\") pod \"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d\" (UID: \"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d\") " Nov 24 14:49:46 crc kubenswrapper[4790]: I1124 14:49:46.719853 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d" (UID: "5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:49:46 crc kubenswrapper[4790]: I1124 14:49:46.719977 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-var-run" (OuterVolumeSpecName: "var-run") pod "5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d" (UID: "5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:49:46 crc kubenswrapper[4790]: I1124 14:49:46.720053 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d" (UID: "5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:49:46 crc kubenswrapper[4790]: I1124 14:49:46.720660 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d" (UID: "5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:49:46 crc kubenswrapper[4790]: I1124 14:49:46.721080 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-scripts" (OuterVolumeSpecName: "scripts") pod "5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d" (UID: "5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:49:46 crc kubenswrapper[4790]: I1124 14:49:46.725722 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-kube-api-access-cjl2p" (OuterVolumeSpecName: "kube-api-access-cjl2p") pod "5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d" (UID: "5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d"). InnerVolumeSpecName "kube-api-access-cjl2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:49:46 crc kubenswrapper[4790]: I1124 14:49:46.822005 4790 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 14:49:46 crc kubenswrapper[4790]: I1124 14:49:46.822065 4790 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:49:46 crc kubenswrapper[4790]: I1124 14:49:46.822088 4790 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-var-run\") on node \"crc\" DevicePath \"\"" Nov 24 14:49:46 crc kubenswrapper[4790]: I1124 14:49:46.822106 4790 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 14:49:46 crc kubenswrapper[4790]: I1124 14:49:46.822124 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:49:46 crc kubenswrapper[4790]: I1124 14:49:46.822142 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjl2p\" (UniqueName: \"kubernetes.io/projected/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d-kube-api-access-cjl2p\") on node \"crc\" DevicePath \"\"" Nov 24 14:49:47 crc kubenswrapper[4790]: I1124 14:49:47.210348 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6p54v-config-n66jn" event={"ID":"5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d","Type":"ContainerDied","Data":"c9b23e7511c4beff79fb815a09d2f57763fc72e6b83599831d7ea3083e0f4553"} Nov 24 14:49:47 crc kubenswrapper[4790]: I1124 14:49:47.210416 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9b23e7511c4beff79fb815a09d2f57763fc72e6b83599831d7ea3083e0f4553" Nov 24 14:49:47 crc kubenswrapper[4790]: I1124 14:49:47.210429 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6p54v-config-n66jn" Nov 24 14:49:47 crc kubenswrapper[4790]: I1124 14:49:47.341438 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-6p54v-config-n66jn"] Nov 24 14:49:47 crc kubenswrapper[4790]: I1124 14:49:47.361749 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-6p54v-config-n66jn"] Nov 24 14:49:47 crc kubenswrapper[4790]: I1124 14:49:47.450142 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-6p54v" Nov 24 14:49:48 crc kubenswrapper[4790]: I1124 14:49:48.333193 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d" path="/var/lib/kubelet/pods/5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d/volumes" Nov 24 14:49:48 crc kubenswrapper[4790]: I1124 14:49:48.583054 4790 scope.go:117] "RemoveContainer" containerID="0d23ee6e35232d5cc608a2b677baea86e7b31006ea39bad6c35a214535941fe4" Nov 24 14:49:48 crc kubenswrapper[4790]: I1124 14:49:48.659020 4790 scope.go:117] "RemoveContainer" containerID="e6f5cad26b70c28d261270b94a0b9ce4f97f19ee078677d537320e82e304e761" Nov 24 14:49:48 crc kubenswrapper[4790]: I1124 14:49:48.756186 4790 scope.go:117] "RemoveContainer" containerID="ccb07244be050d105ef0c39518348f82fbaa508a6212856a2de14a89f191b49b" Nov 24 14:49:48 crc kubenswrapper[4790]: I1124 14:49:48.806605 4790 scope.go:117] "RemoveContainer" containerID="7e81773b13d8d2c24f38bcc52d5bc1472270c6c363df339602ede7da9a78f8e5" Nov 24 14:49:50 crc kubenswrapper[4790]: I1124 14:49:50.194163 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-rsyslog-7rfsl"] Nov 24 14:49:50 crc kubenswrapper[4790]: E1124 14:49:50.195042 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d" containerName="ovn-config" Nov 24 14:49:50 crc kubenswrapper[4790]: I1124 14:49:50.195062 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d" containerName="ovn-config" Nov 24 14:49:50 crc kubenswrapper[4790]: I1124 14:49:50.195305 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="5be5a30f-961d-4b1f-a5b2-0aaad3b9fc4d" containerName="ovn-config" Nov 24 14:49:50 crc kubenswrapper[4790]: I1124 14:49:50.196499 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-7rfsl" Nov 24 14:49:50 crc kubenswrapper[4790]: I1124 14:49:50.199272 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"octavia-hmport-map" Nov 24 14:49:50 crc kubenswrapper[4790]: I1124 14:49:50.199282 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-scripts" Nov 24 14:49:50 crc kubenswrapper[4790]: I1124 14:49:50.200563 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-config-data" Nov 24 14:49:50 crc kubenswrapper[4790]: I1124 14:49:50.206726 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-7rfsl"] Nov 24 14:49:50 crc kubenswrapper[4790]: I1124 14:49:50.325642 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/d749d6ea-42cc-47d0-880e-d079f59b3ad6-hm-ports\") pod \"octavia-rsyslog-7rfsl\" (UID: \"d749d6ea-42cc-47d0-880e-d079f59b3ad6\") " pod="openstack/octavia-rsyslog-7rfsl" Nov 24 14:49:50 crc kubenswrapper[4790]: I1124 14:49:50.325722 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/d749d6ea-42cc-47d0-880e-d079f59b3ad6-config-data-merged\") pod \"octavia-rsyslog-7rfsl\" (UID: \"d749d6ea-42cc-47d0-880e-d079f59b3ad6\") " pod="openstack/octavia-rsyslog-7rfsl" Nov 24 14:49:50 crc kubenswrapper[4790]: I1124 14:49:50.325821 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d749d6ea-42cc-47d0-880e-d079f59b3ad6-config-data\") pod \"octavia-rsyslog-7rfsl\" (UID: \"d749d6ea-42cc-47d0-880e-d079f59b3ad6\") " pod="openstack/octavia-rsyslog-7rfsl" Nov 24 14:49:50 crc kubenswrapper[4790]: I1124 14:49:50.325842 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d749d6ea-42cc-47d0-880e-d079f59b3ad6-scripts\") pod \"octavia-rsyslog-7rfsl\" (UID: \"d749d6ea-42cc-47d0-880e-d079f59b3ad6\") " pod="openstack/octavia-rsyslog-7rfsl" Nov 24 14:49:50 crc kubenswrapper[4790]: I1124 14:49:50.427650 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/d749d6ea-42cc-47d0-880e-d079f59b3ad6-hm-ports\") pod \"octavia-rsyslog-7rfsl\" (UID: \"d749d6ea-42cc-47d0-880e-d079f59b3ad6\") " pod="openstack/octavia-rsyslog-7rfsl" Nov 24 14:49:50 crc kubenswrapper[4790]: I1124 14:49:50.427812 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/d749d6ea-42cc-47d0-880e-d079f59b3ad6-config-data-merged\") pod \"octavia-rsyslog-7rfsl\" (UID: \"d749d6ea-42cc-47d0-880e-d079f59b3ad6\") " pod="openstack/octavia-rsyslog-7rfsl" Nov 24 14:49:50 crc kubenswrapper[4790]: I1124 14:49:50.427961 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d749d6ea-42cc-47d0-880e-d079f59b3ad6-config-data\") pod \"octavia-rsyslog-7rfsl\" (UID: \"d749d6ea-42cc-47d0-880e-d079f59b3ad6\") " pod="openstack/octavia-rsyslog-7rfsl" Nov 24 14:49:50 crc kubenswrapper[4790]: I1124 14:49:50.428006 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d749d6ea-42cc-47d0-880e-d079f59b3ad6-scripts\") pod \"octavia-rsyslog-7rfsl\" (UID: \"d749d6ea-42cc-47d0-880e-d079f59b3ad6\") " pod="openstack/octavia-rsyslog-7rfsl" Nov 24 14:49:50 crc kubenswrapper[4790]: I1124 14:49:50.429043 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/d749d6ea-42cc-47d0-880e-d079f59b3ad6-hm-ports\") pod \"octavia-rsyslog-7rfsl\" (UID: \"d749d6ea-42cc-47d0-880e-d079f59b3ad6\") " pod="openstack/octavia-rsyslog-7rfsl" Nov 24 14:49:50 crc kubenswrapper[4790]: I1124 14:49:50.434164 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/d749d6ea-42cc-47d0-880e-d079f59b3ad6-config-data-merged\") pod \"octavia-rsyslog-7rfsl\" (UID: \"d749d6ea-42cc-47d0-880e-d079f59b3ad6\") " pod="openstack/octavia-rsyslog-7rfsl" Nov 24 14:49:50 crc kubenswrapper[4790]: I1124 14:49:50.437938 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d749d6ea-42cc-47d0-880e-d079f59b3ad6-scripts\") pod \"octavia-rsyslog-7rfsl\" (UID: \"d749d6ea-42cc-47d0-880e-d079f59b3ad6\") " pod="openstack/octavia-rsyslog-7rfsl" Nov 24 14:49:50 crc kubenswrapper[4790]: I1124 14:49:50.438317 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d749d6ea-42cc-47d0-880e-d079f59b3ad6-config-data\") pod \"octavia-rsyslog-7rfsl\" (UID: \"d749d6ea-42cc-47d0-880e-d079f59b3ad6\") " pod="openstack/octavia-rsyslog-7rfsl" Nov 24 14:49:50 crc kubenswrapper[4790]: I1124 14:49:50.530705 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-7rfsl" Nov 24 14:49:50 crc kubenswrapper[4790]: I1124 14:49:50.916447 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-image-upload-59f8cff499-gg8zh"] Nov 24 14:49:50 crc kubenswrapper[4790]: I1124 14:49:50.918406 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-gg8zh" Nov 24 14:49:50 crc kubenswrapper[4790]: I1124 14:49:50.922342 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Nov 24 14:49:50 crc kubenswrapper[4790]: I1124 14:49:50.935022 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-gg8zh"] Nov 24 14:49:51 crc kubenswrapper[4790]: I1124 14:49:51.044004 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b8310be9-af92-4fa0-a6f9-f242ef869fa4-httpd-config\") pod \"octavia-image-upload-59f8cff499-gg8zh\" (UID: \"b8310be9-af92-4fa0-a6f9-f242ef869fa4\") " pod="openstack/octavia-image-upload-59f8cff499-gg8zh" Nov 24 14:49:51 crc kubenswrapper[4790]: I1124 14:49:51.044292 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/b8310be9-af92-4fa0-a6f9-f242ef869fa4-amphora-image\") pod \"octavia-image-upload-59f8cff499-gg8zh\" (UID: \"b8310be9-af92-4fa0-a6f9-f242ef869fa4\") " pod="openstack/octavia-image-upload-59f8cff499-gg8zh" Nov 24 14:49:51 crc kubenswrapper[4790]: I1124 14:49:51.146461 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b8310be9-af92-4fa0-a6f9-f242ef869fa4-httpd-config\") pod \"octavia-image-upload-59f8cff499-gg8zh\" (UID: \"b8310be9-af92-4fa0-a6f9-f242ef869fa4\") " pod="openstack/octavia-image-upload-59f8cff499-gg8zh" Nov 24 14:49:51 crc kubenswrapper[4790]: I1124 14:49:51.146527 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/b8310be9-af92-4fa0-a6f9-f242ef869fa4-amphora-image\") pod \"octavia-image-upload-59f8cff499-gg8zh\" (UID: \"b8310be9-af92-4fa0-a6f9-f242ef869fa4\") " pod="openstack/octavia-image-upload-59f8cff499-gg8zh" Nov 24 14:49:51 crc kubenswrapper[4790]: I1124 14:49:51.147128 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/b8310be9-af92-4fa0-a6f9-f242ef869fa4-amphora-image\") pod \"octavia-image-upload-59f8cff499-gg8zh\" (UID: \"b8310be9-af92-4fa0-a6f9-f242ef869fa4\") " pod="openstack/octavia-image-upload-59f8cff499-gg8zh" Nov 24 14:49:51 crc kubenswrapper[4790]: I1124 14:49:51.155684 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b8310be9-af92-4fa0-a6f9-f242ef869fa4-httpd-config\") pod \"octavia-image-upload-59f8cff499-gg8zh\" (UID: \"b8310be9-af92-4fa0-a6f9-f242ef869fa4\") " pod="openstack/octavia-image-upload-59f8cff499-gg8zh" Nov 24 14:49:51 crc kubenswrapper[4790]: I1124 14:49:51.182872 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-7rfsl"] Nov 24 14:49:51 crc kubenswrapper[4790]: I1124 14:49:51.244785 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-gg8zh" Nov 24 14:49:51 crc kubenswrapper[4790]: I1124 14:49:51.255285 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-7rfsl" event={"ID":"d749d6ea-42cc-47d0-880e-d079f59b3ad6","Type":"ContainerStarted","Data":"721af744a3636b931e0d3e1a40652cc426143b8f6688aa0bead18556fa3d5c42"} Nov 24 14:49:51 crc kubenswrapper[4790]: I1124 14:49:51.734105 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-gg8zh"] Nov 24 14:49:52 crc kubenswrapper[4790]: I1124 14:49:52.265650 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-gg8zh" event={"ID":"b8310be9-af92-4fa0-a6f9-f242ef869fa4","Type":"ContainerStarted","Data":"db5175a1dc0f5aff331b477e35cd701eb945c6f17df227b12397a746e22ab28a"} Nov 24 14:49:54 crc kubenswrapper[4790]: I1124 14:49:54.287470 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-7rfsl" event={"ID":"d749d6ea-42cc-47d0-880e-d079f59b3ad6","Type":"ContainerStarted","Data":"d8997a19bd3f3e2db69647602076b48ca6d0a9e34e2e89dd2dcf7d974dc70d67"} Nov 24 14:49:56 crc kubenswrapper[4790]: I1124 14:49:56.306241 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-sync-qgdgj"] Nov 24 14:49:56 crc kubenswrapper[4790]: I1124 14:49:56.309102 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-qgdgj" Nov 24 14:49:56 crc kubenswrapper[4790]: I1124 14:49:56.312125 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-scripts" Nov 24 14:49:56 crc kubenswrapper[4790]: I1124 14:49:56.357597 4790 generic.go:334] "Generic (PLEG): container finished" podID="d749d6ea-42cc-47d0-880e-d079f59b3ad6" containerID="d8997a19bd3f3e2db69647602076b48ca6d0a9e34e2e89dd2dcf7d974dc70d67" exitCode=0 Nov 24 14:49:56 crc kubenswrapper[4790]: I1124 14:49:56.362728 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-7rfsl" event={"ID":"d749d6ea-42cc-47d0-880e-d079f59b3ad6","Type":"ContainerDied","Data":"d8997a19bd3f3e2db69647602076b48ca6d0a9e34e2e89dd2dcf7d974dc70d67"} Nov 24 14:49:56 crc kubenswrapper[4790]: I1124 14:49:56.362781 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-qgdgj"] Nov 24 14:49:56 crc kubenswrapper[4790]: I1124 14:49:56.465332 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/c5ef870d-41ea-48e7-9a44-3f4a610f0a62-config-data-merged\") pod \"octavia-db-sync-qgdgj\" (UID: \"c5ef870d-41ea-48e7-9a44-3f4a610f0a62\") " pod="openstack/octavia-db-sync-qgdgj" Nov 24 14:49:56 crc kubenswrapper[4790]: I1124 14:49:56.465455 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5ef870d-41ea-48e7-9a44-3f4a610f0a62-scripts\") pod \"octavia-db-sync-qgdgj\" (UID: \"c5ef870d-41ea-48e7-9a44-3f4a610f0a62\") " pod="openstack/octavia-db-sync-qgdgj" Nov 24 14:49:56 crc kubenswrapper[4790]: I1124 14:49:56.466224 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5ef870d-41ea-48e7-9a44-3f4a610f0a62-config-data\") pod \"octavia-db-sync-qgdgj\" (UID: \"c5ef870d-41ea-48e7-9a44-3f4a610f0a62\") " pod="openstack/octavia-db-sync-qgdgj" Nov 24 14:49:56 crc kubenswrapper[4790]: I1124 14:49:56.466298 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5ef870d-41ea-48e7-9a44-3f4a610f0a62-combined-ca-bundle\") pod \"octavia-db-sync-qgdgj\" (UID: \"c5ef870d-41ea-48e7-9a44-3f4a610f0a62\") " pod="openstack/octavia-db-sync-qgdgj" Nov 24 14:49:56 crc kubenswrapper[4790]: I1124 14:49:56.569027 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/c5ef870d-41ea-48e7-9a44-3f4a610f0a62-config-data-merged\") pod \"octavia-db-sync-qgdgj\" (UID: \"c5ef870d-41ea-48e7-9a44-3f4a610f0a62\") " pod="openstack/octavia-db-sync-qgdgj" Nov 24 14:49:56 crc kubenswrapper[4790]: I1124 14:49:56.569354 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5ef870d-41ea-48e7-9a44-3f4a610f0a62-scripts\") pod \"octavia-db-sync-qgdgj\" (UID: \"c5ef870d-41ea-48e7-9a44-3f4a610f0a62\") " pod="openstack/octavia-db-sync-qgdgj" Nov 24 14:49:56 crc kubenswrapper[4790]: I1124 14:49:56.569505 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5ef870d-41ea-48e7-9a44-3f4a610f0a62-config-data\") pod \"octavia-db-sync-qgdgj\" (UID: \"c5ef870d-41ea-48e7-9a44-3f4a610f0a62\") " pod="openstack/octavia-db-sync-qgdgj" Nov 24 14:49:56 crc kubenswrapper[4790]: I1124 14:49:56.569612 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5ef870d-41ea-48e7-9a44-3f4a610f0a62-combined-ca-bundle\") pod \"octavia-db-sync-qgdgj\" (UID: \"c5ef870d-41ea-48e7-9a44-3f4a610f0a62\") " pod="openstack/octavia-db-sync-qgdgj" Nov 24 14:49:56 crc kubenswrapper[4790]: I1124 14:49:56.570117 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/c5ef870d-41ea-48e7-9a44-3f4a610f0a62-config-data-merged\") pod \"octavia-db-sync-qgdgj\" (UID: \"c5ef870d-41ea-48e7-9a44-3f4a610f0a62\") " pod="openstack/octavia-db-sync-qgdgj" Nov 24 14:49:56 crc kubenswrapper[4790]: I1124 14:49:56.578345 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5ef870d-41ea-48e7-9a44-3f4a610f0a62-combined-ca-bundle\") pod \"octavia-db-sync-qgdgj\" (UID: \"c5ef870d-41ea-48e7-9a44-3f4a610f0a62\") " pod="openstack/octavia-db-sync-qgdgj" Nov 24 14:49:56 crc kubenswrapper[4790]: I1124 14:49:56.584623 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5ef870d-41ea-48e7-9a44-3f4a610f0a62-config-data\") pod \"octavia-db-sync-qgdgj\" (UID: \"c5ef870d-41ea-48e7-9a44-3f4a610f0a62\") " pod="openstack/octavia-db-sync-qgdgj" Nov 24 14:49:56 crc kubenswrapper[4790]: I1124 14:49:56.586380 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5ef870d-41ea-48e7-9a44-3f4a610f0a62-scripts\") pod \"octavia-db-sync-qgdgj\" (UID: \"c5ef870d-41ea-48e7-9a44-3f4a610f0a62\") " pod="openstack/octavia-db-sync-qgdgj" Nov 24 14:49:56 crc kubenswrapper[4790]: I1124 14:49:56.644735 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-qgdgj" Nov 24 14:49:57 crc kubenswrapper[4790]: I1124 14:49:57.147996 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-qgdgj"] Nov 24 14:49:57 crc kubenswrapper[4790]: I1124 14:49:57.375306 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-qgdgj" event={"ID":"c5ef870d-41ea-48e7-9a44-3f4a610f0a62","Type":"ContainerStarted","Data":"9adb971e143f83f59505b53ba65f7f8b951a7edfbb0affb96c95b125c67c46a2"} Nov 24 14:49:57 crc kubenswrapper[4790]: I1124 14:49:57.375386 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-qgdgj" event={"ID":"c5ef870d-41ea-48e7-9a44-3f4a610f0a62","Type":"ContainerStarted","Data":"1cf3ccceebf6e420fe6b081f8d1b364484f32789d07525dfaafaa4d05f154fd4"} Nov 24 14:49:58 crc kubenswrapper[4790]: I1124 14:49:58.397998 4790 generic.go:334] "Generic (PLEG): container finished" podID="c5ef870d-41ea-48e7-9a44-3f4a610f0a62" containerID="9adb971e143f83f59505b53ba65f7f8b951a7edfbb0affb96c95b125c67c46a2" exitCode=0 Nov 24 14:49:58 crc kubenswrapper[4790]: I1124 14:49:58.398555 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-qgdgj" event={"ID":"c5ef870d-41ea-48e7-9a44-3f4a610f0a62","Type":"ContainerDied","Data":"9adb971e143f83f59505b53ba65f7f8b951a7edfbb0affb96c95b125c67c46a2"} Nov 24 14:49:59 crc kubenswrapper[4790]: I1124 14:49:59.084943 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-5b99f4ff4b-5rkfz" Nov 24 14:49:59 crc kubenswrapper[4790]: I1124 14:49:59.085333 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-5b99f4ff4b-5rkfz" Nov 24 14:49:59 crc kubenswrapper[4790]: I1124 14:49:59.410563 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-qgdgj" event={"ID":"c5ef870d-41ea-48e7-9a44-3f4a610f0a62","Type":"ContainerStarted","Data":"ce841e968dcaaf518b1267509e0f378bb950ec20d1f40926209f390055d76e84"} Nov 24 14:49:59 crc kubenswrapper[4790]: I1124 14:49:59.413748 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-7rfsl" event={"ID":"d749d6ea-42cc-47d0-880e-d079f59b3ad6","Type":"ContainerStarted","Data":"f7d90bcc42b7fe611a71ae47f934e2f18ce1a5f2a1afdebdd448b094a1fab998"} Nov 24 14:49:59 crc kubenswrapper[4790]: I1124 14:49:59.414105 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-rsyslog-7rfsl" Nov 24 14:49:59 crc kubenswrapper[4790]: I1124 14:49:59.432331 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-db-sync-qgdgj" podStartSLOduration=3.432309755 podStartE2EDuration="3.432309755s" podCreationTimestamp="2025-11-24 14:49:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:49:59.424655285 +0000 UTC m=+5847.804548947" watchObservedRunningTime="2025-11-24 14:49:59.432309755 +0000 UTC m=+5847.812203427" Nov 24 14:49:59 crc kubenswrapper[4790]: I1124 14:49:59.444501 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-rsyslog-7rfsl" podStartSLOduration=2.56785276 podStartE2EDuration="9.444455084s" podCreationTimestamp="2025-11-24 14:49:50 +0000 UTC" firstStartedPulling="2025-11-24 14:49:51.157549456 +0000 UTC m=+5839.537443118" lastFinishedPulling="2025-11-24 14:49:58.03415178 +0000 UTC m=+5846.414045442" observedRunningTime="2025-11-24 14:49:59.440160581 +0000 UTC m=+5847.820054243" watchObservedRunningTime="2025-11-24 14:49:59.444455084 +0000 UTC m=+5847.824348756" Nov 24 14:50:02 crc kubenswrapper[4790]: I1124 14:50:02.449006 4790 generic.go:334] "Generic (PLEG): container finished" podID="c5ef870d-41ea-48e7-9a44-3f4a610f0a62" containerID="ce841e968dcaaf518b1267509e0f378bb950ec20d1f40926209f390055d76e84" exitCode=0 Nov 24 14:50:02 crc kubenswrapper[4790]: I1124 14:50:02.449082 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-qgdgj" event={"ID":"c5ef870d-41ea-48e7-9a44-3f4a610f0a62","Type":"ContainerDied","Data":"ce841e968dcaaf518b1267509e0f378bb950ec20d1f40926209f390055d76e84"} Nov 24 14:50:03 crc kubenswrapper[4790]: I1124 14:50:03.461275 4790 generic.go:334] "Generic (PLEG): container finished" podID="b8310be9-af92-4fa0-a6f9-f242ef869fa4" containerID="16bd19c7b7e3cfacc2546ef655629f90b77fd41b82c1b1e3d599fbd064f6818c" exitCode=0 Nov 24 14:50:03 crc kubenswrapper[4790]: I1124 14:50:03.461485 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-gg8zh" event={"ID":"b8310be9-af92-4fa0-a6f9-f242ef869fa4","Type":"ContainerDied","Data":"16bd19c7b7e3cfacc2546ef655629f90b77fd41b82c1b1e3d599fbd064f6818c"} Nov 24 14:50:03 crc kubenswrapper[4790]: I1124 14:50:03.842306 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-qgdgj" Nov 24 14:50:03 crc kubenswrapper[4790]: I1124 14:50:03.935588 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/c5ef870d-41ea-48e7-9a44-3f4a610f0a62-config-data-merged\") pod \"c5ef870d-41ea-48e7-9a44-3f4a610f0a62\" (UID: \"c5ef870d-41ea-48e7-9a44-3f4a610f0a62\") " Nov 24 14:50:03 crc kubenswrapper[4790]: I1124 14:50:03.936166 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5ef870d-41ea-48e7-9a44-3f4a610f0a62-combined-ca-bundle\") pod \"c5ef870d-41ea-48e7-9a44-3f4a610f0a62\" (UID: \"c5ef870d-41ea-48e7-9a44-3f4a610f0a62\") " Nov 24 14:50:03 crc kubenswrapper[4790]: I1124 14:50:03.936292 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5ef870d-41ea-48e7-9a44-3f4a610f0a62-config-data\") pod \"c5ef870d-41ea-48e7-9a44-3f4a610f0a62\" (UID: \"c5ef870d-41ea-48e7-9a44-3f4a610f0a62\") " Nov 24 14:50:03 crc kubenswrapper[4790]: I1124 14:50:03.936393 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5ef870d-41ea-48e7-9a44-3f4a610f0a62-scripts\") pod \"c5ef870d-41ea-48e7-9a44-3f4a610f0a62\" (UID: \"c5ef870d-41ea-48e7-9a44-3f4a610f0a62\") " Nov 24 14:50:03 crc kubenswrapper[4790]: I1124 14:50:03.941985 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5ef870d-41ea-48e7-9a44-3f4a610f0a62-config-data" (OuterVolumeSpecName: "config-data") pod "c5ef870d-41ea-48e7-9a44-3f4a610f0a62" (UID: "c5ef870d-41ea-48e7-9a44-3f4a610f0a62"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:50:03 crc kubenswrapper[4790]: I1124 14:50:03.962073 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5ef870d-41ea-48e7-9a44-3f4a610f0a62-scripts" (OuterVolumeSpecName: "scripts") pod "c5ef870d-41ea-48e7-9a44-3f4a610f0a62" (UID: "c5ef870d-41ea-48e7-9a44-3f4a610f0a62"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:50:03 crc kubenswrapper[4790]: I1124 14:50:03.965697 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5ef870d-41ea-48e7-9a44-3f4a610f0a62-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c5ef870d-41ea-48e7-9a44-3f4a610f0a62" (UID: "c5ef870d-41ea-48e7-9a44-3f4a610f0a62"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:50:03 crc kubenswrapper[4790]: I1124 14:50:03.980792 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5ef870d-41ea-48e7-9a44-3f4a610f0a62-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "c5ef870d-41ea-48e7-9a44-3f4a610f0a62" (UID: "c5ef870d-41ea-48e7-9a44-3f4a610f0a62"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:50:04 crc kubenswrapper[4790]: I1124 14:50:04.038688 4790 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/c5ef870d-41ea-48e7-9a44-3f4a610f0a62-config-data-merged\") on node \"crc\" DevicePath \"\"" Nov 24 14:50:04 crc kubenswrapper[4790]: I1124 14:50:04.038927 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5ef870d-41ea-48e7-9a44-3f4a610f0a62-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:50:04 crc kubenswrapper[4790]: I1124 14:50:04.039020 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5ef870d-41ea-48e7-9a44-3f4a610f0a62-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:50:04 crc kubenswrapper[4790]: I1124 14:50:04.039076 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5ef870d-41ea-48e7-9a44-3f4a610f0a62-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:50:04 crc kubenswrapper[4790]: I1124 14:50:04.484931 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-qgdgj" event={"ID":"c5ef870d-41ea-48e7-9a44-3f4a610f0a62","Type":"ContainerDied","Data":"1cf3ccceebf6e420fe6b081f8d1b364484f32789d07525dfaafaa4d05f154fd4"} Nov 24 14:50:04 crc kubenswrapper[4790]: I1124 14:50:04.484977 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1cf3ccceebf6e420fe6b081f8d1b364484f32789d07525dfaafaa4d05f154fd4" Nov 24 14:50:04 crc kubenswrapper[4790]: I1124 14:50:04.485110 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-qgdgj" Nov 24 14:50:05 crc kubenswrapper[4790]: I1124 14:50:05.588271 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-rsyslog-7rfsl" Nov 24 14:50:06 crc kubenswrapper[4790]: I1124 14:50:06.518696 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-gg8zh" event={"ID":"b8310be9-af92-4fa0-a6f9-f242ef869fa4","Type":"ContainerStarted","Data":"15e5a31d74c4c30c69536a0f4c1da61589aca601bd1064940c38315f632f7766"} Nov 24 14:50:06 crc kubenswrapper[4790]: I1124 14:50:06.539871 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-image-upload-59f8cff499-gg8zh" podStartSLOduration=2.4587967969999998 podStartE2EDuration="16.539852704s" podCreationTimestamp="2025-11-24 14:49:50 +0000 UTC" firstStartedPulling="2025-11-24 14:49:51.742646764 +0000 UTC m=+5840.122540426" lastFinishedPulling="2025-11-24 14:50:05.823702661 +0000 UTC m=+5854.203596333" observedRunningTime="2025-11-24 14:50:06.53624372 +0000 UTC m=+5854.916137392" watchObservedRunningTime="2025-11-24 14:50:06.539852704 +0000 UTC m=+5854.919746366" Nov 24 14:50:30 crc kubenswrapper[4790]: I1124 14:50:30.657789 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-gg8zh"] Nov 24 14:50:30 crc kubenswrapper[4790]: I1124 14:50:30.658425 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/octavia-image-upload-59f8cff499-gg8zh" podUID="b8310be9-af92-4fa0-a6f9-f242ef869fa4" containerName="octavia-amphora-httpd" containerID="cri-o://15e5a31d74c4c30c69536a0f4c1da61589aca601bd1064940c38315f632f7766" gracePeriod=30 Nov 24 14:50:31 crc kubenswrapper[4790]: I1124 14:50:31.284499 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-gg8zh" Nov 24 14:50:31 crc kubenswrapper[4790]: I1124 14:50:31.364726 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/b8310be9-af92-4fa0-a6f9-f242ef869fa4-amphora-image\") pod \"b8310be9-af92-4fa0-a6f9-f242ef869fa4\" (UID: \"b8310be9-af92-4fa0-a6f9-f242ef869fa4\") " Nov 24 14:50:31 crc kubenswrapper[4790]: I1124 14:50:31.364846 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b8310be9-af92-4fa0-a6f9-f242ef869fa4-httpd-config\") pod \"b8310be9-af92-4fa0-a6f9-f242ef869fa4\" (UID: \"b8310be9-af92-4fa0-a6f9-f242ef869fa4\") " Nov 24 14:50:31 crc kubenswrapper[4790]: I1124 14:50:31.396257 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8310be9-af92-4fa0-a6f9-f242ef869fa4-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "b8310be9-af92-4fa0-a6f9-f242ef869fa4" (UID: "b8310be9-af92-4fa0-a6f9-f242ef869fa4"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:50:31 crc kubenswrapper[4790]: I1124 14:50:31.454793 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8310be9-af92-4fa0-a6f9-f242ef869fa4-amphora-image" (OuterVolumeSpecName: "amphora-image") pod "b8310be9-af92-4fa0-a6f9-f242ef869fa4" (UID: "b8310be9-af92-4fa0-a6f9-f242ef869fa4"). InnerVolumeSpecName "amphora-image". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:50:31 crc kubenswrapper[4790]: I1124 14:50:31.467093 4790 reconciler_common.go:293] "Volume detached for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/b8310be9-af92-4fa0-a6f9-f242ef869fa4-amphora-image\") on node \"crc\" DevicePath \"\"" Nov 24 14:50:31 crc kubenswrapper[4790]: I1124 14:50:31.467160 4790 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b8310be9-af92-4fa0-a6f9-f242ef869fa4-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:50:31 crc kubenswrapper[4790]: I1124 14:50:31.791816 4790 generic.go:334] "Generic (PLEG): container finished" podID="b8310be9-af92-4fa0-a6f9-f242ef869fa4" containerID="15e5a31d74c4c30c69536a0f4c1da61589aca601bd1064940c38315f632f7766" exitCode=0 Nov 24 14:50:31 crc kubenswrapper[4790]: I1124 14:50:31.791873 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-gg8zh" event={"ID":"b8310be9-af92-4fa0-a6f9-f242ef869fa4","Type":"ContainerDied","Data":"15e5a31d74c4c30c69536a0f4c1da61589aca601bd1064940c38315f632f7766"} Nov 24 14:50:31 crc kubenswrapper[4790]: I1124 14:50:31.791903 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-gg8zh" Nov 24 14:50:31 crc kubenswrapper[4790]: I1124 14:50:31.791918 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-gg8zh" event={"ID":"b8310be9-af92-4fa0-a6f9-f242ef869fa4","Type":"ContainerDied","Data":"db5175a1dc0f5aff331b477e35cd701eb945c6f17df227b12397a746e22ab28a"} Nov 24 14:50:31 crc kubenswrapper[4790]: I1124 14:50:31.791939 4790 scope.go:117] "RemoveContainer" containerID="15e5a31d74c4c30c69536a0f4c1da61589aca601bd1064940c38315f632f7766" Nov 24 14:50:31 crc kubenswrapper[4790]: I1124 14:50:31.823302 4790 scope.go:117] "RemoveContainer" containerID="16bd19c7b7e3cfacc2546ef655629f90b77fd41b82c1b1e3d599fbd064f6818c" Nov 24 14:50:31 crc kubenswrapper[4790]: I1124 14:50:31.834905 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-gg8zh"] Nov 24 14:50:31 crc kubenswrapper[4790]: I1124 14:50:31.845996 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-gg8zh"] Nov 24 14:50:31 crc kubenswrapper[4790]: I1124 14:50:31.872955 4790 scope.go:117] "RemoveContainer" containerID="15e5a31d74c4c30c69536a0f4c1da61589aca601bd1064940c38315f632f7766" Nov 24 14:50:31 crc kubenswrapper[4790]: E1124 14:50:31.873724 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15e5a31d74c4c30c69536a0f4c1da61589aca601bd1064940c38315f632f7766\": container with ID starting with 15e5a31d74c4c30c69536a0f4c1da61589aca601bd1064940c38315f632f7766 not found: ID does not exist" containerID="15e5a31d74c4c30c69536a0f4c1da61589aca601bd1064940c38315f632f7766" Nov 24 14:50:31 crc kubenswrapper[4790]: I1124 14:50:31.873766 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15e5a31d74c4c30c69536a0f4c1da61589aca601bd1064940c38315f632f7766"} err="failed to get container status \"15e5a31d74c4c30c69536a0f4c1da61589aca601bd1064940c38315f632f7766\": rpc error: code = NotFound desc = could not find container \"15e5a31d74c4c30c69536a0f4c1da61589aca601bd1064940c38315f632f7766\": container with ID starting with 15e5a31d74c4c30c69536a0f4c1da61589aca601bd1064940c38315f632f7766 not found: ID does not exist" Nov 24 14:50:31 crc kubenswrapper[4790]: I1124 14:50:31.873792 4790 scope.go:117] "RemoveContainer" containerID="16bd19c7b7e3cfacc2546ef655629f90b77fd41b82c1b1e3d599fbd064f6818c" Nov 24 14:50:31 crc kubenswrapper[4790]: E1124 14:50:31.874077 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16bd19c7b7e3cfacc2546ef655629f90b77fd41b82c1b1e3d599fbd064f6818c\": container with ID starting with 16bd19c7b7e3cfacc2546ef655629f90b77fd41b82c1b1e3d599fbd064f6818c not found: ID does not exist" containerID="16bd19c7b7e3cfacc2546ef655629f90b77fd41b82c1b1e3d599fbd064f6818c" Nov 24 14:50:31 crc kubenswrapper[4790]: I1124 14:50:31.874112 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16bd19c7b7e3cfacc2546ef655629f90b77fd41b82c1b1e3d599fbd064f6818c"} err="failed to get container status \"16bd19c7b7e3cfacc2546ef655629f90b77fd41b82c1b1e3d599fbd064f6818c\": rpc error: code = NotFound desc = could not find container \"16bd19c7b7e3cfacc2546ef655629f90b77fd41b82c1b1e3d599fbd064f6818c\": container with ID starting with 16bd19c7b7e3cfacc2546ef655629f90b77fd41b82c1b1e3d599fbd064f6818c not found: ID does not exist" Nov 24 14:50:32 crc kubenswrapper[4790]: I1124 14:50:32.327621 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8310be9-af92-4fa0-a6f9-f242ef869fa4" path="/var/lib/kubelet/pods/b8310be9-af92-4fa0-a6f9-f242ef869fa4/volumes" Nov 24 14:50:35 crc kubenswrapper[4790]: I1124 14:50:35.058593 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-image-upload-59f8cff499-6pxxp"] Nov 24 14:50:35 crc kubenswrapper[4790]: E1124 14:50:35.059329 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5ef870d-41ea-48e7-9a44-3f4a610f0a62" containerName="init" Nov 24 14:50:35 crc kubenswrapper[4790]: I1124 14:50:35.059343 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5ef870d-41ea-48e7-9a44-3f4a610f0a62" containerName="init" Nov 24 14:50:35 crc kubenswrapper[4790]: E1124 14:50:35.059359 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8310be9-af92-4fa0-a6f9-f242ef869fa4" containerName="init" Nov 24 14:50:35 crc kubenswrapper[4790]: I1124 14:50:35.059366 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8310be9-af92-4fa0-a6f9-f242ef869fa4" containerName="init" Nov 24 14:50:35 crc kubenswrapper[4790]: E1124 14:50:35.059394 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8310be9-af92-4fa0-a6f9-f242ef869fa4" containerName="octavia-amphora-httpd" Nov 24 14:50:35 crc kubenswrapper[4790]: I1124 14:50:35.059402 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8310be9-af92-4fa0-a6f9-f242ef869fa4" containerName="octavia-amphora-httpd" Nov 24 14:50:35 crc kubenswrapper[4790]: E1124 14:50:35.059421 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5ef870d-41ea-48e7-9a44-3f4a610f0a62" containerName="octavia-db-sync" Nov 24 14:50:35 crc kubenswrapper[4790]: I1124 14:50:35.059428 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5ef870d-41ea-48e7-9a44-3f4a610f0a62" containerName="octavia-db-sync" Nov 24 14:50:35 crc kubenswrapper[4790]: I1124 14:50:35.059668 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8310be9-af92-4fa0-a6f9-f242ef869fa4" containerName="octavia-amphora-httpd" Nov 24 14:50:35 crc kubenswrapper[4790]: I1124 14:50:35.059695 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5ef870d-41ea-48e7-9a44-3f4a610f0a62" containerName="octavia-db-sync" Nov 24 14:50:35 crc kubenswrapper[4790]: I1124 14:50:35.060787 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-6pxxp" Nov 24 14:50:35 crc kubenswrapper[4790]: I1124 14:50:35.069335 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-6pxxp"] Nov 24 14:50:35 crc kubenswrapper[4790]: I1124 14:50:35.073874 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Nov 24 14:50:35 crc kubenswrapper[4790]: I1124 14:50:35.152404 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/e7f3d559-a9e1-4cb6-bc14-e85e8df77929-amphora-image\") pod \"octavia-image-upload-59f8cff499-6pxxp\" (UID: \"e7f3d559-a9e1-4cb6-bc14-e85e8df77929\") " pod="openstack/octavia-image-upload-59f8cff499-6pxxp" Nov 24 14:50:35 crc kubenswrapper[4790]: I1124 14:50:35.152521 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e7f3d559-a9e1-4cb6-bc14-e85e8df77929-httpd-config\") pod \"octavia-image-upload-59f8cff499-6pxxp\" (UID: \"e7f3d559-a9e1-4cb6-bc14-e85e8df77929\") " pod="openstack/octavia-image-upload-59f8cff499-6pxxp" Nov 24 14:50:35 crc kubenswrapper[4790]: I1124 14:50:35.254304 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e7f3d559-a9e1-4cb6-bc14-e85e8df77929-httpd-config\") pod \"octavia-image-upload-59f8cff499-6pxxp\" (UID: \"e7f3d559-a9e1-4cb6-bc14-e85e8df77929\") " pod="openstack/octavia-image-upload-59f8cff499-6pxxp" Nov 24 14:50:35 crc kubenswrapper[4790]: I1124 14:50:35.254446 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/e7f3d559-a9e1-4cb6-bc14-e85e8df77929-amphora-image\") pod \"octavia-image-upload-59f8cff499-6pxxp\" (UID: \"e7f3d559-a9e1-4cb6-bc14-e85e8df77929\") " pod="openstack/octavia-image-upload-59f8cff499-6pxxp" Nov 24 14:50:35 crc kubenswrapper[4790]: I1124 14:50:35.254940 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/e7f3d559-a9e1-4cb6-bc14-e85e8df77929-amphora-image\") pod \"octavia-image-upload-59f8cff499-6pxxp\" (UID: \"e7f3d559-a9e1-4cb6-bc14-e85e8df77929\") " pod="openstack/octavia-image-upload-59f8cff499-6pxxp" Nov 24 14:50:35 crc kubenswrapper[4790]: I1124 14:50:35.268094 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e7f3d559-a9e1-4cb6-bc14-e85e8df77929-httpd-config\") pod \"octavia-image-upload-59f8cff499-6pxxp\" (UID: \"e7f3d559-a9e1-4cb6-bc14-e85e8df77929\") " pod="openstack/octavia-image-upload-59f8cff499-6pxxp" Nov 24 14:50:35 crc kubenswrapper[4790]: I1124 14:50:35.394191 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-6pxxp" Nov 24 14:50:35 crc kubenswrapper[4790]: I1124 14:50:35.944980 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-6pxxp"] Nov 24 14:50:36 crc kubenswrapper[4790]: I1124 14:50:36.855396 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-6pxxp" event={"ID":"e7f3d559-a9e1-4cb6-bc14-e85e8df77929","Type":"ContainerStarted","Data":"88f43f92ea1c11b1eaa98eb47edbeb1f1a2fcefe051439145158c1710dec54d6"} Nov 24 14:50:37 crc kubenswrapper[4790]: I1124 14:50:37.873997 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-6pxxp" event={"ID":"e7f3d559-a9e1-4cb6-bc14-e85e8df77929","Type":"ContainerStarted","Data":"afd147d62ce1e308704c5b8cdb253bd50f9bb516a1a6e18436b7e8341ebe6ada"} Nov 24 14:50:38 crc kubenswrapper[4790]: I1124 14:50:38.887262 4790 generic.go:334] "Generic (PLEG): container finished" podID="e7f3d559-a9e1-4cb6-bc14-e85e8df77929" containerID="afd147d62ce1e308704c5b8cdb253bd50f9bb516a1a6e18436b7e8341ebe6ada" exitCode=0 Nov 24 14:50:38 crc kubenswrapper[4790]: I1124 14:50:38.887315 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-6pxxp" event={"ID":"e7f3d559-a9e1-4cb6-bc14-e85e8df77929","Type":"ContainerDied","Data":"afd147d62ce1e308704c5b8cdb253bd50f9bb516a1a6e18436b7e8341ebe6ada"} Nov 24 14:50:40 crc kubenswrapper[4790]: I1124 14:50:40.909558 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-6pxxp" event={"ID":"e7f3d559-a9e1-4cb6-bc14-e85e8df77929","Type":"ContainerStarted","Data":"f26eb3c9fd698423c211d47e9d2e15dbef51305b85ae160b6446cf2e54da2c6e"} Nov 24 14:50:40 crc kubenswrapper[4790]: I1124 14:50:40.933280 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-image-upload-59f8cff499-6pxxp" podStartSLOduration=1.9238937470000002 podStartE2EDuration="5.933262544s" podCreationTimestamp="2025-11-24 14:50:35 +0000 UTC" firstStartedPulling="2025-11-24 14:50:35.950130354 +0000 UTC m=+5884.330024026" lastFinishedPulling="2025-11-24 14:50:39.959499141 +0000 UTC m=+5888.339392823" observedRunningTime="2025-11-24 14:50:40.928828356 +0000 UTC m=+5889.308722018" watchObservedRunningTime="2025-11-24 14:50:40.933262544 +0000 UTC m=+5889.313156206" Nov 24 14:50:55 crc kubenswrapper[4790]: I1124 14:50:55.405662 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-healthmanager-n974m"] Nov 24 14:50:55 crc kubenswrapper[4790]: I1124 14:50:55.413749 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-n974m" Nov 24 14:50:55 crc kubenswrapper[4790]: I1124 14:50:55.423262 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-scripts" Nov 24 14:50:55 crc kubenswrapper[4790]: I1124 14:50:55.423639 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-config-data" Nov 24 14:50:55 crc kubenswrapper[4790]: I1124 14:50:55.424107 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-certs-secret" Nov 24 14:50:55 crc kubenswrapper[4790]: I1124 14:50:55.458252 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-n974m"] Nov 24 14:50:55 crc kubenswrapper[4790]: I1124 14:50:55.535777 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/617d4157-f445-4d62-ac72-9d16b5fd52be-amphora-certs\") pod \"octavia-healthmanager-n974m\" (UID: \"617d4157-f445-4d62-ac72-9d16b5fd52be\") " pod="openstack/octavia-healthmanager-n974m" Nov 24 14:50:55 crc kubenswrapper[4790]: I1124 14:50:55.536460 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/617d4157-f445-4d62-ac72-9d16b5fd52be-config-data\") pod \"octavia-healthmanager-n974m\" (UID: \"617d4157-f445-4d62-ac72-9d16b5fd52be\") " pod="openstack/octavia-healthmanager-n974m" Nov 24 14:50:55 crc kubenswrapper[4790]: I1124 14:50:55.536478 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/617d4157-f445-4d62-ac72-9d16b5fd52be-combined-ca-bundle\") pod \"octavia-healthmanager-n974m\" (UID: \"617d4157-f445-4d62-ac72-9d16b5fd52be\") " pod="openstack/octavia-healthmanager-n974m" Nov 24 14:50:55 crc kubenswrapper[4790]: I1124 14:50:55.536556 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/617d4157-f445-4d62-ac72-9d16b5fd52be-hm-ports\") pod \"octavia-healthmanager-n974m\" (UID: \"617d4157-f445-4d62-ac72-9d16b5fd52be\") " pod="openstack/octavia-healthmanager-n974m" Nov 24 14:50:55 crc kubenswrapper[4790]: I1124 14:50:55.536579 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/617d4157-f445-4d62-ac72-9d16b5fd52be-scripts\") pod \"octavia-healthmanager-n974m\" (UID: \"617d4157-f445-4d62-ac72-9d16b5fd52be\") " pod="openstack/octavia-healthmanager-n974m" Nov 24 14:50:55 crc kubenswrapper[4790]: I1124 14:50:55.536640 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/617d4157-f445-4d62-ac72-9d16b5fd52be-config-data-merged\") pod \"octavia-healthmanager-n974m\" (UID: \"617d4157-f445-4d62-ac72-9d16b5fd52be\") " pod="openstack/octavia-healthmanager-n974m" Nov 24 14:50:55 crc kubenswrapper[4790]: I1124 14:50:55.638829 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/617d4157-f445-4d62-ac72-9d16b5fd52be-hm-ports\") pod \"octavia-healthmanager-n974m\" (UID: \"617d4157-f445-4d62-ac72-9d16b5fd52be\") " pod="openstack/octavia-healthmanager-n974m" Nov 24 14:50:55 crc kubenswrapper[4790]: I1124 14:50:55.638891 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/617d4157-f445-4d62-ac72-9d16b5fd52be-scripts\") pod \"octavia-healthmanager-n974m\" (UID: \"617d4157-f445-4d62-ac72-9d16b5fd52be\") " pod="openstack/octavia-healthmanager-n974m" Nov 24 14:50:55 crc kubenswrapper[4790]: I1124 14:50:55.638949 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/617d4157-f445-4d62-ac72-9d16b5fd52be-config-data-merged\") pod \"octavia-healthmanager-n974m\" (UID: \"617d4157-f445-4d62-ac72-9d16b5fd52be\") " pod="openstack/octavia-healthmanager-n974m" Nov 24 14:50:55 crc kubenswrapper[4790]: I1124 14:50:55.639499 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/617d4157-f445-4d62-ac72-9d16b5fd52be-config-data-merged\") pod \"octavia-healthmanager-n974m\" (UID: \"617d4157-f445-4d62-ac72-9d16b5fd52be\") " pod="openstack/octavia-healthmanager-n974m" Nov 24 14:50:55 crc kubenswrapper[4790]: I1124 14:50:55.639981 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/617d4157-f445-4d62-ac72-9d16b5fd52be-amphora-certs\") pod \"octavia-healthmanager-n974m\" (UID: \"617d4157-f445-4d62-ac72-9d16b5fd52be\") " pod="openstack/octavia-healthmanager-n974m" Nov 24 14:50:55 crc kubenswrapper[4790]: I1124 14:50:55.640019 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/617d4157-f445-4d62-ac72-9d16b5fd52be-config-data\") pod \"octavia-healthmanager-n974m\" (UID: \"617d4157-f445-4d62-ac72-9d16b5fd52be\") " pod="openstack/octavia-healthmanager-n974m" Nov 24 14:50:55 crc kubenswrapper[4790]: I1124 14:50:55.640037 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/617d4157-f445-4d62-ac72-9d16b5fd52be-combined-ca-bundle\") pod \"octavia-healthmanager-n974m\" (UID: \"617d4157-f445-4d62-ac72-9d16b5fd52be\") " pod="openstack/octavia-healthmanager-n974m" Nov 24 14:50:55 crc kubenswrapper[4790]: I1124 14:50:55.640173 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/617d4157-f445-4d62-ac72-9d16b5fd52be-hm-ports\") pod \"octavia-healthmanager-n974m\" (UID: \"617d4157-f445-4d62-ac72-9d16b5fd52be\") " pod="openstack/octavia-healthmanager-n974m" Nov 24 14:50:55 crc kubenswrapper[4790]: I1124 14:50:55.644959 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/617d4157-f445-4d62-ac72-9d16b5fd52be-amphora-certs\") pod \"octavia-healthmanager-n974m\" (UID: \"617d4157-f445-4d62-ac72-9d16b5fd52be\") " pod="openstack/octavia-healthmanager-n974m" Nov 24 14:50:55 crc kubenswrapper[4790]: I1124 14:50:55.645836 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/617d4157-f445-4d62-ac72-9d16b5fd52be-scripts\") pod \"octavia-healthmanager-n974m\" (UID: \"617d4157-f445-4d62-ac72-9d16b5fd52be\") " pod="openstack/octavia-healthmanager-n974m" Nov 24 14:50:55 crc kubenswrapper[4790]: I1124 14:50:55.646804 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/617d4157-f445-4d62-ac72-9d16b5fd52be-config-data\") pod \"octavia-healthmanager-n974m\" (UID: \"617d4157-f445-4d62-ac72-9d16b5fd52be\") " pod="openstack/octavia-healthmanager-n974m" Nov 24 14:50:55 crc kubenswrapper[4790]: I1124 14:50:55.663255 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/617d4157-f445-4d62-ac72-9d16b5fd52be-combined-ca-bundle\") pod \"octavia-healthmanager-n974m\" (UID: \"617d4157-f445-4d62-ac72-9d16b5fd52be\") " pod="openstack/octavia-healthmanager-n974m" Nov 24 14:50:55 crc kubenswrapper[4790]: I1124 14:50:55.756767 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-n974m" Nov 24 14:50:56 crc kubenswrapper[4790]: I1124 14:50:56.419970 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-n974m"] Nov 24 14:50:57 crc kubenswrapper[4790]: I1124 14:50:57.048115 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-housekeeping-kfk5m"] Nov 24 14:50:57 crc kubenswrapper[4790]: I1124 14:50:57.050486 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-kfk5m" Nov 24 14:50:57 crc kubenswrapper[4790]: I1124 14:50:57.053426 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-config-data" Nov 24 14:50:57 crc kubenswrapper[4790]: I1124 14:50:57.053453 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-scripts" Nov 24 14:50:57 crc kubenswrapper[4790]: I1124 14:50:57.060963 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-kfk5m"] Nov 24 14:50:57 crc kubenswrapper[4790]: I1124 14:50:57.094232 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-n974m" event={"ID":"617d4157-f445-4d62-ac72-9d16b5fd52be","Type":"ContainerStarted","Data":"eaa689ab1d24e199354dec9751a561519da2d75b9928674f5cc25121d8021a00"} Nov 24 14:50:57 crc kubenswrapper[4790]: I1124 14:50:57.094285 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-n974m" event={"ID":"617d4157-f445-4d62-ac72-9d16b5fd52be","Type":"ContainerStarted","Data":"9c91d598f0c709f7cc6059a046adc906de98cb6ef240df69791c752cb23826e5"} Nov 24 14:50:57 crc kubenswrapper[4790]: I1124 14:50:57.168484 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427-combined-ca-bundle\") pod \"octavia-housekeeping-kfk5m\" (UID: \"2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427\") " pod="openstack/octavia-housekeeping-kfk5m" Nov 24 14:50:57 crc kubenswrapper[4790]: I1124 14:50:57.168532 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427-scripts\") pod \"octavia-housekeeping-kfk5m\" (UID: \"2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427\") " pod="openstack/octavia-housekeeping-kfk5m" Nov 24 14:50:57 crc kubenswrapper[4790]: I1124 14:50:57.168566 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427-config-data-merged\") pod \"octavia-housekeeping-kfk5m\" (UID: \"2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427\") " pod="openstack/octavia-housekeeping-kfk5m" Nov 24 14:50:57 crc kubenswrapper[4790]: I1124 14:50:57.168603 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427-hm-ports\") pod \"octavia-housekeeping-kfk5m\" (UID: \"2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427\") " pod="openstack/octavia-housekeeping-kfk5m" Nov 24 14:50:57 crc kubenswrapper[4790]: I1124 14:50:57.168704 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427-config-data\") pod \"octavia-housekeeping-kfk5m\" (UID: \"2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427\") " pod="openstack/octavia-housekeeping-kfk5m" Nov 24 14:50:57 crc kubenswrapper[4790]: I1124 14:50:57.168826 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427-amphora-certs\") pod \"octavia-housekeeping-kfk5m\" (UID: \"2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427\") " pod="openstack/octavia-housekeeping-kfk5m" Nov 24 14:50:57 crc kubenswrapper[4790]: I1124 14:50:57.270427 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427-combined-ca-bundle\") pod \"octavia-housekeeping-kfk5m\" (UID: \"2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427\") " pod="openstack/octavia-housekeeping-kfk5m" Nov 24 14:50:57 crc kubenswrapper[4790]: I1124 14:50:57.270486 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427-scripts\") pod \"octavia-housekeeping-kfk5m\" (UID: \"2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427\") " pod="openstack/octavia-housekeeping-kfk5m" Nov 24 14:50:57 crc kubenswrapper[4790]: I1124 14:50:57.270523 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427-config-data-merged\") pod \"octavia-housekeeping-kfk5m\" (UID: \"2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427\") " pod="openstack/octavia-housekeeping-kfk5m" Nov 24 14:50:57 crc kubenswrapper[4790]: I1124 14:50:57.270565 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427-hm-ports\") pod \"octavia-housekeeping-kfk5m\" (UID: \"2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427\") " pod="openstack/octavia-housekeeping-kfk5m" Nov 24 14:50:57 crc kubenswrapper[4790]: I1124 14:50:57.270623 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427-config-data\") pod \"octavia-housekeeping-kfk5m\" (UID: \"2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427\") " pod="openstack/octavia-housekeeping-kfk5m" Nov 24 14:50:57 crc kubenswrapper[4790]: I1124 14:50:57.270749 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427-amphora-certs\") pod \"octavia-housekeeping-kfk5m\" (UID: \"2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427\") " pod="openstack/octavia-housekeeping-kfk5m" Nov 24 14:50:57 crc kubenswrapper[4790]: I1124 14:50:57.271459 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427-config-data-merged\") pod \"octavia-housekeeping-kfk5m\" (UID: \"2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427\") " pod="openstack/octavia-housekeeping-kfk5m" Nov 24 14:50:57 crc kubenswrapper[4790]: I1124 14:50:57.272169 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427-hm-ports\") pod \"octavia-housekeeping-kfk5m\" (UID: \"2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427\") " pod="openstack/octavia-housekeeping-kfk5m" Nov 24 14:50:57 crc kubenswrapper[4790]: I1124 14:50:57.277117 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427-combined-ca-bundle\") pod \"octavia-housekeeping-kfk5m\" (UID: \"2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427\") " pod="openstack/octavia-housekeeping-kfk5m" Nov 24 14:50:57 crc kubenswrapper[4790]: I1124 14:50:57.278029 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427-amphora-certs\") pod \"octavia-housekeeping-kfk5m\" (UID: \"2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427\") " pod="openstack/octavia-housekeeping-kfk5m" Nov 24 14:50:57 crc kubenswrapper[4790]: I1124 14:50:57.278489 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427-config-data\") pod \"octavia-housekeeping-kfk5m\" (UID: \"2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427\") " pod="openstack/octavia-housekeeping-kfk5m" Nov 24 14:50:57 crc kubenswrapper[4790]: I1124 14:50:57.302173 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427-scripts\") pod \"octavia-housekeeping-kfk5m\" (UID: \"2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427\") " pod="openstack/octavia-housekeeping-kfk5m" Nov 24 14:50:57 crc kubenswrapper[4790]: I1124 14:50:57.376812 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-kfk5m" Nov 24 14:50:58 crc kubenswrapper[4790]: I1124 14:50:58.101292 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-kfk5m"] Nov 24 14:50:58 crc kubenswrapper[4790]: I1124 14:50:58.232718 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-worker-fkc29"] Nov 24 14:50:58 crc kubenswrapper[4790]: I1124 14:50:58.234748 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-fkc29" Nov 24 14:50:58 crc kubenswrapper[4790]: I1124 14:50:58.242129 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-scripts" Nov 24 14:50:58 crc kubenswrapper[4790]: I1124 14:50:58.242152 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-config-data" Nov 24 14:50:58 crc kubenswrapper[4790]: I1124 14:50:58.246992 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-fkc29"] Nov 24 14:50:58 crc kubenswrapper[4790]: I1124 14:50:58.297465 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/3aa341b2-1c7e-4827-9d80-9a26d5b46fa9-amphora-certs\") pod \"octavia-worker-fkc29\" (UID: \"3aa341b2-1c7e-4827-9d80-9a26d5b46fa9\") " pod="openstack/octavia-worker-fkc29" Nov 24 14:50:58 crc kubenswrapper[4790]: I1124 14:50:58.297553 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3aa341b2-1c7e-4827-9d80-9a26d5b46fa9-config-data\") pod \"octavia-worker-fkc29\" (UID: \"3aa341b2-1c7e-4827-9d80-9a26d5b46fa9\") " pod="openstack/octavia-worker-fkc29" Nov 24 14:50:58 crc kubenswrapper[4790]: I1124 14:50:58.297589 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/3aa341b2-1c7e-4827-9d80-9a26d5b46fa9-config-data-merged\") pod \"octavia-worker-fkc29\" (UID: \"3aa341b2-1c7e-4827-9d80-9a26d5b46fa9\") " pod="openstack/octavia-worker-fkc29" Nov 24 14:50:58 crc kubenswrapper[4790]: I1124 14:50:58.297670 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/3aa341b2-1c7e-4827-9d80-9a26d5b46fa9-hm-ports\") pod \"octavia-worker-fkc29\" (UID: \"3aa341b2-1c7e-4827-9d80-9a26d5b46fa9\") " pod="openstack/octavia-worker-fkc29" Nov 24 14:50:58 crc kubenswrapper[4790]: I1124 14:50:58.297754 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa341b2-1c7e-4827-9d80-9a26d5b46fa9-combined-ca-bundle\") pod \"octavia-worker-fkc29\" (UID: \"3aa341b2-1c7e-4827-9d80-9a26d5b46fa9\") " pod="openstack/octavia-worker-fkc29" Nov 24 14:50:58 crc kubenswrapper[4790]: I1124 14:50:58.297794 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3aa341b2-1c7e-4827-9d80-9a26d5b46fa9-scripts\") pod \"octavia-worker-fkc29\" (UID: \"3aa341b2-1c7e-4827-9d80-9a26d5b46fa9\") " pod="openstack/octavia-worker-fkc29" Nov 24 14:50:58 crc kubenswrapper[4790]: I1124 14:50:58.399213 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/3aa341b2-1c7e-4827-9d80-9a26d5b46fa9-amphora-certs\") pod \"octavia-worker-fkc29\" (UID: \"3aa341b2-1c7e-4827-9d80-9a26d5b46fa9\") " pod="openstack/octavia-worker-fkc29" Nov 24 14:50:58 crc kubenswrapper[4790]: I1124 14:50:58.399284 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3aa341b2-1c7e-4827-9d80-9a26d5b46fa9-config-data\") pod \"octavia-worker-fkc29\" (UID: \"3aa341b2-1c7e-4827-9d80-9a26d5b46fa9\") " pod="openstack/octavia-worker-fkc29" Nov 24 14:50:58 crc kubenswrapper[4790]: I1124 14:50:58.399315 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/3aa341b2-1c7e-4827-9d80-9a26d5b46fa9-config-data-merged\") pod \"octavia-worker-fkc29\" (UID: \"3aa341b2-1c7e-4827-9d80-9a26d5b46fa9\") " pod="openstack/octavia-worker-fkc29" Nov 24 14:50:58 crc kubenswrapper[4790]: I1124 14:50:58.399394 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/3aa341b2-1c7e-4827-9d80-9a26d5b46fa9-hm-ports\") pod \"octavia-worker-fkc29\" (UID: \"3aa341b2-1c7e-4827-9d80-9a26d5b46fa9\") " pod="openstack/octavia-worker-fkc29" Nov 24 14:50:58 crc kubenswrapper[4790]: I1124 14:50:58.399472 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa341b2-1c7e-4827-9d80-9a26d5b46fa9-combined-ca-bundle\") pod \"octavia-worker-fkc29\" (UID: \"3aa341b2-1c7e-4827-9d80-9a26d5b46fa9\") " pod="openstack/octavia-worker-fkc29" Nov 24 14:50:58 crc kubenswrapper[4790]: I1124 14:50:58.399504 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3aa341b2-1c7e-4827-9d80-9a26d5b46fa9-scripts\") pod \"octavia-worker-fkc29\" (UID: \"3aa341b2-1c7e-4827-9d80-9a26d5b46fa9\") " pod="openstack/octavia-worker-fkc29" Nov 24 14:50:58 crc kubenswrapper[4790]: I1124 14:50:58.400417 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/3aa341b2-1c7e-4827-9d80-9a26d5b46fa9-config-data-merged\") pod \"octavia-worker-fkc29\" (UID: \"3aa341b2-1c7e-4827-9d80-9a26d5b46fa9\") " pod="openstack/octavia-worker-fkc29" Nov 24 14:50:58 crc kubenswrapper[4790]: I1124 14:50:58.401004 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/3aa341b2-1c7e-4827-9d80-9a26d5b46fa9-hm-ports\") pod \"octavia-worker-fkc29\" (UID: \"3aa341b2-1c7e-4827-9d80-9a26d5b46fa9\") " pod="openstack/octavia-worker-fkc29" Nov 24 14:50:58 crc kubenswrapper[4790]: I1124 14:50:58.406175 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa341b2-1c7e-4827-9d80-9a26d5b46fa9-combined-ca-bundle\") pod \"octavia-worker-fkc29\" (UID: \"3aa341b2-1c7e-4827-9d80-9a26d5b46fa9\") " pod="openstack/octavia-worker-fkc29" Nov 24 14:50:58 crc kubenswrapper[4790]: I1124 14:50:58.406618 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3aa341b2-1c7e-4827-9d80-9a26d5b46fa9-scripts\") pod \"octavia-worker-fkc29\" (UID: \"3aa341b2-1c7e-4827-9d80-9a26d5b46fa9\") " pod="openstack/octavia-worker-fkc29" Nov 24 14:50:58 crc kubenswrapper[4790]: I1124 14:50:58.413691 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/3aa341b2-1c7e-4827-9d80-9a26d5b46fa9-amphora-certs\") pod \"octavia-worker-fkc29\" (UID: \"3aa341b2-1c7e-4827-9d80-9a26d5b46fa9\") " pod="openstack/octavia-worker-fkc29" Nov 24 14:50:58 crc kubenswrapper[4790]: I1124 14:50:58.413727 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3aa341b2-1c7e-4827-9d80-9a26d5b46fa9-config-data\") pod \"octavia-worker-fkc29\" (UID: \"3aa341b2-1c7e-4827-9d80-9a26d5b46fa9\") " pod="openstack/octavia-worker-fkc29" Nov 24 14:50:58 crc kubenswrapper[4790]: I1124 14:50:58.555756 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-fkc29" Nov 24 14:50:59 crc kubenswrapper[4790]: I1124 14:50:59.115870 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-fkc29"] Nov 24 14:50:59 crc kubenswrapper[4790]: W1124 14:50:59.128111 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3aa341b2_1c7e_4827_9d80_9a26d5b46fa9.slice/crio-8af68dc4b7a614ef8b42eedd1b6999d839f78b4e890bc92e691873e3b9456646 WatchSource:0}: Error finding container 8af68dc4b7a614ef8b42eedd1b6999d839f78b4e890bc92e691873e3b9456646: Status 404 returned error can't find the container with id 8af68dc4b7a614ef8b42eedd1b6999d839f78b4e890bc92e691873e3b9456646 Nov 24 14:50:59 crc kubenswrapper[4790]: I1124 14:50:59.129804 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-kfk5m" event={"ID":"2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427","Type":"ContainerStarted","Data":"9fea6473d48e57be4f22655cfacc052f6ba615c866e43323fc2d1608e7f57b90"} Nov 24 14:50:59 crc kubenswrapper[4790]: I1124 14:50:59.133423 4790 generic.go:334] "Generic (PLEG): container finished" podID="617d4157-f445-4d62-ac72-9d16b5fd52be" containerID="eaa689ab1d24e199354dec9751a561519da2d75b9928674f5cc25121d8021a00" exitCode=0 Nov 24 14:50:59 crc kubenswrapper[4790]: I1124 14:50:59.133504 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-n974m" event={"ID":"617d4157-f445-4d62-ac72-9d16b5fd52be","Type":"ContainerDied","Data":"eaa689ab1d24e199354dec9751a561519da2d75b9928674f5cc25121d8021a00"} Nov 24 14:51:00 crc kubenswrapper[4790]: I1124 14:51:00.152530 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-n974m" event={"ID":"617d4157-f445-4d62-ac72-9d16b5fd52be","Type":"ContainerStarted","Data":"857c90c8eb1aa498e603e5c2cab62d37cda13aab81c6b12b9e08e76f79104951"} Nov 24 14:51:00 crc kubenswrapper[4790]: I1124 14:51:00.154510 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-healthmanager-n974m" Nov 24 14:51:00 crc kubenswrapper[4790]: I1124 14:51:00.175208 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-fkc29" event={"ID":"3aa341b2-1c7e-4827-9d80-9a26d5b46fa9","Type":"ContainerStarted","Data":"8af68dc4b7a614ef8b42eedd1b6999d839f78b4e890bc92e691873e3b9456646"} Nov 24 14:51:00 crc kubenswrapper[4790]: I1124 14:51:00.182649 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-healthmanager-n974m" podStartSLOduration=5.182623171 podStartE2EDuration="5.182623171s" podCreationTimestamp="2025-11-24 14:50:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:51:00.179391948 +0000 UTC m=+5908.559285610" watchObservedRunningTime="2025-11-24 14:51:00.182623171 +0000 UTC m=+5908.562516843" Nov 24 14:51:04 crc kubenswrapper[4790]: I1124 14:51:04.226908 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-kfk5m" event={"ID":"2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427","Type":"ContainerStarted","Data":"18c8c77eab00fe9a6e01993c3d8b31ba01b2fd9f05b6c840615ecb81d6ecec69"} Nov 24 14:51:05 crc kubenswrapper[4790]: I1124 14:51:05.237384 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-fkc29" event={"ID":"3aa341b2-1c7e-4827-9d80-9a26d5b46fa9","Type":"ContainerStarted","Data":"ecd0beb70d5ecc237a95182a447d4289fca4fdb8bb3ac0d673df66c071d26810"} Nov 24 14:51:05 crc kubenswrapper[4790]: I1124 14:51:05.238860 4790 generic.go:334] "Generic (PLEG): container finished" podID="2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427" containerID="18c8c77eab00fe9a6e01993c3d8b31ba01b2fd9f05b6c840615ecb81d6ecec69" exitCode=0 Nov 24 14:51:05 crc kubenswrapper[4790]: I1124 14:51:05.238914 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-kfk5m" event={"ID":"2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427","Type":"ContainerDied","Data":"18c8c77eab00fe9a6e01993c3d8b31ba01b2fd9f05b6c840615ecb81d6ecec69"} Nov 24 14:51:06 crc kubenswrapper[4790]: I1124 14:51:06.272829 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-kfk5m" event={"ID":"2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427","Type":"ContainerStarted","Data":"abe749970daf1d9aa12ef99b3cee3ae1df20e82a9fbeccda6ba95532d95692af"} Nov 24 14:51:06 crc kubenswrapper[4790]: I1124 14:51:06.275107 4790 generic.go:334] "Generic (PLEG): container finished" podID="3aa341b2-1c7e-4827-9d80-9a26d5b46fa9" containerID="ecd0beb70d5ecc237a95182a447d4289fca4fdb8bb3ac0d673df66c071d26810" exitCode=0 Nov 24 14:51:06 crc kubenswrapper[4790]: I1124 14:51:06.275153 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-fkc29" event={"ID":"3aa341b2-1c7e-4827-9d80-9a26d5b46fa9","Type":"ContainerDied","Data":"ecd0beb70d5ecc237a95182a447d4289fca4fdb8bb3ac0d673df66c071d26810"} Nov 24 14:51:06 crc kubenswrapper[4790]: I1124 14:51:06.297381 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-housekeeping-kfk5m" podStartSLOduration=4.509497348 podStartE2EDuration="9.297357178s" podCreationTimestamp="2025-11-24 14:50:57 +0000 UTC" firstStartedPulling="2025-11-24 14:50:58.119478513 +0000 UTC m=+5906.499372185" lastFinishedPulling="2025-11-24 14:51:02.907338343 +0000 UTC m=+5911.287232015" observedRunningTime="2025-11-24 14:51:06.295584578 +0000 UTC m=+5914.675478250" watchObservedRunningTime="2025-11-24 14:51:06.297357178 +0000 UTC m=+5914.677250850" Nov 24 14:51:07 crc kubenswrapper[4790]: I1124 14:51:07.287053 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-fkc29" event={"ID":"3aa341b2-1c7e-4827-9d80-9a26d5b46fa9","Type":"ContainerStarted","Data":"569685d8b63a912a66b7a01cd37a2df545f8ac96e5e6d35a930644df3ea604e2"} Nov 24 14:51:07 crc kubenswrapper[4790]: I1124 14:51:07.287722 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-housekeeping-kfk5m" Nov 24 14:51:07 crc kubenswrapper[4790]: I1124 14:51:07.325972 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-worker-fkc29" podStartSLOduration=3.973666911 podStartE2EDuration="9.325954817s" podCreationTimestamp="2025-11-24 14:50:58 +0000 UTC" firstStartedPulling="2025-11-24 14:50:59.134084219 +0000 UTC m=+5907.513977881" lastFinishedPulling="2025-11-24 14:51:04.486372115 +0000 UTC m=+5912.866265787" observedRunningTime="2025-11-24 14:51:07.319525603 +0000 UTC m=+5915.699419365" watchObservedRunningTime="2025-11-24 14:51:07.325954817 +0000 UTC m=+5915.705848479" Nov 24 14:51:08 crc kubenswrapper[4790]: I1124 14:51:08.296982 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-worker-fkc29" Nov 24 14:51:10 crc kubenswrapper[4790]: I1124 14:51:10.814417 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-healthmanager-n974m" Nov 24 14:51:12 crc kubenswrapper[4790]: I1124 14:51:12.446086 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-housekeeping-kfk5m" Nov 24 14:51:13 crc kubenswrapper[4790]: I1124 14:51:13.602450 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-worker-fkc29" Nov 24 14:51:40 crc kubenswrapper[4790]: I1124 14:51:40.069175 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-60fd-account-create-h4ql2"] Nov 24 14:51:40 crc kubenswrapper[4790]: I1124 14:51:40.085057 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-qvmkw"] Nov 24 14:51:40 crc kubenswrapper[4790]: I1124 14:51:40.099840 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-qvmkw"] Nov 24 14:51:40 crc kubenswrapper[4790]: I1124 14:51:40.110846 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-60fd-account-create-h4ql2"] Nov 24 14:51:40 crc kubenswrapper[4790]: I1124 14:51:40.340468 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7e7c18a-1b25-4d46-8cc5-77110ecc453d" path="/var/lib/kubelet/pods/b7e7c18a-1b25-4d46-8cc5-77110ecc453d/volumes" Nov 24 14:51:40 crc kubenswrapper[4790]: I1124 14:51:40.343716 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6dbd003-4923-4aee-a240-7debd90a4a5b" path="/var/lib/kubelet/pods/e6dbd003-4923-4aee-a240-7debd90a4a5b/volumes" Nov 24 14:51:43 crc kubenswrapper[4790]: I1124 14:51:43.938318 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:51:43 crc kubenswrapper[4790]: I1124 14:51:43.938944 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:51:47 crc kubenswrapper[4790]: I1124 14:51:47.049701 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-bqxsj"] Nov 24 14:51:47 crc kubenswrapper[4790]: I1124 14:51:47.067552 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-bqxsj"] Nov 24 14:51:48 crc kubenswrapper[4790]: I1124 14:51:48.334239 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb099850-58c3-4280-8c66-865463ca19c8" path="/var/lib/kubelet/pods/bb099850-58c3-4280-8c66-865463ca19c8/volumes" Nov 24 14:51:49 crc kubenswrapper[4790]: I1124 14:51:49.239320 4790 scope.go:117] "RemoveContainer" containerID="6588be2e81cf6ac91d6145cc3944b2885e9e86a11a03dbf6ad7632b91ca0821d" Nov 24 14:51:49 crc kubenswrapper[4790]: I1124 14:51:49.332245 4790 scope.go:117] "RemoveContainer" containerID="355778f4672909263ed17cc3bf2bd12f174b827d9d95c64ad4a0d4542fa66ea1" Nov 24 14:51:49 crc kubenswrapper[4790]: I1124 14:51:49.372444 4790 scope.go:117] "RemoveContainer" containerID="0f4d8c1d4bbda20530690ce310bf02b9c1520de832d550f6e6e9cb7c8562249d" Nov 24 14:51:49 crc kubenswrapper[4790]: I1124 14:51:49.456421 4790 scope.go:117] "RemoveContainer" containerID="d7d3f25a0c0ea33125657769d27d12e5822cf9c22a591dfcc4cfcc7f19e7cd10" Nov 24 14:51:49 crc kubenswrapper[4790]: I1124 14:51:49.479976 4790 scope.go:117] "RemoveContainer" containerID="b84fe988a3b86185ef928788e240cae9609c3919645f171e1dc828daefa3244d" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.515443 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-58dfc66457-rqfks"] Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.524234 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-58dfc66457-rqfks" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.527204 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.527448 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-vv697" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.527559 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.527674 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.539357 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-58dfc66457-rqfks"] Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.581899 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.582114 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="4b8cad57-cddc-4273-a1be-498c1172ee79" containerName="glance-log" containerID="cri-o://6b3329faf3f91547a252762a0890d5dba6b36153a7b56704b3a84abd4e8aeeb3" gracePeriod=30 Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.582244 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="4b8cad57-cddc-4273-a1be-498c1172ee79" containerName="glance-httpd" containerID="cri-o://c65baf0b9f2f5600591225f558e76d78f8d8dd850c6290154a8a5b152f89bf66" gracePeriod=30 Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.631258 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f78f1722-568a-4643-9e7b-1d2bb2449456-scripts\") pod \"horizon-58dfc66457-rqfks\" (UID: \"f78f1722-568a-4643-9e7b-1d2bb2449456\") " pod="openstack/horizon-58dfc66457-rqfks" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.631380 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7t98\" (UniqueName: \"kubernetes.io/projected/f78f1722-568a-4643-9e7b-1d2bb2449456-kube-api-access-f7t98\") pod \"horizon-58dfc66457-rqfks\" (UID: \"f78f1722-568a-4643-9e7b-1d2bb2449456\") " pod="openstack/horizon-58dfc66457-rqfks" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.631433 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f78f1722-568a-4643-9e7b-1d2bb2449456-logs\") pod \"horizon-58dfc66457-rqfks\" (UID: \"f78f1722-568a-4643-9e7b-1d2bb2449456\") " pod="openstack/horizon-58dfc66457-rqfks" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.631569 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f78f1722-568a-4643-9e7b-1d2bb2449456-config-data\") pod \"horizon-58dfc66457-rqfks\" (UID: \"f78f1722-568a-4643-9e7b-1d2bb2449456\") " pod="openstack/horizon-58dfc66457-rqfks" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.631635 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f78f1722-568a-4643-9e7b-1d2bb2449456-horizon-secret-key\") pod \"horizon-58dfc66457-rqfks\" (UID: \"f78f1722-568a-4643-9e7b-1d2bb2449456\") " pod="openstack/horizon-58dfc66457-rqfks" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.640633 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.641055 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="82be1eb5-9ab6-471e-baf5-400b4adc92a2" containerName="glance-log" containerID="cri-o://c559ac37f2e2e1a61c0e69a6aa52909d5eebcf31936385ceb56e5d85c035f07e" gracePeriod=30 Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.641193 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="82be1eb5-9ab6-471e-baf5-400b4adc92a2" containerName="glance-httpd" containerID="cri-o://99f6d9838ea1b4bced54c8d65874120944a9a70b875a07da856f63661fe54f0f" gracePeriod=30 Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.678725 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-79f5cf4d49-bhz7p"] Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.680523 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79f5cf4d49-bhz7p" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.710490 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-79f5cf4d49-bhz7p"] Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.735660 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7t98\" (UniqueName: \"kubernetes.io/projected/f78f1722-568a-4643-9e7b-1d2bb2449456-kube-api-access-f7t98\") pod \"horizon-58dfc66457-rqfks\" (UID: \"f78f1722-568a-4643-9e7b-1d2bb2449456\") " pod="openstack/horizon-58dfc66457-rqfks" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.735755 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ceec03ba-936a-4d49-983e-307237454683-logs\") pod \"horizon-79f5cf4d49-bhz7p\" (UID: \"ceec03ba-936a-4d49-983e-307237454683\") " pod="openstack/horizon-79f5cf4d49-bhz7p" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.735808 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f78f1722-568a-4643-9e7b-1d2bb2449456-logs\") pod \"horizon-58dfc66457-rqfks\" (UID: \"f78f1722-568a-4643-9e7b-1d2bb2449456\") " pod="openstack/horizon-58dfc66457-rqfks" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.735841 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ceec03ba-936a-4d49-983e-307237454683-config-data\") pod \"horizon-79f5cf4d49-bhz7p\" (UID: \"ceec03ba-936a-4d49-983e-307237454683\") " pod="openstack/horizon-79f5cf4d49-bhz7p" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.735954 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ceec03ba-936a-4d49-983e-307237454683-scripts\") pod \"horizon-79f5cf4d49-bhz7p\" (UID: \"ceec03ba-936a-4d49-983e-307237454683\") " pod="openstack/horizon-79f5cf4d49-bhz7p" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.736138 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f78f1722-568a-4643-9e7b-1d2bb2449456-config-data\") pod \"horizon-58dfc66457-rqfks\" (UID: \"f78f1722-568a-4643-9e7b-1d2bb2449456\") " pod="openstack/horizon-58dfc66457-rqfks" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.736503 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f78f1722-568a-4643-9e7b-1d2bb2449456-logs\") pod \"horizon-58dfc66457-rqfks\" (UID: \"f78f1722-568a-4643-9e7b-1d2bb2449456\") " pod="openstack/horizon-58dfc66457-rqfks" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.737618 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f78f1722-568a-4643-9e7b-1d2bb2449456-config-data\") pod \"horizon-58dfc66457-rqfks\" (UID: \"f78f1722-568a-4643-9e7b-1d2bb2449456\") " pod="openstack/horizon-58dfc66457-rqfks" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.737668 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ceec03ba-936a-4d49-983e-307237454683-horizon-secret-key\") pod \"horizon-79f5cf4d49-bhz7p\" (UID: \"ceec03ba-936a-4d49-983e-307237454683\") " pod="openstack/horizon-79f5cf4d49-bhz7p" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.738410 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f78f1722-568a-4643-9e7b-1d2bb2449456-horizon-secret-key\") pod \"horizon-58dfc66457-rqfks\" (UID: \"f78f1722-568a-4643-9e7b-1d2bb2449456\") " pod="openstack/horizon-58dfc66457-rqfks" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.738504 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f78f1722-568a-4643-9e7b-1d2bb2449456-scripts\") pod \"horizon-58dfc66457-rqfks\" (UID: \"f78f1722-568a-4643-9e7b-1d2bb2449456\") " pod="openstack/horizon-58dfc66457-rqfks" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.739321 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4vpx\" (UniqueName: \"kubernetes.io/projected/ceec03ba-936a-4d49-983e-307237454683-kube-api-access-r4vpx\") pod \"horizon-79f5cf4d49-bhz7p\" (UID: \"ceec03ba-936a-4d49-983e-307237454683\") " pod="openstack/horizon-79f5cf4d49-bhz7p" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.739505 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f78f1722-568a-4643-9e7b-1d2bb2449456-scripts\") pod \"horizon-58dfc66457-rqfks\" (UID: \"f78f1722-568a-4643-9e7b-1d2bb2449456\") " pod="openstack/horizon-58dfc66457-rqfks" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.747527 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f78f1722-568a-4643-9e7b-1d2bb2449456-horizon-secret-key\") pod \"horizon-58dfc66457-rqfks\" (UID: \"f78f1722-568a-4643-9e7b-1d2bb2449456\") " pod="openstack/horizon-58dfc66457-rqfks" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.755149 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7t98\" (UniqueName: \"kubernetes.io/projected/f78f1722-568a-4643-9e7b-1d2bb2449456-kube-api-access-f7t98\") pod \"horizon-58dfc66457-rqfks\" (UID: \"f78f1722-568a-4643-9e7b-1d2bb2449456\") " pod="openstack/horizon-58dfc66457-rqfks" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.842687 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ceec03ba-936a-4d49-983e-307237454683-horizon-secret-key\") pod \"horizon-79f5cf4d49-bhz7p\" (UID: \"ceec03ba-936a-4d49-983e-307237454683\") " pod="openstack/horizon-79f5cf4d49-bhz7p" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.842800 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4vpx\" (UniqueName: \"kubernetes.io/projected/ceec03ba-936a-4d49-983e-307237454683-kube-api-access-r4vpx\") pod \"horizon-79f5cf4d49-bhz7p\" (UID: \"ceec03ba-936a-4d49-983e-307237454683\") " pod="openstack/horizon-79f5cf4d49-bhz7p" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.842901 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ceec03ba-936a-4d49-983e-307237454683-logs\") pod \"horizon-79f5cf4d49-bhz7p\" (UID: \"ceec03ba-936a-4d49-983e-307237454683\") " pod="openstack/horizon-79f5cf4d49-bhz7p" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.842938 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ceec03ba-936a-4d49-983e-307237454683-config-data\") pod \"horizon-79f5cf4d49-bhz7p\" (UID: \"ceec03ba-936a-4d49-983e-307237454683\") " pod="openstack/horizon-79f5cf4d49-bhz7p" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.842998 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ceec03ba-936a-4d49-983e-307237454683-scripts\") pod \"horizon-79f5cf4d49-bhz7p\" (UID: \"ceec03ba-936a-4d49-983e-307237454683\") " pod="openstack/horizon-79f5cf4d49-bhz7p" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.843649 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ceec03ba-936a-4d49-983e-307237454683-logs\") pod \"horizon-79f5cf4d49-bhz7p\" (UID: \"ceec03ba-936a-4d49-983e-307237454683\") " pod="openstack/horizon-79f5cf4d49-bhz7p" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.844098 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-58dfc66457-rqfks" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.844920 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ceec03ba-936a-4d49-983e-307237454683-scripts\") pod \"horizon-79f5cf4d49-bhz7p\" (UID: \"ceec03ba-936a-4d49-983e-307237454683\") " pod="openstack/horizon-79f5cf4d49-bhz7p" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.845827 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ceec03ba-936a-4d49-983e-307237454683-config-data\") pod \"horizon-79f5cf4d49-bhz7p\" (UID: \"ceec03ba-936a-4d49-983e-307237454683\") " pod="openstack/horizon-79f5cf4d49-bhz7p" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.848515 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ceec03ba-936a-4d49-983e-307237454683-horizon-secret-key\") pod \"horizon-79f5cf4d49-bhz7p\" (UID: \"ceec03ba-936a-4d49-983e-307237454683\") " pod="openstack/horizon-79f5cf4d49-bhz7p" Nov 24 14:52:11 crc kubenswrapper[4790]: I1124 14:52:11.863801 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4vpx\" (UniqueName: \"kubernetes.io/projected/ceec03ba-936a-4d49-983e-307237454683-kube-api-access-r4vpx\") pod \"horizon-79f5cf4d49-bhz7p\" (UID: \"ceec03ba-936a-4d49-983e-307237454683\") " pod="openstack/horizon-79f5cf4d49-bhz7p" Nov 24 14:52:12 crc kubenswrapper[4790]: I1124 14:52:12.010206 4790 generic.go:334] "Generic (PLEG): container finished" podID="82be1eb5-9ab6-471e-baf5-400b4adc92a2" containerID="c559ac37f2e2e1a61c0e69a6aa52909d5eebcf31936385ceb56e5d85c035f07e" exitCode=143 Nov 24 14:52:12 crc kubenswrapper[4790]: I1124 14:52:12.010630 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"82be1eb5-9ab6-471e-baf5-400b4adc92a2","Type":"ContainerDied","Data":"c559ac37f2e2e1a61c0e69a6aa52909d5eebcf31936385ceb56e5d85c035f07e"} Nov 24 14:52:12 crc kubenswrapper[4790]: I1124 14:52:12.016589 4790 generic.go:334] "Generic (PLEG): container finished" podID="4b8cad57-cddc-4273-a1be-498c1172ee79" containerID="6b3329faf3f91547a252762a0890d5dba6b36153a7b56704b3a84abd4e8aeeb3" exitCode=143 Nov 24 14:52:12 crc kubenswrapper[4790]: I1124 14:52:12.016640 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4b8cad57-cddc-4273-a1be-498c1172ee79","Type":"ContainerDied","Data":"6b3329faf3f91547a252762a0890d5dba6b36153a7b56704b3a84abd4e8aeeb3"} Nov 24 14:52:12 crc kubenswrapper[4790]: I1124 14:52:12.040953 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79f5cf4d49-bhz7p" Nov 24 14:52:12 crc kubenswrapper[4790]: I1124 14:52:12.245791 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-58dfc66457-rqfks"] Nov 24 14:52:12 crc kubenswrapper[4790]: I1124 14:52:12.276668 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-67b58d9dc-n6wf7"] Nov 24 14:52:12 crc kubenswrapper[4790]: I1124 14:52:12.278650 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67b58d9dc-n6wf7" Nov 24 14:52:12 crc kubenswrapper[4790]: I1124 14:52:12.301150 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-67b58d9dc-n6wf7"] Nov 24 14:52:12 crc kubenswrapper[4790]: I1124 14:52:12.368175 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/693f3fba-7701-4f36-9a9a-f3b733a9528e-horizon-secret-key\") pod \"horizon-67b58d9dc-n6wf7\" (UID: \"693f3fba-7701-4f36-9a9a-f3b733a9528e\") " pod="openstack/horizon-67b58d9dc-n6wf7" Nov 24 14:52:12 crc kubenswrapper[4790]: I1124 14:52:12.368250 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/693f3fba-7701-4f36-9a9a-f3b733a9528e-logs\") pod \"horizon-67b58d9dc-n6wf7\" (UID: \"693f3fba-7701-4f36-9a9a-f3b733a9528e\") " pod="openstack/horizon-67b58d9dc-n6wf7" Nov 24 14:52:12 crc kubenswrapper[4790]: I1124 14:52:12.368301 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ftkz\" (UniqueName: \"kubernetes.io/projected/693f3fba-7701-4f36-9a9a-f3b733a9528e-kube-api-access-9ftkz\") pod \"horizon-67b58d9dc-n6wf7\" (UID: \"693f3fba-7701-4f36-9a9a-f3b733a9528e\") " pod="openstack/horizon-67b58d9dc-n6wf7" Nov 24 14:52:12 crc kubenswrapper[4790]: I1124 14:52:12.368367 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/693f3fba-7701-4f36-9a9a-f3b733a9528e-scripts\") pod \"horizon-67b58d9dc-n6wf7\" (UID: \"693f3fba-7701-4f36-9a9a-f3b733a9528e\") " pod="openstack/horizon-67b58d9dc-n6wf7" Nov 24 14:52:12 crc kubenswrapper[4790]: I1124 14:52:12.368426 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/693f3fba-7701-4f36-9a9a-f3b733a9528e-config-data\") pod \"horizon-67b58d9dc-n6wf7\" (UID: \"693f3fba-7701-4f36-9a9a-f3b733a9528e\") " pod="openstack/horizon-67b58d9dc-n6wf7" Nov 24 14:52:12 crc kubenswrapper[4790]: I1124 14:52:12.370780 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-58dfc66457-rqfks"] Nov 24 14:52:12 crc kubenswrapper[4790]: I1124 14:52:12.470563 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/693f3fba-7701-4f36-9a9a-f3b733a9528e-scripts\") pod \"horizon-67b58d9dc-n6wf7\" (UID: \"693f3fba-7701-4f36-9a9a-f3b733a9528e\") " pod="openstack/horizon-67b58d9dc-n6wf7" Nov 24 14:52:12 crc kubenswrapper[4790]: I1124 14:52:12.470687 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/693f3fba-7701-4f36-9a9a-f3b733a9528e-config-data\") pod \"horizon-67b58d9dc-n6wf7\" (UID: \"693f3fba-7701-4f36-9a9a-f3b733a9528e\") " pod="openstack/horizon-67b58d9dc-n6wf7" Nov 24 14:52:12 crc kubenswrapper[4790]: I1124 14:52:12.470737 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/693f3fba-7701-4f36-9a9a-f3b733a9528e-horizon-secret-key\") pod \"horizon-67b58d9dc-n6wf7\" (UID: \"693f3fba-7701-4f36-9a9a-f3b733a9528e\") " pod="openstack/horizon-67b58d9dc-n6wf7" Nov 24 14:52:12 crc kubenswrapper[4790]: I1124 14:52:12.470790 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/693f3fba-7701-4f36-9a9a-f3b733a9528e-logs\") pod \"horizon-67b58d9dc-n6wf7\" (UID: \"693f3fba-7701-4f36-9a9a-f3b733a9528e\") " pod="openstack/horizon-67b58d9dc-n6wf7" Nov 24 14:52:12 crc kubenswrapper[4790]: I1124 14:52:12.470930 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ftkz\" (UniqueName: \"kubernetes.io/projected/693f3fba-7701-4f36-9a9a-f3b733a9528e-kube-api-access-9ftkz\") pod \"horizon-67b58d9dc-n6wf7\" (UID: \"693f3fba-7701-4f36-9a9a-f3b733a9528e\") " pod="openstack/horizon-67b58d9dc-n6wf7" Nov 24 14:52:12 crc kubenswrapper[4790]: I1124 14:52:12.472585 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/693f3fba-7701-4f36-9a9a-f3b733a9528e-scripts\") pod \"horizon-67b58d9dc-n6wf7\" (UID: \"693f3fba-7701-4f36-9a9a-f3b733a9528e\") " pod="openstack/horizon-67b58d9dc-n6wf7" Nov 24 14:52:12 crc kubenswrapper[4790]: I1124 14:52:12.472591 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/693f3fba-7701-4f36-9a9a-f3b733a9528e-logs\") pod \"horizon-67b58d9dc-n6wf7\" (UID: \"693f3fba-7701-4f36-9a9a-f3b733a9528e\") " pod="openstack/horizon-67b58d9dc-n6wf7" Nov 24 14:52:12 crc kubenswrapper[4790]: I1124 14:52:12.473921 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/693f3fba-7701-4f36-9a9a-f3b733a9528e-config-data\") pod \"horizon-67b58d9dc-n6wf7\" (UID: \"693f3fba-7701-4f36-9a9a-f3b733a9528e\") " pod="openstack/horizon-67b58d9dc-n6wf7" Nov 24 14:52:12 crc kubenswrapper[4790]: I1124 14:52:12.478364 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/693f3fba-7701-4f36-9a9a-f3b733a9528e-horizon-secret-key\") pod \"horizon-67b58d9dc-n6wf7\" (UID: \"693f3fba-7701-4f36-9a9a-f3b733a9528e\") " pod="openstack/horizon-67b58d9dc-n6wf7" Nov 24 14:52:12 crc kubenswrapper[4790]: I1124 14:52:12.489146 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ftkz\" (UniqueName: \"kubernetes.io/projected/693f3fba-7701-4f36-9a9a-f3b733a9528e-kube-api-access-9ftkz\") pod \"horizon-67b58d9dc-n6wf7\" (UID: \"693f3fba-7701-4f36-9a9a-f3b733a9528e\") " pod="openstack/horizon-67b58d9dc-n6wf7" Nov 24 14:52:12 crc kubenswrapper[4790]: W1124 14:52:12.574253 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podceec03ba_936a_4d49_983e_307237454683.slice/crio-af27caa49df61a03ecb6672883cd59a55080117b0c0979bbff5d1d0c25a21eec WatchSource:0}: Error finding container af27caa49df61a03ecb6672883cd59a55080117b0c0979bbff5d1d0c25a21eec: Status 404 returned error can't find the container with id af27caa49df61a03ecb6672883cd59a55080117b0c0979bbff5d1d0c25a21eec Nov 24 14:52:12 crc kubenswrapper[4790]: I1124 14:52:12.575531 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-79f5cf4d49-bhz7p"] Nov 24 14:52:12 crc kubenswrapper[4790]: I1124 14:52:12.610014 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67b58d9dc-n6wf7" Nov 24 14:52:13 crc kubenswrapper[4790]: I1124 14:52:13.091117 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-58dfc66457-rqfks" event={"ID":"f78f1722-568a-4643-9e7b-1d2bb2449456","Type":"ContainerStarted","Data":"d7387057bdc83a18de55571982040dcc2c95437651945bcf48a7c9082e62bc32"} Nov 24 14:52:13 crc kubenswrapper[4790]: I1124 14:52:13.104108 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79f5cf4d49-bhz7p" event={"ID":"ceec03ba-936a-4d49-983e-307237454683","Type":"ContainerStarted","Data":"af27caa49df61a03ecb6672883cd59a55080117b0c0979bbff5d1d0c25a21eec"} Nov 24 14:52:13 crc kubenswrapper[4790]: I1124 14:52:13.121776 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-67b58d9dc-n6wf7"] Nov 24 14:52:13 crc kubenswrapper[4790]: I1124 14:52:13.958007 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:52:13 crc kubenswrapper[4790]: I1124 14:52:13.958383 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:52:14 crc kubenswrapper[4790]: I1124 14:52:14.113308 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67b58d9dc-n6wf7" event={"ID":"693f3fba-7701-4f36-9a9a-f3b733a9528e","Type":"ContainerStarted","Data":"a3816cba24f9486077834ea3cde17ee385663bbcb74a3154330dd1af27386a6f"} Nov 24 14:52:15 crc kubenswrapper[4790]: I1124 14:52:15.124864 4790 generic.go:334] "Generic (PLEG): container finished" podID="4b8cad57-cddc-4273-a1be-498c1172ee79" containerID="c65baf0b9f2f5600591225f558e76d78f8d8dd850c6290154a8a5b152f89bf66" exitCode=0 Nov 24 14:52:15 crc kubenswrapper[4790]: I1124 14:52:15.124927 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4b8cad57-cddc-4273-a1be-498c1172ee79","Type":"ContainerDied","Data":"c65baf0b9f2f5600591225f558e76d78f8d8dd850c6290154a8a5b152f89bf66"} Nov 24 14:52:15 crc kubenswrapper[4790]: I1124 14:52:15.127982 4790 generic.go:334] "Generic (PLEG): container finished" podID="82be1eb5-9ab6-471e-baf5-400b4adc92a2" containerID="99f6d9838ea1b4bced54c8d65874120944a9a70b875a07da856f63661fe54f0f" exitCode=0 Nov 24 14:52:15 crc kubenswrapper[4790]: I1124 14:52:15.128013 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"82be1eb5-9ab6-471e-baf5-400b4adc92a2","Type":"ContainerDied","Data":"99f6d9838ea1b4bced54c8d65874120944a9a70b875a07da856f63661fe54f0f"} Nov 24 14:52:16 crc kubenswrapper[4790]: I1124 14:52:16.053953 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-7zn9c"] Nov 24 14:52:16 crc kubenswrapper[4790]: I1124 14:52:16.063658 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5df6-account-create-j89gz"] Nov 24 14:52:16 crc kubenswrapper[4790]: I1124 14:52:16.073009 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-7zn9c"] Nov 24 14:52:16 crc kubenswrapper[4790]: I1124 14:52:16.081345 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5df6-account-create-j89gz"] Nov 24 14:52:16 crc kubenswrapper[4790]: I1124 14:52:16.330809 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71846a64-7dcc-41ca-907b-5481e438a3e9" path="/var/lib/kubelet/pods/71846a64-7dcc-41ca-907b-5481e438a3e9/volumes" Nov 24 14:52:16 crc kubenswrapper[4790]: I1124 14:52:16.332795 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2c200d0-9126-40d1-8b7b-d15e646a0f5f" path="/var/lib/kubelet/pods/c2c200d0-9126-40d1-8b7b-d15e646a0f5f/volumes" Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.373564 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.467949 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6wpp\" (UniqueName: \"kubernetes.io/projected/4b8cad57-cddc-4273-a1be-498c1172ee79-kube-api-access-p6wpp\") pod \"4b8cad57-cddc-4273-a1be-498c1172ee79\" (UID: \"4b8cad57-cddc-4273-a1be-498c1172ee79\") " Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.467991 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4b8cad57-cddc-4273-a1be-498c1172ee79-ceph\") pod \"4b8cad57-cddc-4273-a1be-498c1172ee79\" (UID: \"4b8cad57-cddc-4273-a1be-498c1172ee79\") " Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.468205 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b8cad57-cddc-4273-a1be-498c1172ee79-combined-ca-bundle\") pod \"4b8cad57-cddc-4273-a1be-498c1172ee79\" (UID: \"4b8cad57-cddc-4273-a1be-498c1172ee79\") " Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.468272 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4b8cad57-cddc-4273-a1be-498c1172ee79-httpd-run\") pod \"4b8cad57-cddc-4273-a1be-498c1172ee79\" (UID: \"4b8cad57-cddc-4273-a1be-498c1172ee79\") " Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.468307 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b8cad57-cddc-4273-a1be-498c1172ee79-scripts\") pod \"4b8cad57-cddc-4273-a1be-498c1172ee79\" (UID: \"4b8cad57-cddc-4273-a1be-498c1172ee79\") " Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.468403 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b8cad57-cddc-4273-a1be-498c1172ee79-logs\") pod \"4b8cad57-cddc-4273-a1be-498c1172ee79\" (UID: \"4b8cad57-cddc-4273-a1be-498c1172ee79\") " Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.468435 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b8cad57-cddc-4273-a1be-498c1172ee79-config-data\") pod \"4b8cad57-cddc-4273-a1be-498c1172ee79\" (UID: \"4b8cad57-cddc-4273-a1be-498c1172ee79\") " Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.470547 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b8cad57-cddc-4273-a1be-498c1172ee79-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "4b8cad57-cddc-4273-a1be-498c1172ee79" (UID: "4b8cad57-cddc-4273-a1be-498c1172ee79"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.470702 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b8cad57-cddc-4273-a1be-498c1172ee79-logs" (OuterVolumeSpecName: "logs") pod "4b8cad57-cddc-4273-a1be-498c1172ee79" (UID: "4b8cad57-cddc-4273-a1be-498c1172ee79"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.476247 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b8cad57-cddc-4273-a1be-498c1172ee79-kube-api-access-p6wpp" (OuterVolumeSpecName: "kube-api-access-p6wpp") pod "4b8cad57-cddc-4273-a1be-498c1172ee79" (UID: "4b8cad57-cddc-4273-a1be-498c1172ee79"). InnerVolumeSpecName "kube-api-access-p6wpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.476757 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b8cad57-cddc-4273-a1be-498c1172ee79-ceph" (OuterVolumeSpecName: "ceph") pod "4b8cad57-cddc-4273-a1be-498c1172ee79" (UID: "4b8cad57-cddc-4273-a1be-498c1172ee79"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.478352 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b8cad57-cddc-4273-a1be-498c1172ee79-scripts" (OuterVolumeSpecName: "scripts") pod "4b8cad57-cddc-4273-a1be-498c1172ee79" (UID: "4b8cad57-cddc-4273-a1be-498c1172ee79"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.523372 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b8cad57-cddc-4273-a1be-498c1172ee79-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4b8cad57-cddc-4273-a1be-498c1172ee79" (UID: "4b8cad57-cddc-4273-a1be-498c1172ee79"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.570902 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6wpp\" (UniqueName: \"kubernetes.io/projected/4b8cad57-cddc-4273-a1be-498c1172ee79-kube-api-access-p6wpp\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.570932 4790 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4b8cad57-cddc-4273-a1be-498c1172ee79-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.570942 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b8cad57-cddc-4273-a1be-498c1172ee79-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.570950 4790 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4b8cad57-cddc-4273-a1be-498c1172ee79-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.570960 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b8cad57-cddc-4273-a1be-498c1172ee79-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.570968 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b8cad57-cddc-4273-a1be-498c1172ee79-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.571094 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b8cad57-cddc-4273-a1be-498c1172ee79-config-data" (OuterVolumeSpecName: "config-data") pod "4b8cad57-cddc-4273-a1be-498c1172ee79" (UID: "4b8cad57-cddc-4273-a1be-498c1172ee79"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.674221 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b8cad57-cddc-4273-a1be-498c1172ee79-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.792181 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.877316 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82be1eb5-9ab6-471e-baf5-400b4adc92a2-combined-ca-bundle\") pod \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\" (UID: \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\") " Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.877842 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82be1eb5-9ab6-471e-baf5-400b4adc92a2-config-data\") pod \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\" (UID: \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\") " Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.877908 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82be1eb5-9ab6-471e-baf5-400b4adc92a2-scripts\") pod \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\" (UID: \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\") " Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.878011 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/82be1eb5-9ab6-471e-baf5-400b4adc92a2-httpd-run\") pod \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\" (UID: \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\") " Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.878050 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/82be1eb5-9ab6-471e-baf5-400b4adc92a2-ceph\") pod \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\" (UID: \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\") " Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.878144 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82be1eb5-9ab6-471e-baf5-400b4adc92a2-logs\") pod \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\" (UID: \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\") " Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.878180 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bcrjz\" (UniqueName: \"kubernetes.io/projected/82be1eb5-9ab6-471e-baf5-400b4adc92a2-kube-api-access-bcrjz\") pod \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\" (UID: \"82be1eb5-9ab6-471e-baf5-400b4adc92a2\") " Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.879761 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82be1eb5-9ab6-471e-baf5-400b4adc92a2-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "82be1eb5-9ab6-471e-baf5-400b4adc92a2" (UID: "82be1eb5-9ab6-471e-baf5-400b4adc92a2"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.880222 4790 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/82be1eb5-9ab6-471e-baf5-400b4adc92a2-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.880858 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82be1eb5-9ab6-471e-baf5-400b4adc92a2-logs" (OuterVolumeSpecName: "logs") pod "82be1eb5-9ab6-471e-baf5-400b4adc92a2" (UID: "82be1eb5-9ab6-471e-baf5-400b4adc92a2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.883505 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82be1eb5-9ab6-471e-baf5-400b4adc92a2-ceph" (OuterVolumeSpecName: "ceph") pod "82be1eb5-9ab6-471e-baf5-400b4adc92a2" (UID: "82be1eb5-9ab6-471e-baf5-400b4adc92a2"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.885506 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82be1eb5-9ab6-471e-baf5-400b4adc92a2-kube-api-access-bcrjz" (OuterVolumeSpecName: "kube-api-access-bcrjz") pod "82be1eb5-9ab6-471e-baf5-400b4adc92a2" (UID: "82be1eb5-9ab6-471e-baf5-400b4adc92a2"). InnerVolumeSpecName "kube-api-access-bcrjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.891053 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82be1eb5-9ab6-471e-baf5-400b4adc92a2-scripts" (OuterVolumeSpecName: "scripts") pod "82be1eb5-9ab6-471e-baf5-400b4adc92a2" (UID: "82be1eb5-9ab6-471e-baf5-400b4adc92a2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.929062 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82be1eb5-9ab6-471e-baf5-400b4adc92a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "82be1eb5-9ab6-471e-baf5-400b4adc92a2" (UID: "82be1eb5-9ab6-471e-baf5-400b4adc92a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.952795 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82be1eb5-9ab6-471e-baf5-400b4adc92a2-config-data" (OuterVolumeSpecName: "config-data") pod "82be1eb5-9ab6-471e-baf5-400b4adc92a2" (UID: "82be1eb5-9ab6-471e-baf5-400b4adc92a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.981662 4790 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/82be1eb5-9ab6-471e-baf5-400b4adc92a2-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.981706 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82be1eb5-9ab6-471e-baf5-400b4adc92a2-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.981717 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bcrjz\" (UniqueName: \"kubernetes.io/projected/82be1eb5-9ab6-471e-baf5-400b4adc92a2-kube-api-access-bcrjz\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.981730 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82be1eb5-9ab6-471e-baf5-400b4adc92a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.981740 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82be1eb5-9ab6-471e-baf5-400b4adc92a2-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:20 crc kubenswrapper[4790]: I1124 14:52:20.981750 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82be1eb5-9ab6-471e-baf5-400b4adc92a2-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.198956 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79f5cf4d49-bhz7p" event={"ID":"ceec03ba-936a-4d49-983e-307237454683","Type":"ContainerStarted","Data":"3429608e5620adba09bfac7d28a7c297bf04bc207b858b7431e735aa1bd40c36"} Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.199007 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79f5cf4d49-bhz7p" event={"ID":"ceec03ba-936a-4d49-983e-307237454683","Type":"ContainerStarted","Data":"ede38852f5d6ed963db0ac16661982c274841c82f2bf78bb674d2d3b5f0de39d"} Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.200761 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-58dfc66457-rqfks" event={"ID":"f78f1722-568a-4643-9e7b-1d2bb2449456","Type":"ContainerStarted","Data":"691b073e05ba0f932952d9f4e6fc60e065fec6302bd334841a4253caa54a2bce"} Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.200823 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-58dfc66457-rqfks" event={"ID":"f78f1722-568a-4643-9e7b-1d2bb2449456","Type":"ContainerStarted","Data":"9fd79e743941d0f3c4727ccb7d95e21cbc95a5dbb54c5db0c445373886353ef4"} Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.200795 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-58dfc66457-rqfks" podUID="f78f1722-568a-4643-9e7b-1d2bb2449456" containerName="horizon-log" containerID="cri-o://9fd79e743941d0f3c4727ccb7d95e21cbc95a5dbb54c5db0c445373886353ef4" gracePeriod=30 Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.200845 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-58dfc66457-rqfks" podUID="f78f1722-568a-4643-9e7b-1d2bb2449456" containerName="horizon" containerID="cri-o://691b073e05ba0f932952d9f4e6fc60e065fec6302bd334841a4253caa54a2bce" gracePeriod=30 Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.205904 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4b8cad57-cddc-4273-a1be-498c1172ee79","Type":"ContainerDied","Data":"a881c971ddc1a135586fc2b7988aabfaa7c968d7125d6ed5dff19f23c3166fa5"} Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.205947 4790 scope.go:117] "RemoveContainer" containerID="c65baf0b9f2f5600591225f558e76d78f8d8dd850c6290154a8a5b152f89bf66" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.206007 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.208304 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"82be1eb5-9ab6-471e-baf5-400b4adc92a2","Type":"ContainerDied","Data":"a06e368eb0a87c206fdbc844fb69d40ab93bd1924c3e8fd68e58ba1531731205"} Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.208370 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.219936 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67b58d9dc-n6wf7" event={"ID":"693f3fba-7701-4f36-9a9a-f3b733a9528e","Type":"ContainerStarted","Data":"6a291307e0d4114091ab8e92a5a2b1fbfc8ef126006801b9af437486329093ed"} Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.219963 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67b58d9dc-n6wf7" event={"ID":"693f3fba-7701-4f36-9a9a-f3b733a9528e","Type":"ContainerStarted","Data":"3d4c1ac358c0b377fce1b5a6a3f234a637a98fc3d871f5d98db02e7806121d82"} Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.240553 4790 scope.go:117] "RemoveContainer" containerID="6b3329faf3f91547a252762a0890d5dba6b36153a7b56704b3a84abd4e8aeeb3" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.243750 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-79f5cf4d49-bhz7p" podStartSLOduration=2.181006215 podStartE2EDuration="10.243734584s" podCreationTimestamp="2025-11-24 14:52:11 +0000 UTC" firstStartedPulling="2025-11-24 14:52:12.577740616 +0000 UTC m=+5980.957634268" lastFinishedPulling="2025-11-24 14:52:20.640468975 +0000 UTC m=+5989.020362637" observedRunningTime="2025-11-24 14:52:21.230783582 +0000 UTC m=+5989.610677244" watchObservedRunningTime="2025-11-24 14:52:21.243734584 +0000 UTC m=+5989.623628246" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.290354 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.318334 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.327947 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-58dfc66457-rqfks" podStartSLOduration=2.050088375 podStartE2EDuration="10.327914412s" podCreationTimestamp="2025-11-24 14:52:11 +0000 UTC" firstStartedPulling="2025-11-24 14:52:12.361413182 +0000 UTC m=+5980.741306834" lastFinishedPulling="2025-11-24 14:52:20.639239209 +0000 UTC m=+5989.019132871" observedRunningTime="2025-11-24 14:52:21.290454176 +0000 UTC m=+5989.670347848" watchObservedRunningTime="2025-11-24 14:52:21.327914412 +0000 UTC m=+5989.707808084" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.348963 4790 scope.go:117] "RemoveContainer" containerID="99f6d9838ea1b4bced54c8d65874120944a9a70b875a07da856f63661fe54f0f" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.377917 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:52:21 crc kubenswrapper[4790]: E1124 14:52:21.378381 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82be1eb5-9ab6-471e-baf5-400b4adc92a2" containerName="glance-log" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.378394 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="82be1eb5-9ab6-471e-baf5-400b4adc92a2" containerName="glance-log" Nov 24 14:52:21 crc kubenswrapper[4790]: E1124 14:52:21.378404 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b8cad57-cddc-4273-a1be-498c1172ee79" containerName="glance-log" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.378410 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b8cad57-cddc-4273-a1be-498c1172ee79" containerName="glance-log" Nov 24 14:52:21 crc kubenswrapper[4790]: E1124 14:52:21.378428 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82be1eb5-9ab6-471e-baf5-400b4adc92a2" containerName="glance-httpd" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.378435 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="82be1eb5-9ab6-471e-baf5-400b4adc92a2" containerName="glance-httpd" Nov 24 14:52:21 crc kubenswrapper[4790]: E1124 14:52:21.378455 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b8cad57-cddc-4273-a1be-498c1172ee79" containerName="glance-httpd" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.378461 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b8cad57-cddc-4273-a1be-498c1172ee79" containerName="glance-httpd" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.378653 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="82be1eb5-9ab6-471e-baf5-400b4adc92a2" containerName="glance-log" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.378667 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="82be1eb5-9ab6-471e-baf5-400b4adc92a2" containerName="glance-httpd" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.378686 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b8cad57-cddc-4273-a1be-498c1172ee79" containerName="glance-log" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.378697 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b8cad57-cddc-4273-a1be-498c1172ee79" containerName="glance-httpd" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.380031 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.382661 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-kjz5s" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.382687 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.382930 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.387909 4790 scope.go:117] "RemoveContainer" containerID="c559ac37f2e2e1a61c0e69a6aa52909d5eebcf31936385ceb56e5d85c035f07e" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.399935 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.419641 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-67b58d9dc-n6wf7" podStartSLOduration=1.8770858320000001 podStartE2EDuration="9.419618756s" podCreationTimestamp="2025-11-24 14:52:12 +0000 UTC" firstStartedPulling="2025-11-24 14:52:13.131412882 +0000 UTC m=+5981.511306544" lastFinishedPulling="2025-11-24 14:52:20.673945806 +0000 UTC m=+5989.053839468" observedRunningTime="2025-11-24 14:52:21.345985541 +0000 UTC m=+5989.725879223" watchObservedRunningTime="2025-11-24 14:52:21.419618756 +0000 UTC m=+5989.799512418" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.440070 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.453253 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.465094 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.466764 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.468523 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.480603 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.497581 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f53e305e-ffe9-4c30-90bd-c3b4a1134e39-ceph\") pod \"glance-default-external-api-0\" (UID: \"f53e305e-ffe9-4c30-90bd-c3b4a1134e39\") " pod="openstack/glance-default-external-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.497669 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f53e305e-ffe9-4c30-90bd-c3b4a1134e39-config-data\") pod \"glance-default-external-api-0\" (UID: \"f53e305e-ffe9-4c30-90bd-c3b4a1134e39\") " pod="openstack/glance-default-external-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.497783 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/130ad045-32d0-460f-83e8-a357594f53e9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"130ad045-32d0-460f-83e8-a357594f53e9\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.497833 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/130ad045-32d0-460f-83e8-a357594f53e9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"130ad045-32d0-460f-83e8-a357594f53e9\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.498850 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f53e305e-ffe9-4c30-90bd-c3b4a1134e39-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f53e305e-ffe9-4c30-90bd-c3b4a1134e39\") " pod="openstack/glance-default-external-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.498954 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f53e305e-ffe9-4c30-90bd-c3b4a1134e39-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f53e305e-ffe9-4c30-90bd-c3b4a1134e39\") " pod="openstack/glance-default-external-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.499073 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f53e305e-ffe9-4c30-90bd-c3b4a1134e39-scripts\") pod \"glance-default-external-api-0\" (UID: \"f53e305e-ffe9-4c30-90bd-c3b4a1134e39\") " pod="openstack/glance-default-external-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.499203 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/130ad045-32d0-460f-83e8-a357594f53e9-logs\") pod \"glance-default-internal-api-0\" (UID: \"130ad045-32d0-460f-83e8-a357594f53e9\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.499255 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/130ad045-32d0-460f-83e8-a357594f53e9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"130ad045-32d0-460f-83e8-a357594f53e9\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.499305 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvhd7\" (UniqueName: \"kubernetes.io/projected/f53e305e-ffe9-4c30-90bd-c3b4a1134e39-kube-api-access-dvhd7\") pod \"glance-default-external-api-0\" (UID: \"f53e305e-ffe9-4c30-90bd-c3b4a1134e39\") " pod="openstack/glance-default-external-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.499391 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xn4m4\" (UniqueName: \"kubernetes.io/projected/130ad045-32d0-460f-83e8-a357594f53e9-kube-api-access-xn4m4\") pod \"glance-default-internal-api-0\" (UID: \"130ad045-32d0-460f-83e8-a357594f53e9\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.499456 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/130ad045-32d0-460f-83e8-a357594f53e9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"130ad045-32d0-460f-83e8-a357594f53e9\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.499647 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f53e305e-ffe9-4c30-90bd-c3b4a1134e39-logs\") pod \"glance-default-external-api-0\" (UID: \"f53e305e-ffe9-4c30-90bd-c3b4a1134e39\") " pod="openstack/glance-default-external-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.499721 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/130ad045-32d0-460f-83e8-a357594f53e9-ceph\") pod \"glance-default-internal-api-0\" (UID: \"130ad045-32d0-460f-83e8-a357594f53e9\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.602011 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f53e305e-ffe9-4c30-90bd-c3b4a1134e39-logs\") pod \"glance-default-external-api-0\" (UID: \"f53e305e-ffe9-4c30-90bd-c3b4a1134e39\") " pod="openstack/glance-default-external-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.602089 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/130ad045-32d0-460f-83e8-a357594f53e9-ceph\") pod \"glance-default-internal-api-0\" (UID: \"130ad045-32d0-460f-83e8-a357594f53e9\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.602116 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f53e305e-ffe9-4c30-90bd-c3b4a1134e39-ceph\") pod \"glance-default-external-api-0\" (UID: \"f53e305e-ffe9-4c30-90bd-c3b4a1134e39\") " pod="openstack/glance-default-external-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.602144 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f53e305e-ffe9-4c30-90bd-c3b4a1134e39-config-data\") pod \"glance-default-external-api-0\" (UID: \"f53e305e-ffe9-4c30-90bd-c3b4a1134e39\") " pod="openstack/glance-default-external-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.602173 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/130ad045-32d0-460f-83e8-a357594f53e9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"130ad045-32d0-460f-83e8-a357594f53e9\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.602193 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/130ad045-32d0-460f-83e8-a357594f53e9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"130ad045-32d0-460f-83e8-a357594f53e9\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.602227 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f53e305e-ffe9-4c30-90bd-c3b4a1134e39-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f53e305e-ffe9-4c30-90bd-c3b4a1134e39\") " pod="openstack/glance-default-external-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.602247 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f53e305e-ffe9-4c30-90bd-c3b4a1134e39-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f53e305e-ffe9-4c30-90bd-c3b4a1134e39\") " pod="openstack/glance-default-external-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.602271 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f53e305e-ffe9-4c30-90bd-c3b4a1134e39-scripts\") pod \"glance-default-external-api-0\" (UID: \"f53e305e-ffe9-4c30-90bd-c3b4a1134e39\") " pod="openstack/glance-default-external-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.602308 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/130ad045-32d0-460f-83e8-a357594f53e9-logs\") pod \"glance-default-internal-api-0\" (UID: \"130ad045-32d0-460f-83e8-a357594f53e9\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.602324 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/130ad045-32d0-460f-83e8-a357594f53e9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"130ad045-32d0-460f-83e8-a357594f53e9\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.602357 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvhd7\" (UniqueName: \"kubernetes.io/projected/f53e305e-ffe9-4c30-90bd-c3b4a1134e39-kube-api-access-dvhd7\") pod \"glance-default-external-api-0\" (UID: \"f53e305e-ffe9-4c30-90bd-c3b4a1134e39\") " pod="openstack/glance-default-external-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.602381 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xn4m4\" (UniqueName: \"kubernetes.io/projected/130ad045-32d0-460f-83e8-a357594f53e9-kube-api-access-xn4m4\") pod \"glance-default-internal-api-0\" (UID: \"130ad045-32d0-460f-83e8-a357594f53e9\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.602404 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/130ad045-32d0-460f-83e8-a357594f53e9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"130ad045-32d0-460f-83e8-a357594f53e9\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.602514 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f53e305e-ffe9-4c30-90bd-c3b4a1134e39-logs\") pod \"glance-default-external-api-0\" (UID: \"f53e305e-ffe9-4c30-90bd-c3b4a1134e39\") " pod="openstack/glance-default-external-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.603612 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/130ad045-32d0-460f-83e8-a357594f53e9-logs\") pod \"glance-default-internal-api-0\" (UID: \"130ad045-32d0-460f-83e8-a357594f53e9\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.604277 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f53e305e-ffe9-4c30-90bd-c3b4a1134e39-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f53e305e-ffe9-4c30-90bd-c3b4a1134e39\") " pod="openstack/glance-default-external-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.605126 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/130ad045-32d0-460f-83e8-a357594f53e9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"130ad045-32d0-460f-83e8-a357594f53e9\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.607871 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/130ad045-32d0-460f-83e8-a357594f53e9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"130ad045-32d0-460f-83e8-a357594f53e9\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.608138 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f53e305e-ffe9-4c30-90bd-c3b4a1134e39-scripts\") pod \"glance-default-external-api-0\" (UID: \"f53e305e-ffe9-4c30-90bd-c3b4a1134e39\") " pod="openstack/glance-default-external-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.608384 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/130ad045-32d0-460f-83e8-a357594f53e9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"130ad045-32d0-460f-83e8-a357594f53e9\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.609329 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/130ad045-32d0-460f-83e8-a357594f53e9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"130ad045-32d0-460f-83e8-a357594f53e9\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.609828 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f53e305e-ffe9-4c30-90bd-c3b4a1134e39-config-data\") pod \"glance-default-external-api-0\" (UID: \"f53e305e-ffe9-4c30-90bd-c3b4a1134e39\") " pod="openstack/glance-default-external-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.609999 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f53e305e-ffe9-4c30-90bd-c3b4a1134e39-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f53e305e-ffe9-4c30-90bd-c3b4a1134e39\") " pod="openstack/glance-default-external-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.613875 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f53e305e-ffe9-4c30-90bd-c3b4a1134e39-ceph\") pod \"glance-default-external-api-0\" (UID: \"f53e305e-ffe9-4c30-90bd-c3b4a1134e39\") " pod="openstack/glance-default-external-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.615255 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/130ad045-32d0-460f-83e8-a357594f53e9-ceph\") pod \"glance-default-internal-api-0\" (UID: \"130ad045-32d0-460f-83e8-a357594f53e9\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.622550 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xn4m4\" (UniqueName: \"kubernetes.io/projected/130ad045-32d0-460f-83e8-a357594f53e9-kube-api-access-xn4m4\") pod \"glance-default-internal-api-0\" (UID: \"130ad045-32d0-460f-83e8-a357594f53e9\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.623586 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvhd7\" (UniqueName: \"kubernetes.io/projected/f53e305e-ffe9-4c30-90bd-c3b4a1134e39-kube-api-access-dvhd7\") pod \"glance-default-external-api-0\" (UID: \"f53e305e-ffe9-4c30-90bd-c3b4a1134e39\") " pod="openstack/glance-default-external-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.709596 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.784647 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:52:21 crc kubenswrapper[4790]: I1124 14:52:21.845687 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-58dfc66457-rqfks" Nov 24 14:52:22 crc kubenswrapper[4790]: I1124 14:52:22.042235 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-79f5cf4d49-bhz7p" Nov 24 14:52:22 crc kubenswrapper[4790]: I1124 14:52:22.043603 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-79f5cf4d49-bhz7p" Nov 24 14:52:22 crc kubenswrapper[4790]: I1124 14:52:22.268398 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:52:22 crc kubenswrapper[4790]: W1124 14:52:22.280474 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod130ad045_32d0_460f_83e8_a357594f53e9.slice/crio-f001a58c06cef9a5b18e359ed08f759d2de6e8f4e925c6aef283c9dd27471fcc WatchSource:0}: Error finding container f001a58c06cef9a5b18e359ed08f759d2de6e8f4e925c6aef283c9dd27471fcc: Status 404 returned error can't find the container with id f001a58c06cef9a5b18e359ed08f759d2de6e8f4e925c6aef283c9dd27471fcc Nov 24 14:52:22 crc kubenswrapper[4790]: I1124 14:52:22.336057 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b8cad57-cddc-4273-a1be-498c1172ee79" path="/var/lib/kubelet/pods/4b8cad57-cddc-4273-a1be-498c1172ee79/volumes" Nov 24 14:52:22 crc kubenswrapper[4790]: I1124 14:52:22.337316 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82be1eb5-9ab6-471e-baf5-400b4adc92a2" path="/var/lib/kubelet/pods/82be1eb5-9ab6-471e-baf5-400b4adc92a2/volumes" Nov 24 14:52:22 crc kubenswrapper[4790]: I1124 14:52:22.379095 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:52:22 crc kubenswrapper[4790]: I1124 14:52:22.610441 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-67b58d9dc-n6wf7" Nov 24 14:52:22 crc kubenswrapper[4790]: I1124 14:52:22.610616 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-67b58d9dc-n6wf7" Nov 24 14:52:23 crc kubenswrapper[4790]: I1124 14:52:23.261845 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f53e305e-ffe9-4c30-90bd-c3b4a1134e39","Type":"ContainerStarted","Data":"c50e208e5d2daadbd8365d94127aa0c78c838572aab3c9f655f823ae01df6786"} Nov 24 14:52:23 crc kubenswrapper[4790]: I1124 14:52:23.262177 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f53e305e-ffe9-4c30-90bd-c3b4a1134e39","Type":"ContainerStarted","Data":"6ab0b527c2b10be1388f14326decb5fa016714fb1329cc562902cf763d09b1ff"} Nov 24 14:52:23 crc kubenswrapper[4790]: I1124 14:52:23.264770 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"130ad045-32d0-460f-83e8-a357594f53e9","Type":"ContainerStarted","Data":"a060fb030538439b83abb2459fc2e2faf8dc3fc3fed1b6d112970b76efbb353b"} Nov 24 14:52:23 crc kubenswrapper[4790]: I1124 14:52:23.264798 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"130ad045-32d0-460f-83e8-a357594f53e9","Type":"ContainerStarted","Data":"f001a58c06cef9a5b18e359ed08f759d2de6e8f4e925c6aef283c9dd27471fcc"} Nov 24 14:52:24 crc kubenswrapper[4790]: I1124 14:52:24.276654 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f53e305e-ffe9-4c30-90bd-c3b4a1134e39","Type":"ContainerStarted","Data":"59eec03eb7c5febabef7a280fb0b8c92a0da268ecc44a500bdb068f2ba2e4960"} Nov 24 14:52:24 crc kubenswrapper[4790]: I1124 14:52:24.280521 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"130ad045-32d0-460f-83e8-a357594f53e9","Type":"ContainerStarted","Data":"ba959c9a1cc6aa1b2112b3e121e2271917f0f10909df5617b707024c4f62655c"} Nov 24 14:52:24 crc kubenswrapper[4790]: I1124 14:52:24.314533 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.314515519 podStartE2EDuration="3.314515519s" podCreationTimestamp="2025-11-24 14:52:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:52:24.297023137 +0000 UTC m=+5992.676916819" watchObservedRunningTime="2025-11-24 14:52:24.314515519 +0000 UTC m=+5992.694409181" Nov 24 14:52:24 crc kubenswrapper[4790]: I1124 14:52:24.351775 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.351749639 podStartE2EDuration="3.351749639s" podCreationTimestamp="2025-11-24 14:52:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:52:24.332171176 +0000 UTC m=+5992.712064848" watchObservedRunningTime="2025-11-24 14:52:24.351749639 +0000 UTC m=+5992.731643311" Nov 24 14:52:25 crc kubenswrapper[4790]: I1124 14:52:25.040734 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-rw99k"] Nov 24 14:52:25 crc kubenswrapper[4790]: I1124 14:52:25.050272 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-rw99k"] Nov 24 14:52:26 crc kubenswrapper[4790]: I1124 14:52:26.327331 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08" path="/var/lib/kubelet/pods/2a13c0bf-ff7d-4aad-b6dc-ee12cea5ac08/volumes" Nov 24 14:52:31 crc kubenswrapper[4790]: I1124 14:52:31.709833 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 14:52:31 crc kubenswrapper[4790]: I1124 14:52:31.710392 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 14:52:31 crc kubenswrapper[4790]: I1124 14:52:31.748633 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 14:52:31 crc kubenswrapper[4790]: I1124 14:52:31.766172 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 14:52:31 crc kubenswrapper[4790]: I1124 14:52:31.784960 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 14:52:31 crc kubenswrapper[4790]: I1124 14:52:31.785005 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 14:52:31 crc kubenswrapper[4790]: I1124 14:52:31.833134 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 14:52:31 crc kubenswrapper[4790]: I1124 14:52:31.840438 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 14:52:32 crc kubenswrapper[4790]: I1124 14:52:32.043700 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-79f5cf4d49-bhz7p" podUID="ceec03ba-936a-4d49-983e-307237454683" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.111:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.111:8080: connect: connection refused" Nov 24 14:52:32 crc kubenswrapper[4790]: I1124 14:52:32.384581 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 14:52:32 crc kubenswrapper[4790]: I1124 14:52:32.384914 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 14:52:32 crc kubenswrapper[4790]: I1124 14:52:32.384936 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 14:52:32 crc kubenswrapper[4790]: I1124 14:52:32.384965 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 14:52:32 crc kubenswrapper[4790]: I1124 14:52:32.612103 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-67b58d9dc-n6wf7" podUID="693f3fba-7701-4f36-9a9a-f3b733a9528e" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.112:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.112:8080: connect: connection refused" Nov 24 14:52:34 crc kubenswrapper[4790]: I1124 14:52:34.404386 4790 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 14:52:34 crc kubenswrapper[4790]: I1124 14:52:34.404739 4790 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 14:52:34 crc kubenswrapper[4790]: I1124 14:52:34.742401 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 14:52:34 crc kubenswrapper[4790]: I1124 14:52:34.743656 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 14:52:34 crc kubenswrapper[4790]: I1124 14:52:34.853959 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 14:52:34 crc kubenswrapper[4790]: I1124 14:52:34.854352 4790 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 14:52:34 crc kubenswrapper[4790]: I1124 14:52:34.867317 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 14:52:43 crc kubenswrapper[4790]: I1124 14:52:43.778040 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-79f5cf4d49-bhz7p" Nov 24 14:52:43 crc kubenswrapper[4790]: I1124 14:52:43.938407 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:52:43 crc kubenswrapper[4790]: I1124 14:52:43.938480 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:52:43 crc kubenswrapper[4790]: I1124 14:52:43.938532 4790 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 14:52:43 crc kubenswrapper[4790]: I1124 14:52:43.939420 4790 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"28308703817a1d4e0e5344ce19268127c600894f2daa6b921d52a9105f014132"} pod="openshift-machine-config-operator/machine-config-daemon-xz49t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:52:43 crc kubenswrapper[4790]: I1124 14:52:43.939504 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" containerID="cri-o://28308703817a1d4e0e5344ce19268127c600894f2daa6b921d52a9105f014132" gracePeriod=600 Nov 24 14:52:44 crc kubenswrapper[4790]: E1124 14:52:44.132016 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:52:44 crc kubenswrapper[4790]: I1124 14:52:44.358169 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-67b58d9dc-n6wf7" Nov 24 14:52:44 crc kubenswrapper[4790]: I1124 14:52:44.553061 4790 generic.go:334] "Generic (PLEG): container finished" podID="0d73b133-48f0-455f-8f6a-742e633f631a" containerID="28308703817a1d4e0e5344ce19268127c600894f2daa6b921d52a9105f014132" exitCode=0 Nov 24 14:52:44 crc kubenswrapper[4790]: I1124 14:52:44.553115 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerDied","Data":"28308703817a1d4e0e5344ce19268127c600894f2daa6b921d52a9105f014132"} Nov 24 14:52:44 crc kubenswrapper[4790]: I1124 14:52:44.553237 4790 scope.go:117] "RemoveContainer" containerID="cd74b939f536300843ca659100b982ff6a09937cde93ea7f086d7cd5d1e85fc5" Nov 24 14:52:44 crc kubenswrapper[4790]: I1124 14:52:44.553945 4790 scope.go:117] "RemoveContainer" containerID="28308703817a1d4e0e5344ce19268127c600894f2daa6b921d52a9105f014132" Nov 24 14:52:44 crc kubenswrapper[4790]: E1124 14:52:44.554238 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:52:45 crc kubenswrapper[4790]: I1124 14:52:45.447475 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-79f5cf4d49-bhz7p" Nov 24 14:52:46 crc kubenswrapper[4790]: I1124 14:52:46.078479 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-67b58d9dc-n6wf7" Nov 24 14:52:46 crc kubenswrapper[4790]: I1124 14:52:46.158593 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-79f5cf4d49-bhz7p"] Nov 24 14:52:46 crc kubenswrapper[4790]: I1124 14:52:46.159496 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-79f5cf4d49-bhz7p" podUID="ceec03ba-936a-4d49-983e-307237454683" containerName="horizon" containerID="cri-o://3429608e5620adba09bfac7d28a7c297bf04bc207b858b7431e735aa1bd40c36" gracePeriod=30 Nov 24 14:52:46 crc kubenswrapper[4790]: I1124 14:52:46.159667 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-79f5cf4d49-bhz7p" podUID="ceec03ba-936a-4d49-983e-307237454683" containerName="horizon-log" containerID="cri-o://ede38852f5d6ed963db0ac16661982c274841c82f2bf78bb674d2d3b5f0de39d" gracePeriod=30 Nov 24 14:52:49 crc kubenswrapper[4790]: I1124 14:52:49.662281 4790 scope.go:117] "RemoveContainer" containerID="029996d1ea7dc030dc5a485d8ab4c25956d0a826ee1f959513f3b02fd3a01cdb" Nov 24 14:52:49 crc kubenswrapper[4790]: I1124 14:52:49.789912 4790 scope.go:117] "RemoveContainer" containerID="32ee5300164f12d5dd1b26b4aa0bf0fc4ed9444a1482a2967fdead9ddccc1570" Nov 24 14:52:49 crc kubenswrapper[4790]: I1124 14:52:49.963095 4790 scope.go:117] "RemoveContainer" containerID="fe4c4c352cb5e5a2c377b66b28d211014956cf029ce8935dbacb519b6f7f42b7" Nov 24 14:52:50 crc kubenswrapper[4790]: I1124 14:52:50.249181 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="4b8cad57-cddc-4273-a1be-498c1172ee79" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.1.43:9292/healthcheck\": dial tcp 10.217.1.43:9292: i/o timeout (Client.Timeout exceeded while awaiting headers)" Nov 24 14:52:50 crc kubenswrapper[4790]: I1124 14:52:50.250142 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="4b8cad57-cddc-4273-a1be-498c1172ee79" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.1.43:9292/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 14:52:50 crc kubenswrapper[4790]: I1124 14:52:50.633245 4790 generic.go:334] "Generic (PLEG): container finished" podID="ceec03ba-936a-4d49-983e-307237454683" containerID="3429608e5620adba09bfac7d28a7c297bf04bc207b858b7431e735aa1bd40c36" exitCode=0 Nov 24 14:52:50 crc kubenswrapper[4790]: I1124 14:52:50.633361 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79f5cf4d49-bhz7p" event={"ID":"ceec03ba-936a-4d49-983e-307237454683","Type":"ContainerDied","Data":"3429608e5620adba09bfac7d28a7c297bf04bc207b858b7431e735aa1bd40c36"} Nov 24 14:52:51 crc kubenswrapper[4790]: I1124 14:52:51.642904 4790 generic.go:334] "Generic (PLEG): container finished" podID="f78f1722-568a-4643-9e7b-1d2bb2449456" containerID="691b073e05ba0f932952d9f4e6fc60e065fec6302bd334841a4253caa54a2bce" exitCode=137 Nov 24 14:52:51 crc kubenswrapper[4790]: I1124 14:52:51.643179 4790 generic.go:334] "Generic (PLEG): container finished" podID="f78f1722-568a-4643-9e7b-1d2bb2449456" containerID="9fd79e743941d0f3c4727ccb7d95e21cbc95a5dbb54c5db0c445373886353ef4" exitCode=137 Nov 24 14:52:51 crc kubenswrapper[4790]: I1124 14:52:51.643199 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-58dfc66457-rqfks" event={"ID":"f78f1722-568a-4643-9e7b-1d2bb2449456","Type":"ContainerDied","Data":"691b073e05ba0f932952d9f4e6fc60e065fec6302bd334841a4253caa54a2bce"} Nov 24 14:52:51 crc kubenswrapper[4790]: I1124 14:52:51.643223 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-58dfc66457-rqfks" event={"ID":"f78f1722-568a-4643-9e7b-1d2bb2449456","Type":"ContainerDied","Data":"9fd79e743941d0f3c4727ccb7d95e21cbc95a5dbb54c5db0c445373886353ef4"} Nov 24 14:52:51 crc kubenswrapper[4790]: I1124 14:52:51.783935 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-58dfc66457-rqfks" Nov 24 14:52:51 crc kubenswrapper[4790]: I1124 14:52:51.962859 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f78f1722-568a-4643-9e7b-1d2bb2449456-scripts\") pod \"f78f1722-568a-4643-9e7b-1d2bb2449456\" (UID: \"f78f1722-568a-4643-9e7b-1d2bb2449456\") " Nov 24 14:52:51 crc kubenswrapper[4790]: I1124 14:52:51.963133 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7t98\" (UniqueName: \"kubernetes.io/projected/f78f1722-568a-4643-9e7b-1d2bb2449456-kube-api-access-f7t98\") pod \"f78f1722-568a-4643-9e7b-1d2bb2449456\" (UID: \"f78f1722-568a-4643-9e7b-1d2bb2449456\") " Nov 24 14:52:51 crc kubenswrapper[4790]: I1124 14:52:51.963177 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f78f1722-568a-4643-9e7b-1d2bb2449456-logs\") pod \"f78f1722-568a-4643-9e7b-1d2bb2449456\" (UID: \"f78f1722-568a-4643-9e7b-1d2bb2449456\") " Nov 24 14:52:51 crc kubenswrapper[4790]: I1124 14:52:51.963307 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f78f1722-568a-4643-9e7b-1d2bb2449456-config-data\") pod \"f78f1722-568a-4643-9e7b-1d2bb2449456\" (UID: \"f78f1722-568a-4643-9e7b-1d2bb2449456\") " Nov 24 14:52:51 crc kubenswrapper[4790]: I1124 14:52:51.963431 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f78f1722-568a-4643-9e7b-1d2bb2449456-horizon-secret-key\") pod \"f78f1722-568a-4643-9e7b-1d2bb2449456\" (UID: \"f78f1722-568a-4643-9e7b-1d2bb2449456\") " Nov 24 14:52:51 crc kubenswrapper[4790]: I1124 14:52:51.963844 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f78f1722-568a-4643-9e7b-1d2bb2449456-logs" (OuterVolumeSpecName: "logs") pod "f78f1722-568a-4643-9e7b-1d2bb2449456" (UID: "f78f1722-568a-4643-9e7b-1d2bb2449456"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:52:51 crc kubenswrapper[4790]: I1124 14:52:51.978216 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f78f1722-568a-4643-9e7b-1d2bb2449456-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "f78f1722-568a-4643-9e7b-1d2bb2449456" (UID: "f78f1722-568a-4643-9e7b-1d2bb2449456"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:52:51 crc kubenswrapper[4790]: I1124 14:52:51.978235 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f78f1722-568a-4643-9e7b-1d2bb2449456-kube-api-access-f7t98" (OuterVolumeSpecName: "kube-api-access-f7t98") pod "f78f1722-568a-4643-9e7b-1d2bb2449456" (UID: "f78f1722-568a-4643-9e7b-1d2bb2449456"). InnerVolumeSpecName "kube-api-access-f7t98". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:52:51 crc kubenswrapper[4790]: I1124 14:52:51.990659 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f78f1722-568a-4643-9e7b-1d2bb2449456-scripts" (OuterVolumeSpecName: "scripts") pod "f78f1722-568a-4643-9e7b-1d2bb2449456" (UID: "f78f1722-568a-4643-9e7b-1d2bb2449456"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:52:52 crc kubenswrapper[4790]: I1124 14:52:52.003771 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f78f1722-568a-4643-9e7b-1d2bb2449456-config-data" (OuterVolumeSpecName: "config-data") pod "f78f1722-568a-4643-9e7b-1d2bb2449456" (UID: "f78f1722-568a-4643-9e7b-1d2bb2449456"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:52:52 crc kubenswrapper[4790]: I1124 14:52:52.043168 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-79f5cf4d49-bhz7p" podUID="ceec03ba-936a-4d49-983e-307237454683" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.111:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.111:8080: connect: connection refused" Nov 24 14:52:52 crc kubenswrapper[4790]: I1124 14:52:52.065824 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7t98\" (UniqueName: \"kubernetes.io/projected/f78f1722-568a-4643-9e7b-1d2bb2449456-kube-api-access-f7t98\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:52 crc kubenswrapper[4790]: I1124 14:52:52.065890 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f78f1722-568a-4643-9e7b-1d2bb2449456-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:52 crc kubenswrapper[4790]: I1124 14:52:52.065906 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f78f1722-568a-4643-9e7b-1d2bb2449456-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:52 crc kubenswrapper[4790]: I1124 14:52:52.065921 4790 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f78f1722-568a-4643-9e7b-1d2bb2449456-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:52 crc kubenswrapper[4790]: I1124 14:52:52.065930 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f78f1722-568a-4643-9e7b-1d2bb2449456-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:52 crc kubenswrapper[4790]: I1124 14:52:52.655733 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-58dfc66457-rqfks" event={"ID":"f78f1722-568a-4643-9e7b-1d2bb2449456","Type":"ContainerDied","Data":"d7387057bdc83a18de55571982040dcc2c95437651945bcf48a7c9082e62bc32"} Nov 24 14:52:52 crc kubenswrapper[4790]: I1124 14:52:52.655783 4790 scope.go:117] "RemoveContainer" containerID="691b073e05ba0f932952d9f4e6fc60e065fec6302bd334841a4253caa54a2bce" Nov 24 14:52:52 crc kubenswrapper[4790]: I1124 14:52:52.655926 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-58dfc66457-rqfks" Nov 24 14:52:52 crc kubenswrapper[4790]: I1124 14:52:52.687191 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-58dfc66457-rqfks"] Nov 24 14:52:52 crc kubenswrapper[4790]: I1124 14:52:52.696341 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-58dfc66457-rqfks"] Nov 24 14:52:52 crc kubenswrapper[4790]: I1124 14:52:52.866901 4790 scope.go:117] "RemoveContainer" containerID="9fd79e743941d0f3c4727ccb7d95e21cbc95a5dbb54c5db0c445373886353ef4" Nov 24 14:52:54 crc kubenswrapper[4790]: I1124 14:52:54.325837 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f78f1722-568a-4643-9e7b-1d2bb2449456" path="/var/lib/kubelet/pods/f78f1722-568a-4643-9e7b-1d2bb2449456/volumes" Nov 24 14:52:56 crc kubenswrapper[4790]: I1124 14:52:56.314783 4790 scope.go:117] "RemoveContainer" containerID="28308703817a1d4e0e5344ce19268127c600894f2daa6b921d52a9105f014132" Nov 24 14:52:56 crc kubenswrapper[4790]: E1124 14:52:56.315494 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:53:02 crc kubenswrapper[4790]: I1124 14:53:02.041936 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-79f5cf4d49-bhz7p" podUID="ceec03ba-936a-4d49-983e-307237454683" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.111:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.111:8080: connect: connection refused" Nov 24 14:53:07 crc kubenswrapper[4790]: I1124 14:53:07.050938 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-6b9f-account-create-wb2xr"] Nov 24 14:53:07 crc kubenswrapper[4790]: I1124 14:53:07.066137 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-6b9f-account-create-wb2xr"] Nov 24 14:53:07 crc kubenswrapper[4790]: I1124 14:53:07.092353 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-dpfn5"] Nov 24 14:53:07 crc kubenswrapper[4790]: I1124 14:53:07.106560 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-dpfn5"] Nov 24 14:53:07 crc kubenswrapper[4790]: I1124 14:53:07.316416 4790 scope.go:117] "RemoveContainer" containerID="28308703817a1d4e0e5344ce19268127c600894f2daa6b921d52a9105f014132" Nov 24 14:53:07 crc kubenswrapper[4790]: E1124 14:53:07.317066 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:53:08 crc kubenswrapper[4790]: I1124 14:53:08.336124 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e5cfb13-d03c-4981-b260-2ae89c41f0bf" path="/var/lib/kubelet/pods/2e5cfb13-d03c-4981-b260-2ae89c41f0bf/volumes" Nov 24 14:53:08 crc kubenswrapper[4790]: I1124 14:53:08.338822 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda25cdc-1a37-44a2-8314-461f0f1de55a" path="/var/lib/kubelet/pods/fda25cdc-1a37-44a2-8314-461f0f1de55a/volumes" Nov 24 14:53:12 crc kubenswrapper[4790]: I1124 14:53:12.042758 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-79f5cf4d49-bhz7p" podUID="ceec03ba-936a-4d49-983e-307237454683" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.111:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.111:8080: connect: connection refused" Nov 24 14:53:12 crc kubenswrapper[4790]: I1124 14:53:12.043433 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-79f5cf4d49-bhz7p" Nov 24 14:53:15 crc kubenswrapper[4790]: I1124 14:53:15.044013 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-tslzm"] Nov 24 14:53:15 crc kubenswrapper[4790]: I1124 14:53:15.054760 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-tslzm"] Nov 24 14:53:16 crc kubenswrapper[4790]: I1124 14:53:16.341427 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b7cfed5-a39c-4bee-b427-f74b2af51e73" path="/var/lib/kubelet/pods/0b7cfed5-a39c-4bee-b427-f74b2af51e73/volumes" Nov 24 14:53:16 crc kubenswrapper[4790]: I1124 14:53:16.628857 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79f5cf4d49-bhz7p" Nov 24 14:53:16 crc kubenswrapper[4790]: I1124 14:53:16.762711 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ceec03ba-936a-4d49-983e-307237454683-logs\") pod \"ceec03ba-936a-4d49-983e-307237454683\" (UID: \"ceec03ba-936a-4d49-983e-307237454683\") " Nov 24 14:53:16 crc kubenswrapper[4790]: I1124 14:53:16.762808 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ceec03ba-936a-4d49-983e-307237454683-config-data\") pod \"ceec03ba-936a-4d49-983e-307237454683\" (UID: \"ceec03ba-936a-4d49-983e-307237454683\") " Nov 24 14:53:16 crc kubenswrapper[4790]: I1124 14:53:16.762909 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4vpx\" (UniqueName: \"kubernetes.io/projected/ceec03ba-936a-4d49-983e-307237454683-kube-api-access-r4vpx\") pod \"ceec03ba-936a-4d49-983e-307237454683\" (UID: \"ceec03ba-936a-4d49-983e-307237454683\") " Nov 24 14:53:16 crc kubenswrapper[4790]: I1124 14:53:16.763017 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ceec03ba-936a-4d49-983e-307237454683-scripts\") pod \"ceec03ba-936a-4d49-983e-307237454683\" (UID: \"ceec03ba-936a-4d49-983e-307237454683\") " Nov 24 14:53:16 crc kubenswrapper[4790]: I1124 14:53:16.763137 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ceec03ba-936a-4d49-983e-307237454683-horizon-secret-key\") pod \"ceec03ba-936a-4d49-983e-307237454683\" (UID: \"ceec03ba-936a-4d49-983e-307237454683\") " Nov 24 14:53:16 crc kubenswrapper[4790]: I1124 14:53:16.763148 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ceec03ba-936a-4d49-983e-307237454683-logs" (OuterVolumeSpecName: "logs") pod "ceec03ba-936a-4d49-983e-307237454683" (UID: "ceec03ba-936a-4d49-983e-307237454683"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:53:16 crc kubenswrapper[4790]: I1124 14:53:16.763623 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ceec03ba-936a-4d49-983e-307237454683-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:53:16 crc kubenswrapper[4790]: I1124 14:53:16.767986 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ceec03ba-936a-4d49-983e-307237454683-kube-api-access-r4vpx" (OuterVolumeSpecName: "kube-api-access-r4vpx") pod "ceec03ba-936a-4d49-983e-307237454683" (UID: "ceec03ba-936a-4d49-983e-307237454683"). InnerVolumeSpecName "kube-api-access-r4vpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:53:16 crc kubenswrapper[4790]: I1124 14:53:16.769242 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ceec03ba-936a-4d49-983e-307237454683-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "ceec03ba-936a-4d49-983e-307237454683" (UID: "ceec03ba-936a-4d49-983e-307237454683"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:53:16 crc kubenswrapper[4790]: I1124 14:53:16.787908 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ceec03ba-936a-4d49-983e-307237454683-scripts" (OuterVolumeSpecName: "scripts") pod "ceec03ba-936a-4d49-983e-307237454683" (UID: "ceec03ba-936a-4d49-983e-307237454683"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:53:16 crc kubenswrapper[4790]: I1124 14:53:16.797925 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ceec03ba-936a-4d49-983e-307237454683-config-data" (OuterVolumeSpecName: "config-data") pod "ceec03ba-936a-4d49-983e-307237454683" (UID: "ceec03ba-936a-4d49-983e-307237454683"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:53:16 crc kubenswrapper[4790]: I1124 14:53:16.865956 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ceec03ba-936a-4d49-983e-307237454683-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:53:16 crc kubenswrapper[4790]: I1124 14:53:16.866021 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4vpx\" (UniqueName: \"kubernetes.io/projected/ceec03ba-936a-4d49-983e-307237454683-kube-api-access-r4vpx\") on node \"crc\" DevicePath \"\"" Nov 24 14:53:16 crc kubenswrapper[4790]: I1124 14:53:16.866046 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ceec03ba-936a-4d49-983e-307237454683-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:53:16 crc kubenswrapper[4790]: I1124 14:53:16.866067 4790 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ceec03ba-936a-4d49-983e-307237454683-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:53:16 crc kubenswrapper[4790]: I1124 14:53:16.936822 4790 generic.go:334] "Generic (PLEG): container finished" podID="ceec03ba-936a-4d49-983e-307237454683" containerID="ede38852f5d6ed963db0ac16661982c274841c82f2bf78bb674d2d3b5f0de39d" exitCode=137 Nov 24 14:53:16 crc kubenswrapper[4790]: I1124 14:53:16.936944 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79f5cf4d49-bhz7p" Nov 24 14:53:16 crc kubenswrapper[4790]: I1124 14:53:16.936963 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79f5cf4d49-bhz7p" event={"ID":"ceec03ba-936a-4d49-983e-307237454683","Type":"ContainerDied","Data":"ede38852f5d6ed963db0ac16661982c274841c82f2bf78bb674d2d3b5f0de39d"} Nov 24 14:53:16 crc kubenswrapper[4790]: I1124 14:53:16.937378 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79f5cf4d49-bhz7p" event={"ID":"ceec03ba-936a-4d49-983e-307237454683","Type":"ContainerDied","Data":"af27caa49df61a03ecb6672883cd59a55080117b0c0979bbff5d1d0c25a21eec"} Nov 24 14:53:16 crc kubenswrapper[4790]: I1124 14:53:16.937401 4790 scope.go:117] "RemoveContainer" containerID="3429608e5620adba09bfac7d28a7c297bf04bc207b858b7431e735aa1bd40c36" Nov 24 14:53:16 crc kubenswrapper[4790]: I1124 14:53:16.972217 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-79f5cf4d49-bhz7p"] Nov 24 14:53:16 crc kubenswrapper[4790]: I1124 14:53:16.982606 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-79f5cf4d49-bhz7p"] Nov 24 14:53:17 crc kubenswrapper[4790]: I1124 14:53:17.151959 4790 scope.go:117] "RemoveContainer" containerID="ede38852f5d6ed963db0ac16661982c274841c82f2bf78bb674d2d3b5f0de39d" Nov 24 14:53:17 crc kubenswrapper[4790]: I1124 14:53:17.193547 4790 scope.go:117] "RemoveContainer" containerID="3429608e5620adba09bfac7d28a7c297bf04bc207b858b7431e735aa1bd40c36" Nov 24 14:53:17 crc kubenswrapper[4790]: E1124 14:53:17.194655 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3429608e5620adba09bfac7d28a7c297bf04bc207b858b7431e735aa1bd40c36\": container with ID starting with 3429608e5620adba09bfac7d28a7c297bf04bc207b858b7431e735aa1bd40c36 not found: ID does not exist" containerID="3429608e5620adba09bfac7d28a7c297bf04bc207b858b7431e735aa1bd40c36" Nov 24 14:53:17 crc kubenswrapper[4790]: I1124 14:53:17.194733 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3429608e5620adba09bfac7d28a7c297bf04bc207b858b7431e735aa1bd40c36"} err="failed to get container status \"3429608e5620adba09bfac7d28a7c297bf04bc207b858b7431e735aa1bd40c36\": rpc error: code = NotFound desc = could not find container \"3429608e5620adba09bfac7d28a7c297bf04bc207b858b7431e735aa1bd40c36\": container with ID starting with 3429608e5620adba09bfac7d28a7c297bf04bc207b858b7431e735aa1bd40c36 not found: ID does not exist" Nov 24 14:53:17 crc kubenswrapper[4790]: I1124 14:53:17.194781 4790 scope.go:117] "RemoveContainer" containerID="ede38852f5d6ed963db0ac16661982c274841c82f2bf78bb674d2d3b5f0de39d" Nov 24 14:53:17 crc kubenswrapper[4790]: E1124 14:53:17.195487 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ede38852f5d6ed963db0ac16661982c274841c82f2bf78bb674d2d3b5f0de39d\": container with ID starting with ede38852f5d6ed963db0ac16661982c274841c82f2bf78bb674d2d3b5f0de39d not found: ID does not exist" containerID="ede38852f5d6ed963db0ac16661982c274841c82f2bf78bb674d2d3b5f0de39d" Nov 24 14:53:17 crc kubenswrapper[4790]: I1124 14:53:17.195567 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ede38852f5d6ed963db0ac16661982c274841c82f2bf78bb674d2d3b5f0de39d"} err="failed to get container status \"ede38852f5d6ed963db0ac16661982c274841c82f2bf78bb674d2d3b5f0de39d\": rpc error: code = NotFound desc = could not find container \"ede38852f5d6ed963db0ac16661982c274841c82f2bf78bb674d2d3b5f0de39d\": container with ID starting with ede38852f5d6ed963db0ac16661982c274841c82f2bf78bb674d2d3b5f0de39d not found: ID does not exist" Nov 24 14:53:18 crc kubenswrapper[4790]: I1124 14:53:18.338382 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ceec03ba-936a-4d49-983e-307237454683" path="/var/lib/kubelet/pods/ceec03ba-936a-4d49-983e-307237454683/volumes" Nov 24 14:53:20 crc kubenswrapper[4790]: I1124 14:53:20.315567 4790 scope.go:117] "RemoveContainer" containerID="28308703817a1d4e0e5344ce19268127c600894f2daa6b921d52a9105f014132" Nov 24 14:53:20 crc kubenswrapper[4790]: E1124 14:53:20.316312 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:53:29 crc kubenswrapper[4790]: I1124 14:53:29.517304 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-8487684449-pkmn4"] Nov 24 14:53:29 crc kubenswrapper[4790]: E1124 14:53:29.518839 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ceec03ba-936a-4d49-983e-307237454683" containerName="horizon" Nov 24 14:53:29 crc kubenswrapper[4790]: I1124 14:53:29.518957 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="ceec03ba-936a-4d49-983e-307237454683" containerName="horizon" Nov 24 14:53:29 crc kubenswrapper[4790]: E1124 14:53:29.519029 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f78f1722-568a-4643-9e7b-1d2bb2449456" containerName="horizon-log" Nov 24 14:53:29 crc kubenswrapper[4790]: I1124 14:53:29.519079 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="f78f1722-568a-4643-9e7b-1d2bb2449456" containerName="horizon-log" Nov 24 14:53:29 crc kubenswrapper[4790]: E1124 14:53:29.519146 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ceec03ba-936a-4d49-983e-307237454683" containerName="horizon-log" Nov 24 14:53:29 crc kubenswrapper[4790]: I1124 14:53:29.519201 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="ceec03ba-936a-4d49-983e-307237454683" containerName="horizon-log" Nov 24 14:53:29 crc kubenswrapper[4790]: E1124 14:53:29.519263 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f78f1722-568a-4643-9e7b-1d2bb2449456" containerName="horizon" Nov 24 14:53:29 crc kubenswrapper[4790]: I1124 14:53:29.519320 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="f78f1722-568a-4643-9e7b-1d2bb2449456" containerName="horizon" Nov 24 14:53:29 crc kubenswrapper[4790]: I1124 14:53:29.520698 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="f78f1722-568a-4643-9e7b-1d2bb2449456" containerName="horizon" Nov 24 14:53:29 crc kubenswrapper[4790]: I1124 14:53:29.520795 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="ceec03ba-936a-4d49-983e-307237454683" containerName="horizon" Nov 24 14:53:29 crc kubenswrapper[4790]: I1124 14:53:29.520904 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="ceec03ba-936a-4d49-983e-307237454683" containerName="horizon-log" Nov 24 14:53:29 crc kubenswrapper[4790]: I1124 14:53:29.520976 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="f78f1722-568a-4643-9e7b-1d2bb2449456" containerName="horizon-log" Nov 24 14:53:29 crc kubenswrapper[4790]: I1124 14:53:29.522204 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8487684449-pkmn4" Nov 24 14:53:29 crc kubenswrapper[4790]: I1124 14:53:29.544444 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-8487684449-pkmn4"] Nov 24 14:53:29 crc kubenswrapper[4790]: I1124 14:53:29.678308 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a96b1a40-c526-4796-9023-fd8f4f699323-scripts\") pod \"horizon-8487684449-pkmn4\" (UID: \"a96b1a40-c526-4796-9023-fd8f4f699323\") " pod="openstack/horizon-8487684449-pkmn4" Nov 24 14:53:29 crc kubenswrapper[4790]: I1124 14:53:29.678350 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a96b1a40-c526-4796-9023-fd8f4f699323-config-data\") pod \"horizon-8487684449-pkmn4\" (UID: \"a96b1a40-c526-4796-9023-fd8f4f699323\") " pod="openstack/horizon-8487684449-pkmn4" Nov 24 14:53:29 crc kubenswrapper[4790]: I1124 14:53:29.678382 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a96b1a40-c526-4796-9023-fd8f4f699323-horizon-secret-key\") pod \"horizon-8487684449-pkmn4\" (UID: \"a96b1a40-c526-4796-9023-fd8f4f699323\") " pod="openstack/horizon-8487684449-pkmn4" Nov 24 14:53:29 crc kubenswrapper[4790]: I1124 14:53:29.679096 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4tc7\" (UniqueName: \"kubernetes.io/projected/a96b1a40-c526-4796-9023-fd8f4f699323-kube-api-access-c4tc7\") pod \"horizon-8487684449-pkmn4\" (UID: \"a96b1a40-c526-4796-9023-fd8f4f699323\") " pod="openstack/horizon-8487684449-pkmn4" Nov 24 14:53:29 crc kubenswrapper[4790]: I1124 14:53:29.679406 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a96b1a40-c526-4796-9023-fd8f4f699323-logs\") pod \"horizon-8487684449-pkmn4\" (UID: \"a96b1a40-c526-4796-9023-fd8f4f699323\") " pod="openstack/horizon-8487684449-pkmn4" Nov 24 14:53:29 crc kubenswrapper[4790]: I1124 14:53:29.781857 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a96b1a40-c526-4796-9023-fd8f4f699323-scripts\") pod \"horizon-8487684449-pkmn4\" (UID: \"a96b1a40-c526-4796-9023-fd8f4f699323\") " pod="openstack/horizon-8487684449-pkmn4" Nov 24 14:53:29 crc kubenswrapper[4790]: I1124 14:53:29.782346 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a96b1a40-c526-4796-9023-fd8f4f699323-config-data\") pod \"horizon-8487684449-pkmn4\" (UID: \"a96b1a40-c526-4796-9023-fd8f4f699323\") " pod="openstack/horizon-8487684449-pkmn4" Nov 24 14:53:29 crc kubenswrapper[4790]: I1124 14:53:29.784522 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a96b1a40-c526-4796-9023-fd8f4f699323-horizon-secret-key\") pod \"horizon-8487684449-pkmn4\" (UID: \"a96b1a40-c526-4796-9023-fd8f4f699323\") " pod="openstack/horizon-8487684449-pkmn4" Nov 24 14:53:29 crc kubenswrapper[4790]: I1124 14:53:29.784416 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a96b1a40-c526-4796-9023-fd8f4f699323-config-data\") pod \"horizon-8487684449-pkmn4\" (UID: \"a96b1a40-c526-4796-9023-fd8f4f699323\") " pod="openstack/horizon-8487684449-pkmn4" Nov 24 14:53:29 crc kubenswrapper[4790]: I1124 14:53:29.782642 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a96b1a40-c526-4796-9023-fd8f4f699323-scripts\") pod \"horizon-8487684449-pkmn4\" (UID: \"a96b1a40-c526-4796-9023-fd8f4f699323\") " pod="openstack/horizon-8487684449-pkmn4" Nov 24 14:53:29 crc kubenswrapper[4790]: I1124 14:53:29.786985 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4tc7\" (UniqueName: \"kubernetes.io/projected/a96b1a40-c526-4796-9023-fd8f4f699323-kube-api-access-c4tc7\") pod \"horizon-8487684449-pkmn4\" (UID: \"a96b1a40-c526-4796-9023-fd8f4f699323\") " pod="openstack/horizon-8487684449-pkmn4" Nov 24 14:53:29 crc kubenswrapper[4790]: I1124 14:53:29.787411 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a96b1a40-c526-4796-9023-fd8f4f699323-logs\") pod \"horizon-8487684449-pkmn4\" (UID: \"a96b1a40-c526-4796-9023-fd8f4f699323\") " pod="openstack/horizon-8487684449-pkmn4" Nov 24 14:53:29 crc kubenswrapper[4790]: I1124 14:53:29.788204 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a96b1a40-c526-4796-9023-fd8f4f699323-logs\") pod \"horizon-8487684449-pkmn4\" (UID: \"a96b1a40-c526-4796-9023-fd8f4f699323\") " pod="openstack/horizon-8487684449-pkmn4" Nov 24 14:53:29 crc kubenswrapper[4790]: I1124 14:53:29.794503 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a96b1a40-c526-4796-9023-fd8f4f699323-horizon-secret-key\") pod \"horizon-8487684449-pkmn4\" (UID: \"a96b1a40-c526-4796-9023-fd8f4f699323\") " pod="openstack/horizon-8487684449-pkmn4" Nov 24 14:53:29 crc kubenswrapper[4790]: I1124 14:53:29.820431 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4tc7\" (UniqueName: \"kubernetes.io/projected/a96b1a40-c526-4796-9023-fd8f4f699323-kube-api-access-c4tc7\") pod \"horizon-8487684449-pkmn4\" (UID: \"a96b1a40-c526-4796-9023-fd8f4f699323\") " pod="openstack/horizon-8487684449-pkmn4" Nov 24 14:53:29 crc kubenswrapper[4790]: I1124 14:53:29.844375 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8487684449-pkmn4" Nov 24 14:53:30 crc kubenswrapper[4790]: I1124 14:53:30.387948 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-8487684449-pkmn4"] Nov 24 14:53:30 crc kubenswrapper[4790]: I1124 14:53:30.799777 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-pd5mv"] Nov 24 14:53:30 crc kubenswrapper[4790]: I1124 14:53:30.801754 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-pd5mv" Nov 24 14:53:30 crc kubenswrapper[4790]: I1124 14:53:30.807760 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-ff8b-account-create-fw2ks"] Nov 24 14:53:30 crc kubenswrapper[4790]: I1124 14:53:30.809114 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-ff8b-account-create-fw2ks" Nov 24 14:53:30 crc kubenswrapper[4790]: I1124 14:53:30.810900 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Nov 24 14:53:30 crc kubenswrapper[4790]: I1124 14:53:30.822914 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-pd5mv"] Nov 24 14:53:30 crc kubenswrapper[4790]: I1124 14:53:30.857438 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-ff8b-account-create-fw2ks"] Nov 24 14:53:30 crc kubenswrapper[4790]: I1124 14:53:30.906352 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7tbq\" (UniqueName: \"kubernetes.io/projected/6501aed1-881c-481d-abd1-92da3c7d0121-kube-api-access-m7tbq\") pod \"heat-ff8b-account-create-fw2ks\" (UID: \"6501aed1-881c-481d-abd1-92da3c7d0121\") " pod="openstack/heat-ff8b-account-create-fw2ks" Nov 24 14:53:30 crc kubenswrapper[4790]: I1124 14:53:30.906575 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6501aed1-881c-481d-abd1-92da3c7d0121-operator-scripts\") pod \"heat-ff8b-account-create-fw2ks\" (UID: \"6501aed1-881c-481d-abd1-92da3c7d0121\") " pod="openstack/heat-ff8b-account-create-fw2ks" Nov 24 14:53:30 crc kubenswrapper[4790]: I1124 14:53:30.906733 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f152e218-4ab5-45a5-a87b-c8d4178c0f20-operator-scripts\") pod \"heat-db-create-pd5mv\" (UID: \"f152e218-4ab5-45a5-a87b-c8d4178c0f20\") " pod="openstack/heat-db-create-pd5mv" Nov 24 14:53:30 crc kubenswrapper[4790]: I1124 14:53:30.906931 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9khs\" (UniqueName: \"kubernetes.io/projected/f152e218-4ab5-45a5-a87b-c8d4178c0f20-kube-api-access-x9khs\") pod \"heat-db-create-pd5mv\" (UID: \"f152e218-4ab5-45a5-a87b-c8d4178c0f20\") " pod="openstack/heat-db-create-pd5mv" Nov 24 14:53:31 crc kubenswrapper[4790]: I1124 14:53:31.008643 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f152e218-4ab5-45a5-a87b-c8d4178c0f20-operator-scripts\") pod \"heat-db-create-pd5mv\" (UID: \"f152e218-4ab5-45a5-a87b-c8d4178c0f20\") " pod="openstack/heat-db-create-pd5mv" Nov 24 14:53:31 crc kubenswrapper[4790]: I1124 14:53:31.008726 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9khs\" (UniqueName: \"kubernetes.io/projected/f152e218-4ab5-45a5-a87b-c8d4178c0f20-kube-api-access-x9khs\") pod \"heat-db-create-pd5mv\" (UID: \"f152e218-4ab5-45a5-a87b-c8d4178c0f20\") " pod="openstack/heat-db-create-pd5mv" Nov 24 14:53:31 crc kubenswrapper[4790]: I1124 14:53:31.008845 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7tbq\" (UniqueName: \"kubernetes.io/projected/6501aed1-881c-481d-abd1-92da3c7d0121-kube-api-access-m7tbq\") pod \"heat-ff8b-account-create-fw2ks\" (UID: \"6501aed1-881c-481d-abd1-92da3c7d0121\") " pod="openstack/heat-ff8b-account-create-fw2ks" Nov 24 14:53:31 crc kubenswrapper[4790]: I1124 14:53:31.009125 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6501aed1-881c-481d-abd1-92da3c7d0121-operator-scripts\") pod \"heat-ff8b-account-create-fw2ks\" (UID: \"6501aed1-881c-481d-abd1-92da3c7d0121\") " pod="openstack/heat-ff8b-account-create-fw2ks" Nov 24 14:53:31 crc kubenswrapper[4790]: I1124 14:53:31.010362 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f152e218-4ab5-45a5-a87b-c8d4178c0f20-operator-scripts\") pod \"heat-db-create-pd5mv\" (UID: \"f152e218-4ab5-45a5-a87b-c8d4178c0f20\") " pod="openstack/heat-db-create-pd5mv" Nov 24 14:53:31 crc kubenswrapper[4790]: I1124 14:53:31.011943 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6501aed1-881c-481d-abd1-92da3c7d0121-operator-scripts\") pod \"heat-ff8b-account-create-fw2ks\" (UID: \"6501aed1-881c-481d-abd1-92da3c7d0121\") " pod="openstack/heat-ff8b-account-create-fw2ks" Nov 24 14:53:31 crc kubenswrapper[4790]: I1124 14:53:31.043207 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7tbq\" (UniqueName: \"kubernetes.io/projected/6501aed1-881c-481d-abd1-92da3c7d0121-kube-api-access-m7tbq\") pod \"heat-ff8b-account-create-fw2ks\" (UID: \"6501aed1-881c-481d-abd1-92da3c7d0121\") " pod="openstack/heat-ff8b-account-create-fw2ks" Nov 24 14:53:31 crc kubenswrapper[4790]: I1124 14:53:31.063575 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9khs\" (UniqueName: \"kubernetes.io/projected/f152e218-4ab5-45a5-a87b-c8d4178c0f20-kube-api-access-x9khs\") pod \"heat-db-create-pd5mv\" (UID: \"f152e218-4ab5-45a5-a87b-c8d4178c0f20\") " pod="openstack/heat-db-create-pd5mv" Nov 24 14:53:31 crc kubenswrapper[4790]: I1124 14:53:31.139381 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8487684449-pkmn4" event={"ID":"a96b1a40-c526-4796-9023-fd8f4f699323","Type":"ContainerStarted","Data":"13d006e63c1a7eaee27892e39ad5816dae81ae1a4eb48c62cdbd40218fdd236c"} Nov 24 14:53:31 crc kubenswrapper[4790]: I1124 14:53:31.139431 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8487684449-pkmn4" event={"ID":"a96b1a40-c526-4796-9023-fd8f4f699323","Type":"ContainerStarted","Data":"e90c802085e88ed4c6ed0e806ab5ceb438d97862ab2e571a093b0dd957048d8c"} Nov 24 14:53:31 crc kubenswrapper[4790]: I1124 14:53:31.139441 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8487684449-pkmn4" event={"ID":"a96b1a40-c526-4796-9023-fd8f4f699323","Type":"ContainerStarted","Data":"75fe02bb0db392003d6b9150583360f1b55488978e79fcb11c9ded4c9d4b9588"} Nov 24 14:53:31 crc kubenswrapper[4790]: I1124 14:53:31.164607 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-8487684449-pkmn4" podStartSLOduration=2.164587491 podStartE2EDuration="2.164587491s" podCreationTimestamp="2025-11-24 14:53:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:53:31.156749316 +0000 UTC m=+6059.536642978" watchObservedRunningTime="2025-11-24 14:53:31.164587491 +0000 UTC m=+6059.544481153" Nov 24 14:53:31 crc kubenswrapper[4790]: I1124 14:53:31.181284 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-pd5mv" Nov 24 14:53:31 crc kubenswrapper[4790]: I1124 14:53:31.189606 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-ff8b-account-create-fw2ks" Nov 24 14:53:31 crc kubenswrapper[4790]: I1124 14:53:31.315463 4790 scope.go:117] "RemoveContainer" containerID="28308703817a1d4e0e5344ce19268127c600894f2daa6b921d52a9105f014132" Nov 24 14:53:31 crc kubenswrapper[4790]: E1124 14:53:31.316010 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:53:31 crc kubenswrapper[4790]: I1124 14:53:31.716184 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-pd5mv"] Nov 24 14:53:31 crc kubenswrapper[4790]: I1124 14:53:31.736981 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-ff8b-account-create-fw2ks"] Nov 24 14:53:32 crc kubenswrapper[4790]: I1124 14:53:32.150198 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-ff8b-account-create-fw2ks" event={"ID":"6501aed1-881c-481d-abd1-92da3c7d0121","Type":"ContainerStarted","Data":"ee8bb4fd17e401091383fb7fb4fa7bf0909ed573b8dde432e905058966c643b3"} Nov 24 14:53:32 crc kubenswrapper[4790]: I1124 14:53:32.150581 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-ff8b-account-create-fw2ks" event={"ID":"6501aed1-881c-481d-abd1-92da3c7d0121","Type":"ContainerStarted","Data":"bb419a9bc6aa3cf1bee37aeed5c3becf932767fba0d8ad2f6c661a23ad4f9883"} Nov 24 14:53:32 crc kubenswrapper[4790]: I1124 14:53:32.153073 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-pd5mv" event={"ID":"f152e218-4ab5-45a5-a87b-c8d4178c0f20","Type":"ContainerStarted","Data":"f85f79c2bf1bdb2d3e9824fb8e78a4de6856485471122600b3b2c511fa64f873"} Nov 24 14:53:32 crc kubenswrapper[4790]: I1124 14:53:32.153126 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-pd5mv" event={"ID":"f152e218-4ab5-45a5-a87b-c8d4178c0f20","Type":"ContainerStarted","Data":"ef01f45efee2868eaaace1f81ef052d3d9332f5060d92aa8d05d3807828f81e8"} Nov 24 14:53:32 crc kubenswrapper[4790]: I1124 14:53:32.171845 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-ff8b-account-create-fw2ks" podStartSLOduration=2.171824196 podStartE2EDuration="2.171824196s" podCreationTimestamp="2025-11-24 14:53:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:53:32.168145271 +0000 UTC m=+6060.548038953" watchObservedRunningTime="2025-11-24 14:53:32.171824196 +0000 UTC m=+6060.551717858" Nov 24 14:53:32 crc kubenswrapper[4790]: I1124 14:53:32.186122 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-create-pd5mv" podStartSLOduration=2.186100526 podStartE2EDuration="2.186100526s" podCreationTimestamp="2025-11-24 14:53:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:53:32.184041737 +0000 UTC m=+6060.563935409" watchObservedRunningTime="2025-11-24 14:53:32.186100526 +0000 UTC m=+6060.565994198" Nov 24 14:53:33 crc kubenswrapper[4790]: I1124 14:53:33.170627 4790 generic.go:334] "Generic (PLEG): container finished" podID="6501aed1-881c-481d-abd1-92da3c7d0121" containerID="ee8bb4fd17e401091383fb7fb4fa7bf0909ed573b8dde432e905058966c643b3" exitCode=0 Nov 24 14:53:33 crc kubenswrapper[4790]: I1124 14:53:33.170668 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-ff8b-account-create-fw2ks" event={"ID":"6501aed1-881c-481d-abd1-92da3c7d0121","Type":"ContainerDied","Data":"ee8bb4fd17e401091383fb7fb4fa7bf0909ed573b8dde432e905058966c643b3"} Nov 24 14:53:33 crc kubenswrapper[4790]: I1124 14:53:33.181374 4790 generic.go:334] "Generic (PLEG): container finished" podID="f152e218-4ab5-45a5-a87b-c8d4178c0f20" containerID="f85f79c2bf1bdb2d3e9824fb8e78a4de6856485471122600b3b2c511fa64f873" exitCode=0 Nov 24 14:53:33 crc kubenswrapper[4790]: I1124 14:53:33.181432 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-pd5mv" event={"ID":"f152e218-4ab5-45a5-a87b-c8d4178c0f20","Type":"ContainerDied","Data":"f85f79c2bf1bdb2d3e9824fb8e78a4de6856485471122600b3b2c511fa64f873"} Nov 24 14:53:34 crc kubenswrapper[4790]: I1124 14:53:34.615643 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-ff8b-account-create-fw2ks" Nov 24 14:53:34 crc kubenswrapper[4790]: I1124 14:53:34.631404 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-pd5mv" Nov 24 14:53:34 crc kubenswrapper[4790]: I1124 14:53:34.798951 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7tbq\" (UniqueName: \"kubernetes.io/projected/6501aed1-881c-481d-abd1-92da3c7d0121-kube-api-access-m7tbq\") pod \"6501aed1-881c-481d-abd1-92da3c7d0121\" (UID: \"6501aed1-881c-481d-abd1-92da3c7d0121\") " Nov 24 14:53:34 crc kubenswrapper[4790]: I1124 14:53:34.799046 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6501aed1-881c-481d-abd1-92da3c7d0121-operator-scripts\") pod \"6501aed1-881c-481d-abd1-92da3c7d0121\" (UID: \"6501aed1-881c-481d-abd1-92da3c7d0121\") " Nov 24 14:53:34 crc kubenswrapper[4790]: I1124 14:53:34.799129 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f152e218-4ab5-45a5-a87b-c8d4178c0f20-operator-scripts\") pod \"f152e218-4ab5-45a5-a87b-c8d4178c0f20\" (UID: \"f152e218-4ab5-45a5-a87b-c8d4178c0f20\") " Nov 24 14:53:34 crc kubenswrapper[4790]: I1124 14:53:34.799298 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9khs\" (UniqueName: \"kubernetes.io/projected/f152e218-4ab5-45a5-a87b-c8d4178c0f20-kube-api-access-x9khs\") pod \"f152e218-4ab5-45a5-a87b-c8d4178c0f20\" (UID: \"f152e218-4ab5-45a5-a87b-c8d4178c0f20\") " Nov 24 14:53:34 crc kubenswrapper[4790]: I1124 14:53:34.800180 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f152e218-4ab5-45a5-a87b-c8d4178c0f20-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f152e218-4ab5-45a5-a87b-c8d4178c0f20" (UID: "f152e218-4ab5-45a5-a87b-c8d4178c0f20"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:53:34 crc kubenswrapper[4790]: I1124 14:53:34.800196 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6501aed1-881c-481d-abd1-92da3c7d0121-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6501aed1-881c-481d-abd1-92da3c7d0121" (UID: "6501aed1-881c-481d-abd1-92da3c7d0121"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:53:34 crc kubenswrapper[4790]: I1124 14:53:34.804984 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6501aed1-881c-481d-abd1-92da3c7d0121-kube-api-access-m7tbq" (OuterVolumeSpecName: "kube-api-access-m7tbq") pod "6501aed1-881c-481d-abd1-92da3c7d0121" (UID: "6501aed1-881c-481d-abd1-92da3c7d0121"). InnerVolumeSpecName "kube-api-access-m7tbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:53:34 crc kubenswrapper[4790]: I1124 14:53:34.805733 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f152e218-4ab5-45a5-a87b-c8d4178c0f20-kube-api-access-x9khs" (OuterVolumeSpecName: "kube-api-access-x9khs") pod "f152e218-4ab5-45a5-a87b-c8d4178c0f20" (UID: "f152e218-4ab5-45a5-a87b-c8d4178c0f20"). InnerVolumeSpecName "kube-api-access-x9khs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:53:34 crc kubenswrapper[4790]: I1124 14:53:34.901261 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7tbq\" (UniqueName: \"kubernetes.io/projected/6501aed1-881c-481d-abd1-92da3c7d0121-kube-api-access-m7tbq\") on node \"crc\" DevicePath \"\"" Nov 24 14:53:34 crc kubenswrapper[4790]: I1124 14:53:34.901307 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6501aed1-881c-481d-abd1-92da3c7d0121-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:53:34 crc kubenswrapper[4790]: I1124 14:53:34.901319 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f152e218-4ab5-45a5-a87b-c8d4178c0f20-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:53:34 crc kubenswrapper[4790]: I1124 14:53:34.901330 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9khs\" (UniqueName: \"kubernetes.io/projected/f152e218-4ab5-45a5-a87b-c8d4178c0f20-kube-api-access-x9khs\") on node \"crc\" DevicePath \"\"" Nov 24 14:53:35 crc kubenswrapper[4790]: I1124 14:53:35.199741 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-ff8b-account-create-fw2ks" event={"ID":"6501aed1-881c-481d-abd1-92da3c7d0121","Type":"ContainerDied","Data":"bb419a9bc6aa3cf1bee37aeed5c3becf932767fba0d8ad2f6c661a23ad4f9883"} Nov 24 14:53:35 crc kubenswrapper[4790]: I1124 14:53:35.199788 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb419a9bc6aa3cf1bee37aeed5c3becf932767fba0d8ad2f6c661a23ad4f9883" Nov 24 14:53:35 crc kubenswrapper[4790]: I1124 14:53:35.199791 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-ff8b-account-create-fw2ks" Nov 24 14:53:35 crc kubenswrapper[4790]: I1124 14:53:35.202684 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-pd5mv" event={"ID":"f152e218-4ab5-45a5-a87b-c8d4178c0f20","Type":"ContainerDied","Data":"ef01f45efee2868eaaace1f81ef052d3d9332f5060d92aa8d05d3807828f81e8"} Nov 24 14:53:35 crc kubenswrapper[4790]: I1124 14:53:35.202735 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef01f45efee2868eaaace1f81ef052d3d9332f5060d92aa8d05d3807828f81e8" Nov 24 14:53:35 crc kubenswrapper[4790]: I1124 14:53:35.202738 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-pd5mv" Nov 24 14:53:36 crc kubenswrapper[4790]: I1124 14:53:36.065424 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-j42q5"] Nov 24 14:53:36 crc kubenswrapper[4790]: E1124 14:53:36.066210 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6501aed1-881c-481d-abd1-92da3c7d0121" containerName="mariadb-account-create" Nov 24 14:53:36 crc kubenswrapper[4790]: I1124 14:53:36.066225 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="6501aed1-881c-481d-abd1-92da3c7d0121" containerName="mariadb-account-create" Nov 24 14:53:36 crc kubenswrapper[4790]: E1124 14:53:36.066275 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f152e218-4ab5-45a5-a87b-c8d4178c0f20" containerName="mariadb-database-create" Nov 24 14:53:36 crc kubenswrapper[4790]: I1124 14:53:36.066285 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="f152e218-4ab5-45a5-a87b-c8d4178c0f20" containerName="mariadb-database-create" Nov 24 14:53:36 crc kubenswrapper[4790]: I1124 14:53:36.066560 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="f152e218-4ab5-45a5-a87b-c8d4178c0f20" containerName="mariadb-database-create" Nov 24 14:53:36 crc kubenswrapper[4790]: I1124 14:53:36.066583 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="6501aed1-881c-481d-abd1-92da3c7d0121" containerName="mariadb-account-create" Nov 24 14:53:36 crc kubenswrapper[4790]: I1124 14:53:36.067484 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-j42q5" Nov 24 14:53:36 crc kubenswrapper[4790]: I1124 14:53:36.069849 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-pkmvn" Nov 24 14:53:36 crc kubenswrapper[4790]: I1124 14:53:36.070084 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Nov 24 14:53:36 crc kubenswrapper[4790]: I1124 14:53:36.095422 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-j42q5"] Nov 24 14:53:36 crc kubenswrapper[4790]: I1124 14:53:36.225388 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d55c02d-16d4-41aa-9ac5-98236ea35b71-config-data\") pod \"heat-db-sync-j42q5\" (UID: \"9d55c02d-16d4-41aa-9ac5-98236ea35b71\") " pod="openstack/heat-db-sync-j42q5" Nov 24 14:53:36 crc kubenswrapper[4790]: I1124 14:53:36.225968 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d55c02d-16d4-41aa-9ac5-98236ea35b71-combined-ca-bundle\") pod \"heat-db-sync-j42q5\" (UID: \"9d55c02d-16d4-41aa-9ac5-98236ea35b71\") " pod="openstack/heat-db-sync-j42q5" Nov 24 14:53:36 crc kubenswrapper[4790]: I1124 14:53:36.226015 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrw2x\" (UniqueName: \"kubernetes.io/projected/9d55c02d-16d4-41aa-9ac5-98236ea35b71-kube-api-access-hrw2x\") pod \"heat-db-sync-j42q5\" (UID: \"9d55c02d-16d4-41aa-9ac5-98236ea35b71\") " pod="openstack/heat-db-sync-j42q5" Nov 24 14:53:36 crc kubenswrapper[4790]: I1124 14:53:36.328041 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d55c02d-16d4-41aa-9ac5-98236ea35b71-combined-ca-bundle\") pod \"heat-db-sync-j42q5\" (UID: \"9d55c02d-16d4-41aa-9ac5-98236ea35b71\") " pod="openstack/heat-db-sync-j42q5" Nov 24 14:53:36 crc kubenswrapper[4790]: I1124 14:53:36.328099 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrw2x\" (UniqueName: \"kubernetes.io/projected/9d55c02d-16d4-41aa-9ac5-98236ea35b71-kube-api-access-hrw2x\") pod \"heat-db-sync-j42q5\" (UID: \"9d55c02d-16d4-41aa-9ac5-98236ea35b71\") " pod="openstack/heat-db-sync-j42q5" Nov 24 14:53:36 crc kubenswrapper[4790]: I1124 14:53:36.328181 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d55c02d-16d4-41aa-9ac5-98236ea35b71-config-data\") pod \"heat-db-sync-j42q5\" (UID: \"9d55c02d-16d4-41aa-9ac5-98236ea35b71\") " pod="openstack/heat-db-sync-j42q5" Nov 24 14:53:36 crc kubenswrapper[4790]: I1124 14:53:36.336367 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d55c02d-16d4-41aa-9ac5-98236ea35b71-config-data\") pod \"heat-db-sync-j42q5\" (UID: \"9d55c02d-16d4-41aa-9ac5-98236ea35b71\") " pod="openstack/heat-db-sync-j42q5" Nov 24 14:53:36 crc kubenswrapper[4790]: I1124 14:53:36.347571 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d55c02d-16d4-41aa-9ac5-98236ea35b71-combined-ca-bundle\") pod \"heat-db-sync-j42q5\" (UID: \"9d55c02d-16d4-41aa-9ac5-98236ea35b71\") " pod="openstack/heat-db-sync-j42q5" Nov 24 14:53:36 crc kubenswrapper[4790]: I1124 14:53:36.361213 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrw2x\" (UniqueName: \"kubernetes.io/projected/9d55c02d-16d4-41aa-9ac5-98236ea35b71-kube-api-access-hrw2x\") pod \"heat-db-sync-j42q5\" (UID: \"9d55c02d-16d4-41aa-9ac5-98236ea35b71\") " pod="openstack/heat-db-sync-j42q5" Nov 24 14:53:36 crc kubenswrapper[4790]: I1124 14:53:36.386819 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-j42q5" Nov 24 14:53:36 crc kubenswrapper[4790]: I1124 14:53:36.868447 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-j42q5"] Nov 24 14:53:37 crc kubenswrapper[4790]: I1124 14:53:37.220806 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-j42q5" event={"ID":"9d55c02d-16d4-41aa-9ac5-98236ea35b71","Type":"ContainerStarted","Data":"47aa5f5e808d1d7d4c71829fe28e81183f214de960f17a045a9ccc1c647f09c0"} Nov 24 14:53:39 crc kubenswrapper[4790]: I1124 14:53:39.844507 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-8487684449-pkmn4" Nov 24 14:53:39 crc kubenswrapper[4790]: I1124 14:53:39.845039 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-8487684449-pkmn4" Nov 24 14:53:44 crc kubenswrapper[4790]: I1124 14:53:44.299827 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-j42q5" event={"ID":"9d55c02d-16d4-41aa-9ac5-98236ea35b71","Type":"ContainerStarted","Data":"ec9bedbc7b636fd77da187df1c0e895ea495bc019f0ff22ce501ab052a36d412"} Nov 24 14:53:44 crc kubenswrapper[4790]: I1124 14:53:44.314285 4790 scope.go:117] "RemoveContainer" containerID="28308703817a1d4e0e5344ce19268127c600894f2daa6b921d52a9105f014132" Nov 24 14:53:44 crc kubenswrapper[4790]: E1124 14:53:44.314528 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:53:44 crc kubenswrapper[4790]: I1124 14:53:44.333141 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-j42q5" podStartSLOduration=1.180968303 podStartE2EDuration="8.333120673s" podCreationTimestamp="2025-11-24 14:53:36 +0000 UTC" firstStartedPulling="2025-11-24 14:53:36.871770361 +0000 UTC m=+6065.251664023" lastFinishedPulling="2025-11-24 14:53:44.023922721 +0000 UTC m=+6072.403816393" observedRunningTime="2025-11-24 14:53:44.327679447 +0000 UTC m=+6072.707573119" watchObservedRunningTime="2025-11-24 14:53:44.333120673 +0000 UTC m=+6072.713014335" Nov 24 14:53:46 crc kubenswrapper[4790]: I1124 14:53:46.057692 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-6251-account-create-9fg9t"] Nov 24 14:53:46 crc kubenswrapper[4790]: I1124 14:53:46.072330 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-2qmwz"] Nov 24 14:53:46 crc kubenswrapper[4790]: I1124 14:53:46.084340 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-2qmwz"] Nov 24 14:53:46 crc kubenswrapper[4790]: I1124 14:53:46.092901 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-6251-account-create-9fg9t"] Nov 24 14:53:46 crc kubenswrapper[4790]: I1124 14:53:46.327310 4790 generic.go:334] "Generic (PLEG): container finished" podID="9d55c02d-16d4-41aa-9ac5-98236ea35b71" containerID="ec9bedbc7b636fd77da187df1c0e895ea495bc019f0ff22ce501ab052a36d412" exitCode=0 Nov 24 14:53:46 crc kubenswrapper[4790]: I1124 14:53:46.329903 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80e59caa-53d1-4697-b987-e93cba1bf143" path="/var/lib/kubelet/pods/80e59caa-53d1-4697-b987-e93cba1bf143/volumes" Nov 24 14:53:46 crc kubenswrapper[4790]: I1124 14:53:46.332037 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c460e843-5c12-44dd-8245-e58ef2ea3546" path="/var/lib/kubelet/pods/c460e843-5c12-44dd-8245-e58ef2ea3546/volumes" Nov 24 14:53:46 crc kubenswrapper[4790]: I1124 14:53:46.343413 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-j42q5" event={"ID":"9d55c02d-16d4-41aa-9ac5-98236ea35b71","Type":"ContainerDied","Data":"ec9bedbc7b636fd77da187df1c0e895ea495bc019f0ff22ce501ab052a36d412"} Nov 24 14:53:47 crc kubenswrapper[4790]: I1124 14:53:47.757544 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-j42q5" Nov 24 14:53:47 crc kubenswrapper[4790]: I1124 14:53:47.898420 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d55c02d-16d4-41aa-9ac5-98236ea35b71-combined-ca-bundle\") pod \"9d55c02d-16d4-41aa-9ac5-98236ea35b71\" (UID: \"9d55c02d-16d4-41aa-9ac5-98236ea35b71\") " Nov 24 14:53:47 crc kubenswrapper[4790]: I1124 14:53:47.898712 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d55c02d-16d4-41aa-9ac5-98236ea35b71-config-data\") pod \"9d55c02d-16d4-41aa-9ac5-98236ea35b71\" (UID: \"9d55c02d-16d4-41aa-9ac5-98236ea35b71\") " Nov 24 14:53:47 crc kubenswrapper[4790]: I1124 14:53:47.898792 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrw2x\" (UniqueName: \"kubernetes.io/projected/9d55c02d-16d4-41aa-9ac5-98236ea35b71-kube-api-access-hrw2x\") pod \"9d55c02d-16d4-41aa-9ac5-98236ea35b71\" (UID: \"9d55c02d-16d4-41aa-9ac5-98236ea35b71\") " Nov 24 14:53:47 crc kubenswrapper[4790]: I1124 14:53:47.920525 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d55c02d-16d4-41aa-9ac5-98236ea35b71-kube-api-access-hrw2x" (OuterVolumeSpecName: "kube-api-access-hrw2x") pod "9d55c02d-16d4-41aa-9ac5-98236ea35b71" (UID: "9d55c02d-16d4-41aa-9ac5-98236ea35b71"). InnerVolumeSpecName "kube-api-access-hrw2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:53:47 crc kubenswrapper[4790]: I1124 14:53:47.945867 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d55c02d-16d4-41aa-9ac5-98236ea35b71-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9d55c02d-16d4-41aa-9ac5-98236ea35b71" (UID: "9d55c02d-16d4-41aa-9ac5-98236ea35b71"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:53:48 crc kubenswrapper[4790]: I1124 14:53:48.001052 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrw2x\" (UniqueName: \"kubernetes.io/projected/9d55c02d-16d4-41aa-9ac5-98236ea35b71-kube-api-access-hrw2x\") on node \"crc\" DevicePath \"\"" Nov 24 14:53:48 crc kubenswrapper[4790]: I1124 14:53:48.001080 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d55c02d-16d4-41aa-9ac5-98236ea35b71-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:53:48 crc kubenswrapper[4790]: I1124 14:53:48.007270 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d55c02d-16d4-41aa-9ac5-98236ea35b71-config-data" (OuterVolumeSpecName: "config-data") pod "9d55c02d-16d4-41aa-9ac5-98236ea35b71" (UID: "9d55c02d-16d4-41aa-9ac5-98236ea35b71"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:53:48 crc kubenswrapper[4790]: I1124 14:53:48.103350 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d55c02d-16d4-41aa-9ac5-98236ea35b71-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:53:48 crc kubenswrapper[4790]: I1124 14:53:48.366295 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-j42q5" event={"ID":"9d55c02d-16d4-41aa-9ac5-98236ea35b71","Type":"ContainerDied","Data":"47aa5f5e808d1d7d4c71829fe28e81183f214de960f17a045a9ccc1c647f09c0"} Nov 24 14:53:48 crc kubenswrapper[4790]: I1124 14:53:48.366355 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47aa5f5e808d1d7d4c71829fe28e81183f214de960f17a045a9ccc1c647f09c0" Nov 24 14:53:48 crc kubenswrapper[4790]: I1124 14:53:48.366420 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-j42q5" Nov 24 14:53:49 crc kubenswrapper[4790]: I1124 14:53:49.655690 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-76c78d699-44kq8"] Nov 24 14:53:49 crc kubenswrapper[4790]: E1124 14:53:49.657760 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d55c02d-16d4-41aa-9ac5-98236ea35b71" containerName="heat-db-sync" Nov 24 14:53:49 crc kubenswrapper[4790]: I1124 14:53:49.657786 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d55c02d-16d4-41aa-9ac5-98236ea35b71" containerName="heat-db-sync" Nov 24 14:53:49 crc kubenswrapper[4790]: I1124 14:53:49.658077 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d55c02d-16d4-41aa-9ac5-98236ea35b71" containerName="heat-db-sync" Nov 24 14:53:49 crc kubenswrapper[4790]: I1124 14:53:49.658925 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-76c78d699-44kq8" Nov 24 14:53:49 crc kubenswrapper[4790]: I1124 14:53:49.661930 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Nov 24 14:53:49 crc kubenswrapper[4790]: I1124 14:53:49.663071 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-pkmvn" Nov 24 14:53:49 crc kubenswrapper[4790]: I1124 14:53:49.672122 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Nov 24 14:53:49 crc kubenswrapper[4790]: I1124 14:53:49.677350 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-76c78d699-44kq8"] Nov 24 14:53:49 crc kubenswrapper[4790]: I1124 14:53:49.769271 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-746456fccd-k95mp"] Nov 24 14:53:49 crc kubenswrapper[4790]: I1124 14:53:49.770581 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-746456fccd-k95mp" Nov 24 14:53:49 crc kubenswrapper[4790]: I1124 14:53:49.772203 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Nov 24 14:53:49 crc kubenswrapper[4790]: I1124 14:53:49.836014 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-746456fccd-k95mp"] Nov 24 14:53:49 crc kubenswrapper[4790]: I1124 14:53:49.841641 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5f0775f-7428-4cda-aa40-8e8b50df6c81-combined-ca-bundle\") pod \"heat-engine-76c78d699-44kq8\" (UID: \"f5f0775f-7428-4cda-aa40-8e8b50df6c81\") " pod="openstack/heat-engine-76c78d699-44kq8" Nov 24 14:53:49 crc kubenswrapper[4790]: I1124 14:53:49.841678 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5f0775f-7428-4cda-aa40-8e8b50df6c81-config-data\") pod \"heat-engine-76c78d699-44kq8\" (UID: \"f5f0775f-7428-4cda-aa40-8e8b50df6c81\") " pod="openstack/heat-engine-76c78d699-44kq8" Nov 24 14:53:49 crc kubenswrapper[4790]: I1124 14:53:49.841719 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krt9x\" (UniqueName: \"kubernetes.io/projected/f5f0775f-7428-4cda-aa40-8e8b50df6c81-kube-api-access-krt9x\") pod \"heat-engine-76c78d699-44kq8\" (UID: \"f5f0775f-7428-4cda-aa40-8e8b50df6c81\") " pod="openstack/heat-engine-76c78d699-44kq8" Nov 24 14:53:49 crc kubenswrapper[4790]: I1124 14:53:49.841758 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f5f0775f-7428-4cda-aa40-8e8b50df6c81-config-data-custom\") pod \"heat-engine-76c78d699-44kq8\" (UID: \"f5f0775f-7428-4cda-aa40-8e8b50df6c81\") " pod="openstack/heat-engine-76c78d699-44kq8" Nov 24 14:53:49 crc kubenswrapper[4790]: I1124 14:53:49.906326 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-688db64b5b-xg9f7"] Nov 24 14:53:49 crc kubenswrapper[4790]: I1124 14:53:49.907699 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-688db64b5b-xg9f7" Nov 24 14:53:49 crc kubenswrapper[4790]: I1124 14:53:49.912552 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Nov 24 14:53:49 crc kubenswrapper[4790]: I1124 14:53:49.942036 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-688db64b5b-xg9f7"] Nov 24 14:53:49 crc kubenswrapper[4790]: I1124 14:53:49.943026 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9d135e7-5d36-4403-826e-e75f7d654ce9-combined-ca-bundle\") pod \"heat-api-746456fccd-k95mp\" (UID: \"c9d135e7-5d36-4403-826e-e75f7d654ce9\") " pod="openstack/heat-api-746456fccd-k95mp" Nov 24 14:53:49 crc kubenswrapper[4790]: I1124 14:53:49.943078 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c9d135e7-5d36-4403-826e-e75f7d654ce9-config-data-custom\") pod \"heat-api-746456fccd-k95mp\" (UID: \"c9d135e7-5d36-4403-826e-e75f7d654ce9\") " pod="openstack/heat-api-746456fccd-k95mp" Nov 24 14:53:49 crc kubenswrapper[4790]: I1124 14:53:49.943338 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5f0775f-7428-4cda-aa40-8e8b50df6c81-combined-ca-bundle\") pod \"heat-engine-76c78d699-44kq8\" (UID: \"f5f0775f-7428-4cda-aa40-8e8b50df6c81\") " pod="openstack/heat-engine-76c78d699-44kq8" Nov 24 14:53:49 crc kubenswrapper[4790]: I1124 14:53:49.943374 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5f0775f-7428-4cda-aa40-8e8b50df6c81-config-data\") pod \"heat-engine-76c78d699-44kq8\" (UID: \"f5f0775f-7428-4cda-aa40-8e8b50df6c81\") " pod="openstack/heat-engine-76c78d699-44kq8" Nov 24 14:53:49 crc kubenswrapper[4790]: I1124 14:53:49.943455 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krt9x\" (UniqueName: \"kubernetes.io/projected/f5f0775f-7428-4cda-aa40-8e8b50df6c81-kube-api-access-krt9x\") pod \"heat-engine-76c78d699-44kq8\" (UID: \"f5f0775f-7428-4cda-aa40-8e8b50df6c81\") " pod="openstack/heat-engine-76c78d699-44kq8" Nov 24 14:53:49 crc kubenswrapper[4790]: I1124 14:53:49.943535 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f5f0775f-7428-4cda-aa40-8e8b50df6c81-config-data-custom\") pod \"heat-engine-76c78d699-44kq8\" (UID: \"f5f0775f-7428-4cda-aa40-8e8b50df6c81\") " pod="openstack/heat-engine-76c78d699-44kq8" Nov 24 14:53:49 crc kubenswrapper[4790]: I1124 14:53:49.943572 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5c59\" (UniqueName: \"kubernetes.io/projected/c9d135e7-5d36-4403-826e-e75f7d654ce9-kube-api-access-x5c59\") pod \"heat-api-746456fccd-k95mp\" (UID: \"c9d135e7-5d36-4403-826e-e75f7d654ce9\") " pod="openstack/heat-api-746456fccd-k95mp" Nov 24 14:53:49 crc kubenswrapper[4790]: I1124 14:53:49.943597 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9d135e7-5d36-4403-826e-e75f7d654ce9-config-data\") pod \"heat-api-746456fccd-k95mp\" (UID: \"c9d135e7-5d36-4403-826e-e75f7d654ce9\") " pod="openstack/heat-api-746456fccd-k95mp" Nov 24 14:53:49 crc kubenswrapper[4790]: I1124 14:53:49.952679 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5f0775f-7428-4cda-aa40-8e8b50df6c81-combined-ca-bundle\") pod \"heat-engine-76c78d699-44kq8\" (UID: \"f5f0775f-7428-4cda-aa40-8e8b50df6c81\") " pod="openstack/heat-engine-76c78d699-44kq8" Nov 24 14:53:49 crc kubenswrapper[4790]: I1124 14:53:49.958876 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f5f0775f-7428-4cda-aa40-8e8b50df6c81-config-data-custom\") pod \"heat-engine-76c78d699-44kq8\" (UID: \"f5f0775f-7428-4cda-aa40-8e8b50df6c81\") " pod="openstack/heat-engine-76c78d699-44kq8" Nov 24 14:53:49 crc kubenswrapper[4790]: I1124 14:53:49.969517 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5f0775f-7428-4cda-aa40-8e8b50df6c81-config-data\") pod \"heat-engine-76c78d699-44kq8\" (UID: \"f5f0775f-7428-4cda-aa40-8e8b50df6c81\") " pod="openstack/heat-engine-76c78d699-44kq8" Nov 24 14:53:49 crc kubenswrapper[4790]: I1124 14:53:49.971189 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krt9x\" (UniqueName: \"kubernetes.io/projected/f5f0775f-7428-4cda-aa40-8e8b50df6c81-kube-api-access-krt9x\") pod \"heat-engine-76c78d699-44kq8\" (UID: \"f5f0775f-7428-4cda-aa40-8e8b50df6c81\") " pod="openstack/heat-engine-76c78d699-44kq8" Nov 24 14:53:49 crc kubenswrapper[4790]: I1124 14:53:49.983566 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-76c78d699-44kq8" Nov 24 14:53:50 crc kubenswrapper[4790]: I1124 14:53:50.045338 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcnmv\" (UniqueName: \"kubernetes.io/projected/3a294679-0487-4813-8948-97732fe25595-kube-api-access-dcnmv\") pod \"heat-cfnapi-688db64b5b-xg9f7\" (UID: \"3a294679-0487-4813-8948-97732fe25595\") " pod="openstack/heat-cfnapi-688db64b5b-xg9f7" Nov 24 14:53:50 crc kubenswrapper[4790]: I1124 14:53:50.045618 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5c59\" (UniqueName: \"kubernetes.io/projected/c9d135e7-5d36-4403-826e-e75f7d654ce9-kube-api-access-x5c59\") pod \"heat-api-746456fccd-k95mp\" (UID: \"c9d135e7-5d36-4403-826e-e75f7d654ce9\") " pod="openstack/heat-api-746456fccd-k95mp" Nov 24 14:53:50 crc kubenswrapper[4790]: I1124 14:53:50.045642 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9d135e7-5d36-4403-826e-e75f7d654ce9-config-data\") pod \"heat-api-746456fccd-k95mp\" (UID: \"c9d135e7-5d36-4403-826e-e75f7d654ce9\") " pod="openstack/heat-api-746456fccd-k95mp" Nov 24 14:53:50 crc kubenswrapper[4790]: I1124 14:53:50.045667 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a294679-0487-4813-8948-97732fe25595-combined-ca-bundle\") pod \"heat-cfnapi-688db64b5b-xg9f7\" (UID: \"3a294679-0487-4813-8948-97732fe25595\") " pod="openstack/heat-cfnapi-688db64b5b-xg9f7" Nov 24 14:53:50 crc kubenswrapper[4790]: I1124 14:53:50.045695 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a294679-0487-4813-8948-97732fe25595-config-data\") pod \"heat-cfnapi-688db64b5b-xg9f7\" (UID: \"3a294679-0487-4813-8948-97732fe25595\") " pod="openstack/heat-cfnapi-688db64b5b-xg9f7" Nov 24 14:53:50 crc kubenswrapper[4790]: I1124 14:53:50.045740 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9d135e7-5d36-4403-826e-e75f7d654ce9-combined-ca-bundle\") pod \"heat-api-746456fccd-k95mp\" (UID: \"c9d135e7-5d36-4403-826e-e75f7d654ce9\") " pod="openstack/heat-api-746456fccd-k95mp" Nov 24 14:53:50 crc kubenswrapper[4790]: I1124 14:53:50.045772 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c9d135e7-5d36-4403-826e-e75f7d654ce9-config-data-custom\") pod \"heat-api-746456fccd-k95mp\" (UID: \"c9d135e7-5d36-4403-826e-e75f7d654ce9\") " pod="openstack/heat-api-746456fccd-k95mp" Nov 24 14:53:50 crc kubenswrapper[4790]: I1124 14:53:50.045795 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a294679-0487-4813-8948-97732fe25595-config-data-custom\") pod \"heat-cfnapi-688db64b5b-xg9f7\" (UID: \"3a294679-0487-4813-8948-97732fe25595\") " pod="openstack/heat-cfnapi-688db64b5b-xg9f7" Nov 24 14:53:50 crc kubenswrapper[4790]: I1124 14:53:50.050177 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9d135e7-5d36-4403-826e-e75f7d654ce9-combined-ca-bundle\") pod \"heat-api-746456fccd-k95mp\" (UID: \"c9d135e7-5d36-4403-826e-e75f7d654ce9\") " pod="openstack/heat-api-746456fccd-k95mp" Nov 24 14:53:50 crc kubenswrapper[4790]: I1124 14:53:50.050364 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c9d135e7-5d36-4403-826e-e75f7d654ce9-config-data-custom\") pod \"heat-api-746456fccd-k95mp\" (UID: \"c9d135e7-5d36-4403-826e-e75f7d654ce9\") " pod="openstack/heat-api-746456fccd-k95mp" Nov 24 14:53:50 crc kubenswrapper[4790]: I1124 14:53:50.057926 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9d135e7-5d36-4403-826e-e75f7d654ce9-config-data\") pod \"heat-api-746456fccd-k95mp\" (UID: \"c9d135e7-5d36-4403-826e-e75f7d654ce9\") " pod="openstack/heat-api-746456fccd-k95mp" Nov 24 14:53:50 crc kubenswrapper[4790]: I1124 14:53:50.067522 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5c59\" (UniqueName: \"kubernetes.io/projected/c9d135e7-5d36-4403-826e-e75f7d654ce9-kube-api-access-x5c59\") pod \"heat-api-746456fccd-k95mp\" (UID: \"c9d135e7-5d36-4403-826e-e75f7d654ce9\") " pod="openstack/heat-api-746456fccd-k95mp" Nov 24 14:53:50 crc kubenswrapper[4790]: I1124 14:53:50.120977 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-746456fccd-k95mp" Nov 24 14:53:50 crc kubenswrapper[4790]: I1124 14:53:50.148060 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcnmv\" (UniqueName: \"kubernetes.io/projected/3a294679-0487-4813-8948-97732fe25595-kube-api-access-dcnmv\") pod \"heat-cfnapi-688db64b5b-xg9f7\" (UID: \"3a294679-0487-4813-8948-97732fe25595\") " pod="openstack/heat-cfnapi-688db64b5b-xg9f7" Nov 24 14:53:50 crc kubenswrapper[4790]: I1124 14:53:50.148168 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a294679-0487-4813-8948-97732fe25595-combined-ca-bundle\") pod \"heat-cfnapi-688db64b5b-xg9f7\" (UID: \"3a294679-0487-4813-8948-97732fe25595\") " pod="openstack/heat-cfnapi-688db64b5b-xg9f7" Nov 24 14:53:50 crc kubenswrapper[4790]: I1124 14:53:50.148221 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a294679-0487-4813-8948-97732fe25595-config-data\") pod \"heat-cfnapi-688db64b5b-xg9f7\" (UID: \"3a294679-0487-4813-8948-97732fe25595\") " pod="openstack/heat-cfnapi-688db64b5b-xg9f7" Nov 24 14:53:50 crc kubenswrapper[4790]: I1124 14:53:50.148267 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a294679-0487-4813-8948-97732fe25595-config-data-custom\") pod \"heat-cfnapi-688db64b5b-xg9f7\" (UID: \"3a294679-0487-4813-8948-97732fe25595\") " pod="openstack/heat-cfnapi-688db64b5b-xg9f7" Nov 24 14:53:50 crc kubenswrapper[4790]: I1124 14:53:50.154619 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a294679-0487-4813-8948-97732fe25595-config-data-custom\") pod \"heat-cfnapi-688db64b5b-xg9f7\" (UID: \"3a294679-0487-4813-8948-97732fe25595\") " pod="openstack/heat-cfnapi-688db64b5b-xg9f7" Nov 24 14:53:50 crc kubenswrapper[4790]: I1124 14:53:50.155194 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a294679-0487-4813-8948-97732fe25595-config-data\") pod \"heat-cfnapi-688db64b5b-xg9f7\" (UID: \"3a294679-0487-4813-8948-97732fe25595\") " pod="openstack/heat-cfnapi-688db64b5b-xg9f7" Nov 24 14:53:50 crc kubenswrapper[4790]: I1124 14:53:50.155946 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a294679-0487-4813-8948-97732fe25595-combined-ca-bundle\") pod \"heat-cfnapi-688db64b5b-xg9f7\" (UID: \"3a294679-0487-4813-8948-97732fe25595\") " pod="openstack/heat-cfnapi-688db64b5b-xg9f7" Nov 24 14:53:50 crc kubenswrapper[4790]: I1124 14:53:50.166813 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcnmv\" (UniqueName: \"kubernetes.io/projected/3a294679-0487-4813-8948-97732fe25595-kube-api-access-dcnmv\") pod \"heat-cfnapi-688db64b5b-xg9f7\" (UID: \"3a294679-0487-4813-8948-97732fe25595\") " pod="openstack/heat-cfnapi-688db64b5b-xg9f7" Nov 24 14:53:50 crc kubenswrapper[4790]: I1124 14:53:50.178396 4790 scope.go:117] "RemoveContainer" containerID="a87adc1cd91559216c434df0a21f1568e0d2f3b6b62c1ed294ed2e61f3e21d60" Nov 24 14:53:50 crc kubenswrapper[4790]: I1124 14:53:50.227154 4790 scope.go:117] "RemoveContainer" containerID="1113d3dcf1845850aa7b42b43e945aac7f247747af2f5dc144c9265ca98d2bd4" Nov 24 14:53:50 crc kubenswrapper[4790]: I1124 14:53:50.234254 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-688db64b5b-xg9f7" Nov 24 14:53:50 crc kubenswrapper[4790]: I1124 14:53:50.257023 4790 scope.go:117] "RemoveContainer" containerID="b979bb516a5898b201f6b42009b4d7a638f44ab9c7b9e0e4bfef7783795882eb" Nov 24 14:53:50 crc kubenswrapper[4790]: I1124 14:53:50.315731 4790 scope.go:117] "RemoveContainer" containerID="0a63b354c131cec2adc6fe916b132e0b8847ff1811f52a961791676583a7da0e" Nov 24 14:53:50 crc kubenswrapper[4790]: I1124 14:53:50.398346 4790 scope.go:117] "RemoveContainer" containerID="b816cb66ae39bd0fb01fd1d5efb8914832138243781b2438acf9996a2c387fce" Nov 24 14:53:50 crc kubenswrapper[4790]: I1124 14:53:50.514859 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-76c78d699-44kq8"] Nov 24 14:53:50 crc kubenswrapper[4790]: I1124 14:53:50.713035 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-746456fccd-k95mp"] Nov 24 14:53:50 crc kubenswrapper[4790]: I1124 14:53:50.827131 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-688db64b5b-xg9f7"] Nov 24 14:53:50 crc kubenswrapper[4790]: W1124 14:53:50.844620 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3a294679_0487_4813_8948_97732fe25595.slice/crio-d793aae5e2d36e89dfe7b3eb92afca378f3675e7cbe0803baed2774d3ab9993c WatchSource:0}: Error finding container d793aae5e2d36e89dfe7b3eb92afca378f3675e7cbe0803baed2774d3ab9993c: Status 404 returned error can't find the container with id d793aae5e2d36e89dfe7b3eb92afca378f3675e7cbe0803baed2774d3ab9993c Nov 24 14:53:51 crc kubenswrapper[4790]: I1124 14:53:51.418656 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-688db64b5b-xg9f7" event={"ID":"3a294679-0487-4813-8948-97732fe25595","Type":"ContainerStarted","Data":"d793aae5e2d36e89dfe7b3eb92afca378f3675e7cbe0803baed2774d3ab9993c"} Nov 24 14:53:51 crc kubenswrapper[4790]: I1124 14:53:51.423569 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-746456fccd-k95mp" event={"ID":"c9d135e7-5d36-4403-826e-e75f7d654ce9","Type":"ContainerStarted","Data":"7e2e2999c79f9d6f1a4b34209e980a3e6137dbf0cbcb0a1f232129ce676885c8"} Nov 24 14:53:51 crc kubenswrapper[4790]: I1124 14:53:51.429988 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-76c78d699-44kq8" event={"ID":"f5f0775f-7428-4cda-aa40-8e8b50df6c81","Type":"ContainerStarted","Data":"23fb5a95a0f0919ab90e97fe82e3ebc95cbcddc0997aee1f050d7fb64946a4ff"} Nov 24 14:53:51 crc kubenswrapper[4790]: I1124 14:53:51.430031 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-76c78d699-44kq8" event={"ID":"f5f0775f-7428-4cda-aa40-8e8b50df6c81","Type":"ContainerStarted","Data":"3d418149af3717bcb98a5fb01e6544ccac8f5a770f106149afbbe6c4dada240f"} Nov 24 14:53:51 crc kubenswrapper[4790]: I1124 14:53:51.431549 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-76c78d699-44kq8" Nov 24 14:53:51 crc kubenswrapper[4790]: I1124 14:53:51.456395 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-76c78d699-44kq8" podStartSLOduration=2.456377873 podStartE2EDuration="2.456377873s" podCreationTimestamp="2025-11-24 14:53:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:53:51.4520977 +0000 UTC m=+6079.831991362" watchObservedRunningTime="2025-11-24 14:53:51.456377873 +0000 UTC m=+6079.836271535" Nov 24 14:53:51 crc kubenswrapper[4790]: I1124 14:53:51.908086 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-8487684449-pkmn4" Nov 24 14:53:53 crc kubenswrapper[4790]: I1124 14:53:53.842507 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-8487684449-pkmn4" Nov 24 14:53:53 crc kubenswrapper[4790]: I1124 14:53:53.907439 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-67b58d9dc-n6wf7"] Nov 24 14:53:53 crc kubenswrapper[4790]: I1124 14:53:53.907656 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-67b58d9dc-n6wf7" podUID="693f3fba-7701-4f36-9a9a-f3b733a9528e" containerName="horizon-log" containerID="cri-o://3d4c1ac358c0b377fce1b5a6a3f234a637a98fc3d871f5d98db02e7806121d82" gracePeriod=30 Nov 24 14:53:53 crc kubenswrapper[4790]: I1124 14:53:53.907871 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-67b58d9dc-n6wf7" podUID="693f3fba-7701-4f36-9a9a-f3b733a9528e" containerName="horizon" containerID="cri-o://6a291307e0d4114091ab8e92a5a2b1fbfc8ef126006801b9af437486329093ed" gracePeriod=30 Nov 24 14:53:54 crc kubenswrapper[4790]: I1124 14:53:54.460851 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-746456fccd-k95mp" event={"ID":"c9d135e7-5d36-4403-826e-e75f7d654ce9","Type":"ContainerStarted","Data":"dd4039aa12e391c5268f1f4d4728dc3eed599eb1e517d8da7e64f0b6f8661a8e"} Nov 24 14:53:54 crc kubenswrapper[4790]: I1124 14:53:54.461476 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-746456fccd-k95mp" Nov 24 14:53:54 crc kubenswrapper[4790]: I1124 14:53:54.462547 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-688db64b5b-xg9f7" event={"ID":"3a294679-0487-4813-8948-97732fe25595","Type":"ContainerStarted","Data":"74a7b3e97c97296f76a64b6f4baec75991cb9ca1c7c8d1023c77fa34029d2c1d"} Nov 24 14:53:54 crc kubenswrapper[4790]: I1124 14:53:54.462721 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-688db64b5b-xg9f7" Nov 24 14:53:54 crc kubenswrapper[4790]: I1124 14:53:54.483041 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-746456fccd-k95mp" podStartSLOduration=3.106024756 podStartE2EDuration="5.48302437s" podCreationTimestamp="2025-11-24 14:53:49 +0000 UTC" firstStartedPulling="2025-11-24 14:53:50.763386776 +0000 UTC m=+6079.143280438" lastFinishedPulling="2025-11-24 14:53:53.14038639 +0000 UTC m=+6081.520280052" observedRunningTime="2025-11-24 14:53:54.481070174 +0000 UTC m=+6082.860963846" watchObservedRunningTime="2025-11-24 14:53:54.48302437 +0000 UTC m=+6082.862918042" Nov 24 14:53:54 crc kubenswrapper[4790]: I1124 14:53:54.533394 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-688db64b5b-xg9f7" podStartSLOduration=3.247703684 podStartE2EDuration="5.533378296s" podCreationTimestamp="2025-11-24 14:53:49 +0000 UTC" firstStartedPulling="2025-11-24 14:53:50.850791536 +0000 UTC m=+6079.230685198" lastFinishedPulling="2025-11-24 14:53:53.136466148 +0000 UTC m=+6081.516359810" observedRunningTime="2025-11-24 14:53:54.521643399 +0000 UTC m=+6082.901537071" watchObservedRunningTime="2025-11-24 14:53:54.533378296 +0000 UTC m=+6082.913271958" Nov 24 14:53:56 crc kubenswrapper[4790]: I1124 14:53:56.034344 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-x66l4"] Nov 24 14:53:56 crc kubenswrapper[4790]: I1124 14:53:56.047196 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-x66l4"] Nov 24 14:53:56 crc kubenswrapper[4790]: I1124 14:53:56.366257 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e31ed6a6-8482-4dc5-9da6-46dd87a73053" path="/var/lib/kubelet/pods/e31ed6a6-8482-4dc5-9da6-46dd87a73053/volumes" Nov 24 14:53:57 crc kubenswrapper[4790]: I1124 14:53:57.496933 4790 generic.go:334] "Generic (PLEG): container finished" podID="693f3fba-7701-4f36-9a9a-f3b733a9528e" containerID="6a291307e0d4114091ab8e92a5a2b1fbfc8ef126006801b9af437486329093ed" exitCode=0 Nov 24 14:53:57 crc kubenswrapper[4790]: I1124 14:53:57.497446 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67b58d9dc-n6wf7" event={"ID":"693f3fba-7701-4f36-9a9a-f3b733a9528e","Type":"ContainerDied","Data":"6a291307e0d4114091ab8e92a5a2b1fbfc8ef126006801b9af437486329093ed"} Nov 24 14:53:59 crc kubenswrapper[4790]: I1124 14:53:59.316047 4790 scope.go:117] "RemoveContainer" containerID="28308703817a1d4e0e5344ce19268127c600894f2daa6b921d52a9105f014132" Nov 24 14:53:59 crc kubenswrapper[4790]: E1124 14:53:59.318103 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:54:01 crc kubenswrapper[4790]: I1124 14:54:01.445627 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-746456fccd-k95mp" Nov 24 14:54:01 crc kubenswrapper[4790]: I1124 14:54:01.590537 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-688db64b5b-xg9f7" Nov 24 14:54:02 crc kubenswrapper[4790]: I1124 14:54:02.610598 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-67b58d9dc-n6wf7" podUID="693f3fba-7701-4f36-9a9a-f3b733a9528e" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.112:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.112:8080: connect: connection refused" Nov 24 14:54:09 crc kubenswrapper[4790]: I1124 14:54:09.583746 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xl5lf"] Nov 24 14:54:09 crc kubenswrapper[4790]: I1124 14:54:09.586799 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xl5lf" Nov 24 14:54:09 crc kubenswrapper[4790]: I1124 14:54:09.611505 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xl5lf"] Nov 24 14:54:09 crc kubenswrapper[4790]: I1124 14:54:09.762021 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53d798fe-64a3-4159-9707-6298d4c459a3-utilities\") pod \"certified-operators-xl5lf\" (UID: \"53d798fe-64a3-4159-9707-6298d4c459a3\") " pod="openshift-marketplace/certified-operators-xl5lf" Nov 24 14:54:09 crc kubenswrapper[4790]: I1124 14:54:09.762081 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53d798fe-64a3-4159-9707-6298d4c459a3-catalog-content\") pod \"certified-operators-xl5lf\" (UID: \"53d798fe-64a3-4159-9707-6298d4c459a3\") " pod="openshift-marketplace/certified-operators-xl5lf" Nov 24 14:54:09 crc kubenswrapper[4790]: I1124 14:54:09.762157 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlzv9\" (UniqueName: \"kubernetes.io/projected/53d798fe-64a3-4159-9707-6298d4c459a3-kube-api-access-zlzv9\") pod \"certified-operators-xl5lf\" (UID: \"53d798fe-64a3-4159-9707-6298d4c459a3\") " pod="openshift-marketplace/certified-operators-xl5lf" Nov 24 14:54:09 crc kubenswrapper[4790]: I1124 14:54:09.864492 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlzv9\" (UniqueName: \"kubernetes.io/projected/53d798fe-64a3-4159-9707-6298d4c459a3-kube-api-access-zlzv9\") pod \"certified-operators-xl5lf\" (UID: \"53d798fe-64a3-4159-9707-6298d4c459a3\") " pod="openshift-marketplace/certified-operators-xl5lf" Nov 24 14:54:09 crc kubenswrapper[4790]: I1124 14:54:09.864769 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53d798fe-64a3-4159-9707-6298d4c459a3-utilities\") pod \"certified-operators-xl5lf\" (UID: \"53d798fe-64a3-4159-9707-6298d4c459a3\") " pod="openshift-marketplace/certified-operators-xl5lf" Nov 24 14:54:09 crc kubenswrapper[4790]: I1124 14:54:09.864829 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53d798fe-64a3-4159-9707-6298d4c459a3-catalog-content\") pod \"certified-operators-xl5lf\" (UID: \"53d798fe-64a3-4159-9707-6298d4c459a3\") " pod="openshift-marketplace/certified-operators-xl5lf" Nov 24 14:54:09 crc kubenswrapper[4790]: I1124 14:54:09.865287 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53d798fe-64a3-4159-9707-6298d4c459a3-utilities\") pod \"certified-operators-xl5lf\" (UID: \"53d798fe-64a3-4159-9707-6298d4c459a3\") " pod="openshift-marketplace/certified-operators-xl5lf" Nov 24 14:54:09 crc kubenswrapper[4790]: I1124 14:54:09.865379 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53d798fe-64a3-4159-9707-6298d4c459a3-catalog-content\") pod \"certified-operators-xl5lf\" (UID: \"53d798fe-64a3-4159-9707-6298d4c459a3\") " pod="openshift-marketplace/certified-operators-xl5lf" Nov 24 14:54:09 crc kubenswrapper[4790]: I1124 14:54:09.886805 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlzv9\" (UniqueName: \"kubernetes.io/projected/53d798fe-64a3-4159-9707-6298d4c459a3-kube-api-access-zlzv9\") pod \"certified-operators-xl5lf\" (UID: \"53d798fe-64a3-4159-9707-6298d4c459a3\") " pod="openshift-marketplace/certified-operators-xl5lf" Nov 24 14:54:09 crc kubenswrapper[4790]: I1124 14:54:09.914463 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xl5lf" Nov 24 14:54:10 crc kubenswrapper[4790]: I1124 14:54:10.028616 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-76c78d699-44kq8" Nov 24 14:54:10 crc kubenswrapper[4790]: I1124 14:54:10.252075 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xl5lf"] Nov 24 14:54:10 crc kubenswrapper[4790]: I1124 14:54:10.318272 4790 scope.go:117] "RemoveContainer" containerID="28308703817a1d4e0e5344ce19268127c600894f2daa6b921d52a9105f014132" Nov 24 14:54:10 crc kubenswrapper[4790]: E1124 14:54:10.318478 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:54:10 crc kubenswrapper[4790]: I1124 14:54:10.823398 4790 generic.go:334] "Generic (PLEG): container finished" podID="53d798fe-64a3-4159-9707-6298d4c459a3" containerID="d7f41c7538a521805fa93ecc1ab3ec6bfbd7d294460d182f7317c3ea28ddb4dc" exitCode=0 Nov 24 14:54:10 crc kubenswrapper[4790]: I1124 14:54:10.823677 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xl5lf" event={"ID":"53d798fe-64a3-4159-9707-6298d4c459a3","Type":"ContainerDied","Data":"d7f41c7538a521805fa93ecc1ab3ec6bfbd7d294460d182f7317c3ea28ddb4dc"} Nov 24 14:54:10 crc kubenswrapper[4790]: I1124 14:54:10.823750 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xl5lf" event={"ID":"53d798fe-64a3-4159-9707-6298d4c459a3","Type":"ContainerStarted","Data":"be246cead2e0afae3e86ee1ad710e431d37cf0cf2709ed29a2ee9dcfeb980ba9"} Nov 24 14:54:12 crc kubenswrapper[4790]: I1124 14:54:12.612041 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-67b58d9dc-n6wf7" podUID="693f3fba-7701-4f36-9a9a-f3b733a9528e" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.112:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.112:8080: connect: connection refused" Nov 24 14:54:12 crc kubenswrapper[4790]: I1124 14:54:12.848913 4790 generic.go:334] "Generic (PLEG): container finished" podID="53d798fe-64a3-4159-9707-6298d4c459a3" containerID="f3c2d57645c2ca0251dccfa8d9d7a7c3d7822271ae50e388b24af23c96f556e7" exitCode=0 Nov 24 14:54:12 crc kubenswrapper[4790]: I1124 14:54:12.848969 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xl5lf" event={"ID":"53d798fe-64a3-4159-9707-6298d4c459a3","Type":"ContainerDied","Data":"f3c2d57645c2ca0251dccfa8d9d7a7c3d7822271ae50e388b24af23c96f556e7"} Nov 24 14:54:13 crc kubenswrapper[4790]: I1124 14:54:13.860486 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xl5lf" event={"ID":"53d798fe-64a3-4159-9707-6298d4c459a3","Type":"ContainerStarted","Data":"9b9465c643ebff1a8d5312676a33187d460a1b04a28a5790bf643f9b51933ec4"} Nov 24 14:54:13 crc kubenswrapper[4790]: I1124 14:54:13.890611 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xl5lf" podStartSLOduration=2.422899393 podStartE2EDuration="4.890591982s" podCreationTimestamp="2025-11-24 14:54:09 +0000 UTC" firstStartedPulling="2025-11-24 14:54:10.82581618 +0000 UTC m=+6099.205709872" lastFinishedPulling="2025-11-24 14:54:13.293508799 +0000 UTC m=+6101.673402461" observedRunningTime="2025-11-24 14:54:13.881538711 +0000 UTC m=+6102.261432373" watchObservedRunningTime="2025-11-24 14:54:13.890591982 +0000 UTC m=+6102.270485634" Nov 24 14:54:19 crc kubenswrapper[4790]: I1124 14:54:19.914619 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xl5lf" Nov 24 14:54:19 crc kubenswrapper[4790]: I1124 14:54:19.916765 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xl5lf" Nov 24 14:54:20 crc kubenswrapper[4790]: I1124 14:54:20.960100 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-xl5lf" podUID="53d798fe-64a3-4159-9707-6298d4c459a3" containerName="registry-server" probeResult="failure" output=< Nov 24 14:54:20 crc kubenswrapper[4790]: timeout: failed to connect service ":50051" within 1s Nov 24 14:54:20 crc kubenswrapper[4790]: > Nov 24 14:54:21 crc kubenswrapper[4790]: I1124 14:54:21.315435 4790 scope.go:117] "RemoveContainer" containerID="28308703817a1d4e0e5344ce19268127c600894f2daa6b921d52a9105f014132" Nov 24 14:54:21 crc kubenswrapper[4790]: E1124 14:54:21.315994 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:54:22 crc kubenswrapper[4790]: I1124 14:54:22.613343 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-67b58d9dc-n6wf7" podUID="693f3fba-7701-4f36-9a9a-f3b733a9528e" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.112:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.112:8080: connect: connection refused" Nov 24 14:54:22 crc kubenswrapper[4790]: I1124 14:54:22.613836 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-67b58d9dc-n6wf7" Nov 24 14:54:23 crc kubenswrapper[4790]: I1124 14:54:23.971940 4790 generic.go:334] "Generic (PLEG): container finished" podID="693f3fba-7701-4f36-9a9a-f3b733a9528e" containerID="3d4c1ac358c0b377fce1b5a6a3f234a637a98fc3d871f5d98db02e7806121d82" exitCode=137 Nov 24 14:54:23 crc kubenswrapper[4790]: I1124 14:54:23.972201 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67b58d9dc-n6wf7" event={"ID":"693f3fba-7701-4f36-9a9a-f3b733a9528e","Type":"ContainerDied","Data":"3d4c1ac358c0b377fce1b5a6a3f234a637a98fc3d871f5d98db02e7806121d82"} Nov 24 14:54:24 crc kubenswrapper[4790]: I1124 14:54:24.303617 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67b58d9dc-n6wf7" Nov 24 14:54:24 crc kubenswrapper[4790]: I1124 14:54:24.409876 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/693f3fba-7701-4f36-9a9a-f3b733a9528e-scripts\") pod \"693f3fba-7701-4f36-9a9a-f3b733a9528e\" (UID: \"693f3fba-7701-4f36-9a9a-f3b733a9528e\") " Nov 24 14:54:24 crc kubenswrapper[4790]: I1124 14:54:24.410173 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ftkz\" (UniqueName: \"kubernetes.io/projected/693f3fba-7701-4f36-9a9a-f3b733a9528e-kube-api-access-9ftkz\") pod \"693f3fba-7701-4f36-9a9a-f3b733a9528e\" (UID: \"693f3fba-7701-4f36-9a9a-f3b733a9528e\") " Nov 24 14:54:24 crc kubenswrapper[4790]: I1124 14:54:24.410203 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/693f3fba-7701-4f36-9a9a-f3b733a9528e-config-data\") pod \"693f3fba-7701-4f36-9a9a-f3b733a9528e\" (UID: \"693f3fba-7701-4f36-9a9a-f3b733a9528e\") " Nov 24 14:54:24 crc kubenswrapper[4790]: I1124 14:54:24.410269 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/693f3fba-7701-4f36-9a9a-f3b733a9528e-horizon-secret-key\") pod \"693f3fba-7701-4f36-9a9a-f3b733a9528e\" (UID: \"693f3fba-7701-4f36-9a9a-f3b733a9528e\") " Nov 24 14:54:24 crc kubenswrapper[4790]: I1124 14:54:24.410292 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/693f3fba-7701-4f36-9a9a-f3b733a9528e-logs\") pod \"693f3fba-7701-4f36-9a9a-f3b733a9528e\" (UID: \"693f3fba-7701-4f36-9a9a-f3b733a9528e\") " Nov 24 14:54:24 crc kubenswrapper[4790]: I1124 14:54:24.411185 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/693f3fba-7701-4f36-9a9a-f3b733a9528e-logs" (OuterVolumeSpecName: "logs") pod "693f3fba-7701-4f36-9a9a-f3b733a9528e" (UID: "693f3fba-7701-4f36-9a9a-f3b733a9528e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:54:24 crc kubenswrapper[4790]: I1124 14:54:24.412215 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/693f3fba-7701-4f36-9a9a-f3b733a9528e-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:54:24 crc kubenswrapper[4790]: I1124 14:54:24.419100 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/693f3fba-7701-4f36-9a9a-f3b733a9528e-kube-api-access-9ftkz" (OuterVolumeSpecName: "kube-api-access-9ftkz") pod "693f3fba-7701-4f36-9a9a-f3b733a9528e" (UID: "693f3fba-7701-4f36-9a9a-f3b733a9528e"). InnerVolumeSpecName "kube-api-access-9ftkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:54:24 crc kubenswrapper[4790]: I1124 14:54:24.422439 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/693f3fba-7701-4f36-9a9a-f3b733a9528e-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "693f3fba-7701-4f36-9a9a-f3b733a9528e" (UID: "693f3fba-7701-4f36-9a9a-f3b733a9528e"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:54:24 crc kubenswrapper[4790]: I1124 14:54:24.437584 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/693f3fba-7701-4f36-9a9a-f3b733a9528e-scripts" (OuterVolumeSpecName: "scripts") pod "693f3fba-7701-4f36-9a9a-f3b733a9528e" (UID: "693f3fba-7701-4f36-9a9a-f3b733a9528e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:54:24 crc kubenswrapper[4790]: I1124 14:54:24.443289 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/693f3fba-7701-4f36-9a9a-f3b733a9528e-config-data" (OuterVolumeSpecName: "config-data") pod "693f3fba-7701-4f36-9a9a-f3b733a9528e" (UID: "693f3fba-7701-4f36-9a9a-f3b733a9528e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:54:24 crc kubenswrapper[4790]: I1124 14:54:24.516046 4790 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/693f3fba-7701-4f36-9a9a-f3b733a9528e-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:54:24 crc kubenswrapper[4790]: I1124 14:54:24.516118 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/693f3fba-7701-4f36-9a9a-f3b733a9528e-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:54:24 crc kubenswrapper[4790]: I1124 14:54:24.516148 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ftkz\" (UniqueName: \"kubernetes.io/projected/693f3fba-7701-4f36-9a9a-f3b733a9528e-kube-api-access-9ftkz\") on node \"crc\" DevicePath \"\"" Nov 24 14:54:24 crc kubenswrapper[4790]: I1124 14:54:24.516173 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/693f3fba-7701-4f36-9a9a-f3b733a9528e-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:54:24 crc kubenswrapper[4790]: I1124 14:54:24.983695 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67b58d9dc-n6wf7" event={"ID":"693f3fba-7701-4f36-9a9a-f3b733a9528e","Type":"ContainerDied","Data":"a3816cba24f9486077834ea3cde17ee385663bbcb74a3154330dd1af27386a6f"} Nov 24 14:54:24 crc kubenswrapper[4790]: I1124 14:54:24.983750 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67b58d9dc-n6wf7" Nov 24 14:54:24 crc kubenswrapper[4790]: I1124 14:54:24.983754 4790 scope.go:117] "RemoveContainer" containerID="6a291307e0d4114091ab8e92a5a2b1fbfc8ef126006801b9af437486329093ed" Nov 24 14:54:25 crc kubenswrapper[4790]: I1124 14:54:25.018381 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-67b58d9dc-n6wf7"] Nov 24 14:54:25 crc kubenswrapper[4790]: I1124 14:54:25.027509 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-67b58d9dc-n6wf7"] Nov 24 14:54:25 crc kubenswrapper[4790]: I1124 14:54:25.164875 4790 scope.go:117] "RemoveContainer" containerID="3d4c1ac358c0b377fce1b5a6a3f234a637a98fc3d871f5d98db02e7806121d82" Nov 24 14:54:26 crc kubenswrapper[4790]: I1124 14:54:26.335615 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="693f3fba-7701-4f36-9a9a-f3b733a9528e" path="/var/lib/kubelet/pods/693f3fba-7701-4f36-9a9a-f3b733a9528e/volumes" Nov 24 14:54:29 crc kubenswrapper[4790]: I1124 14:54:29.967758 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xl5lf" Nov 24 14:54:30 crc kubenswrapper[4790]: I1124 14:54:30.035435 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xl5lf" Nov 24 14:54:30 crc kubenswrapper[4790]: I1124 14:54:30.211622 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xl5lf"] Nov 24 14:54:31 crc kubenswrapper[4790]: I1124 14:54:31.051083 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xl5lf" podUID="53d798fe-64a3-4159-9707-6298d4c459a3" containerName="registry-server" containerID="cri-o://9b9465c643ebff1a8d5312676a33187d460a1b04a28a5790bf643f9b51933ec4" gracePeriod=2 Nov 24 14:54:31 crc kubenswrapper[4790]: I1124 14:54:31.539026 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xl5lf" Nov 24 14:54:31 crc kubenswrapper[4790]: I1124 14:54:31.674266 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53d798fe-64a3-4159-9707-6298d4c459a3-utilities\") pod \"53d798fe-64a3-4159-9707-6298d4c459a3\" (UID: \"53d798fe-64a3-4159-9707-6298d4c459a3\") " Nov 24 14:54:31 crc kubenswrapper[4790]: I1124 14:54:31.674551 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53d798fe-64a3-4159-9707-6298d4c459a3-catalog-content\") pod \"53d798fe-64a3-4159-9707-6298d4c459a3\" (UID: \"53d798fe-64a3-4159-9707-6298d4c459a3\") " Nov 24 14:54:31 crc kubenswrapper[4790]: I1124 14:54:31.674637 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlzv9\" (UniqueName: \"kubernetes.io/projected/53d798fe-64a3-4159-9707-6298d4c459a3-kube-api-access-zlzv9\") pod \"53d798fe-64a3-4159-9707-6298d4c459a3\" (UID: \"53d798fe-64a3-4159-9707-6298d4c459a3\") " Nov 24 14:54:31 crc kubenswrapper[4790]: I1124 14:54:31.674989 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53d798fe-64a3-4159-9707-6298d4c459a3-utilities" (OuterVolumeSpecName: "utilities") pod "53d798fe-64a3-4159-9707-6298d4c459a3" (UID: "53d798fe-64a3-4159-9707-6298d4c459a3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:54:31 crc kubenswrapper[4790]: I1124 14:54:31.675318 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53d798fe-64a3-4159-9707-6298d4c459a3-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:54:31 crc kubenswrapper[4790]: I1124 14:54:31.681567 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53d798fe-64a3-4159-9707-6298d4c459a3-kube-api-access-zlzv9" (OuterVolumeSpecName: "kube-api-access-zlzv9") pod "53d798fe-64a3-4159-9707-6298d4c459a3" (UID: "53d798fe-64a3-4159-9707-6298d4c459a3"). InnerVolumeSpecName "kube-api-access-zlzv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:54:31 crc kubenswrapper[4790]: I1124 14:54:31.734756 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53d798fe-64a3-4159-9707-6298d4c459a3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "53d798fe-64a3-4159-9707-6298d4c459a3" (UID: "53d798fe-64a3-4159-9707-6298d4c459a3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:54:31 crc kubenswrapper[4790]: I1124 14:54:31.776649 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53d798fe-64a3-4159-9707-6298d4c459a3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:54:31 crc kubenswrapper[4790]: I1124 14:54:31.776688 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlzv9\" (UniqueName: \"kubernetes.io/projected/53d798fe-64a3-4159-9707-6298d4c459a3-kube-api-access-zlzv9\") on node \"crc\" DevicePath \"\"" Nov 24 14:54:32 crc kubenswrapper[4790]: I1124 14:54:32.074005 4790 generic.go:334] "Generic (PLEG): container finished" podID="53d798fe-64a3-4159-9707-6298d4c459a3" containerID="9b9465c643ebff1a8d5312676a33187d460a1b04a28a5790bf643f9b51933ec4" exitCode=0 Nov 24 14:54:32 crc kubenswrapper[4790]: I1124 14:54:32.074164 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xl5lf" Nov 24 14:54:32 crc kubenswrapper[4790]: I1124 14:54:32.074249 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xl5lf" event={"ID":"53d798fe-64a3-4159-9707-6298d4c459a3","Type":"ContainerDied","Data":"9b9465c643ebff1a8d5312676a33187d460a1b04a28a5790bf643f9b51933ec4"} Nov 24 14:54:32 crc kubenswrapper[4790]: I1124 14:54:32.074531 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xl5lf" event={"ID":"53d798fe-64a3-4159-9707-6298d4c459a3","Type":"ContainerDied","Data":"be246cead2e0afae3e86ee1ad710e431d37cf0cf2709ed29a2ee9dcfeb980ba9"} Nov 24 14:54:32 crc kubenswrapper[4790]: I1124 14:54:32.074564 4790 scope.go:117] "RemoveContainer" containerID="9b9465c643ebff1a8d5312676a33187d460a1b04a28a5790bf643f9b51933ec4" Nov 24 14:54:32 crc kubenswrapper[4790]: I1124 14:54:32.105683 4790 scope.go:117] "RemoveContainer" containerID="f3c2d57645c2ca0251dccfa8d9d7a7c3d7822271ae50e388b24af23c96f556e7" Nov 24 14:54:32 crc kubenswrapper[4790]: I1124 14:54:32.128509 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xl5lf"] Nov 24 14:54:32 crc kubenswrapper[4790]: I1124 14:54:32.140152 4790 scope.go:117] "RemoveContainer" containerID="d7f41c7538a521805fa93ecc1ab3ec6bfbd7d294460d182f7317c3ea28ddb4dc" Nov 24 14:54:32 crc kubenswrapper[4790]: I1124 14:54:32.142233 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xl5lf"] Nov 24 14:54:32 crc kubenswrapper[4790]: I1124 14:54:32.208279 4790 scope.go:117] "RemoveContainer" containerID="9b9465c643ebff1a8d5312676a33187d460a1b04a28a5790bf643f9b51933ec4" Nov 24 14:54:32 crc kubenswrapper[4790]: E1124 14:54:32.208678 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b9465c643ebff1a8d5312676a33187d460a1b04a28a5790bf643f9b51933ec4\": container with ID starting with 9b9465c643ebff1a8d5312676a33187d460a1b04a28a5790bf643f9b51933ec4 not found: ID does not exist" containerID="9b9465c643ebff1a8d5312676a33187d460a1b04a28a5790bf643f9b51933ec4" Nov 24 14:54:32 crc kubenswrapper[4790]: I1124 14:54:32.208710 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b9465c643ebff1a8d5312676a33187d460a1b04a28a5790bf643f9b51933ec4"} err="failed to get container status \"9b9465c643ebff1a8d5312676a33187d460a1b04a28a5790bf643f9b51933ec4\": rpc error: code = NotFound desc = could not find container \"9b9465c643ebff1a8d5312676a33187d460a1b04a28a5790bf643f9b51933ec4\": container with ID starting with 9b9465c643ebff1a8d5312676a33187d460a1b04a28a5790bf643f9b51933ec4 not found: ID does not exist" Nov 24 14:54:32 crc kubenswrapper[4790]: I1124 14:54:32.208730 4790 scope.go:117] "RemoveContainer" containerID="f3c2d57645c2ca0251dccfa8d9d7a7c3d7822271ae50e388b24af23c96f556e7" Nov 24 14:54:32 crc kubenswrapper[4790]: E1124 14:54:32.209169 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3c2d57645c2ca0251dccfa8d9d7a7c3d7822271ae50e388b24af23c96f556e7\": container with ID starting with f3c2d57645c2ca0251dccfa8d9d7a7c3d7822271ae50e388b24af23c96f556e7 not found: ID does not exist" containerID="f3c2d57645c2ca0251dccfa8d9d7a7c3d7822271ae50e388b24af23c96f556e7" Nov 24 14:54:32 crc kubenswrapper[4790]: I1124 14:54:32.209239 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3c2d57645c2ca0251dccfa8d9d7a7c3d7822271ae50e388b24af23c96f556e7"} err="failed to get container status \"f3c2d57645c2ca0251dccfa8d9d7a7c3d7822271ae50e388b24af23c96f556e7\": rpc error: code = NotFound desc = could not find container \"f3c2d57645c2ca0251dccfa8d9d7a7c3d7822271ae50e388b24af23c96f556e7\": container with ID starting with f3c2d57645c2ca0251dccfa8d9d7a7c3d7822271ae50e388b24af23c96f556e7 not found: ID does not exist" Nov 24 14:54:32 crc kubenswrapper[4790]: I1124 14:54:32.209282 4790 scope.go:117] "RemoveContainer" containerID="d7f41c7538a521805fa93ecc1ab3ec6bfbd7d294460d182f7317c3ea28ddb4dc" Nov 24 14:54:32 crc kubenswrapper[4790]: E1124 14:54:32.210031 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7f41c7538a521805fa93ecc1ab3ec6bfbd7d294460d182f7317c3ea28ddb4dc\": container with ID starting with d7f41c7538a521805fa93ecc1ab3ec6bfbd7d294460d182f7317c3ea28ddb4dc not found: ID does not exist" containerID="d7f41c7538a521805fa93ecc1ab3ec6bfbd7d294460d182f7317c3ea28ddb4dc" Nov 24 14:54:32 crc kubenswrapper[4790]: I1124 14:54:32.210091 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7f41c7538a521805fa93ecc1ab3ec6bfbd7d294460d182f7317c3ea28ddb4dc"} err="failed to get container status \"d7f41c7538a521805fa93ecc1ab3ec6bfbd7d294460d182f7317c3ea28ddb4dc\": rpc error: code = NotFound desc = could not find container \"d7f41c7538a521805fa93ecc1ab3ec6bfbd7d294460d182f7317c3ea28ddb4dc\": container with ID starting with d7f41c7538a521805fa93ecc1ab3ec6bfbd7d294460d182f7317c3ea28ddb4dc not found: ID does not exist" Nov 24 14:54:32 crc kubenswrapper[4790]: I1124 14:54:32.330704 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53d798fe-64a3-4159-9707-6298d4c459a3" path="/var/lib/kubelet/pods/53d798fe-64a3-4159-9707-6298d4c459a3/volumes" Nov 24 14:54:35 crc kubenswrapper[4790]: I1124 14:54:35.316921 4790 scope.go:117] "RemoveContainer" containerID="28308703817a1d4e0e5344ce19268127c600894f2daa6b921d52a9105f014132" Nov 24 14:54:35 crc kubenswrapper[4790]: E1124 14:54:35.317523 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:54:45 crc kubenswrapper[4790]: I1124 14:54:45.399152 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8"] Nov 24 14:54:45 crc kubenswrapper[4790]: E1124 14:54:45.400106 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="693f3fba-7701-4f36-9a9a-f3b733a9528e" containerName="horizon-log" Nov 24 14:54:45 crc kubenswrapper[4790]: I1124 14:54:45.400123 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="693f3fba-7701-4f36-9a9a-f3b733a9528e" containerName="horizon-log" Nov 24 14:54:45 crc kubenswrapper[4790]: E1124 14:54:45.400146 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="693f3fba-7701-4f36-9a9a-f3b733a9528e" containerName="horizon" Nov 24 14:54:45 crc kubenswrapper[4790]: I1124 14:54:45.400154 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="693f3fba-7701-4f36-9a9a-f3b733a9528e" containerName="horizon" Nov 24 14:54:45 crc kubenswrapper[4790]: E1124 14:54:45.400176 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53d798fe-64a3-4159-9707-6298d4c459a3" containerName="registry-server" Nov 24 14:54:45 crc kubenswrapper[4790]: I1124 14:54:45.400186 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="53d798fe-64a3-4159-9707-6298d4c459a3" containerName="registry-server" Nov 24 14:54:45 crc kubenswrapper[4790]: E1124 14:54:45.400217 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53d798fe-64a3-4159-9707-6298d4c459a3" containerName="extract-utilities" Nov 24 14:54:45 crc kubenswrapper[4790]: I1124 14:54:45.400226 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="53d798fe-64a3-4159-9707-6298d4c459a3" containerName="extract-utilities" Nov 24 14:54:45 crc kubenswrapper[4790]: E1124 14:54:45.400247 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53d798fe-64a3-4159-9707-6298d4c459a3" containerName="extract-content" Nov 24 14:54:45 crc kubenswrapper[4790]: I1124 14:54:45.400257 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="53d798fe-64a3-4159-9707-6298d4c459a3" containerName="extract-content" Nov 24 14:54:45 crc kubenswrapper[4790]: I1124 14:54:45.400482 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="693f3fba-7701-4f36-9a9a-f3b733a9528e" containerName="horizon" Nov 24 14:54:45 crc kubenswrapper[4790]: I1124 14:54:45.400515 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="693f3fba-7701-4f36-9a9a-f3b733a9528e" containerName="horizon-log" Nov 24 14:54:45 crc kubenswrapper[4790]: I1124 14:54:45.400539 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="53d798fe-64a3-4159-9707-6298d4c459a3" containerName="registry-server" Nov 24 14:54:45 crc kubenswrapper[4790]: I1124 14:54:45.402313 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8" Nov 24 14:54:45 crc kubenswrapper[4790]: I1124 14:54:45.414817 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 14:54:45 crc kubenswrapper[4790]: I1124 14:54:45.419146 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8"] Nov 24 14:54:45 crc kubenswrapper[4790]: I1124 14:54:45.485617 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9dd5c102-10b0-4a96-8a12-6e1ca4f63c30-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8\" (UID: \"9dd5c102-10b0-4a96-8a12-6e1ca4f63c30\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8" Nov 24 14:54:45 crc kubenswrapper[4790]: I1124 14:54:45.485696 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzp4s\" (UniqueName: \"kubernetes.io/projected/9dd5c102-10b0-4a96-8a12-6e1ca4f63c30-kube-api-access-mzp4s\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8\" (UID: \"9dd5c102-10b0-4a96-8a12-6e1ca4f63c30\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8" Nov 24 14:54:45 crc kubenswrapper[4790]: I1124 14:54:45.485746 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9dd5c102-10b0-4a96-8a12-6e1ca4f63c30-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8\" (UID: \"9dd5c102-10b0-4a96-8a12-6e1ca4f63c30\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8" Nov 24 14:54:45 crc kubenswrapper[4790]: I1124 14:54:45.587343 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzp4s\" (UniqueName: \"kubernetes.io/projected/9dd5c102-10b0-4a96-8a12-6e1ca4f63c30-kube-api-access-mzp4s\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8\" (UID: \"9dd5c102-10b0-4a96-8a12-6e1ca4f63c30\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8" Nov 24 14:54:45 crc kubenswrapper[4790]: I1124 14:54:45.587448 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9dd5c102-10b0-4a96-8a12-6e1ca4f63c30-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8\" (UID: \"9dd5c102-10b0-4a96-8a12-6e1ca4f63c30\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8" Nov 24 14:54:45 crc kubenswrapper[4790]: I1124 14:54:45.587593 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9dd5c102-10b0-4a96-8a12-6e1ca4f63c30-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8\" (UID: \"9dd5c102-10b0-4a96-8a12-6e1ca4f63c30\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8" Nov 24 14:54:45 crc kubenswrapper[4790]: I1124 14:54:45.587901 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9dd5c102-10b0-4a96-8a12-6e1ca4f63c30-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8\" (UID: \"9dd5c102-10b0-4a96-8a12-6e1ca4f63c30\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8" Nov 24 14:54:45 crc kubenswrapper[4790]: I1124 14:54:45.588100 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9dd5c102-10b0-4a96-8a12-6e1ca4f63c30-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8\" (UID: \"9dd5c102-10b0-4a96-8a12-6e1ca4f63c30\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8" Nov 24 14:54:45 crc kubenswrapper[4790]: I1124 14:54:45.607377 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzp4s\" (UniqueName: \"kubernetes.io/projected/9dd5c102-10b0-4a96-8a12-6e1ca4f63c30-kube-api-access-mzp4s\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8\" (UID: \"9dd5c102-10b0-4a96-8a12-6e1ca4f63c30\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8" Nov 24 14:54:45 crc kubenswrapper[4790]: I1124 14:54:45.734227 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8" Nov 24 14:54:46 crc kubenswrapper[4790]: I1124 14:54:46.221479 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8"] Nov 24 14:54:47 crc kubenswrapper[4790]: I1124 14:54:47.236614 4790 generic.go:334] "Generic (PLEG): container finished" podID="9dd5c102-10b0-4a96-8a12-6e1ca4f63c30" containerID="e9da90d71cfa5272eceac9b64ae76c16257b846339d871a0d3784289c329fad6" exitCode=0 Nov 24 14:54:47 crc kubenswrapper[4790]: I1124 14:54:47.237528 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8" event={"ID":"9dd5c102-10b0-4a96-8a12-6e1ca4f63c30","Type":"ContainerDied","Data":"e9da90d71cfa5272eceac9b64ae76c16257b846339d871a0d3784289c329fad6"} Nov 24 14:54:47 crc kubenswrapper[4790]: I1124 14:54:47.237721 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8" event={"ID":"9dd5c102-10b0-4a96-8a12-6e1ca4f63c30","Type":"ContainerStarted","Data":"e4e5464794aec3f287ac5560a15e578ef1e396ca3e302c278d1d59f8a83b982f"} Nov 24 14:54:47 crc kubenswrapper[4790]: I1124 14:54:47.242394 4790 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 14:54:47 crc kubenswrapper[4790]: I1124 14:54:47.316077 4790 scope.go:117] "RemoveContainer" containerID="28308703817a1d4e0e5344ce19268127c600894f2daa6b921d52a9105f014132" Nov 24 14:54:47 crc kubenswrapper[4790]: E1124 14:54:47.316584 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:54:49 crc kubenswrapper[4790]: I1124 14:54:49.263260 4790 generic.go:334] "Generic (PLEG): container finished" podID="9dd5c102-10b0-4a96-8a12-6e1ca4f63c30" containerID="241b77aa975816be8eee563d63dfa315efb1703ff0f78df89bbb2c6998790b7d" exitCode=0 Nov 24 14:54:49 crc kubenswrapper[4790]: I1124 14:54:49.264074 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8" event={"ID":"9dd5c102-10b0-4a96-8a12-6e1ca4f63c30","Type":"ContainerDied","Data":"241b77aa975816be8eee563d63dfa315efb1703ff0f78df89bbb2c6998790b7d"} Nov 24 14:54:50 crc kubenswrapper[4790]: I1124 14:54:50.279397 4790 generic.go:334] "Generic (PLEG): container finished" podID="9dd5c102-10b0-4a96-8a12-6e1ca4f63c30" containerID="52ff8f26c9c211947241bfee3decdd8cf411ad7ab71f09985095ed7d3be87aca" exitCode=0 Nov 24 14:54:50 crc kubenswrapper[4790]: I1124 14:54:50.279534 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8" event={"ID":"9dd5c102-10b0-4a96-8a12-6e1ca4f63c30","Type":"ContainerDied","Data":"52ff8f26c9c211947241bfee3decdd8cf411ad7ab71f09985095ed7d3be87aca"} Nov 24 14:54:50 crc kubenswrapper[4790]: I1124 14:54:50.803608 4790 scope.go:117] "RemoveContainer" containerID="d3d13bedb07d91e6537015645cdcf203b6eb953e219100e13cd04d174716439b" Nov 24 14:54:51 crc kubenswrapper[4790]: I1124 14:54:51.711979 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8" Nov 24 14:54:51 crc kubenswrapper[4790]: I1124 14:54:51.861362 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9dd5c102-10b0-4a96-8a12-6e1ca4f63c30-util\") pod \"9dd5c102-10b0-4a96-8a12-6e1ca4f63c30\" (UID: \"9dd5c102-10b0-4a96-8a12-6e1ca4f63c30\") " Nov 24 14:54:51 crc kubenswrapper[4790]: I1124 14:54:51.861450 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9dd5c102-10b0-4a96-8a12-6e1ca4f63c30-bundle\") pod \"9dd5c102-10b0-4a96-8a12-6e1ca4f63c30\" (UID: \"9dd5c102-10b0-4a96-8a12-6e1ca4f63c30\") " Nov 24 14:54:51 crc kubenswrapper[4790]: I1124 14:54:51.861478 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mzp4s\" (UniqueName: \"kubernetes.io/projected/9dd5c102-10b0-4a96-8a12-6e1ca4f63c30-kube-api-access-mzp4s\") pod \"9dd5c102-10b0-4a96-8a12-6e1ca4f63c30\" (UID: \"9dd5c102-10b0-4a96-8a12-6e1ca4f63c30\") " Nov 24 14:54:51 crc kubenswrapper[4790]: I1124 14:54:51.865287 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9dd5c102-10b0-4a96-8a12-6e1ca4f63c30-bundle" (OuterVolumeSpecName: "bundle") pod "9dd5c102-10b0-4a96-8a12-6e1ca4f63c30" (UID: "9dd5c102-10b0-4a96-8a12-6e1ca4f63c30"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:54:51 crc kubenswrapper[4790]: I1124 14:54:51.872802 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dd5c102-10b0-4a96-8a12-6e1ca4f63c30-kube-api-access-mzp4s" (OuterVolumeSpecName: "kube-api-access-mzp4s") pod "9dd5c102-10b0-4a96-8a12-6e1ca4f63c30" (UID: "9dd5c102-10b0-4a96-8a12-6e1ca4f63c30"). InnerVolumeSpecName "kube-api-access-mzp4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:54:51 crc kubenswrapper[4790]: I1124 14:54:51.965464 4790 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9dd5c102-10b0-4a96-8a12-6e1ca4f63c30-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:54:51 crc kubenswrapper[4790]: I1124 14:54:51.965499 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mzp4s\" (UniqueName: \"kubernetes.io/projected/9dd5c102-10b0-4a96-8a12-6e1ca4f63c30-kube-api-access-mzp4s\") on node \"crc\" DevicePath \"\"" Nov 24 14:54:51 crc kubenswrapper[4790]: I1124 14:54:51.998061 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9dd5c102-10b0-4a96-8a12-6e1ca4f63c30-util" (OuterVolumeSpecName: "util") pod "9dd5c102-10b0-4a96-8a12-6e1ca4f63c30" (UID: "9dd5c102-10b0-4a96-8a12-6e1ca4f63c30"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:54:52 crc kubenswrapper[4790]: I1124 14:54:52.068527 4790 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9dd5c102-10b0-4a96-8a12-6e1ca4f63c30-util\") on node \"crc\" DevicePath \"\"" Nov 24 14:54:52 crc kubenswrapper[4790]: I1124 14:54:52.312982 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8" event={"ID":"9dd5c102-10b0-4a96-8a12-6e1ca4f63c30","Type":"ContainerDied","Data":"e4e5464794aec3f287ac5560a15e578ef1e396ca3e302c278d1d59f8a83b982f"} Nov 24 14:54:52 crc kubenswrapper[4790]: I1124 14:54:52.313050 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4e5464794aec3f287ac5560a15e578ef1e396ca3e302c278d1d59f8a83b982f" Nov 24 14:54:52 crc kubenswrapper[4790]: I1124 14:54:52.313132 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8" Nov 24 14:54:53 crc kubenswrapper[4790]: I1124 14:54:53.055117 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-b690-account-create-jkmsr"] Nov 24 14:54:53 crc kubenswrapper[4790]: I1124 14:54:53.072414 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-j98nt"] Nov 24 14:54:53 crc kubenswrapper[4790]: I1124 14:54:53.082793 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-rg74q"] Nov 24 14:54:53 crc kubenswrapper[4790]: I1124 14:54:53.094418 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-j98nt"] Nov 24 14:54:53 crc kubenswrapper[4790]: I1124 14:54:53.105187 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-rg74q"] Nov 24 14:54:53 crc kubenswrapper[4790]: I1124 14:54:53.116106 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-b690-account-create-jkmsr"] Nov 24 14:54:54 crc kubenswrapper[4790]: I1124 14:54:54.038974 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-c1a6-account-create-lwdnx"] Nov 24 14:54:54 crc kubenswrapper[4790]: I1124 14:54:54.049602 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-c1a6-account-create-lwdnx"] Nov 24 14:54:54 crc kubenswrapper[4790]: I1124 14:54:54.057912 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-rm2jx"] Nov 24 14:54:54 crc kubenswrapper[4790]: I1124 14:54:54.065994 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-3cfe-account-create-7wftp"] Nov 24 14:54:54 crc kubenswrapper[4790]: I1124 14:54:54.074555 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-rm2jx"] Nov 24 14:54:54 crc kubenswrapper[4790]: I1124 14:54:54.081686 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-3cfe-account-create-7wftp"] Nov 24 14:54:54 crc kubenswrapper[4790]: I1124 14:54:54.330772 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="054b0bf7-15a9-4475-9c53-f5e5baf24518" path="/var/lib/kubelet/pods/054b0bf7-15a9-4475-9c53-f5e5baf24518/volumes" Nov 24 14:54:54 crc kubenswrapper[4790]: I1124 14:54:54.332630 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4574a286-1090-4721-8e45-12a7ee77e76e" path="/var/lib/kubelet/pods/4574a286-1090-4721-8e45-12a7ee77e76e/volumes" Nov 24 14:54:54 crc kubenswrapper[4790]: I1124 14:54:54.333500 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bfcc365-af75-4e73-9980-bce20e886b88" path="/var/lib/kubelet/pods/4bfcc365-af75-4e73-9980-bce20e886b88/volumes" Nov 24 14:54:54 crc kubenswrapper[4790]: I1124 14:54:54.334359 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c63e605-bedd-4a88-b557-f8581dbf5513" path="/var/lib/kubelet/pods/5c63e605-bedd-4a88-b557-f8581dbf5513/volumes" Nov 24 14:54:54 crc kubenswrapper[4790]: I1124 14:54:54.335895 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64f433c4-6665-42d3-90bd-19d5648fccf6" path="/var/lib/kubelet/pods/64f433c4-6665-42d3-90bd-19d5648fccf6/volumes" Nov 24 14:54:54 crc kubenswrapper[4790]: I1124 14:54:54.337532 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99ad8afe-4fb4-4823-a699-c99e3efb75fb" path="/var/lib/kubelet/pods/99ad8afe-4fb4-4823-a699-c99e3efb75fb/volumes" Nov 24 14:55:00 crc kubenswrapper[4790]: I1124 14:55:00.315013 4790 scope.go:117] "RemoveContainer" containerID="28308703817a1d4e0e5344ce19268127c600894f2daa6b921d52a9105f014132" Nov 24 14:55:00 crc kubenswrapper[4790]: E1124 14:55:00.315701 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:55:03 crc kubenswrapper[4790]: I1124 14:55:03.805087 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-gqvr8"] Nov 24 14:55:03 crc kubenswrapper[4790]: E1124 14:55:03.807812 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dd5c102-10b0-4a96-8a12-6e1ca4f63c30" containerName="util" Nov 24 14:55:03 crc kubenswrapper[4790]: I1124 14:55:03.807963 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dd5c102-10b0-4a96-8a12-6e1ca4f63c30" containerName="util" Nov 24 14:55:03 crc kubenswrapper[4790]: E1124 14:55:03.807984 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dd5c102-10b0-4a96-8a12-6e1ca4f63c30" containerName="pull" Nov 24 14:55:03 crc kubenswrapper[4790]: I1124 14:55:03.807991 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dd5c102-10b0-4a96-8a12-6e1ca4f63c30" containerName="pull" Nov 24 14:55:03 crc kubenswrapper[4790]: E1124 14:55:03.808021 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dd5c102-10b0-4a96-8a12-6e1ca4f63c30" containerName="extract" Nov 24 14:55:03 crc kubenswrapper[4790]: I1124 14:55:03.808027 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dd5c102-10b0-4a96-8a12-6e1ca4f63c30" containerName="extract" Nov 24 14:55:03 crc kubenswrapper[4790]: I1124 14:55:03.808225 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dd5c102-10b0-4a96-8a12-6e1ca4f63c30" containerName="extract" Nov 24 14:55:03 crc kubenswrapper[4790]: I1124 14:55:03.808875 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-gqvr8" Nov 24 14:55:03 crc kubenswrapper[4790]: I1124 14:55:03.810848 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Nov 24 14:55:03 crc kubenswrapper[4790]: I1124 14:55:03.815387 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-md2wp" Nov 24 14:55:03 crc kubenswrapper[4790]: I1124 14:55:03.816257 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Nov 24 14:55:03 crc kubenswrapper[4790]: I1124 14:55:03.816744 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-gqvr8"] Nov 24 14:55:03 crc kubenswrapper[4790]: I1124 14:55:03.914467 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6cbcb9c64b-c4f6l"] Nov 24 14:55:03 crc kubenswrapper[4790]: I1124 14:55:03.918179 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cbcb9c64b-c4f6l" Nov 24 14:55:03 crc kubenswrapper[4790]: I1124 14:55:03.921544 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Nov 24 14:55:03 crc kubenswrapper[4790]: I1124 14:55:03.921589 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-qxbnk" Nov 24 14:55:03 crc kubenswrapper[4790]: I1124 14:55:03.929432 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pz4l5\" (UniqueName: \"kubernetes.io/projected/84415f57-fc06-4b4e-9ecc-f0f1d32f8ea1-kube-api-access-pz4l5\") pod \"obo-prometheus-operator-668cf9dfbb-gqvr8\" (UID: \"84415f57-fc06-4b4e-9ecc-f0f1d32f8ea1\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-gqvr8" Nov 24 14:55:03 crc kubenswrapper[4790]: I1124 14:55:03.931440 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6cbcb9c64b-jx7cm"] Nov 24 14:55:03 crc kubenswrapper[4790]: I1124 14:55:03.933119 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cbcb9c64b-jx7cm" Nov 24 14:55:03 crc kubenswrapper[4790]: I1124 14:55:03.949371 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6cbcb9c64b-c4f6l"] Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.007584 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6cbcb9c64b-jx7cm"] Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.032156 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bc0609a2-7ee5-4fed-b741-b9bc0f85b6bb-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6cbcb9c64b-c4f6l\" (UID: \"bc0609a2-7ee5-4fed-b741-b9bc0f85b6bb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cbcb9c64b-c4f6l" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.032220 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/55eaed3b-e46e-4a6e-81b1-123e1a2f7d99-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6cbcb9c64b-jx7cm\" (UID: \"55eaed3b-e46e-4a6e-81b1-123e1a2f7d99\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cbcb9c64b-jx7cm" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.032263 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pz4l5\" (UniqueName: \"kubernetes.io/projected/84415f57-fc06-4b4e-9ecc-f0f1d32f8ea1-kube-api-access-pz4l5\") pod \"obo-prometheus-operator-668cf9dfbb-gqvr8\" (UID: \"84415f57-fc06-4b4e-9ecc-f0f1d32f8ea1\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-gqvr8" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.032318 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/55eaed3b-e46e-4a6e-81b1-123e1a2f7d99-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6cbcb9c64b-jx7cm\" (UID: \"55eaed3b-e46e-4a6e-81b1-123e1a2f7d99\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cbcb9c64b-jx7cm" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.032378 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bc0609a2-7ee5-4fed-b741-b9bc0f85b6bb-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6cbcb9c64b-c4f6l\" (UID: \"bc0609a2-7ee5-4fed-b741-b9bc0f85b6bb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cbcb9c64b-c4f6l" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.050737 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-2w7vx"] Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.055134 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pz4l5\" (UniqueName: \"kubernetes.io/projected/84415f57-fc06-4b4e-9ecc-f0f1d32f8ea1-kube-api-access-pz4l5\") pod \"obo-prometheus-operator-668cf9dfbb-gqvr8\" (UID: \"84415f57-fc06-4b4e-9ecc-f0f1d32f8ea1\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-gqvr8" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.061504 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-2w7vx"] Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.134684 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/55eaed3b-e46e-4a6e-81b1-123e1a2f7d99-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6cbcb9c64b-jx7cm\" (UID: \"55eaed3b-e46e-4a6e-81b1-123e1a2f7d99\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cbcb9c64b-jx7cm" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.134805 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bc0609a2-7ee5-4fed-b741-b9bc0f85b6bb-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6cbcb9c64b-c4f6l\" (UID: \"bc0609a2-7ee5-4fed-b741-b9bc0f85b6bb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cbcb9c64b-c4f6l" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.134941 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bc0609a2-7ee5-4fed-b741-b9bc0f85b6bb-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6cbcb9c64b-c4f6l\" (UID: \"bc0609a2-7ee5-4fed-b741-b9bc0f85b6bb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cbcb9c64b-c4f6l" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.134987 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/55eaed3b-e46e-4a6e-81b1-123e1a2f7d99-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6cbcb9c64b-jx7cm\" (UID: \"55eaed3b-e46e-4a6e-81b1-123e1a2f7d99\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cbcb9c64b-jx7cm" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.138707 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-gqvr8" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.141994 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-t6r5w"] Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.143755 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-t6r5w" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.150991 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-zds2r" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.151286 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.166376 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bc0609a2-7ee5-4fed-b741-b9bc0f85b6bb-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6cbcb9c64b-c4f6l\" (UID: \"bc0609a2-7ee5-4fed-b741-b9bc0f85b6bb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cbcb9c64b-c4f6l" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.166690 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bc0609a2-7ee5-4fed-b741-b9bc0f85b6bb-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6cbcb9c64b-c4f6l\" (UID: \"bc0609a2-7ee5-4fed-b741-b9bc0f85b6bb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cbcb9c64b-c4f6l" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.172838 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-t6r5w"] Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.174042 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/55eaed3b-e46e-4a6e-81b1-123e1a2f7d99-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6cbcb9c64b-jx7cm\" (UID: \"55eaed3b-e46e-4a6e-81b1-123e1a2f7d99\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cbcb9c64b-jx7cm" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.186566 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/55eaed3b-e46e-4a6e-81b1-123e1a2f7d99-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6cbcb9c64b-jx7cm\" (UID: \"55eaed3b-e46e-4a6e-81b1-123e1a2f7d99\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cbcb9c64b-jx7cm" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.237303 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cr8zh\" (UniqueName: \"kubernetes.io/projected/70de2641-0cdd-4ebe-9a2b-5aa4efa44c72-kube-api-access-cr8zh\") pod \"observability-operator-d8bb48f5d-t6r5w\" (UID: \"70de2641-0cdd-4ebe-9a2b-5aa4efa44c72\") " pod="openshift-operators/observability-operator-d8bb48f5d-t6r5w" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.237659 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/70de2641-0cdd-4ebe-9a2b-5aa4efa44c72-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-t6r5w\" (UID: \"70de2641-0cdd-4ebe-9a2b-5aa4efa44c72\") " pod="openshift-operators/observability-operator-d8bb48f5d-t6r5w" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.250444 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cbcb9c64b-c4f6l" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.286316 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cbcb9c64b-jx7cm" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.341428 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cr8zh\" (UniqueName: \"kubernetes.io/projected/70de2641-0cdd-4ebe-9a2b-5aa4efa44c72-kube-api-access-cr8zh\") pod \"observability-operator-d8bb48f5d-t6r5w\" (UID: \"70de2641-0cdd-4ebe-9a2b-5aa4efa44c72\") " pod="openshift-operators/observability-operator-d8bb48f5d-t6r5w" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.341488 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/70de2641-0cdd-4ebe-9a2b-5aa4efa44c72-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-t6r5w\" (UID: \"70de2641-0cdd-4ebe-9a2b-5aa4efa44c72\") " pod="openshift-operators/observability-operator-d8bb48f5d-t6r5w" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.350254 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/70de2641-0cdd-4ebe-9a2b-5aa4efa44c72-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-t6r5w\" (UID: \"70de2641-0cdd-4ebe-9a2b-5aa4efa44c72\") " pod="openshift-operators/observability-operator-d8bb48f5d-t6r5w" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.383233 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cr8zh\" (UniqueName: \"kubernetes.io/projected/70de2641-0cdd-4ebe-9a2b-5aa4efa44c72-kube-api-access-cr8zh\") pod \"observability-operator-d8bb48f5d-t6r5w\" (UID: \"70de2641-0cdd-4ebe-9a2b-5aa4efa44c72\") " pod="openshift-operators/observability-operator-d8bb48f5d-t6r5w" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.426825 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce20586f-aa65-426b-940e-22a8a12cd16d" path="/var/lib/kubelet/pods/ce20586f-aa65-426b-940e-22a8a12cd16d/volumes" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.427416 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-ntlws"] Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.428729 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-ntlws"] Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.428816 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-ntlws" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.446096 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-2lnp4" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.559464 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/1dd8ee5b-488a-41e3-ba58-e424899180ae-openshift-service-ca\") pod \"perses-operator-5446b9c989-ntlws\" (UID: \"1dd8ee5b-488a-41e3-ba58-e424899180ae\") " pod="openshift-operators/perses-operator-5446b9c989-ntlws" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.559745 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxj8k\" (UniqueName: \"kubernetes.io/projected/1dd8ee5b-488a-41e3-ba58-e424899180ae-kube-api-access-pxj8k\") pod \"perses-operator-5446b9c989-ntlws\" (UID: \"1dd8ee5b-488a-41e3-ba58-e424899180ae\") " pod="openshift-operators/perses-operator-5446b9c989-ntlws" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.669429 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxj8k\" (UniqueName: \"kubernetes.io/projected/1dd8ee5b-488a-41e3-ba58-e424899180ae-kube-api-access-pxj8k\") pod \"perses-operator-5446b9c989-ntlws\" (UID: \"1dd8ee5b-488a-41e3-ba58-e424899180ae\") " pod="openshift-operators/perses-operator-5446b9c989-ntlws" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.669601 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/1dd8ee5b-488a-41e3-ba58-e424899180ae-openshift-service-ca\") pod \"perses-operator-5446b9c989-ntlws\" (UID: \"1dd8ee5b-488a-41e3-ba58-e424899180ae\") " pod="openshift-operators/perses-operator-5446b9c989-ntlws" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.670530 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/1dd8ee5b-488a-41e3-ba58-e424899180ae-openshift-service-ca\") pod \"perses-operator-5446b9c989-ntlws\" (UID: \"1dd8ee5b-488a-41e3-ba58-e424899180ae\") " pod="openshift-operators/perses-operator-5446b9c989-ntlws" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.674609 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-t6r5w" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.679680 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-gqvr8"] Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.741913 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxj8k\" (UniqueName: \"kubernetes.io/projected/1dd8ee5b-488a-41e3-ba58-e424899180ae-kube-api-access-pxj8k\") pod \"perses-operator-5446b9c989-ntlws\" (UID: \"1dd8ee5b-488a-41e3-ba58-e424899180ae\") " pod="openshift-operators/perses-operator-5446b9c989-ntlws" Nov 24 14:55:04 crc kubenswrapper[4790]: I1124 14:55:04.783661 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-ntlws" Nov 24 14:55:05 crc kubenswrapper[4790]: I1124 14:55:05.269248 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6cbcb9c64b-c4f6l"] Nov 24 14:55:05 crc kubenswrapper[4790]: I1124 14:55:05.386004 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6cbcb9c64b-jx7cm"] Nov 24 14:55:05 crc kubenswrapper[4790]: I1124 14:55:05.477677 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cbcb9c64b-jx7cm" event={"ID":"55eaed3b-e46e-4a6e-81b1-123e1a2f7d99","Type":"ContainerStarted","Data":"067c307e5edb9fb9053750f1a463bbf108eb54b6304739f73732d76dbb382c9d"} Nov 24 14:55:05 crc kubenswrapper[4790]: I1124 14:55:05.480128 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cbcb9c64b-c4f6l" event={"ID":"bc0609a2-7ee5-4fed-b741-b9bc0f85b6bb","Type":"ContainerStarted","Data":"129e6eca6396ae8c3b05916bd39f75ca55937684a4a3bc757f47e07fcf8fd306"} Nov 24 14:55:05 crc kubenswrapper[4790]: I1124 14:55:05.484931 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-gqvr8" event={"ID":"84415f57-fc06-4b4e-9ecc-f0f1d32f8ea1","Type":"ContainerStarted","Data":"29579a9289e3fe8981508991c9033f8131b31ab9030312691c01c329f7468bf0"} Nov 24 14:55:05 crc kubenswrapper[4790]: I1124 14:55:05.496616 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-ntlws"] Nov 24 14:55:05 crc kubenswrapper[4790]: W1124 14:55:05.506104 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70de2641_0cdd_4ebe_9a2b_5aa4efa44c72.slice/crio-e34fda12cc06a1bb80a4ae63cdb29e097ea871649a45b895d8e8369859d0cfa1 WatchSource:0}: Error finding container e34fda12cc06a1bb80a4ae63cdb29e097ea871649a45b895d8e8369859d0cfa1: Status 404 returned error can't find the container with id e34fda12cc06a1bb80a4ae63cdb29e097ea871649a45b895d8e8369859d0cfa1 Nov 24 14:55:05 crc kubenswrapper[4790]: I1124 14:55:05.508646 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-t6r5w"] Nov 24 14:55:06 crc kubenswrapper[4790]: I1124 14:55:06.533197 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-ntlws" event={"ID":"1dd8ee5b-488a-41e3-ba58-e424899180ae","Type":"ContainerStarted","Data":"8087743a98d28f8124da4a15e04f4a8224e7c169502f2fb91ddba0d0b136203f"} Nov 24 14:55:06 crc kubenswrapper[4790]: I1124 14:55:06.542520 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-t6r5w" event={"ID":"70de2641-0cdd-4ebe-9a2b-5aa4efa44c72","Type":"ContainerStarted","Data":"e34fda12cc06a1bb80a4ae63cdb29e097ea871649a45b895d8e8369859d0cfa1"} Nov 24 14:55:11 crc kubenswrapper[4790]: I1124 14:55:11.015975 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-v7xkz"] Nov 24 14:55:11 crc kubenswrapper[4790]: I1124 14:55:11.019569 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v7xkz" Nov 24 14:55:11 crc kubenswrapper[4790]: I1124 14:55:11.040144 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v7xkz"] Nov 24 14:55:11 crc kubenswrapper[4790]: I1124 14:55:11.126581 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33662038-1049-4380-804a-1ce66fd13db7-catalog-content\") pod \"redhat-marketplace-v7xkz\" (UID: \"33662038-1049-4380-804a-1ce66fd13db7\") " pod="openshift-marketplace/redhat-marketplace-v7xkz" Nov 24 14:55:11 crc kubenswrapper[4790]: I1124 14:55:11.126655 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnfmh\" (UniqueName: \"kubernetes.io/projected/33662038-1049-4380-804a-1ce66fd13db7-kube-api-access-nnfmh\") pod \"redhat-marketplace-v7xkz\" (UID: \"33662038-1049-4380-804a-1ce66fd13db7\") " pod="openshift-marketplace/redhat-marketplace-v7xkz" Nov 24 14:55:11 crc kubenswrapper[4790]: I1124 14:55:11.126702 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33662038-1049-4380-804a-1ce66fd13db7-utilities\") pod \"redhat-marketplace-v7xkz\" (UID: \"33662038-1049-4380-804a-1ce66fd13db7\") " pod="openshift-marketplace/redhat-marketplace-v7xkz" Nov 24 14:55:11 crc kubenswrapper[4790]: I1124 14:55:11.228595 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33662038-1049-4380-804a-1ce66fd13db7-catalog-content\") pod \"redhat-marketplace-v7xkz\" (UID: \"33662038-1049-4380-804a-1ce66fd13db7\") " pod="openshift-marketplace/redhat-marketplace-v7xkz" Nov 24 14:55:11 crc kubenswrapper[4790]: I1124 14:55:11.228732 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnfmh\" (UniqueName: \"kubernetes.io/projected/33662038-1049-4380-804a-1ce66fd13db7-kube-api-access-nnfmh\") pod \"redhat-marketplace-v7xkz\" (UID: \"33662038-1049-4380-804a-1ce66fd13db7\") " pod="openshift-marketplace/redhat-marketplace-v7xkz" Nov 24 14:55:11 crc kubenswrapper[4790]: I1124 14:55:11.228851 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33662038-1049-4380-804a-1ce66fd13db7-utilities\") pod \"redhat-marketplace-v7xkz\" (UID: \"33662038-1049-4380-804a-1ce66fd13db7\") " pod="openshift-marketplace/redhat-marketplace-v7xkz" Nov 24 14:55:11 crc kubenswrapper[4790]: I1124 14:55:11.229271 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33662038-1049-4380-804a-1ce66fd13db7-catalog-content\") pod \"redhat-marketplace-v7xkz\" (UID: \"33662038-1049-4380-804a-1ce66fd13db7\") " pod="openshift-marketplace/redhat-marketplace-v7xkz" Nov 24 14:55:11 crc kubenswrapper[4790]: I1124 14:55:11.229302 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33662038-1049-4380-804a-1ce66fd13db7-utilities\") pod \"redhat-marketplace-v7xkz\" (UID: \"33662038-1049-4380-804a-1ce66fd13db7\") " pod="openshift-marketplace/redhat-marketplace-v7xkz" Nov 24 14:55:11 crc kubenswrapper[4790]: I1124 14:55:11.253524 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnfmh\" (UniqueName: \"kubernetes.io/projected/33662038-1049-4380-804a-1ce66fd13db7-kube-api-access-nnfmh\") pod \"redhat-marketplace-v7xkz\" (UID: \"33662038-1049-4380-804a-1ce66fd13db7\") " pod="openshift-marketplace/redhat-marketplace-v7xkz" Nov 24 14:55:11 crc kubenswrapper[4790]: I1124 14:55:11.352590 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v7xkz" Nov 24 14:55:12 crc kubenswrapper[4790]: I1124 14:55:12.328050 4790 scope.go:117] "RemoveContainer" containerID="28308703817a1d4e0e5344ce19268127c600894f2daa6b921d52a9105f014132" Nov 24 14:55:12 crc kubenswrapper[4790]: E1124 14:55:12.328615 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:55:14 crc kubenswrapper[4790]: I1124 14:55:14.641548 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v7xkz"] Nov 24 14:55:14 crc kubenswrapper[4790]: I1124 14:55:14.656294 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cbcb9c64b-c4f6l" event={"ID":"bc0609a2-7ee5-4fed-b741-b9bc0f85b6bb","Type":"ContainerStarted","Data":"c702247d56a75ef5bea0018d40804c8dc5afd8f36894c8a4951f0182d84d934f"} Nov 24 14:55:14 crc kubenswrapper[4790]: I1124 14:55:14.671180 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-ntlws" event={"ID":"1dd8ee5b-488a-41e3-ba58-e424899180ae","Type":"ContainerStarted","Data":"959fb63d981effe02a5c1625e38bfed7aa9eadd4225bd3944c280cafac06a9f8"} Nov 24 14:55:14 crc kubenswrapper[4790]: I1124 14:55:14.671270 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-ntlws" Nov 24 14:55:14 crc kubenswrapper[4790]: I1124 14:55:14.675995 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cbcb9c64b-c4f6l" podStartSLOduration=3.062134987 podStartE2EDuration="11.675978422s" podCreationTimestamp="2025-11-24 14:55:03 +0000 UTC" firstStartedPulling="2025-11-24 14:55:05.266939331 +0000 UTC m=+6153.646833003" lastFinishedPulling="2025-11-24 14:55:13.880782776 +0000 UTC m=+6162.260676438" observedRunningTime="2025-11-24 14:55:14.674809993 +0000 UTC m=+6163.054703655" watchObservedRunningTime="2025-11-24 14:55:14.675978422 +0000 UTC m=+6163.055872084" Nov 24 14:55:14 crc kubenswrapper[4790]: I1124 14:55:14.677477 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-t6r5w" event={"ID":"70de2641-0cdd-4ebe-9a2b-5aa4efa44c72","Type":"ContainerStarted","Data":"5824e9e01c76feece33067dcbd30f1e5041539fceef18ae24250f44376fe8410"} Nov 24 14:55:14 crc kubenswrapper[4790]: I1124 14:55:14.678996 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-t6r5w" Nov 24 14:55:14 crc kubenswrapper[4790]: I1124 14:55:14.679156 4790 patch_prober.go:28] interesting pod/observability-operator-d8bb48f5d-t6r5w container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.1.127:8081/healthz\": dial tcp 10.217.1.127:8081: connect: connection refused" start-of-body= Nov 24 14:55:14 crc kubenswrapper[4790]: I1124 14:55:14.679183 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-d8bb48f5d-t6r5w" podUID="70de2641-0cdd-4ebe-9a2b-5aa4efa44c72" containerName="operator" probeResult="failure" output="Get \"http://10.217.1.127:8081/healthz\": dial tcp 10.217.1.127:8081: connect: connection refused" Nov 24 14:55:14 crc kubenswrapper[4790]: I1124 14:55:14.694242 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cbcb9c64b-jx7cm" event={"ID":"55eaed3b-e46e-4a6e-81b1-123e1a2f7d99","Type":"ContainerStarted","Data":"a9603deea4f22caeddc3afc21e8c2b99d32e20a5d3b6075e9ebda3d9f305aa17"} Nov 24 14:55:14 crc kubenswrapper[4790]: I1124 14:55:14.717325 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-ntlws" podStartSLOduration=2.335445919 podStartE2EDuration="10.717300536s" podCreationTimestamp="2025-11-24 14:55:04 +0000 UTC" firstStartedPulling="2025-11-24 14:55:05.508055929 +0000 UTC m=+6153.887949591" lastFinishedPulling="2025-11-24 14:55:13.889910546 +0000 UTC m=+6162.269804208" observedRunningTime="2025-11-24 14:55:14.702378489 +0000 UTC m=+6163.082272151" watchObservedRunningTime="2025-11-24 14:55:14.717300536 +0000 UTC m=+6163.097194198" Nov 24 14:55:14 crc kubenswrapper[4790]: I1124 14:55:14.738786 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-t6r5w" podStartSLOduration=2.160204874 podStartE2EDuration="10.738764658s" podCreationTimestamp="2025-11-24 14:55:04 +0000 UTC" firstStartedPulling="2025-11-24 14:55:05.509732841 +0000 UTC m=+6153.889626503" lastFinishedPulling="2025-11-24 14:55:14.088292625 +0000 UTC m=+6162.468186287" observedRunningTime="2025-11-24 14:55:14.737979958 +0000 UTC m=+6163.117873620" watchObservedRunningTime="2025-11-24 14:55:14.738764658 +0000 UTC m=+6163.118658320" Nov 24 14:55:14 crc kubenswrapper[4790]: I1124 14:55:14.767285 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6cbcb9c64b-jx7cm" podStartSLOduration=3.339550662 podStartE2EDuration="11.767264077s" podCreationTimestamp="2025-11-24 14:55:03 +0000 UTC" firstStartedPulling="2025-11-24 14:55:05.392379998 +0000 UTC m=+6153.772273660" lastFinishedPulling="2025-11-24 14:55:13.820093413 +0000 UTC m=+6162.199987075" observedRunningTime="2025-11-24 14:55:14.757735577 +0000 UTC m=+6163.137629239" watchObservedRunningTime="2025-11-24 14:55:14.767264077 +0000 UTC m=+6163.147157739" Nov 24 14:55:15 crc kubenswrapper[4790]: I1124 14:55:15.708165 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-gqvr8" event={"ID":"84415f57-fc06-4b4e-9ecc-f0f1d32f8ea1","Type":"ContainerStarted","Data":"553652e95858472341be8c68c09aa17c55b12476e0ea460213510bdb0043827e"} Nov 24 14:55:15 crc kubenswrapper[4790]: I1124 14:55:15.710267 4790 generic.go:334] "Generic (PLEG): container finished" podID="33662038-1049-4380-804a-1ce66fd13db7" containerID="2017ab431c0705f7a464fc6a3033bb6312f0d9fb4b202198caed6be1ba577b36" exitCode=0 Nov 24 14:55:15 crc kubenswrapper[4790]: I1124 14:55:15.710323 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v7xkz" event={"ID":"33662038-1049-4380-804a-1ce66fd13db7","Type":"ContainerDied","Data":"2017ab431c0705f7a464fc6a3033bb6312f0d9fb4b202198caed6be1ba577b36"} Nov 24 14:55:15 crc kubenswrapper[4790]: I1124 14:55:15.710530 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v7xkz" event={"ID":"33662038-1049-4380-804a-1ce66fd13db7","Type":"ContainerStarted","Data":"bdf680c73f46e872ccc05eeb1382b2920dbbf9fb83dab68873f17cf0bdf2cf99"} Nov 24 14:55:15 crc kubenswrapper[4790]: I1124 14:55:15.712353 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-t6r5w" Nov 24 14:55:15 crc kubenswrapper[4790]: I1124 14:55:15.749152 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-gqvr8" podStartSLOduration=3.723065025 podStartE2EDuration="12.749130907s" podCreationTimestamp="2025-11-24 14:55:03 +0000 UTC" firstStartedPulling="2025-11-24 14:55:04.876190826 +0000 UTC m=+6153.256084488" lastFinishedPulling="2025-11-24 14:55:13.902256708 +0000 UTC m=+6162.282150370" observedRunningTime="2025-11-24 14:55:15.72944241 +0000 UTC m=+6164.109336082" watchObservedRunningTime="2025-11-24 14:55:15.749130907 +0000 UTC m=+6164.129024579" Nov 24 14:55:16 crc kubenswrapper[4790]: I1124 14:55:16.721642 4790 generic.go:334] "Generic (PLEG): container finished" podID="33662038-1049-4380-804a-1ce66fd13db7" containerID="3f15a721b100c376d986283df9271dd98af54c99d32f88b55b9e5916d76d5511" exitCode=0 Nov 24 14:55:16 crc kubenswrapper[4790]: I1124 14:55:16.721726 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v7xkz" event={"ID":"33662038-1049-4380-804a-1ce66fd13db7","Type":"ContainerDied","Data":"3f15a721b100c376d986283df9271dd98af54c99d32f88b55b9e5916d76d5511"} Nov 24 14:55:17 crc kubenswrapper[4790]: I1124 14:55:17.734186 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v7xkz" event={"ID":"33662038-1049-4380-804a-1ce66fd13db7","Type":"ContainerStarted","Data":"d5cb2bb1b93c40424bb48a6a0e7f76518b73206f6e2011df2fd1ab5f2673ff95"} Nov 24 14:55:21 crc kubenswrapper[4790]: I1124 14:55:21.354545 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-v7xkz" Nov 24 14:55:21 crc kubenswrapper[4790]: I1124 14:55:21.355582 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-v7xkz" Nov 24 14:55:21 crc kubenswrapper[4790]: I1124 14:55:21.425014 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-v7xkz" Nov 24 14:55:21 crc kubenswrapper[4790]: I1124 14:55:21.456099 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-v7xkz" podStartSLOduration=10.040562434 podStartE2EDuration="11.456068201s" podCreationTimestamp="2025-11-24 14:55:10 +0000 UTC" firstStartedPulling="2025-11-24 14:55:15.712379059 +0000 UTC m=+6164.092272731" lastFinishedPulling="2025-11-24 14:55:17.127884836 +0000 UTC m=+6165.507778498" observedRunningTime="2025-11-24 14:55:17.769381572 +0000 UTC m=+6166.149275254" watchObservedRunningTime="2025-11-24 14:55:21.456068201 +0000 UTC m=+6169.835961873" Nov 24 14:55:22 crc kubenswrapper[4790]: I1124 14:55:22.218625 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vqnlr"] Nov 24 14:55:22 crc kubenswrapper[4790]: I1124 14:55:22.229987 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vqnlr" Nov 24 14:55:22 crc kubenswrapper[4790]: I1124 14:55:22.269291 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vqnlr"] Nov 24 14:55:22 crc kubenswrapper[4790]: I1124 14:55:22.278594 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fp8hx\" (UniqueName: \"kubernetes.io/projected/6fbf1db8-8f67-429c-97c9-e97714898b0e-kube-api-access-fp8hx\") pod \"redhat-operators-vqnlr\" (UID: \"6fbf1db8-8f67-429c-97c9-e97714898b0e\") " pod="openshift-marketplace/redhat-operators-vqnlr" Nov 24 14:55:22 crc kubenswrapper[4790]: I1124 14:55:22.278994 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fbf1db8-8f67-429c-97c9-e97714898b0e-catalog-content\") pod \"redhat-operators-vqnlr\" (UID: \"6fbf1db8-8f67-429c-97c9-e97714898b0e\") " pod="openshift-marketplace/redhat-operators-vqnlr" Nov 24 14:55:22 crc kubenswrapper[4790]: I1124 14:55:22.279229 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fbf1db8-8f67-429c-97c9-e97714898b0e-utilities\") pod \"redhat-operators-vqnlr\" (UID: \"6fbf1db8-8f67-429c-97c9-e97714898b0e\") " pod="openshift-marketplace/redhat-operators-vqnlr" Nov 24 14:55:22 crc kubenswrapper[4790]: I1124 14:55:22.385183 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fp8hx\" (UniqueName: \"kubernetes.io/projected/6fbf1db8-8f67-429c-97c9-e97714898b0e-kube-api-access-fp8hx\") pod \"redhat-operators-vqnlr\" (UID: \"6fbf1db8-8f67-429c-97c9-e97714898b0e\") " pod="openshift-marketplace/redhat-operators-vqnlr" Nov 24 14:55:22 crc kubenswrapper[4790]: I1124 14:55:22.385852 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fbf1db8-8f67-429c-97c9-e97714898b0e-catalog-content\") pod \"redhat-operators-vqnlr\" (UID: \"6fbf1db8-8f67-429c-97c9-e97714898b0e\") " pod="openshift-marketplace/redhat-operators-vqnlr" Nov 24 14:55:22 crc kubenswrapper[4790]: I1124 14:55:22.386002 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fbf1db8-8f67-429c-97c9-e97714898b0e-utilities\") pod \"redhat-operators-vqnlr\" (UID: \"6fbf1db8-8f67-429c-97c9-e97714898b0e\") " pod="openshift-marketplace/redhat-operators-vqnlr" Nov 24 14:55:22 crc kubenswrapper[4790]: I1124 14:55:22.386543 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fbf1db8-8f67-429c-97c9-e97714898b0e-utilities\") pod \"redhat-operators-vqnlr\" (UID: \"6fbf1db8-8f67-429c-97c9-e97714898b0e\") " pod="openshift-marketplace/redhat-operators-vqnlr" Nov 24 14:55:22 crc kubenswrapper[4790]: I1124 14:55:22.386740 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fbf1db8-8f67-429c-97c9-e97714898b0e-catalog-content\") pod \"redhat-operators-vqnlr\" (UID: \"6fbf1db8-8f67-429c-97c9-e97714898b0e\") " pod="openshift-marketplace/redhat-operators-vqnlr" Nov 24 14:55:22 crc kubenswrapper[4790]: I1124 14:55:22.433355 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fp8hx\" (UniqueName: \"kubernetes.io/projected/6fbf1db8-8f67-429c-97c9-e97714898b0e-kube-api-access-fp8hx\") pod \"redhat-operators-vqnlr\" (UID: \"6fbf1db8-8f67-429c-97c9-e97714898b0e\") " pod="openshift-marketplace/redhat-operators-vqnlr" Nov 24 14:55:22 crc kubenswrapper[4790]: I1124 14:55:22.577598 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vqnlr" Nov 24 14:55:23 crc kubenswrapper[4790]: W1124 14:55:23.302654 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6fbf1db8_8f67_429c_97c9_e97714898b0e.slice/crio-9929df02e103dc2a97ecfa0d48f084854715817bc3161d276828afd439e7ed4d WatchSource:0}: Error finding container 9929df02e103dc2a97ecfa0d48f084854715817bc3161d276828afd439e7ed4d: Status 404 returned error can't find the container with id 9929df02e103dc2a97ecfa0d48f084854715817bc3161d276828afd439e7ed4d Nov 24 14:55:23 crc kubenswrapper[4790]: I1124 14:55:23.321164 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vqnlr"] Nov 24 14:55:23 crc kubenswrapper[4790]: I1124 14:55:23.796425 4790 generic.go:334] "Generic (PLEG): container finished" podID="6fbf1db8-8f67-429c-97c9-e97714898b0e" containerID="aa8b4355207d9e157b8e068c5ae0ee0ca10dc9bf9aed40ed86ebe8347e1f279c" exitCode=0 Nov 24 14:55:23 crc kubenswrapper[4790]: I1124 14:55:23.796478 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vqnlr" event={"ID":"6fbf1db8-8f67-429c-97c9-e97714898b0e","Type":"ContainerDied","Data":"aa8b4355207d9e157b8e068c5ae0ee0ca10dc9bf9aed40ed86ebe8347e1f279c"} Nov 24 14:55:23 crc kubenswrapper[4790]: I1124 14:55:23.796503 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vqnlr" event={"ID":"6fbf1db8-8f67-429c-97c9-e97714898b0e","Type":"ContainerStarted","Data":"9929df02e103dc2a97ecfa0d48f084854715817bc3161d276828afd439e7ed4d"} Nov 24 14:55:24 crc kubenswrapper[4790]: I1124 14:55:24.044080 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-w7f6n"] Nov 24 14:55:24 crc kubenswrapper[4790]: I1124 14:55:24.052734 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-w7f6n"] Nov 24 14:55:24 crc kubenswrapper[4790]: I1124 14:55:24.061960 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zv4x8"] Nov 24 14:55:24 crc kubenswrapper[4790]: I1124 14:55:24.072946 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zv4x8"] Nov 24 14:55:24 crc kubenswrapper[4790]: I1124 14:55:24.337316 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="873b8408-4a0a-42dd-bb32-5e3cba17a985" path="/var/lib/kubelet/pods/873b8408-4a0a-42dd-bb32-5e3cba17a985/volumes" Nov 24 14:55:24 crc kubenswrapper[4790]: I1124 14:55:24.339345 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8596f2b-213d-4a3f-940a-9cc7421e85be" path="/var/lib/kubelet/pods/b8596f2b-213d-4a3f-940a-9cc7421e85be/volumes" Nov 24 14:55:24 crc kubenswrapper[4790]: I1124 14:55:24.786564 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-ntlws" Nov 24 14:55:25 crc kubenswrapper[4790]: I1124 14:55:25.827899 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vqnlr" event={"ID":"6fbf1db8-8f67-429c-97c9-e97714898b0e","Type":"ContainerStarted","Data":"742e97e5aac502e3bc676c0c4da51dd8b1049dd9c52ae758218c5c24533f59c5"} Nov 24 14:55:26 crc kubenswrapper[4790]: I1124 14:55:26.314367 4790 scope.go:117] "RemoveContainer" containerID="28308703817a1d4e0e5344ce19268127c600894f2daa6b921d52a9105f014132" Nov 24 14:55:26 crc kubenswrapper[4790]: E1124 14:55:26.314870 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:55:27 crc kubenswrapper[4790]: I1124 14:55:27.691035 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 24 14:55:27 crc kubenswrapper[4790]: I1124 14:55:27.691421 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="b5399fd5-dda4-4a13-954f-76a99b1577a3" containerName="openstackclient" containerID="cri-o://51ce6763398e66fcf80424e574cd3875e846b87aaa373c269f452945a5f9b9ce" gracePeriod=2 Nov 24 14:55:27 crc kubenswrapper[4790]: I1124 14:55:27.709048 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 24 14:55:27 crc kubenswrapper[4790]: I1124 14:55:27.771570 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 24 14:55:27 crc kubenswrapper[4790]: E1124 14:55:27.772031 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5399fd5-dda4-4a13-954f-76a99b1577a3" containerName="openstackclient" Nov 24 14:55:27 crc kubenswrapper[4790]: I1124 14:55:27.772053 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5399fd5-dda4-4a13-954f-76a99b1577a3" containerName="openstackclient" Nov 24 14:55:27 crc kubenswrapper[4790]: I1124 14:55:27.772278 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5399fd5-dda4-4a13-954f-76a99b1577a3" containerName="openstackclient" Nov 24 14:55:27 crc kubenswrapper[4790]: I1124 14:55:27.773378 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 14:55:27 crc kubenswrapper[4790]: I1124 14:55:27.777477 4790 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="b5399fd5-dda4-4a13-954f-76a99b1577a3" podUID="a6d18574-10ed-4b11-b5a9-eeed45e66cb2" Nov 24 14:55:27 crc kubenswrapper[4790]: I1124 14:55:27.790096 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 14:55:27 crc kubenswrapper[4790]: I1124 14:55:27.908706 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a6d18574-10ed-4b11-b5a9-eeed45e66cb2-openstack-config-secret\") pod \"openstackclient\" (UID: \"a6d18574-10ed-4b11-b5a9-eeed45e66cb2\") " pod="openstack/openstackclient" Nov 24 14:55:27 crc kubenswrapper[4790]: I1124 14:55:27.909095 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvjxk\" (UniqueName: \"kubernetes.io/projected/a6d18574-10ed-4b11-b5a9-eeed45e66cb2-kube-api-access-qvjxk\") pod \"openstackclient\" (UID: \"a6d18574-10ed-4b11-b5a9-eeed45e66cb2\") " pod="openstack/openstackclient" Nov 24 14:55:27 crc kubenswrapper[4790]: I1124 14:55:27.909202 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a6d18574-10ed-4b11-b5a9-eeed45e66cb2-openstack-config\") pod \"openstackclient\" (UID: \"a6d18574-10ed-4b11-b5a9-eeed45e66cb2\") " pod="openstack/openstackclient" Nov 24 14:55:27 crc kubenswrapper[4790]: I1124 14:55:27.960980 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 14:55:27 crc kubenswrapper[4790]: I1124 14:55:27.962563 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 14:55:27 crc kubenswrapper[4790]: I1124 14:55:27.978806 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 14:55:27 crc kubenswrapper[4790]: I1124 14:55:27.982206 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-6fzwc" Nov 24 14:55:28 crc kubenswrapper[4790]: I1124 14:55:28.021116 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a6d18574-10ed-4b11-b5a9-eeed45e66cb2-openstack-config-secret\") pod \"openstackclient\" (UID: \"a6d18574-10ed-4b11-b5a9-eeed45e66cb2\") " pod="openstack/openstackclient" Nov 24 14:55:28 crc kubenswrapper[4790]: I1124 14:55:28.021246 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvjxk\" (UniqueName: \"kubernetes.io/projected/a6d18574-10ed-4b11-b5a9-eeed45e66cb2-kube-api-access-qvjxk\") pod \"openstackclient\" (UID: \"a6d18574-10ed-4b11-b5a9-eeed45e66cb2\") " pod="openstack/openstackclient" Nov 24 14:55:28 crc kubenswrapper[4790]: I1124 14:55:28.021292 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a6d18574-10ed-4b11-b5a9-eeed45e66cb2-openstack-config\") pod \"openstackclient\" (UID: \"a6d18574-10ed-4b11-b5a9-eeed45e66cb2\") " pod="openstack/openstackclient" Nov 24 14:55:28 crc kubenswrapper[4790]: I1124 14:55:28.022215 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a6d18574-10ed-4b11-b5a9-eeed45e66cb2-openstack-config\") pod \"openstackclient\" (UID: \"a6d18574-10ed-4b11-b5a9-eeed45e66cb2\") " pod="openstack/openstackclient" Nov 24 14:55:28 crc kubenswrapper[4790]: I1124 14:55:28.078819 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a6d18574-10ed-4b11-b5a9-eeed45e66cb2-openstack-config-secret\") pod \"openstackclient\" (UID: \"a6d18574-10ed-4b11-b5a9-eeed45e66cb2\") " pod="openstack/openstackclient" Nov 24 14:55:28 crc kubenswrapper[4790]: I1124 14:55:28.114914 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvjxk\" (UniqueName: \"kubernetes.io/projected/a6d18574-10ed-4b11-b5a9-eeed45e66cb2-kube-api-access-qvjxk\") pod \"openstackclient\" (UID: \"a6d18574-10ed-4b11-b5a9-eeed45e66cb2\") " pod="openstack/openstackclient" Nov 24 14:55:28 crc kubenswrapper[4790]: I1124 14:55:28.126463 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wk9x4\" (UniqueName: \"kubernetes.io/projected/582a457e-86b9-42e9-8a6a-99598f5e85c9-kube-api-access-wk9x4\") pod \"kube-state-metrics-0\" (UID: \"582a457e-86b9-42e9-8a6a-99598f5e85c9\") " pod="openstack/kube-state-metrics-0" Nov 24 14:55:28 crc kubenswrapper[4790]: I1124 14:55:28.230233 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wk9x4\" (UniqueName: \"kubernetes.io/projected/582a457e-86b9-42e9-8a6a-99598f5e85c9-kube-api-access-wk9x4\") pod \"kube-state-metrics-0\" (UID: \"582a457e-86b9-42e9-8a6a-99598f5e85c9\") " pod="openstack/kube-state-metrics-0" Nov 24 14:55:28 crc kubenswrapper[4790]: I1124 14:55:28.296846 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wk9x4\" (UniqueName: \"kubernetes.io/projected/582a457e-86b9-42e9-8a6a-99598f5e85c9-kube-api-access-wk9x4\") pod \"kube-state-metrics-0\" (UID: \"582a457e-86b9-42e9-8a6a-99598f5e85c9\") " pod="openstack/kube-state-metrics-0" Nov 24 14:55:28 crc kubenswrapper[4790]: I1124 14:55:28.405806 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 14:55:28 crc kubenswrapper[4790]: I1124 14:55:28.580175 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 14:55:28 crc kubenswrapper[4790]: I1124 14:55:28.892974 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Nov 24 14:55:28 crc kubenswrapper[4790]: I1124 14:55:28.899861 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Nov 24 14:55:28 crc kubenswrapper[4790]: I1124 14:55:28.914926 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-cluster-tls-config" Nov 24 14:55:28 crc kubenswrapper[4790]: I1124 14:55:28.915594 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Nov 24 14:55:28 crc kubenswrapper[4790]: I1124 14:55:28.915790 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-qz4f8" Nov 24 14:55:28 crc kubenswrapper[4790]: I1124 14:55:28.915846 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Nov 24 14:55:28 crc kubenswrapper[4790]: I1124 14:55:28.915905 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Nov 24 14:55:28 crc kubenswrapper[4790]: I1124 14:55:28.924449 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Nov 24 14:55:28 crc kubenswrapper[4790]: I1124 14:55:28.994471 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/35216e6d-676a-48f1-aeda-e379d7f6ddb4-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"35216e6d-676a-48f1-aeda-e379d7f6ddb4\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:55:28 crc kubenswrapper[4790]: I1124 14:55:28.994517 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/35216e6d-676a-48f1-aeda-e379d7f6ddb4-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"35216e6d-676a-48f1-aeda-e379d7f6ddb4\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:55:28 crc kubenswrapper[4790]: I1124 14:55:28.994560 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/35216e6d-676a-48f1-aeda-e379d7f6ddb4-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"35216e6d-676a-48f1-aeda-e379d7f6ddb4\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:55:28 crc kubenswrapper[4790]: I1124 14:55:28.994594 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/35216e6d-676a-48f1-aeda-e379d7f6ddb4-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"35216e6d-676a-48f1-aeda-e379d7f6ddb4\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:55:28 crc kubenswrapper[4790]: I1124 14:55:28.994629 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gm8jm\" (UniqueName: \"kubernetes.io/projected/35216e6d-676a-48f1-aeda-e379d7f6ddb4-kube-api-access-gm8jm\") pod \"alertmanager-metric-storage-0\" (UID: \"35216e6d-676a-48f1-aeda-e379d7f6ddb4\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:55:28 crc kubenswrapper[4790]: I1124 14:55:28.994652 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/35216e6d-676a-48f1-aeda-e379d7f6ddb4-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"35216e6d-676a-48f1-aeda-e379d7f6ddb4\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:55:28 crc kubenswrapper[4790]: I1124 14:55:28.994684 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/35216e6d-676a-48f1-aeda-e379d7f6ddb4-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"35216e6d-676a-48f1-aeda-e379d7f6ddb4\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.096016 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/35216e6d-676a-48f1-aeda-e379d7f6ddb4-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"35216e6d-676a-48f1-aeda-e379d7f6ddb4\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.096085 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/35216e6d-676a-48f1-aeda-e379d7f6ddb4-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"35216e6d-676a-48f1-aeda-e379d7f6ddb4\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.096134 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/35216e6d-676a-48f1-aeda-e379d7f6ddb4-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"35216e6d-676a-48f1-aeda-e379d7f6ddb4\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.096171 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gm8jm\" (UniqueName: \"kubernetes.io/projected/35216e6d-676a-48f1-aeda-e379d7f6ddb4-kube-api-access-gm8jm\") pod \"alertmanager-metric-storage-0\" (UID: \"35216e6d-676a-48f1-aeda-e379d7f6ddb4\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.096193 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/35216e6d-676a-48f1-aeda-e379d7f6ddb4-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"35216e6d-676a-48f1-aeda-e379d7f6ddb4\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.096222 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/35216e6d-676a-48f1-aeda-e379d7f6ddb4-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"35216e6d-676a-48f1-aeda-e379d7f6ddb4\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.096325 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/35216e6d-676a-48f1-aeda-e379d7f6ddb4-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"35216e6d-676a-48f1-aeda-e379d7f6ddb4\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.102419 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/35216e6d-676a-48f1-aeda-e379d7f6ddb4-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"35216e6d-676a-48f1-aeda-e379d7f6ddb4\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.111229 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/35216e6d-676a-48f1-aeda-e379d7f6ddb4-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"35216e6d-676a-48f1-aeda-e379d7f6ddb4\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.115131 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/35216e6d-676a-48f1-aeda-e379d7f6ddb4-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"35216e6d-676a-48f1-aeda-e379d7f6ddb4\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.140961 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gm8jm\" (UniqueName: \"kubernetes.io/projected/35216e6d-676a-48f1-aeda-e379d7f6ddb4-kube-api-access-gm8jm\") pod \"alertmanager-metric-storage-0\" (UID: \"35216e6d-676a-48f1-aeda-e379d7f6ddb4\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.145906 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/35216e6d-676a-48f1-aeda-e379d7f6ddb4-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"35216e6d-676a-48f1-aeda-e379d7f6ddb4\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.151230 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/35216e6d-676a-48f1-aeda-e379d7f6ddb4-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"35216e6d-676a-48f1-aeda-e379d7f6ddb4\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.151467 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/35216e6d-676a-48f1-aeda-e379d7f6ddb4-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"35216e6d-676a-48f1-aeda-e379d7f6ddb4\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.275349 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.318393 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.320766 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.330132 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.330388 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.330531 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-ssk9z" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.330712 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.331618 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.331830 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.373584 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.489512 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.510572 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/09da3379-5d69-4618-ad5c-baa983f57405-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"09da3379-5d69-4618-ad5c-baa983f57405\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.510648 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/09da3379-5d69-4618-ad5c-baa983f57405-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"09da3379-5d69-4618-ad5c-baa983f57405\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.510679 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/09da3379-5d69-4618-ad5c-baa983f57405-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"09da3379-5d69-4618-ad5c-baa983f57405\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.510704 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/09da3379-5d69-4618-ad5c-baa983f57405-config\") pod \"prometheus-metric-storage-0\" (UID: \"09da3379-5d69-4618-ad5c-baa983f57405\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.510791 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/09da3379-5d69-4618-ad5c-baa983f57405-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"09da3379-5d69-4618-ad5c-baa983f57405\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.510855 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjp96\" (UniqueName: \"kubernetes.io/projected/09da3379-5d69-4618-ad5c-baa983f57405-kube-api-access-vjp96\") pod \"prometheus-metric-storage-0\" (UID: \"09da3379-5d69-4618-ad5c-baa983f57405\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.510946 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/09da3379-5d69-4618-ad5c-baa983f57405-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"09da3379-5d69-4618-ad5c-baa983f57405\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.510969 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-0aa90a71-0ce3-488e-8a52-4cf348f32818\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0aa90a71-0ce3-488e-8a52-4cf348f32818\") pod \"prometheus-metric-storage-0\" (UID: \"09da3379-5d69-4618-ad5c-baa983f57405\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.535527 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.617028 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjp96\" (UniqueName: \"kubernetes.io/projected/09da3379-5d69-4618-ad5c-baa983f57405-kube-api-access-vjp96\") pod \"prometheus-metric-storage-0\" (UID: \"09da3379-5d69-4618-ad5c-baa983f57405\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.617323 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/09da3379-5d69-4618-ad5c-baa983f57405-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"09da3379-5d69-4618-ad5c-baa983f57405\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.617348 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-0aa90a71-0ce3-488e-8a52-4cf348f32818\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0aa90a71-0ce3-488e-8a52-4cf348f32818\") pod \"prometheus-metric-storage-0\" (UID: \"09da3379-5d69-4618-ad5c-baa983f57405\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.617429 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/09da3379-5d69-4618-ad5c-baa983f57405-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"09da3379-5d69-4618-ad5c-baa983f57405\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.617476 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/09da3379-5d69-4618-ad5c-baa983f57405-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"09da3379-5d69-4618-ad5c-baa983f57405\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.617495 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/09da3379-5d69-4618-ad5c-baa983f57405-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"09da3379-5d69-4618-ad5c-baa983f57405\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.617514 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/09da3379-5d69-4618-ad5c-baa983f57405-config\") pod \"prometheus-metric-storage-0\" (UID: \"09da3379-5d69-4618-ad5c-baa983f57405\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.617567 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/09da3379-5d69-4618-ad5c-baa983f57405-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"09da3379-5d69-4618-ad5c-baa983f57405\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.618634 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/09da3379-5d69-4618-ad5c-baa983f57405-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"09da3379-5d69-4618-ad5c-baa983f57405\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.629536 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/09da3379-5d69-4618-ad5c-baa983f57405-config\") pod \"prometheus-metric-storage-0\" (UID: \"09da3379-5d69-4618-ad5c-baa983f57405\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.630883 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/09da3379-5d69-4618-ad5c-baa983f57405-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"09da3379-5d69-4618-ad5c-baa983f57405\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.634552 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/09da3379-5d69-4618-ad5c-baa983f57405-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"09da3379-5d69-4618-ad5c-baa983f57405\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.636255 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/09da3379-5d69-4618-ad5c-baa983f57405-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"09da3379-5d69-4618-ad5c-baa983f57405\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.641781 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/09da3379-5d69-4618-ad5c-baa983f57405-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"09da3379-5d69-4618-ad5c-baa983f57405\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.681691 4790 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.681750 4790 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-0aa90a71-0ce3-488e-8a52-4cf348f32818\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0aa90a71-0ce3-488e-8a52-4cf348f32818\") pod \"prometheus-metric-storage-0\" (UID: \"09da3379-5d69-4618-ad5c-baa983f57405\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/53e291418aa15661fe8e1d27d747b97ed99dd7240c2163720e18b1cb9071773a/globalmount\"" pod="openstack/prometheus-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.750590 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjp96\" (UniqueName: \"kubernetes.io/projected/09da3379-5d69-4618-ad5c-baa983f57405-kube-api-access-vjp96\") pod \"prometheus-metric-storage-0\" (UID: \"09da3379-5d69-4618-ad5c-baa983f57405\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.854628 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-0aa90a71-0ce3-488e-8a52-4cf348f32818\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0aa90a71-0ce3-488e-8a52-4cf348f32818\") pod \"prometheus-metric-storage-0\" (UID: \"09da3379-5d69-4618-ad5c-baa983f57405\") " pod="openstack/prometheus-metric-storage-0" Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.968166 4790 generic.go:334] "Generic (PLEG): container finished" podID="b5399fd5-dda4-4a13-954f-76a99b1577a3" containerID="51ce6763398e66fcf80424e574cd3875e846b87aaa373c269f452945a5f9b9ce" exitCode=137 Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.972212 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"a6d18574-10ed-4b11-b5a9-eeed45e66cb2","Type":"ContainerStarted","Data":"7c2cd05a069dae3d739ba20c70fe2d60cd578e4864cc145e79d967dc298820f5"} Nov 24 14:55:29 crc kubenswrapper[4790]: I1124 14:55:29.975488 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"582a457e-86b9-42e9-8a6a-99598f5e85c9","Type":"ContainerStarted","Data":"03afd4ef6909ccb9e6db42b01bb29a0cfa0ee5f9498d991bbcea40371e30e067"} Nov 24 14:55:30 crc kubenswrapper[4790]: I1124 14:55:30.043239 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 24 14:55:30 crc kubenswrapper[4790]: I1124 14:55:30.289809 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Nov 24 14:55:30 crc kubenswrapper[4790]: I1124 14:55:30.921084 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 14:55:31 crc kubenswrapper[4790]: I1124 14:55:31.000253 4790 scope.go:117] "RemoveContainer" containerID="51ce6763398e66fcf80424e574cd3875e846b87aaa373c269f452945a5f9b9ce" Nov 24 14:55:31 crc kubenswrapper[4790]: I1124 14:55:31.000275 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 14:55:31 crc kubenswrapper[4790]: I1124 14:55:31.010002 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"a6d18574-10ed-4b11-b5a9-eeed45e66cb2","Type":"ContainerStarted","Data":"9b5cdf5afafb8fde6c82ac85472eecea79406e1817bc2a95e3bf16ce1c784c74"} Nov 24 14:55:31 crc kubenswrapper[4790]: I1124 14:55:31.016947 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 24 14:55:31 crc kubenswrapper[4790]: I1124 14:55:31.019734 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"35216e6d-676a-48f1-aeda-e379d7f6ddb4","Type":"ContainerStarted","Data":"dc212abc2c8f2ad2a25a3c5fa46d856e859580a2d87b7dff8d7e49b5832298a4"} Nov 24 14:55:31 crc kubenswrapper[4790]: I1124 14:55:31.032482 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rm9l9\" (UniqueName: \"kubernetes.io/projected/b5399fd5-dda4-4a13-954f-76a99b1577a3-kube-api-access-rm9l9\") pod \"b5399fd5-dda4-4a13-954f-76a99b1577a3\" (UID: \"b5399fd5-dda4-4a13-954f-76a99b1577a3\") " Nov 24 14:55:31 crc kubenswrapper[4790]: I1124 14:55:31.033980 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b5399fd5-dda4-4a13-954f-76a99b1577a3-openstack-config\") pod \"b5399fd5-dda4-4a13-954f-76a99b1577a3\" (UID: \"b5399fd5-dda4-4a13-954f-76a99b1577a3\") " Nov 24 14:55:31 crc kubenswrapper[4790]: I1124 14:55:31.036395 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b5399fd5-dda4-4a13-954f-76a99b1577a3-openstack-config-secret\") pod \"b5399fd5-dda4-4a13-954f-76a99b1577a3\" (UID: \"b5399fd5-dda4-4a13-954f-76a99b1577a3\") " Nov 24 14:55:31 crc kubenswrapper[4790]: I1124 14:55:31.041021 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=4.041001663 podStartE2EDuration="4.041001663s" podCreationTimestamp="2025-11-24 14:55:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:55:31.024577599 +0000 UTC m=+6179.404471261" watchObservedRunningTime="2025-11-24 14:55:31.041001663 +0000 UTC m=+6179.420895315" Nov 24 14:55:31 crc kubenswrapper[4790]: I1124 14:55:31.043084 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.824977701 podStartE2EDuration="4.043075205s" podCreationTimestamp="2025-11-24 14:55:27 +0000 UTC" firstStartedPulling="2025-11-24 14:55:29.494200151 +0000 UTC m=+6177.874093813" lastFinishedPulling="2025-11-24 14:55:30.712297655 +0000 UTC m=+6179.092191317" observedRunningTime="2025-11-24 14:55:31.03891422 +0000 UTC m=+6179.418807892" watchObservedRunningTime="2025-11-24 14:55:31.043075205 +0000 UTC m=+6179.422968867" Nov 24 14:55:31 crc kubenswrapper[4790]: I1124 14:55:31.046038 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5399fd5-dda4-4a13-954f-76a99b1577a3-kube-api-access-rm9l9" (OuterVolumeSpecName: "kube-api-access-rm9l9") pod "b5399fd5-dda4-4a13-954f-76a99b1577a3" (UID: "b5399fd5-dda4-4a13-954f-76a99b1577a3"). InnerVolumeSpecName "kube-api-access-rm9l9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:55:31 crc kubenswrapper[4790]: I1124 14:55:31.122747 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5399fd5-dda4-4a13-954f-76a99b1577a3-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "b5399fd5-dda4-4a13-954f-76a99b1577a3" (UID: "b5399fd5-dda4-4a13-954f-76a99b1577a3"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:55:31 crc kubenswrapper[4790]: I1124 14:55:31.140770 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5399fd5-dda4-4a13-954f-76a99b1577a3-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "b5399fd5-dda4-4a13-954f-76a99b1577a3" (UID: "b5399fd5-dda4-4a13-954f-76a99b1577a3"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:55:31 crc kubenswrapper[4790]: I1124 14:55:31.147525 4790 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b5399fd5-dda4-4a13-954f-76a99b1577a3-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 24 14:55:31 crc kubenswrapper[4790]: I1124 14:55:31.147563 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rm9l9\" (UniqueName: \"kubernetes.io/projected/b5399fd5-dda4-4a13-954f-76a99b1577a3-kube-api-access-rm9l9\") on node \"crc\" DevicePath \"\"" Nov 24 14:55:31 crc kubenswrapper[4790]: I1124 14:55:31.147577 4790 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b5399fd5-dda4-4a13-954f-76a99b1577a3-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:55:31 crc kubenswrapper[4790]: I1124 14:55:31.319409 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 14:55:31 crc kubenswrapper[4790]: I1124 14:55:31.322329 4790 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="b5399fd5-dda4-4a13-954f-76a99b1577a3" podUID="a6d18574-10ed-4b11-b5a9-eeed45e66cb2" Nov 24 14:55:31 crc kubenswrapper[4790]: W1124 14:55:31.369550 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09da3379_5d69_4618_ad5c_baa983f57405.slice/crio-20ad7f7e526982c7ca1340e8b482e9ef39f950d4baad48b3c9684f85783328d7 WatchSource:0}: Error finding container 20ad7f7e526982c7ca1340e8b482e9ef39f950d4baad48b3c9684f85783328d7: Status 404 returned error can't find the container with id 20ad7f7e526982c7ca1340e8b482e9ef39f950d4baad48b3c9684f85783328d7 Nov 24 14:55:31 crc kubenswrapper[4790]: I1124 14:55:31.434084 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-v7xkz" Nov 24 14:55:32 crc kubenswrapper[4790]: I1124 14:55:32.037040 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"09da3379-5d69-4618-ad5c-baa983f57405","Type":"ContainerStarted","Data":"20ad7f7e526982c7ca1340e8b482e9ef39f950d4baad48b3c9684f85783328d7"} Nov 24 14:55:32 crc kubenswrapper[4790]: I1124 14:55:32.039452 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"582a457e-86b9-42e9-8a6a-99598f5e85c9","Type":"ContainerStarted","Data":"9f49b5d3750d71e885214a19f8cdb20e3aca9422ca6b6c7d4b2413405589dd91"} Nov 24 14:55:32 crc kubenswrapper[4790]: I1124 14:55:32.337914 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5399fd5-dda4-4a13-954f-76a99b1577a3" path="/var/lib/kubelet/pods/b5399fd5-dda4-4a13-954f-76a99b1577a3/volumes" Nov 24 14:55:33 crc kubenswrapper[4790]: I1124 14:55:33.050966 4790 generic.go:334] "Generic (PLEG): container finished" podID="6fbf1db8-8f67-429c-97c9-e97714898b0e" containerID="742e97e5aac502e3bc676c0c4da51dd8b1049dd9c52ae758218c5c24533f59c5" exitCode=0 Nov 24 14:55:33 crc kubenswrapper[4790]: I1124 14:55:33.051023 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vqnlr" event={"ID":"6fbf1db8-8f67-429c-97c9-e97714898b0e","Type":"ContainerDied","Data":"742e97e5aac502e3bc676c0c4da51dd8b1049dd9c52ae758218c5c24533f59c5"} Nov 24 14:55:35 crc kubenswrapper[4790]: I1124 14:55:35.084984 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vqnlr" event={"ID":"6fbf1db8-8f67-429c-97c9-e97714898b0e","Type":"ContainerStarted","Data":"b2f00cf0b52b771e7d12ba762349bde9037f0865c61c11025523ce3fca2690c3"} Nov 24 14:55:35 crc kubenswrapper[4790]: I1124 14:55:35.102071 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vqnlr" podStartSLOduration=2.993428198 podStartE2EDuration="13.102052553s" podCreationTimestamp="2025-11-24 14:55:22 +0000 UTC" firstStartedPulling="2025-11-24 14:55:23.798953662 +0000 UTC m=+6172.178847334" lastFinishedPulling="2025-11-24 14:55:33.907578027 +0000 UTC m=+6182.287471689" observedRunningTime="2025-11-24 14:55:35.10032687 +0000 UTC m=+6183.480220522" watchObservedRunningTime="2025-11-24 14:55:35.102052553 +0000 UTC m=+6183.481946215" Nov 24 14:55:35 crc kubenswrapper[4790]: I1124 14:55:35.606857 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v7xkz"] Nov 24 14:55:35 crc kubenswrapper[4790]: I1124 14:55:35.607517 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-v7xkz" podUID="33662038-1049-4380-804a-1ce66fd13db7" containerName="registry-server" containerID="cri-o://d5cb2bb1b93c40424bb48a6a0e7f76518b73206f6e2011df2fd1ab5f2673ff95" gracePeriod=2 Nov 24 14:55:36 crc kubenswrapper[4790]: I1124 14:55:36.100557 4790 generic.go:334] "Generic (PLEG): container finished" podID="33662038-1049-4380-804a-1ce66fd13db7" containerID="d5cb2bb1b93c40424bb48a6a0e7f76518b73206f6e2011df2fd1ab5f2673ff95" exitCode=0 Nov 24 14:55:36 crc kubenswrapper[4790]: I1124 14:55:36.100625 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v7xkz" event={"ID":"33662038-1049-4380-804a-1ce66fd13db7","Type":"ContainerDied","Data":"d5cb2bb1b93c40424bb48a6a0e7f76518b73206f6e2011df2fd1ab5f2673ff95"} Nov 24 14:55:36 crc kubenswrapper[4790]: I1124 14:55:36.100979 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v7xkz" event={"ID":"33662038-1049-4380-804a-1ce66fd13db7","Type":"ContainerDied","Data":"bdf680c73f46e872ccc05eeb1382b2920dbbf9fb83dab68873f17cf0bdf2cf99"} Nov 24 14:55:36 crc kubenswrapper[4790]: I1124 14:55:36.101010 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bdf680c73f46e872ccc05eeb1382b2920dbbf9fb83dab68873f17cf0bdf2cf99" Nov 24 14:55:36 crc kubenswrapper[4790]: I1124 14:55:36.243619 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v7xkz" Nov 24 14:55:36 crc kubenswrapper[4790]: I1124 14:55:36.369780 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33662038-1049-4380-804a-1ce66fd13db7-catalog-content\") pod \"33662038-1049-4380-804a-1ce66fd13db7\" (UID: \"33662038-1049-4380-804a-1ce66fd13db7\") " Nov 24 14:55:36 crc kubenswrapper[4790]: I1124 14:55:36.370062 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnfmh\" (UniqueName: \"kubernetes.io/projected/33662038-1049-4380-804a-1ce66fd13db7-kube-api-access-nnfmh\") pod \"33662038-1049-4380-804a-1ce66fd13db7\" (UID: \"33662038-1049-4380-804a-1ce66fd13db7\") " Nov 24 14:55:36 crc kubenswrapper[4790]: I1124 14:55:36.370099 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33662038-1049-4380-804a-1ce66fd13db7-utilities\") pod \"33662038-1049-4380-804a-1ce66fd13db7\" (UID: \"33662038-1049-4380-804a-1ce66fd13db7\") " Nov 24 14:55:36 crc kubenswrapper[4790]: I1124 14:55:36.374207 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33662038-1049-4380-804a-1ce66fd13db7-utilities" (OuterVolumeSpecName: "utilities") pod "33662038-1049-4380-804a-1ce66fd13db7" (UID: "33662038-1049-4380-804a-1ce66fd13db7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:55:36 crc kubenswrapper[4790]: I1124 14:55:36.378116 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33662038-1049-4380-804a-1ce66fd13db7-kube-api-access-nnfmh" (OuterVolumeSpecName: "kube-api-access-nnfmh") pod "33662038-1049-4380-804a-1ce66fd13db7" (UID: "33662038-1049-4380-804a-1ce66fd13db7"). InnerVolumeSpecName "kube-api-access-nnfmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:55:36 crc kubenswrapper[4790]: I1124 14:55:36.396292 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33662038-1049-4380-804a-1ce66fd13db7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "33662038-1049-4380-804a-1ce66fd13db7" (UID: "33662038-1049-4380-804a-1ce66fd13db7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:55:36 crc kubenswrapper[4790]: I1124 14:55:36.472240 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33662038-1049-4380-804a-1ce66fd13db7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:55:36 crc kubenswrapper[4790]: I1124 14:55:36.472282 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnfmh\" (UniqueName: \"kubernetes.io/projected/33662038-1049-4380-804a-1ce66fd13db7-kube-api-access-nnfmh\") on node \"crc\" DevicePath \"\"" Nov 24 14:55:36 crc kubenswrapper[4790]: I1124 14:55:36.472297 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33662038-1049-4380-804a-1ce66fd13db7-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:55:37 crc kubenswrapper[4790]: I1124 14:55:37.111198 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v7xkz" Nov 24 14:55:37 crc kubenswrapper[4790]: I1124 14:55:37.181694 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v7xkz"] Nov 24 14:55:37 crc kubenswrapper[4790]: I1124 14:55:37.193490 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-v7xkz"] Nov 24 14:55:37 crc kubenswrapper[4790]: I1124 14:55:37.316208 4790 scope.go:117] "RemoveContainer" containerID="28308703817a1d4e0e5344ce19268127c600894f2daa6b921d52a9105f014132" Nov 24 14:55:37 crc kubenswrapper[4790]: E1124 14:55:37.316583 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:55:38 crc kubenswrapper[4790]: I1124 14:55:38.038816 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-68pwg"] Nov 24 14:55:38 crc kubenswrapper[4790]: I1124 14:55:38.054452 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-68pwg"] Nov 24 14:55:38 crc kubenswrapper[4790]: I1124 14:55:38.122344 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"09da3379-5d69-4618-ad5c-baa983f57405","Type":"ContainerStarted","Data":"719bb0dbb76aec2e4568c8b26c67a272c73aec2919c3bdadde878e07919ef769"} Nov 24 14:55:38 crc kubenswrapper[4790]: I1124 14:55:38.124595 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"35216e6d-676a-48f1-aeda-e379d7f6ddb4","Type":"ContainerStarted","Data":"7b763ebbd6f968b7911a4e02d3f05fc0901f9af685c1ba3177052cbaf663a9f8"} Nov 24 14:55:38 crc kubenswrapper[4790]: I1124 14:55:38.336192 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33662038-1049-4380-804a-1ce66fd13db7" path="/var/lib/kubelet/pods/33662038-1049-4380-804a-1ce66fd13db7/volumes" Nov 24 14:55:38 crc kubenswrapper[4790]: I1124 14:55:38.341779 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7581458b-2c21-4297-af62-05fd707bb201" path="/var/lib/kubelet/pods/7581458b-2c21-4297-af62-05fd707bb201/volumes" Nov 24 14:55:38 crc kubenswrapper[4790]: I1124 14:55:38.594099 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 24 14:55:42 crc kubenswrapper[4790]: I1124 14:55:42.577661 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vqnlr" Nov 24 14:55:42 crc kubenswrapper[4790]: I1124 14:55:42.578378 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vqnlr" Nov 24 14:55:42 crc kubenswrapper[4790]: I1124 14:55:42.653606 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vqnlr" Nov 24 14:55:43 crc kubenswrapper[4790]: I1124 14:55:43.251475 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vqnlr" Nov 24 14:55:43 crc kubenswrapper[4790]: I1124 14:55:43.299941 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vqnlr"] Nov 24 14:55:45 crc kubenswrapper[4790]: I1124 14:55:45.227824 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vqnlr" podUID="6fbf1db8-8f67-429c-97c9-e97714898b0e" containerName="registry-server" containerID="cri-o://b2f00cf0b52b771e7d12ba762349bde9037f0865c61c11025523ce3fca2690c3" gracePeriod=2 Nov 24 14:55:45 crc kubenswrapper[4790]: I1124 14:55:45.880723 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vqnlr" Nov 24 14:55:45 crc kubenswrapper[4790]: I1124 14:55:45.902196 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fbf1db8-8f67-429c-97c9-e97714898b0e-utilities\") pod \"6fbf1db8-8f67-429c-97c9-e97714898b0e\" (UID: \"6fbf1db8-8f67-429c-97c9-e97714898b0e\") " Nov 24 14:55:45 crc kubenswrapper[4790]: I1124 14:55:45.902412 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fp8hx\" (UniqueName: \"kubernetes.io/projected/6fbf1db8-8f67-429c-97c9-e97714898b0e-kube-api-access-fp8hx\") pod \"6fbf1db8-8f67-429c-97c9-e97714898b0e\" (UID: \"6fbf1db8-8f67-429c-97c9-e97714898b0e\") " Nov 24 14:55:45 crc kubenswrapper[4790]: I1124 14:55:45.902466 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fbf1db8-8f67-429c-97c9-e97714898b0e-catalog-content\") pod \"6fbf1db8-8f67-429c-97c9-e97714898b0e\" (UID: \"6fbf1db8-8f67-429c-97c9-e97714898b0e\") " Nov 24 14:55:45 crc kubenswrapper[4790]: I1124 14:55:45.903310 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6fbf1db8-8f67-429c-97c9-e97714898b0e-utilities" (OuterVolumeSpecName: "utilities") pod "6fbf1db8-8f67-429c-97c9-e97714898b0e" (UID: "6fbf1db8-8f67-429c-97c9-e97714898b0e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:55:45 crc kubenswrapper[4790]: I1124 14:55:45.909655 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fbf1db8-8f67-429c-97c9-e97714898b0e-kube-api-access-fp8hx" (OuterVolumeSpecName: "kube-api-access-fp8hx") pod "6fbf1db8-8f67-429c-97c9-e97714898b0e" (UID: "6fbf1db8-8f67-429c-97c9-e97714898b0e"). InnerVolumeSpecName "kube-api-access-fp8hx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:55:46 crc kubenswrapper[4790]: I1124 14:55:46.005456 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fp8hx\" (UniqueName: \"kubernetes.io/projected/6fbf1db8-8f67-429c-97c9-e97714898b0e-kube-api-access-fp8hx\") on node \"crc\" DevicePath \"\"" Nov 24 14:55:46 crc kubenswrapper[4790]: I1124 14:55:46.005496 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fbf1db8-8f67-429c-97c9-e97714898b0e-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:55:46 crc kubenswrapper[4790]: I1124 14:55:46.014253 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6fbf1db8-8f67-429c-97c9-e97714898b0e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6fbf1db8-8f67-429c-97c9-e97714898b0e" (UID: "6fbf1db8-8f67-429c-97c9-e97714898b0e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:55:46 crc kubenswrapper[4790]: I1124 14:55:46.108843 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fbf1db8-8f67-429c-97c9-e97714898b0e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:55:46 crc kubenswrapper[4790]: I1124 14:55:46.238642 4790 generic.go:334] "Generic (PLEG): container finished" podID="09da3379-5d69-4618-ad5c-baa983f57405" containerID="719bb0dbb76aec2e4568c8b26c67a272c73aec2919c3bdadde878e07919ef769" exitCode=0 Nov 24 14:55:46 crc kubenswrapper[4790]: I1124 14:55:46.238766 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"09da3379-5d69-4618-ad5c-baa983f57405","Type":"ContainerDied","Data":"719bb0dbb76aec2e4568c8b26c67a272c73aec2919c3bdadde878e07919ef769"} Nov 24 14:55:46 crc kubenswrapper[4790]: I1124 14:55:46.240984 4790 generic.go:334] "Generic (PLEG): container finished" podID="35216e6d-676a-48f1-aeda-e379d7f6ddb4" containerID="7b763ebbd6f968b7911a4e02d3f05fc0901f9af685c1ba3177052cbaf663a9f8" exitCode=0 Nov 24 14:55:46 crc kubenswrapper[4790]: I1124 14:55:46.241115 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"35216e6d-676a-48f1-aeda-e379d7f6ddb4","Type":"ContainerDied","Data":"7b763ebbd6f968b7911a4e02d3f05fc0901f9af685c1ba3177052cbaf663a9f8"} Nov 24 14:55:46 crc kubenswrapper[4790]: I1124 14:55:46.244270 4790 generic.go:334] "Generic (PLEG): container finished" podID="6fbf1db8-8f67-429c-97c9-e97714898b0e" containerID="b2f00cf0b52b771e7d12ba762349bde9037f0865c61c11025523ce3fca2690c3" exitCode=0 Nov 24 14:55:46 crc kubenswrapper[4790]: I1124 14:55:46.244317 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vqnlr" event={"ID":"6fbf1db8-8f67-429c-97c9-e97714898b0e","Type":"ContainerDied","Data":"b2f00cf0b52b771e7d12ba762349bde9037f0865c61c11025523ce3fca2690c3"} Nov 24 14:55:46 crc kubenswrapper[4790]: I1124 14:55:46.244348 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vqnlr" event={"ID":"6fbf1db8-8f67-429c-97c9-e97714898b0e","Type":"ContainerDied","Data":"9929df02e103dc2a97ecfa0d48f084854715817bc3161d276828afd439e7ed4d"} Nov 24 14:55:46 crc kubenswrapper[4790]: I1124 14:55:46.244376 4790 scope.go:117] "RemoveContainer" containerID="b2f00cf0b52b771e7d12ba762349bde9037f0865c61c11025523ce3fca2690c3" Nov 24 14:55:46 crc kubenswrapper[4790]: I1124 14:55:46.244561 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vqnlr" Nov 24 14:55:46 crc kubenswrapper[4790]: I1124 14:55:46.306956 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vqnlr"] Nov 24 14:55:46 crc kubenswrapper[4790]: I1124 14:55:46.307753 4790 scope.go:117] "RemoveContainer" containerID="742e97e5aac502e3bc676c0c4da51dd8b1049dd9c52ae758218c5c24533f59c5" Nov 24 14:55:46 crc kubenswrapper[4790]: I1124 14:55:46.335853 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vqnlr"] Nov 24 14:55:46 crc kubenswrapper[4790]: I1124 14:55:46.343798 4790 scope.go:117] "RemoveContainer" containerID="aa8b4355207d9e157b8e068c5ae0ee0ca10dc9bf9aed40ed86ebe8347e1f279c" Nov 24 14:55:46 crc kubenswrapper[4790]: I1124 14:55:46.402434 4790 scope.go:117] "RemoveContainer" containerID="b2f00cf0b52b771e7d12ba762349bde9037f0865c61c11025523ce3fca2690c3" Nov 24 14:55:46 crc kubenswrapper[4790]: E1124 14:55:46.403113 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2f00cf0b52b771e7d12ba762349bde9037f0865c61c11025523ce3fca2690c3\": container with ID starting with b2f00cf0b52b771e7d12ba762349bde9037f0865c61c11025523ce3fca2690c3 not found: ID does not exist" containerID="b2f00cf0b52b771e7d12ba762349bde9037f0865c61c11025523ce3fca2690c3" Nov 24 14:55:46 crc kubenswrapper[4790]: I1124 14:55:46.403166 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2f00cf0b52b771e7d12ba762349bde9037f0865c61c11025523ce3fca2690c3"} err="failed to get container status \"b2f00cf0b52b771e7d12ba762349bde9037f0865c61c11025523ce3fca2690c3\": rpc error: code = NotFound desc = could not find container \"b2f00cf0b52b771e7d12ba762349bde9037f0865c61c11025523ce3fca2690c3\": container with ID starting with b2f00cf0b52b771e7d12ba762349bde9037f0865c61c11025523ce3fca2690c3 not found: ID does not exist" Nov 24 14:55:46 crc kubenswrapper[4790]: I1124 14:55:46.403198 4790 scope.go:117] "RemoveContainer" containerID="742e97e5aac502e3bc676c0c4da51dd8b1049dd9c52ae758218c5c24533f59c5" Nov 24 14:55:46 crc kubenswrapper[4790]: E1124 14:55:46.404038 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"742e97e5aac502e3bc676c0c4da51dd8b1049dd9c52ae758218c5c24533f59c5\": container with ID starting with 742e97e5aac502e3bc676c0c4da51dd8b1049dd9c52ae758218c5c24533f59c5 not found: ID does not exist" containerID="742e97e5aac502e3bc676c0c4da51dd8b1049dd9c52ae758218c5c24533f59c5" Nov 24 14:55:46 crc kubenswrapper[4790]: I1124 14:55:46.404076 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"742e97e5aac502e3bc676c0c4da51dd8b1049dd9c52ae758218c5c24533f59c5"} err="failed to get container status \"742e97e5aac502e3bc676c0c4da51dd8b1049dd9c52ae758218c5c24533f59c5\": rpc error: code = NotFound desc = could not find container \"742e97e5aac502e3bc676c0c4da51dd8b1049dd9c52ae758218c5c24533f59c5\": container with ID starting with 742e97e5aac502e3bc676c0c4da51dd8b1049dd9c52ae758218c5c24533f59c5 not found: ID does not exist" Nov 24 14:55:46 crc kubenswrapper[4790]: I1124 14:55:46.404100 4790 scope.go:117] "RemoveContainer" containerID="aa8b4355207d9e157b8e068c5ae0ee0ca10dc9bf9aed40ed86ebe8347e1f279c" Nov 24 14:55:46 crc kubenswrapper[4790]: E1124 14:55:46.404552 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa8b4355207d9e157b8e068c5ae0ee0ca10dc9bf9aed40ed86ebe8347e1f279c\": container with ID starting with aa8b4355207d9e157b8e068c5ae0ee0ca10dc9bf9aed40ed86ebe8347e1f279c not found: ID does not exist" containerID="aa8b4355207d9e157b8e068c5ae0ee0ca10dc9bf9aed40ed86ebe8347e1f279c" Nov 24 14:55:46 crc kubenswrapper[4790]: I1124 14:55:46.404594 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa8b4355207d9e157b8e068c5ae0ee0ca10dc9bf9aed40ed86ebe8347e1f279c"} err="failed to get container status \"aa8b4355207d9e157b8e068c5ae0ee0ca10dc9bf9aed40ed86ebe8347e1f279c\": rpc error: code = NotFound desc = could not find container \"aa8b4355207d9e157b8e068c5ae0ee0ca10dc9bf9aed40ed86ebe8347e1f279c\": container with ID starting with aa8b4355207d9e157b8e068c5ae0ee0ca10dc9bf9aed40ed86ebe8347e1f279c not found: ID does not exist" Nov 24 14:55:48 crc kubenswrapper[4790]: I1124 14:55:48.329023 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fbf1db8-8f67-429c-97c9-e97714898b0e" path="/var/lib/kubelet/pods/6fbf1db8-8f67-429c-97c9-e97714898b0e/volumes" Nov 24 14:55:49 crc kubenswrapper[4790]: I1124 14:55:49.284021 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"35216e6d-676a-48f1-aeda-e379d7f6ddb4","Type":"ContainerStarted","Data":"bf90f2265d9b3ec04813eae940a8aec52eea6e0d9068c186946e86df7905c1d4"} Nov 24 14:55:50 crc kubenswrapper[4790]: I1124 14:55:50.930651 4790 scope.go:117] "RemoveContainer" containerID="7d4b9f6a11c41644138576f5ccb753d918c145085f323cc28912ba50cced709b" Nov 24 14:55:51 crc kubenswrapper[4790]: I1124 14:55:51.688320 4790 scope.go:117] "RemoveContainer" containerID="cb7a331fbf30381c6e7a979dd1f51750b18f7a4083a30da259de3cc4a9286e29" Nov 24 14:55:51 crc kubenswrapper[4790]: I1124 14:55:51.735669 4790 scope.go:117] "RemoveContainer" containerID="20a60ed2404a9b78c22ff3b4d82a3742b2ceeda85aa4919fff5d62f3e933ade4" Nov 24 14:55:51 crc kubenswrapper[4790]: I1124 14:55:51.803067 4790 scope.go:117] "RemoveContainer" containerID="7617c44da6fa13317bdbd19520037809f78c2aa5cc3ee058ec952ff2b9802314" Nov 24 14:55:51 crc kubenswrapper[4790]: I1124 14:55:51.945794 4790 scope.go:117] "RemoveContainer" containerID="c108fdf41c12d6c658e87825b230bc0807a4e4df5eaa6e1b7735027278a07b20" Nov 24 14:55:51 crc kubenswrapper[4790]: I1124 14:55:51.990257 4790 scope.go:117] "RemoveContainer" containerID="f4132b7f7a234c5c9e8140136b0824fcdd3d69623d363a43024e4ccfa3f7c0b7" Nov 24 14:55:52 crc kubenswrapper[4790]: I1124 14:55:52.042082 4790 scope.go:117] "RemoveContainer" containerID="fcebf0df908b2a9cd9189fe5ede67e477dff4ce9334dc82d06f074beb6bcf297" Nov 24 14:55:52 crc kubenswrapper[4790]: I1124 14:55:52.081856 4790 scope.go:117] "RemoveContainer" containerID="c1d044acc60c098abe9db072b72ce2adfa61682892f24054ec2fd9a97c4ccef6" Nov 24 14:55:52 crc kubenswrapper[4790]: I1124 14:55:52.132626 4790 scope.go:117] "RemoveContainer" containerID="ef4eb8d324481200a647ad7872c0e232d63e2484a2666846c3c212a4e8c4912e" Nov 24 14:55:52 crc kubenswrapper[4790]: I1124 14:55:52.155107 4790 scope.go:117] "RemoveContainer" containerID="dd6f816e5859dda855280e67acfa0fd8a91290d029ab7b064470cf2a6b471786" Nov 24 14:55:52 crc kubenswrapper[4790]: I1124 14:55:52.181032 4790 scope.go:117] "RemoveContainer" containerID="d330b7dd28854cd58a09d86e5bda959ea7b9d02af5cae00a4bb708ac7457bcf0" Nov 24 14:55:52 crc kubenswrapper[4790]: I1124 14:55:52.326016 4790 scope.go:117] "RemoveContainer" containerID="28308703817a1d4e0e5344ce19268127c600894f2daa6b921d52a9105f014132" Nov 24 14:55:52 crc kubenswrapper[4790]: E1124 14:55:52.326309 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:55:52 crc kubenswrapper[4790]: I1124 14:55:52.331296 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"09da3379-5d69-4618-ad5c-baa983f57405","Type":"ContainerStarted","Data":"ef908f6bac4fceb88ab2051ced1d5abc5692ad7f5ce1f63ad325e4933dd14161"} Nov 24 14:55:55 crc kubenswrapper[4790]: I1124 14:55:55.371170 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"35216e6d-676a-48f1-aeda-e379d7f6ddb4","Type":"ContainerStarted","Data":"84315422a9d02a43f7589d1a8c3f150eea35268cd0514718216809b0d79ae688"} Nov 24 14:55:55 crc kubenswrapper[4790]: I1124 14:55:55.371737 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Nov 24 14:55:55 crc kubenswrapper[4790]: I1124 14:55:55.377442 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Nov 24 14:55:55 crc kubenswrapper[4790]: I1124 14:55:55.401140 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=9.454150818 podStartE2EDuration="27.401122007s" podCreationTimestamp="2025-11-24 14:55:28 +0000 UTC" firstStartedPulling="2025-11-24 14:55:30.682201605 +0000 UTC m=+6179.062095267" lastFinishedPulling="2025-11-24 14:55:48.629172794 +0000 UTC m=+6197.009066456" observedRunningTime="2025-11-24 14:55:55.397019534 +0000 UTC m=+6203.776913226" watchObservedRunningTime="2025-11-24 14:55:55.401122007 +0000 UTC m=+6203.781015669" Nov 24 14:55:57 crc kubenswrapper[4790]: I1124 14:55:57.672026 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"09da3379-5d69-4618-ad5c-baa983f57405","Type":"ContainerStarted","Data":"361f745c8f863c8ffa2be9a8181a1e1164f793c3cf0001ceb3b69588a2e55daf"} Nov 24 14:56:00 crc kubenswrapper[4790]: I1124 14:56:00.702524 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"09da3379-5d69-4618-ad5c-baa983f57405","Type":"ContainerStarted","Data":"5d82460a56a20c6dd25afbd4b32b10f1addcd37312744042e7304f1215549d8c"} Nov 24 14:56:00 crc kubenswrapper[4790]: I1124 14:56:00.737117 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=3.630282081 podStartE2EDuration="32.737095885s" podCreationTimestamp="2025-11-24 14:55:28 +0000 UTC" firstStartedPulling="2025-11-24 14:55:31.37356888 +0000 UTC m=+6179.753462552" lastFinishedPulling="2025-11-24 14:56:00.480382694 +0000 UTC m=+6208.860276356" observedRunningTime="2025-11-24 14:56:00.730632602 +0000 UTC m=+6209.110526294" watchObservedRunningTime="2025-11-24 14:56:00.737095885 +0000 UTC m=+6209.116989547" Nov 24 14:56:05 crc kubenswrapper[4790]: I1124 14:56:05.044415 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Nov 24 14:56:05 crc kubenswrapper[4790]: I1124 14:56:05.316032 4790 scope.go:117] "RemoveContainer" containerID="28308703817a1d4e0e5344ce19268127c600894f2daa6b921d52a9105f014132" Nov 24 14:56:05 crc kubenswrapper[4790]: E1124 14:56:05.316350 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.197752 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:56:08 crc kubenswrapper[4790]: E1124 14:56:08.198543 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33662038-1049-4380-804a-1ce66fd13db7" containerName="registry-server" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.198617 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="33662038-1049-4380-804a-1ce66fd13db7" containerName="registry-server" Nov 24 14:56:08 crc kubenswrapper[4790]: E1124 14:56:08.198632 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fbf1db8-8f67-429c-97c9-e97714898b0e" containerName="extract-content" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.198638 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fbf1db8-8f67-429c-97c9-e97714898b0e" containerName="extract-content" Nov 24 14:56:08 crc kubenswrapper[4790]: E1124 14:56:08.198654 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33662038-1049-4380-804a-1ce66fd13db7" containerName="extract-content" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.198661 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="33662038-1049-4380-804a-1ce66fd13db7" containerName="extract-content" Nov 24 14:56:08 crc kubenswrapper[4790]: E1124 14:56:08.198672 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fbf1db8-8f67-429c-97c9-e97714898b0e" containerName="registry-server" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.198679 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fbf1db8-8f67-429c-97c9-e97714898b0e" containerName="registry-server" Nov 24 14:56:08 crc kubenswrapper[4790]: E1124 14:56:08.198703 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fbf1db8-8f67-429c-97c9-e97714898b0e" containerName="extract-utilities" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.198710 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fbf1db8-8f67-429c-97c9-e97714898b0e" containerName="extract-utilities" Nov 24 14:56:08 crc kubenswrapper[4790]: E1124 14:56:08.198723 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33662038-1049-4380-804a-1ce66fd13db7" containerName="extract-utilities" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.198728 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="33662038-1049-4380-804a-1ce66fd13db7" containerName="extract-utilities" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.198934 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="33662038-1049-4380-804a-1ce66fd13db7" containerName="registry-server" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.198942 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fbf1db8-8f67-429c-97c9-e97714898b0e" containerName="registry-server" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.200703 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.205737 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.207417 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/debad20a-2518-49f1-a1d6-49f5c39d7316-scripts\") pod \"ceilometer-0\" (UID: \"debad20a-2518-49f1-a1d6-49f5c39d7316\") " pod="openstack/ceilometer-0" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.207467 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/debad20a-2518-49f1-a1d6-49f5c39d7316-config-data\") pod \"ceilometer-0\" (UID: \"debad20a-2518-49f1-a1d6-49f5c39d7316\") " pod="openstack/ceilometer-0" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.207575 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/debad20a-2518-49f1-a1d6-49f5c39d7316-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"debad20a-2518-49f1-a1d6-49f5c39d7316\") " pod="openstack/ceilometer-0" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.207623 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/debad20a-2518-49f1-a1d6-49f5c39d7316-run-httpd\") pod \"ceilometer-0\" (UID: \"debad20a-2518-49f1-a1d6-49f5c39d7316\") " pod="openstack/ceilometer-0" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.207661 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/debad20a-2518-49f1-a1d6-49f5c39d7316-log-httpd\") pod \"ceilometer-0\" (UID: \"debad20a-2518-49f1-a1d6-49f5c39d7316\") " pod="openstack/ceilometer-0" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.207685 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7ql7\" (UniqueName: \"kubernetes.io/projected/debad20a-2518-49f1-a1d6-49f5c39d7316-kube-api-access-t7ql7\") pod \"ceilometer-0\" (UID: \"debad20a-2518-49f1-a1d6-49f5c39d7316\") " pod="openstack/ceilometer-0" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.207750 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/debad20a-2518-49f1-a1d6-49f5c39d7316-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"debad20a-2518-49f1-a1d6-49f5c39d7316\") " pod="openstack/ceilometer-0" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.207865 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.213304 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.309299 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/debad20a-2518-49f1-a1d6-49f5c39d7316-scripts\") pod \"ceilometer-0\" (UID: \"debad20a-2518-49f1-a1d6-49f5c39d7316\") " pod="openstack/ceilometer-0" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.309435 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/debad20a-2518-49f1-a1d6-49f5c39d7316-config-data\") pod \"ceilometer-0\" (UID: \"debad20a-2518-49f1-a1d6-49f5c39d7316\") " pod="openstack/ceilometer-0" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.309633 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/debad20a-2518-49f1-a1d6-49f5c39d7316-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"debad20a-2518-49f1-a1d6-49f5c39d7316\") " pod="openstack/ceilometer-0" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.309682 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/debad20a-2518-49f1-a1d6-49f5c39d7316-run-httpd\") pod \"ceilometer-0\" (UID: \"debad20a-2518-49f1-a1d6-49f5c39d7316\") " pod="openstack/ceilometer-0" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.309783 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/debad20a-2518-49f1-a1d6-49f5c39d7316-log-httpd\") pod \"ceilometer-0\" (UID: \"debad20a-2518-49f1-a1d6-49f5c39d7316\") " pod="openstack/ceilometer-0" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.309821 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7ql7\" (UniqueName: \"kubernetes.io/projected/debad20a-2518-49f1-a1d6-49f5c39d7316-kube-api-access-t7ql7\") pod \"ceilometer-0\" (UID: \"debad20a-2518-49f1-a1d6-49f5c39d7316\") " pod="openstack/ceilometer-0" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.309975 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/debad20a-2518-49f1-a1d6-49f5c39d7316-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"debad20a-2518-49f1-a1d6-49f5c39d7316\") " pod="openstack/ceilometer-0" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.310160 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/debad20a-2518-49f1-a1d6-49f5c39d7316-run-httpd\") pod \"ceilometer-0\" (UID: \"debad20a-2518-49f1-a1d6-49f5c39d7316\") " pod="openstack/ceilometer-0" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.310331 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/debad20a-2518-49f1-a1d6-49f5c39d7316-log-httpd\") pod \"ceilometer-0\" (UID: \"debad20a-2518-49f1-a1d6-49f5c39d7316\") " pod="openstack/ceilometer-0" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.315312 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/debad20a-2518-49f1-a1d6-49f5c39d7316-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"debad20a-2518-49f1-a1d6-49f5c39d7316\") " pod="openstack/ceilometer-0" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.315704 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/debad20a-2518-49f1-a1d6-49f5c39d7316-config-data\") pod \"ceilometer-0\" (UID: \"debad20a-2518-49f1-a1d6-49f5c39d7316\") " pod="openstack/ceilometer-0" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.338371 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/debad20a-2518-49f1-a1d6-49f5c39d7316-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"debad20a-2518-49f1-a1d6-49f5c39d7316\") " pod="openstack/ceilometer-0" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.338562 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/debad20a-2518-49f1-a1d6-49f5c39d7316-scripts\") pod \"ceilometer-0\" (UID: \"debad20a-2518-49f1-a1d6-49f5c39d7316\") " pod="openstack/ceilometer-0" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.339838 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7ql7\" (UniqueName: \"kubernetes.io/projected/debad20a-2518-49f1-a1d6-49f5c39d7316-kube-api-access-t7ql7\") pod \"ceilometer-0\" (UID: \"debad20a-2518-49f1-a1d6-49f5c39d7316\") " pod="openstack/ceilometer-0" Nov 24 14:56:08 crc kubenswrapper[4790]: I1124 14:56:08.528497 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:56:09 crc kubenswrapper[4790]: I1124 14:56:09.008006 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:56:09 crc kubenswrapper[4790]: I1124 14:56:09.814140 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"debad20a-2518-49f1-a1d6-49f5c39d7316","Type":"ContainerStarted","Data":"926712c26d54a03dbf190d2dd69a4c2871a2ac13304b21eb29921e964c369938"} Nov 24 14:56:09 crc kubenswrapper[4790]: I1124 14:56:09.814813 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"debad20a-2518-49f1-a1d6-49f5c39d7316","Type":"ContainerStarted","Data":"b9cd64e2e72d1ce70af37458424de602baeeffd6e5a1422026269a8da6b8f6f5"} Nov 24 14:56:11 crc kubenswrapper[4790]: I1124 14:56:11.841136 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"debad20a-2518-49f1-a1d6-49f5c39d7316","Type":"ContainerStarted","Data":"b42d440385b953399bc07de8bdb6974ec06cd9fdc735d457dea6aae49461fd49"} Nov 24 14:56:12 crc kubenswrapper[4790]: I1124 14:56:12.858285 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"debad20a-2518-49f1-a1d6-49f5c39d7316","Type":"ContainerStarted","Data":"9eebc6a326e6ecd25981550fadc6f071d57b5a734d8c279c22add9c347695246"} Nov 24 14:56:14 crc kubenswrapper[4790]: I1124 14:56:14.880507 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"debad20a-2518-49f1-a1d6-49f5c39d7316","Type":"ContainerStarted","Data":"f584a595a9deda7c2c4a69d6b484fe1900aad04db58629cb657e5497cfab3c04"} Nov 24 14:56:14 crc kubenswrapper[4790]: I1124 14:56:14.882180 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 14:56:14 crc kubenswrapper[4790]: I1124 14:56:14.917438 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.274338923 podStartE2EDuration="6.917411428s" podCreationTimestamp="2025-11-24 14:56:08 +0000 UTC" firstStartedPulling="2025-11-24 14:56:09.021934474 +0000 UTC m=+6217.401828156" lastFinishedPulling="2025-11-24 14:56:13.665006999 +0000 UTC m=+6222.044900661" observedRunningTime="2025-11-24 14:56:14.904441801 +0000 UTC m=+6223.284335503" watchObservedRunningTime="2025-11-24 14:56:14.917411428 +0000 UTC m=+6223.297305080" Nov 24 14:56:15 crc kubenswrapper[4790]: I1124 14:56:15.044546 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Nov 24 14:56:15 crc kubenswrapper[4790]: I1124 14:56:15.046615 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Nov 24 14:56:15 crc kubenswrapper[4790]: I1124 14:56:15.891167 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Nov 24 14:56:16 crc kubenswrapper[4790]: I1124 14:56:16.314925 4790 scope.go:117] "RemoveContainer" containerID="28308703817a1d4e0e5344ce19268127c600894f2daa6b921d52a9105f014132" Nov 24 14:56:16 crc kubenswrapper[4790]: E1124 14:56:16.315210 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:56:18 crc kubenswrapper[4790]: I1124 14:56:18.901406 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-jtsf4"] Nov 24 14:56:18 crc kubenswrapper[4790]: I1124 14:56:18.903547 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-jtsf4" Nov 24 14:56:18 crc kubenswrapper[4790]: I1124 14:56:18.920168 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-jtsf4"] Nov 24 14:56:19 crc kubenswrapper[4790]: I1124 14:56:19.012608 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-bcf5-account-create-ktm8m"] Nov 24 14:56:19 crc kubenswrapper[4790]: I1124 14:56:19.014510 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-bcf5-account-create-ktm8m" Nov 24 14:56:19 crc kubenswrapper[4790]: I1124 14:56:19.018011 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Nov 24 14:56:19 crc kubenswrapper[4790]: I1124 14:56:19.024570 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-bcf5-account-create-ktm8m"] Nov 24 14:56:19 crc kubenswrapper[4790]: I1124 14:56:19.047796 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-642xl\" (UniqueName: \"kubernetes.io/projected/f8b24900-d7bb-4af6-9279-4b93f117e15e-kube-api-access-642xl\") pod \"aodh-db-create-jtsf4\" (UID: \"f8b24900-d7bb-4af6-9279-4b93f117e15e\") " pod="openstack/aodh-db-create-jtsf4" Nov 24 14:56:19 crc kubenswrapper[4790]: I1124 14:56:19.047930 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f8b24900-d7bb-4af6-9279-4b93f117e15e-operator-scripts\") pod \"aodh-db-create-jtsf4\" (UID: \"f8b24900-d7bb-4af6-9279-4b93f117e15e\") " pod="openstack/aodh-db-create-jtsf4" Nov 24 14:56:19 crc kubenswrapper[4790]: I1124 14:56:19.150198 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-642xl\" (UniqueName: \"kubernetes.io/projected/f8b24900-d7bb-4af6-9279-4b93f117e15e-kube-api-access-642xl\") pod \"aodh-db-create-jtsf4\" (UID: \"f8b24900-d7bb-4af6-9279-4b93f117e15e\") " pod="openstack/aodh-db-create-jtsf4" Nov 24 14:56:19 crc kubenswrapper[4790]: I1124 14:56:19.150350 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f8b24900-d7bb-4af6-9279-4b93f117e15e-operator-scripts\") pod \"aodh-db-create-jtsf4\" (UID: \"f8b24900-d7bb-4af6-9279-4b93f117e15e\") " pod="openstack/aodh-db-create-jtsf4" Nov 24 14:56:19 crc kubenswrapper[4790]: I1124 14:56:19.150387 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7d7f7\" (UniqueName: \"kubernetes.io/projected/236160ca-0716-4cfe-9ae2-bd706d30cc3c-kube-api-access-7d7f7\") pod \"aodh-bcf5-account-create-ktm8m\" (UID: \"236160ca-0716-4cfe-9ae2-bd706d30cc3c\") " pod="openstack/aodh-bcf5-account-create-ktm8m" Nov 24 14:56:19 crc kubenswrapper[4790]: I1124 14:56:19.150430 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/236160ca-0716-4cfe-9ae2-bd706d30cc3c-operator-scripts\") pod \"aodh-bcf5-account-create-ktm8m\" (UID: \"236160ca-0716-4cfe-9ae2-bd706d30cc3c\") " pod="openstack/aodh-bcf5-account-create-ktm8m" Nov 24 14:56:19 crc kubenswrapper[4790]: I1124 14:56:19.151183 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f8b24900-d7bb-4af6-9279-4b93f117e15e-operator-scripts\") pod \"aodh-db-create-jtsf4\" (UID: \"f8b24900-d7bb-4af6-9279-4b93f117e15e\") " pod="openstack/aodh-db-create-jtsf4" Nov 24 14:56:19 crc kubenswrapper[4790]: I1124 14:56:19.171900 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-642xl\" (UniqueName: \"kubernetes.io/projected/f8b24900-d7bb-4af6-9279-4b93f117e15e-kube-api-access-642xl\") pod \"aodh-db-create-jtsf4\" (UID: \"f8b24900-d7bb-4af6-9279-4b93f117e15e\") " pod="openstack/aodh-db-create-jtsf4" Nov 24 14:56:19 crc kubenswrapper[4790]: I1124 14:56:19.228802 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-jtsf4" Nov 24 14:56:19 crc kubenswrapper[4790]: I1124 14:56:19.252118 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7d7f7\" (UniqueName: \"kubernetes.io/projected/236160ca-0716-4cfe-9ae2-bd706d30cc3c-kube-api-access-7d7f7\") pod \"aodh-bcf5-account-create-ktm8m\" (UID: \"236160ca-0716-4cfe-9ae2-bd706d30cc3c\") " pod="openstack/aodh-bcf5-account-create-ktm8m" Nov 24 14:56:19 crc kubenswrapper[4790]: I1124 14:56:19.252190 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/236160ca-0716-4cfe-9ae2-bd706d30cc3c-operator-scripts\") pod \"aodh-bcf5-account-create-ktm8m\" (UID: \"236160ca-0716-4cfe-9ae2-bd706d30cc3c\") " pod="openstack/aodh-bcf5-account-create-ktm8m" Nov 24 14:56:19 crc kubenswrapper[4790]: I1124 14:56:19.253223 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/236160ca-0716-4cfe-9ae2-bd706d30cc3c-operator-scripts\") pod \"aodh-bcf5-account-create-ktm8m\" (UID: \"236160ca-0716-4cfe-9ae2-bd706d30cc3c\") " pod="openstack/aodh-bcf5-account-create-ktm8m" Nov 24 14:56:19 crc kubenswrapper[4790]: I1124 14:56:19.271752 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7d7f7\" (UniqueName: \"kubernetes.io/projected/236160ca-0716-4cfe-9ae2-bd706d30cc3c-kube-api-access-7d7f7\") pod \"aodh-bcf5-account-create-ktm8m\" (UID: \"236160ca-0716-4cfe-9ae2-bd706d30cc3c\") " pod="openstack/aodh-bcf5-account-create-ktm8m" Nov 24 14:56:19 crc kubenswrapper[4790]: I1124 14:56:19.335008 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-bcf5-account-create-ktm8m" Nov 24 14:56:19 crc kubenswrapper[4790]: I1124 14:56:19.782029 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-jtsf4"] Nov 24 14:56:19 crc kubenswrapper[4790]: I1124 14:56:19.952979 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-jtsf4" event={"ID":"f8b24900-d7bb-4af6-9279-4b93f117e15e","Type":"ContainerStarted","Data":"561aa7d7aff49971b47f10420dfb423870d1e905734ed6e099359d1056c172fd"} Nov 24 14:56:19 crc kubenswrapper[4790]: W1124 14:56:19.961352 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod236160ca_0716_4cfe_9ae2_bd706d30cc3c.slice/crio-b5a002ea1f228193bfa7d5f38bdbdd12765486e802d554459919602985b20b8a WatchSource:0}: Error finding container b5a002ea1f228193bfa7d5f38bdbdd12765486e802d554459919602985b20b8a: Status 404 returned error can't find the container with id b5a002ea1f228193bfa7d5f38bdbdd12765486e802d554459919602985b20b8a Nov 24 14:56:19 crc kubenswrapper[4790]: I1124 14:56:19.974372 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-bcf5-account-create-ktm8m"] Nov 24 14:56:20 crc kubenswrapper[4790]: I1124 14:56:20.964405 4790 generic.go:334] "Generic (PLEG): container finished" podID="f8b24900-d7bb-4af6-9279-4b93f117e15e" containerID="6feeb517f1e72804474985e9df49115d56dc0d1c4c5125f8d6c86fff4b9dddab" exitCode=0 Nov 24 14:56:20 crc kubenswrapper[4790]: I1124 14:56:20.964511 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-jtsf4" event={"ID":"f8b24900-d7bb-4af6-9279-4b93f117e15e","Type":"ContainerDied","Data":"6feeb517f1e72804474985e9df49115d56dc0d1c4c5125f8d6c86fff4b9dddab"} Nov 24 14:56:20 crc kubenswrapper[4790]: I1124 14:56:20.966207 4790 generic.go:334] "Generic (PLEG): container finished" podID="236160ca-0716-4cfe-9ae2-bd706d30cc3c" containerID="d3bb2790edd355c76c55d6b5837eaa5ce4ab7eea196da627ba2299094fa002e2" exitCode=0 Nov 24 14:56:20 crc kubenswrapper[4790]: I1124 14:56:20.966247 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-bcf5-account-create-ktm8m" event={"ID":"236160ca-0716-4cfe-9ae2-bd706d30cc3c","Type":"ContainerDied","Data":"d3bb2790edd355c76c55d6b5837eaa5ce4ab7eea196da627ba2299094fa002e2"} Nov 24 14:56:20 crc kubenswrapper[4790]: I1124 14:56:20.966277 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-bcf5-account-create-ktm8m" event={"ID":"236160ca-0716-4cfe-9ae2-bd706d30cc3c","Type":"ContainerStarted","Data":"b5a002ea1f228193bfa7d5f38bdbdd12765486e802d554459919602985b20b8a"} Nov 24 14:56:22 crc kubenswrapper[4790]: I1124 14:56:22.054268 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-f478-account-create-lpf9g"] Nov 24 14:56:22 crc kubenswrapper[4790]: I1124 14:56:22.066697 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-vf46c"] Nov 24 14:56:22 crc kubenswrapper[4790]: I1124 14:56:22.085019 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-f478-account-create-lpf9g"] Nov 24 14:56:22 crc kubenswrapper[4790]: I1124 14:56:22.111154 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-vf46c"] Nov 24 14:56:22 crc kubenswrapper[4790]: I1124 14:56:22.339754 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b185743-a8f8-43bc-bdd3-c62e5275309e" path="/var/lib/kubelet/pods/9b185743-a8f8-43bc-bdd3-c62e5275309e/volumes" Nov 24 14:56:22 crc kubenswrapper[4790]: I1124 14:56:22.340686 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c65f65d6-1e55-46a4-a310-ae5fa73ae854" path="/var/lib/kubelet/pods/c65f65d6-1e55-46a4-a310-ae5fa73ae854/volumes" Nov 24 14:56:22 crc kubenswrapper[4790]: I1124 14:56:22.448027 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-bcf5-account-create-ktm8m" Nov 24 14:56:22 crc kubenswrapper[4790]: I1124 14:56:22.452027 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-jtsf4" Nov 24 14:56:22 crc kubenswrapper[4790]: I1124 14:56:22.625138 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7d7f7\" (UniqueName: \"kubernetes.io/projected/236160ca-0716-4cfe-9ae2-bd706d30cc3c-kube-api-access-7d7f7\") pod \"236160ca-0716-4cfe-9ae2-bd706d30cc3c\" (UID: \"236160ca-0716-4cfe-9ae2-bd706d30cc3c\") " Nov 24 14:56:22 crc kubenswrapper[4790]: I1124 14:56:22.625289 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-642xl\" (UniqueName: \"kubernetes.io/projected/f8b24900-d7bb-4af6-9279-4b93f117e15e-kube-api-access-642xl\") pod \"f8b24900-d7bb-4af6-9279-4b93f117e15e\" (UID: \"f8b24900-d7bb-4af6-9279-4b93f117e15e\") " Nov 24 14:56:22 crc kubenswrapper[4790]: I1124 14:56:22.625326 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/236160ca-0716-4cfe-9ae2-bd706d30cc3c-operator-scripts\") pod \"236160ca-0716-4cfe-9ae2-bd706d30cc3c\" (UID: \"236160ca-0716-4cfe-9ae2-bd706d30cc3c\") " Nov 24 14:56:22 crc kubenswrapper[4790]: I1124 14:56:22.625403 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f8b24900-d7bb-4af6-9279-4b93f117e15e-operator-scripts\") pod \"f8b24900-d7bb-4af6-9279-4b93f117e15e\" (UID: \"f8b24900-d7bb-4af6-9279-4b93f117e15e\") " Nov 24 14:56:22 crc kubenswrapper[4790]: I1124 14:56:22.625950 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/236160ca-0716-4cfe-9ae2-bd706d30cc3c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "236160ca-0716-4cfe-9ae2-bd706d30cc3c" (UID: "236160ca-0716-4cfe-9ae2-bd706d30cc3c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:56:22 crc kubenswrapper[4790]: I1124 14:56:22.626105 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8b24900-d7bb-4af6-9279-4b93f117e15e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f8b24900-d7bb-4af6-9279-4b93f117e15e" (UID: "f8b24900-d7bb-4af6-9279-4b93f117e15e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:56:22 crc kubenswrapper[4790]: I1124 14:56:22.626122 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/236160ca-0716-4cfe-9ae2-bd706d30cc3c-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:56:22 crc kubenswrapper[4790]: I1124 14:56:22.632211 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/236160ca-0716-4cfe-9ae2-bd706d30cc3c-kube-api-access-7d7f7" (OuterVolumeSpecName: "kube-api-access-7d7f7") pod "236160ca-0716-4cfe-9ae2-bd706d30cc3c" (UID: "236160ca-0716-4cfe-9ae2-bd706d30cc3c"). InnerVolumeSpecName "kube-api-access-7d7f7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:56:22 crc kubenswrapper[4790]: I1124 14:56:22.632406 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8b24900-d7bb-4af6-9279-4b93f117e15e-kube-api-access-642xl" (OuterVolumeSpecName: "kube-api-access-642xl") pod "f8b24900-d7bb-4af6-9279-4b93f117e15e" (UID: "f8b24900-d7bb-4af6-9279-4b93f117e15e"). InnerVolumeSpecName "kube-api-access-642xl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:56:22 crc kubenswrapper[4790]: I1124 14:56:22.728168 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f8b24900-d7bb-4af6-9279-4b93f117e15e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:56:22 crc kubenswrapper[4790]: I1124 14:56:22.728205 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7d7f7\" (UniqueName: \"kubernetes.io/projected/236160ca-0716-4cfe-9ae2-bd706d30cc3c-kube-api-access-7d7f7\") on node \"crc\" DevicePath \"\"" Nov 24 14:56:22 crc kubenswrapper[4790]: I1124 14:56:22.728220 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-642xl\" (UniqueName: \"kubernetes.io/projected/f8b24900-d7bb-4af6-9279-4b93f117e15e-kube-api-access-642xl\") on node \"crc\" DevicePath \"\"" Nov 24 14:56:22 crc kubenswrapper[4790]: I1124 14:56:22.988352 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-jtsf4" Nov 24 14:56:22 crc kubenswrapper[4790]: I1124 14:56:22.988397 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-jtsf4" event={"ID":"f8b24900-d7bb-4af6-9279-4b93f117e15e","Type":"ContainerDied","Data":"561aa7d7aff49971b47f10420dfb423870d1e905734ed6e099359d1056c172fd"} Nov 24 14:56:22 crc kubenswrapper[4790]: I1124 14:56:22.988503 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="561aa7d7aff49971b47f10420dfb423870d1e905734ed6e099359d1056c172fd" Nov 24 14:56:22 crc kubenswrapper[4790]: I1124 14:56:22.990684 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-bcf5-account-create-ktm8m" event={"ID":"236160ca-0716-4cfe-9ae2-bd706d30cc3c","Type":"ContainerDied","Data":"b5a002ea1f228193bfa7d5f38bdbdd12765486e802d554459919602985b20b8a"} Nov 24 14:56:22 crc kubenswrapper[4790]: I1124 14:56:22.990726 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b5a002ea1f228193bfa7d5f38bdbdd12765486e802d554459919602985b20b8a" Nov 24 14:56:22 crc kubenswrapper[4790]: I1124 14:56:22.990749 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-bcf5-account-create-ktm8m" Nov 24 14:56:24 crc kubenswrapper[4790]: I1124 14:56:24.375240 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-qjttx"] Nov 24 14:56:24 crc kubenswrapper[4790]: E1124 14:56:24.375989 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8b24900-d7bb-4af6-9279-4b93f117e15e" containerName="mariadb-database-create" Nov 24 14:56:24 crc kubenswrapper[4790]: I1124 14:56:24.376006 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8b24900-d7bb-4af6-9279-4b93f117e15e" containerName="mariadb-database-create" Nov 24 14:56:24 crc kubenswrapper[4790]: E1124 14:56:24.376035 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="236160ca-0716-4cfe-9ae2-bd706d30cc3c" containerName="mariadb-account-create" Nov 24 14:56:24 crc kubenswrapper[4790]: I1124 14:56:24.376043 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="236160ca-0716-4cfe-9ae2-bd706d30cc3c" containerName="mariadb-account-create" Nov 24 14:56:24 crc kubenswrapper[4790]: I1124 14:56:24.376306 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="236160ca-0716-4cfe-9ae2-bd706d30cc3c" containerName="mariadb-account-create" Nov 24 14:56:24 crc kubenswrapper[4790]: I1124 14:56:24.376325 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8b24900-d7bb-4af6-9279-4b93f117e15e" containerName="mariadb-database-create" Nov 24 14:56:24 crc kubenswrapper[4790]: I1124 14:56:24.377239 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-qjttx" Nov 24 14:56:24 crc kubenswrapper[4790]: I1124 14:56:24.379671 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 24 14:56:24 crc kubenswrapper[4790]: I1124 14:56:24.379720 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 24 14:56:24 crc kubenswrapper[4790]: I1124 14:56:24.379852 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-77ld6" Nov 24 14:56:24 crc kubenswrapper[4790]: I1124 14:56:24.383540 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 14:56:24 crc kubenswrapper[4790]: I1124 14:56:24.387504 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-qjttx"] Nov 24 14:56:24 crc kubenswrapper[4790]: I1124 14:56:24.573978 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzn6f\" (UniqueName: \"kubernetes.io/projected/df033470-00fb-45ac-920a-fa5d75d69d73-kube-api-access-rzn6f\") pod \"aodh-db-sync-qjttx\" (UID: \"df033470-00fb-45ac-920a-fa5d75d69d73\") " pod="openstack/aodh-db-sync-qjttx" Nov 24 14:56:24 crc kubenswrapper[4790]: I1124 14:56:24.574054 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df033470-00fb-45ac-920a-fa5d75d69d73-combined-ca-bundle\") pod \"aodh-db-sync-qjttx\" (UID: \"df033470-00fb-45ac-920a-fa5d75d69d73\") " pod="openstack/aodh-db-sync-qjttx" Nov 24 14:56:24 crc kubenswrapper[4790]: I1124 14:56:24.574312 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df033470-00fb-45ac-920a-fa5d75d69d73-config-data\") pod \"aodh-db-sync-qjttx\" (UID: \"df033470-00fb-45ac-920a-fa5d75d69d73\") " pod="openstack/aodh-db-sync-qjttx" Nov 24 14:56:24 crc kubenswrapper[4790]: I1124 14:56:24.574445 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df033470-00fb-45ac-920a-fa5d75d69d73-scripts\") pod \"aodh-db-sync-qjttx\" (UID: \"df033470-00fb-45ac-920a-fa5d75d69d73\") " pod="openstack/aodh-db-sync-qjttx" Nov 24 14:56:24 crc kubenswrapper[4790]: I1124 14:56:24.676930 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df033470-00fb-45ac-920a-fa5d75d69d73-combined-ca-bundle\") pod \"aodh-db-sync-qjttx\" (UID: \"df033470-00fb-45ac-920a-fa5d75d69d73\") " pod="openstack/aodh-db-sync-qjttx" Nov 24 14:56:24 crc kubenswrapper[4790]: I1124 14:56:24.677336 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df033470-00fb-45ac-920a-fa5d75d69d73-config-data\") pod \"aodh-db-sync-qjttx\" (UID: \"df033470-00fb-45ac-920a-fa5d75d69d73\") " pod="openstack/aodh-db-sync-qjttx" Nov 24 14:56:24 crc kubenswrapper[4790]: I1124 14:56:24.677494 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df033470-00fb-45ac-920a-fa5d75d69d73-scripts\") pod \"aodh-db-sync-qjttx\" (UID: \"df033470-00fb-45ac-920a-fa5d75d69d73\") " pod="openstack/aodh-db-sync-qjttx" Nov 24 14:56:24 crc kubenswrapper[4790]: I1124 14:56:24.677647 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzn6f\" (UniqueName: \"kubernetes.io/projected/df033470-00fb-45ac-920a-fa5d75d69d73-kube-api-access-rzn6f\") pod \"aodh-db-sync-qjttx\" (UID: \"df033470-00fb-45ac-920a-fa5d75d69d73\") " pod="openstack/aodh-db-sync-qjttx" Nov 24 14:56:24 crc kubenswrapper[4790]: I1124 14:56:24.684857 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df033470-00fb-45ac-920a-fa5d75d69d73-config-data\") pod \"aodh-db-sync-qjttx\" (UID: \"df033470-00fb-45ac-920a-fa5d75d69d73\") " pod="openstack/aodh-db-sync-qjttx" Nov 24 14:56:24 crc kubenswrapper[4790]: I1124 14:56:24.685151 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df033470-00fb-45ac-920a-fa5d75d69d73-combined-ca-bundle\") pod \"aodh-db-sync-qjttx\" (UID: \"df033470-00fb-45ac-920a-fa5d75d69d73\") " pod="openstack/aodh-db-sync-qjttx" Nov 24 14:56:24 crc kubenswrapper[4790]: I1124 14:56:24.689272 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df033470-00fb-45ac-920a-fa5d75d69d73-scripts\") pod \"aodh-db-sync-qjttx\" (UID: \"df033470-00fb-45ac-920a-fa5d75d69d73\") " pod="openstack/aodh-db-sync-qjttx" Nov 24 14:56:24 crc kubenswrapper[4790]: I1124 14:56:24.694111 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzn6f\" (UniqueName: \"kubernetes.io/projected/df033470-00fb-45ac-920a-fa5d75d69d73-kube-api-access-rzn6f\") pod \"aodh-db-sync-qjttx\" (UID: \"df033470-00fb-45ac-920a-fa5d75d69d73\") " pod="openstack/aodh-db-sync-qjttx" Nov 24 14:56:24 crc kubenswrapper[4790]: I1124 14:56:24.696820 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-qjttx" Nov 24 14:56:25 crc kubenswrapper[4790]: I1124 14:56:25.224282 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-qjttx"] Nov 24 14:56:26 crc kubenswrapper[4790]: I1124 14:56:26.020723 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-qjttx" event={"ID":"df033470-00fb-45ac-920a-fa5d75d69d73","Type":"ContainerStarted","Data":"ab897be8dc8c5b6202500f0522aa19b5ce7c8f50827a5a86968bbb32c23291ec"} Nov 24 14:56:30 crc kubenswrapper[4790]: I1124 14:56:30.058843 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-tjhvv"] Nov 24 14:56:30 crc kubenswrapper[4790]: I1124 14:56:30.074864 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-tjhvv"] Nov 24 14:56:30 crc kubenswrapper[4790]: I1124 14:56:30.102766 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-qjttx" event={"ID":"df033470-00fb-45ac-920a-fa5d75d69d73","Type":"ContainerStarted","Data":"0f55a96bfe303125f6f43218805e6ec19ad7633c4a6b92a1d332475a7f046501"} Nov 24 14:56:30 crc kubenswrapper[4790]: I1124 14:56:30.123316 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-qjttx" podStartSLOduration=2.152493502 podStartE2EDuration="6.123300344s" podCreationTimestamp="2025-11-24 14:56:24 +0000 UTC" firstStartedPulling="2025-11-24 14:56:25.232549506 +0000 UTC m=+6233.612443178" lastFinishedPulling="2025-11-24 14:56:29.203356368 +0000 UTC m=+6237.583250020" observedRunningTime="2025-11-24 14:56:30.120510244 +0000 UTC m=+6238.500403906" watchObservedRunningTime="2025-11-24 14:56:30.123300344 +0000 UTC m=+6238.503194006" Nov 24 14:56:30 crc kubenswrapper[4790]: I1124 14:56:30.327193 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f89e3f88-652b-4467-9a73-dbcf6514e674" path="/var/lib/kubelet/pods/f89e3f88-652b-4467-9a73-dbcf6514e674/volumes" Nov 24 14:56:31 crc kubenswrapper[4790]: I1124 14:56:31.315526 4790 scope.go:117] "RemoveContainer" containerID="28308703817a1d4e0e5344ce19268127c600894f2daa6b921d52a9105f014132" Nov 24 14:56:31 crc kubenswrapper[4790]: E1124 14:56:31.316300 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:56:32 crc kubenswrapper[4790]: I1124 14:56:32.128574 4790 generic.go:334] "Generic (PLEG): container finished" podID="df033470-00fb-45ac-920a-fa5d75d69d73" containerID="0f55a96bfe303125f6f43218805e6ec19ad7633c4a6b92a1d332475a7f046501" exitCode=0 Nov 24 14:56:32 crc kubenswrapper[4790]: I1124 14:56:32.128676 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-qjttx" event={"ID":"df033470-00fb-45ac-920a-fa5d75d69d73","Type":"ContainerDied","Data":"0f55a96bfe303125f6f43218805e6ec19ad7633c4a6b92a1d332475a7f046501"} Nov 24 14:56:33 crc kubenswrapper[4790]: I1124 14:56:33.475087 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-qjttx" Nov 24 14:56:33 crc kubenswrapper[4790]: I1124 14:56:33.671406 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rzn6f\" (UniqueName: \"kubernetes.io/projected/df033470-00fb-45ac-920a-fa5d75d69d73-kube-api-access-rzn6f\") pod \"df033470-00fb-45ac-920a-fa5d75d69d73\" (UID: \"df033470-00fb-45ac-920a-fa5d75d69d73\") " Nov 24 14:56:33 crc kubenswrapper[4790]: I1124 14:56:33.671521 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df033470-00fb-45ac-920a-fa5d75d69d73-scripts\") pod \"df033470-00fb-45ac-920a-fa5d75d69d73\" (UID: \"df033470-00fb-45ac-920a-fa5d75d69d73\") " Nov 24 14:56:33 crc kubenswrapper[4790]: I1124 14:56:33.671618 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df033470-00fb-45ac-920a-fa5d75d69d73-config-data\") pod \"df033470-00fb-45ac-920a-fa5d75d69d73\" (UID: \"df033470-00fb-45ac-920a-fa5d75d69d73\") " Nov 24 14:56:33 crc kubenswrapper[4790]: I1124 14:56:33.671661 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df033470-00fb-45ac-920a-fa5d75d69d73-combined-ca-bundle\") pod \"df033470-00fb-45ac-920a-fa5d75d69d73\" (UID: \"df033470-00fb-45ac-920a-fa5d75d69d73\") " Nov 24 14:56:33 crc kubenswrapper[4790]: I1124 14:56:33.677117 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df033470-00fb-45ac-920a-fa5d75d69d73-kube-api-access-rzn6f" (OuterVolumeSpecName: "kube-api-access-rzn6f") pod "df033470-00fb-45ac-920a-fa5d75d69d73" (UID: "df033470-00fb-45ac-920a-fa5d75d69d73"). InnerVolumeSpecName "kube-api-access-rzn6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:56:33 crc kubenswrapper[4790]: I1124 14:56:33.678465 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df033470-00fb-45ac-920a-fa5d75d69d73-scripts" (OuterVolumeSpecName: "scripts") pod "df033470-00fb-45ac-920a-fa5d75d69d73" (UID: "df033470-00fb-45ac-920a-fa5d75d69d73"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:56:33 crc kubenswrapper[4790]: I1124 14:56:33.702793 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df033470-00fb-45ac-920a-fa5d75d69d73-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "df033470-00fb-45ac-920a-fa5d75d69d73" (UID: "df033470-00fb-45ac-920a-fa5d75d69d73"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:56:33 crc kubenswrapper[4790]: I1124 14:56:33.709185 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df033470-00fb-45ac-920a-fa5d75d69d73-config-data" (OuterVolumeSpecName: "config-data") pod "df033470-00fb-45ac-920a-fa5d75d69d73" (UID: "df033470-00fb-45ac-920a-fa5d75d69d73"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:56:33 crc kubenswrapper[4790]: I1124 14:56:33.775755 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df033470-00fb-45ac-920a-fa5d75d69d73-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:56:33 crc kubenswrapper[4790]: I1124 14:56:33.775797 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df033470-00fb-45ac-920a-fa5d75d69d73-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:56:33 crc kubenswrapper[4790]: I1124 14:56:33.775810 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rzn6f\" (UniqueName: \"kubernetes.io/projected/df033470-00fb-45ac-920a-fa5d75d69d73-kube-api-access-rzn6f\") on node \"crc\" DevicePath \"\"" Nov 24 14:56:33 crc kubenswrapper[4790]: I1124 14:56:33.775821 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df033470-00fb-45ac-920a-fa5d75d69d73-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:56:34 crc kubenswrapper[4790]: I1124 14:56:34.151723 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-qjttx" event={"ID":"df033470-00fb-45ac-920a-fa5d75d69d73","Type":"ContainerDied","Data":"ab897be8dc8c5b6202500f0522aa19b5ce7c8f50827a5a86968bbb32c23291ec"} Nov 24 14:56:34 crc kubenswrapper[4790]: I1124 14:56:34.151778 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab897be8dc8c5b6202500f0522aa19b5ce7c8f50827a5a86968bbb32c23291ec" Nov 24 14:56:34 crc kubenswrapper[4790]: I1124 14:56:34.151837 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-qjttx" Nov 24 14:56:34 crc kubenswrapper[4790]: I1124 14:56:34.487791 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Nov 24 14:56:34 crc kubenswrapper[4790]: E1124 14:56:34.488315 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df033470-00fb-45ac-920a-fa5d75d69d73" containerName="aodh-db-sync" Nov 24 14:56:34 crc kubenswrapper[4790]: I1124 14:56:34.488331 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="df033470-00fb-45ac-920a-fa5d75d69d73" containerName="aodh-db-sync" Nov 24 14:56:34 crc kubenswrapper[4790]: I1124 14:56:34.488629 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="df033470-00fb-45ac-920a-fa5d75d69d73" containerName="aodh-db-sync" Nov 24 14:56:34 crc kubenswrapper[4790]: I1124 14:56:34.491117 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 24 14:56:34 crc kubenswrapper[4790]: I1124 14:56:34.497402 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 24 14:56:34 crc kubenswrapper[4790]: I1124 14:56:34.498035 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 24 14:56:34 crc kubenswrapper[4790]: I1124 14:56:34.498249 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-77ld6" Nov 24 14:56:34 crc kubenswrapper[4790]: I1124 14:56:34.527296 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 24 14:56:34 crc kubenswrapper[4790]: I1124 14:56:34.693642 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n96mg\" (UniqueName: \"kubernetes.io/projected/d0c3fa90-a58b-4a63-836a-35744ec234df-kube-api-access-n96mg\") pod \"aodh-0\" (UID: \"d0c3fa90-a58b-4a63-836a-35744ec234df\") " pod="openstack/aodh-0" Nov 24 14:56:34 crc kubenswrapper[4790]: I1124 14:56:34.693831 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0c3fa90-a58b-4a63-836a-35744ec234df-config-data\") pod \"aodh-0\" (UID: \"d0c3fa90-a58b-4a63-836a-35744ec234df\") " pod="openstack/aodh-0" Nov 24 14:56:34 crc kubenswrapper[4790]: I1124 14:56:34.693927 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0c3fa90-a58b-4a63-836a-35744ec234df-combined-ca-bundle\") pod \"aodh-0\" (UID: \"d0c3fa90-a58b-4a63-836a-35744ec234df\") " pod="openstack/aodh-0" Nov 24 14:56:34 crc kubenswrapper[4790]: I1124 14:56:34.694053 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0c3fa90-a58b-4a63-836a-35744ec234df-scripts\") pod \"aodh-0\" (UID: \"d0c3fa90-a58b-4a63-836a-35744ec234df\") " pod="openstack/aodh-0" Nov 24 14:56:34 crc kubenswrapper[4790]: I1124 14:56:34.795410 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n96mg\" (UniqueName: \"kubernetes.io/projected/d0c3fa90-a58b-4a63-836a-35744ec234df-kube-api-access-n96mg\") pod \"aodh-0\" (UID: \"d0c3fa90-a58b-4a63-836a-35744ec234df\") " pod="openstack/aodh-0" Nov 24 14:56:34 crc kubenswrapper[4790]: I1124 14:56:34.795524 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0c3fa90-a58b-4a63-836a-35744ec234df-config-data\") pod \"aodh-0\" (UID: \"d0c3fa90-a58b-4a63-836a-35744ec234df\") " pod="openstack/aodh-0" Nov 24 14:56:34 crc kubenswrapper[4790]: I1124 14:56:34.795563 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0c3fa90-a58b-4a63-836a-35744ec234df-combined-ca-bundle\") pod \"aodh-0\" (UID: \"d0c3fa90-a58b-4a63-836a-35744ec234df\") " pod="openstack/aodh-0" Nov 24 14:56:34 crc kubenswrapper[4790]: I1124 14:56:34.795625 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0c3fa90-a58b-4a63-836a-35744ec234df-scripts\") pod \"aodh-0\" (UID: \"d0c3fa90-a58b-4a63-836a-35744ec234df\") " pod="openstack/aodh-0" Nov 24 14:56:34 crc kubenswrapper[4790]: I1124 14:56:34.799586 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0c3fa90-a58b-4a63-836a-35744ec234df-combined-ca-bundle\") pod \"aodh-0\" (UID: \"d0c3fa90-a58b-4a63-836a-35744ec234df\") " pod="openstack/aodh-0" Nov 24 14:56:34 crc kubenswrapper[4790]: I1124 14:56:34.800144 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0c3fa90-a58b-4a63-836a-35744ec234df-config-data\") pod \"aodh-0\" (UID: \"d0c3fa90-a58b-4a63-836a-35744ec234df\") " pod="openstack/aodh-0" Nov 24 14:56:34 crc kubenswrapper[4790]: I1124 14:56:34.802381 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0c3fa90-a58b-4a63-836a-35744ec234df-scripts\") pod \"aodh-0\" (UID: \"d0c3fa90-a58b-4a63-836a-35744ec234df\") " pod="openstack/aodh-0" Nov 24 14:56:34 crc kubenswrapper[4790]: I1124 14:56:34.819922 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n96mg\" (UniqueName: \"kubernetes.io/projected/d0c3fa90-a58b-4a63-836a-35744ec234df-kube-api-access-n96mg\") pod \"aodh-0\" (UID: \"d0c3fa90-a58b-4a63-836a-35744ec234df\") " pod="openstack/aodh-0" Nov 24 14:56:35 crc kubenswrapper[4790]: I1124 14:56:35.112959 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 24 14:56:35 crc kubenswrapper[4790]: I1124 14:56:35.674985 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 24 14:56:36 crc kubenswrapper[4790]: I1124 14:56:36.176671 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d0c3fa90-a58b-4a63-836a-35744ec234df","Type":"ContainerStarted","Data":"956855b5b2a0868579352a68406d8385a1f08da4ff597ca4d1346e210212c904"} Nov 24 14:56:36 crc kubenswrapper[4790]: I1124 14:56:36.821373 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:56:36 crc kubenswrapper[4790]: I1124 14:56:36.822028 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="debad20a-2518-49f1-a1d6-49f5c39d7316" containerName="ceilometer-central-agent" containerID="cri-o://926712c26d54a03dbf190d2dd69a4c2871a2ac13304b21eb29921e964c369938" gracePeriod=30 Nov 24 14:56:36 crc kubenswrapper[4790]: I1124 14:56:36.822083 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="debad20a-2518-49f1-a1d6-49f5c39d7316" containerName="sg-core" containerID="cri-o://9eebc6a326e6ecd25981550fadc6f071d57b5a734d8c279c22add9c347695246" gracePeriod=30 Nov 24 14:56:36 crc kubenswrapper[4790]: I1124 14:56:36.822158 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="debad20a-2518-49f1-a1d6-49f5c39d7316" containerName="proxy-httpd" containerID="cri-o://f584a595a9deda7c2c4a69d6b484fe1900aad04db58629cb657e5497cfab3c04" gracePeriod=30 Nov 24 14:56:36 crc kubenswrapper[4790]: I1124 14:56:36.822177 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="debad20a-2518-49f1-a1d6-49f5c39d7316" containerName="ceilometer-notification-agent" containerID="cri-o://b42d440385b953399bc07de8bdb6974ec06cd9fdc735d457dea6aae49461fd49" gracePeriod=30 Nov 24 14:56:36 crc kubenswrapper[4790]: I1124 14:56:36.832105 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="debad20a-2518-49f1-a1d6-49f5c39d7316" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.1.135:3000/\": EOF" Nov 24 14:56:37 crc kubenswrapper[4790]: I1124 14:56:37.195697 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d0c3fa90-a58b-4a63-836a-35744ec234df","Type":"ContainerStarted","Data":"374880d30bfe0f34714a716193b6144be3607ef746aa94cc363d4f822c3c57eb"} Nov 24 14:56:37 crc kubenswrapper[4790]: I1124 14:56:37.198595 4790 generic.go:334] "Generic (PLEG): container finished" podID="debad20a-2518-49f1-a1d6-49f5c39d7316" containerID="f584a595a9deda7c2c4a69d6b484fe1900aad04db58629cb657e5497cfab3c04" exitCode=0 Nov 24 14:56:37 crc kubenswrapper[4790]: I1124 14:56:37.198622 4790 generic.go:334] "Generic (PLEG): container finished" podID="debad20a-2518-49f1-a1d6-49f5c39d7316" containerID="9eebc6a326e6ecd25981550fadc6f071d57b5a734d8c279c22add9c347695246" exitCode=2 Nov 24 14:56:37 crc kubenswrapper[4790]: I1124 14:56:37.198641 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"debad20a-2518-49f1-a1d6-49f5c39d7316","Type":"ContainerDied","Data":"f584a595a9deda7c2c4a69d6b484fe1900aad04db58629cb657e5497cfab3c04"} Nov 24 14:56:37 crc kubenswrapper[4790]: I1124 14:56:37.198660 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"debad20a-2518-49f1-a1d6-49f5c39d7316","Type":"ContainerDied","Data":"9eebc6a326e6ecd25981550fadc6f071d57b5a734d8c279c22add9c347695246"} Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.052229 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.171515 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/debad20a-2518-49f1-a1d6-49f5c39d7316-run-httpd\") pod \"debad20a-2518-49f1-a1d6-49f5c39d7316\" (UID: \"debad20a-2518-49f1-a1d6-49f5c39d7316\") " Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.171733 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/debad20a-2518-49f1-a1d6-49f5c39d7316-log-httpd\") pod \"debad20a-2518-49f1-a1d6-49f5c39d7316\" (UID: \"debad20a-2518-49f1-a1d6-49f5c39d7316\") " Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.171791 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/debad20a-2518-49f1-a1d6-49f5c39d7316-combined-ca-bundle\") pod \"debad20a-2518-49f1-a1d6-49f5c39d7316\" (UID: \"debad20a-2518-49f1-a1d6-49f5c39d7316\") " Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.171816 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7ql7\" (UniqueName: \"kubernetes.io/projected/debad20a-2518-49f1-a1d6-49f5c39d7316-kube-api-access-t7ql7\") pod \"debad20a-2518-49f1-a1d6-49f5c39d7316\" (UID: \"debad20a-2518-49f1-a1d6-49f5c39d7316\") " Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.171864 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/debad20a-2518-49f1-a1d6-49f5c39d7316-config-data\") pod \"debad20a-2518-49f1-a1d6-49f5c39d7316\" (UID: \"debad20a-2518-49f1-a1d6-49f5c39d7316\") " Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.171935 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/debad20a-2518-49f1-a1d6-49f5c39d7316-sg-core-conf-yaml\") pod \"debad20a-2518-49f1-a1d6-49f5c39d7316\" (UID: \"debad20a-2518-49f1-a1d6-49f5c39d7316\") " Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.172014 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/debad20a-2518-49f1-a1d6-49f5c39d7316-scripts\") pod \"debad20a-2518-49f1-a1d6-49f5c39d7316\" (UID: \"debad20a-2518-49f1-a1d6-49f5c39d7316\") " Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.174492 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/debad20a-2518-49f1-a1d6-49f5c39d7316-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "debad20a-2518-49f1-a1d6-49f5c39d7316" (UID: "debad20a-2518-49f1-a1d6-49f5c39d7316"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.174955 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/debad20a-2518-49f1-a1d6-49f5c39d7316-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "debad20a-2518-49f1-a1d6-49f5c39d7316" (UID: "debad20a-2518-49f1-a1d6-49f5c39d7316"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.176605 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/debad20a-2518-49f1-a1d6-49f5c39d7316-scripts" (OuterVolumeSpecName: "scripts") pod "debad20a-2518-49f1-a1d6-49f5c39d7316" (UID: "debad20a-2518-49f1-a1d6-49f5c39d7316"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.177363 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/debad20a-2518-49f1-a1d6-49f5c39d7316-kube-api-access-t7ql7" (OuterVolumeSpecName: "kube-api-access-t7ql7") pod "debad20a-2518-49f1-a1d6-49f5c39d7316" (UID: "debad20a-2518-49f1-a1d6-49f5c39d7316"). InnerVolumeSpecName "kube-api-access-t7ql7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.212695 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d0c3fa90-a58b-4a63-836a-35744ec234df","Type":"ContainerStarted","Data":"561c7e3af013da8c1e30d490c48d6a896d001f099028def59e44313dcd62e431"} Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.213288 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/debad20a-2518-49f1-a1d6-49f5c39d7316-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "debad20a-2518-49f1-a1d6-49f5c39d7316" (UID: "debad20a-2518-49f1-a1d6-49f5c39d7316"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.215984 4790 generic.go:334] "Generic (PLEG): container finished" podID="debad20a-2518-49f1-a1d6-49f5c39d7316" containerID="b42d440385b953399bc07de8bdb6974ec06cd9fdc735d457dea6aae49461fd49" exitCode=0 Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.216242 4790 generic.go:334] "Generic (PLEG): container finished" podID="debad20a-2518-49f1-a1d6-49f5c39d7316" containerID="926712c26d54a03dbf190d2dd69a4c2871a2ac13304b21eb29921e964c369938" exitCode=0 Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.216216 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"debad20a-2518-49f1-a1d6-49f5c39d7316","Type":"ContainerDied","Data":"b42d440385b953399bc07de8bdb6974ec06cd9fdc735d457dea6aae49461fd49"} Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.216423 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"debad20a-2518-49f1-a1d6-49f5c39d7316","Type":"ContainerDied","Data":"926712c26d54a03dbf190d2dd69a4c2871a2ac13304b21eb29921e964c369938"} Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.216487 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"debad20a-2518-49f1-a1d6-49f5c39d7316","Type":"ContainerDied","Data":"b9cd64e2e72d1ce70af37458424de602baeeffd6e5a1422026269a8da6b8f6f5"} Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.216553 4790 scope.go:117] "RemoveContainer" containerID="f584a595a9deda7c2c4a69d6b484fe1900aad04db58629cb657e5497cfab3c04" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.216200 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.242183 4790 scope.go:117] "RemoveContainer" containerID="9eebc6a326e6ecd25981550fadc6f071d57b5a734d8c279c22add9c347695246" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.261963 4790 scope.go:117] "RemoveContainer" containerID="b42d440385b953399bc07de8bdb6974ec06cd9fdc735d457dea6aae49461fd49" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.266492 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/debad20a-2518-49f1-a1d6-49f5c39d7316-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "debad20a-2518-49f1-a1d6-49f5c39d7316" (UID: "debad20a-2518-49f1-a1d6-49f5c39d7316"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.274384 4790 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/debad20a-2518-49f1-a1d6-49f5c39d7316-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.274576 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/debad20a-2518-49f1-a1d6-49f5c39d7316-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.274600 4790 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/debad20a-2518-49f1-a1d6-49f5c39d7316-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.274629 4790 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/debad20a-2518-49f1-a1d6-49f5c39d7316-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.274640 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/debad20a-2518-49f1-a1d6-49f5c39d7316-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.274665 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7ql7\" (UniqueName: \"kubernetes.io/projected/debad20a-2518-49f1-a1d6-49f5c39d7316-kube-api-access-t7ql7\") on node \"crc\" DevicePath \"\"" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.287489 4790 scope.go:117] "RemoveContainer" containerID="926712c26d54a03dbf190d2dd69a4c2871a2ac13304b21eb29921e964c369938" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.298611 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/debad20a-2518-49f1-a1d6-49f5c39d7316-config-data" (OuterVolumeSpecName: "config-data") pod "debad20a-2518-49f1-a1d6-49f5c39d7316" (UID: "debad20a-2518-49f1-a1d6-49f5c39d7316"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.309155 4790 scope.go:117] "RemoveContainer" containerID="f584a595a9deda7c2c4a69d6b484fe1900aad04db58629cb657e5497cfab3c04" Nov 24 14:56:38 crc kubenswrapper[4790]: E1124 14:56:38.309706 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f584a595a9deda7c2c4a69d6b484fe1900aad04db58629cb657e5497cfab3c04\": container with ID starting with f584a595a9deda7c2c4a69d6b484fe1900aad04db58629cb657e5497cfab3c04 not found: ID does not exist" containerID="f584a595a9deda7c2c4a69d6b484fe1900aad04db58629cb657e5497cfab3c04" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.309745 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f584a595a9deda7c2c4a69d6b484fe1900aad04db58629cb657e5497cfab3c04"} err="failed to get container status \"f584a595a9deda7c2c4a69d6b484fe1900aad04db58629cb657e5497cfab3c04\": rpc error: code = NotFound desc = could not find container \"f584a595a9deda7c2c4a69d6b484fe1900aad04db58629cb657e5497cfab3c04\": container with ID starting with f584a595a9deda7c2c4a69d6b484fe1900aad04db58629cb657e5497cfab3c04 not found: ID does not exist" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.309769 4790 scope.go:117] "RemoveContainer" containerID="9eebc6a326e6ecd25981550fadc6f071d57b5a734d8c279c22add9c347695246" Nov 24 14:56:38 crc kubenswrapper[4790]: E1124 14:56:38.310005 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9eebc6a326e6ecd25981550fadc6f071d57b5a734d8c279c22add9c347695246\": container with ID starting with 9eebc6a326e6ecd25981550fadc6f071d57b5a734d8c279c22add9c347695246 not found: ID does not exist" containerID="9eebc6a326e6ecd25981550fadc6f071d57b5a734d8c279c22add9c347695246" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.310032 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9eebc6a326e6ecd25981550fadc6f071d57b5a734d8c279c22add9c347695246"} err="failed to get container status \"9eebc6a326e6ecd25981550fadc6f071d57b5a734d8c279c22add9c347695246\": rpc error: code = NotFound desc = could not find container \"9eebc6a326e6ecd25981550fadc6f071d57b5a734d8c279c22add9c347695246\": container with ID starting with 9eebc6a326e6ecd25981550fadc6f071d57b5a734d8c279c22add9c347695246 not found: ID does not exist" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.310050 4790 scope.go:117] "RemoveContainer" containerID="b42d440385b953399bc07de8bdb6974ec06cd9fdc735d457dea6aae49461fd49" Nov 24 14:56:38 crc kubenswrapper[4790]: E1124 14:56:38.310283 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b42d440385b953399bc07de8bdb6974ec06cd9fdc735d457dea6aae49461fd49\": container with ID starting with b42d440385b953399bc07de8bdb6974ec06cd9fdc735d457dea6aae49461fd49 not found: ID does not exist" containerID="b42d440385b953399bc07de8bdb6974ec06cd9fdc735d457dea6aae49461fd49" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.310305 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b42d440385b953399bc07de8bdb6974ec06cd9fdc735d457dea6aae49461fd49"} err="failed to get container status \"b42d440385b953399bc07de8bdb6974ec06cd9fdc735d457dea6aae49461fd49\": rpc error: code = NotFound desc = could not find container \"b42d440385b953399bc07de8bdb6974ec06cd9fdc735d457dea6aae49461fd49\": container with ID starting with b42d440385b953399bc07de8bdb6974ec06cd9fdc735d457dea6aae49461fd49 not found: ID does not exist" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.310317 4790 scope.go:117] "RemoveContainer" containerID="926712c26d54a03dbf190d2dd69a4c2871a2ac13304b21eb29921e964c369938" Nov 24 14:56:38 crc kubenswrapper[4790]: E1124 14:56:38.320095 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"926712c26d54a03dbf190d2dd69a4c2871a2ac13304b21eb29921e964c369938\": container with ID starting with 926712c26d54a03dbf190d2dd69a4c2871a2ac13304b21eb29921e964c369938 not found: ID does not exist" containerID="926712c26d54a03dbf190d2dd69a4c2871a2ac13304b21eb29921e964c369938" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.320153 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"926712c26d54a03dbf190d2dd69a4c2871a2ac13304b21eb29921e964c369938"} err="failed to get container status \"926712c26d54a03dbf190d2dd69a4c2871a2ac13304b21eb29921e964c369938\": rpc error: code = NotFound desc = could not find container \"926712c26d54a03dbf190d2dd69a4c2871a2ac13304b21eb29921e964c369938\": container with ID starting with 926712c26d54a03dbf190d2dd69a4c2871a2ac13304b21eb29921e964c369938 not found: ID does not exist" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.320193 4790 scope.go:117] "RemoveContainer" containerID="f584a595a9deda7c2c4a69d6b484fe1900aad04db58629cb657e5497cfab3c04" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.321220 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f584a595a9deda7c2c4a69d6b484fe1900aad04db58629cb657e5497cfab3c04"} err="failed to get container status \"f584a595a9deda7c2c4a69d6b484fe1900aad04db58629cb657e5497cfab3c04\": rpc error: code = NotFound desc = could not find container \"f584a595a9deda7c2c4a69d6b484fe1900aad04db58629cb657e5497cfab3c04\": container with ID starting with f584a595a9deda7c2c4a69d6b484fe1900aad04db58629cb657e5497cfab3c04 not found: ID does not exist" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.321270 4790 scope.go:117] "RemoveContainer" containerID="9eebc6a326e6ecd25981550fadc6f071d57b5a734d8c279c22add9c347695246" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.322545 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9eebc6a326e6ecd25981550fadc6f071d57b5a734d8c279c22add9c347695246"} err="failed to get container status \"9eebc6a326e6ecd25981550fadc6f071d57b5a734d8c279c22add9c347695246\": rpc error: code = NotFound desc = could not find container \"9eebc6a326e6ecd25981550fadc6f071d57b5a734d8c279c22add9c347695246\": container with ID starting with 9eebc6a326e6ecd25981550fadc6f071d57b5a734d8c279c22add9c347695246 not found: ID does not exist" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.322575 4790 scope.go:117] "RemoveContainer" containerID="b42d440385b953399bc07de8bdb6974ec06cd9fdc735d457dea6aae49461fd49" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.322809 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b42d440385b953399bc07de8bdb6974ec06cd9fdc735d457dea6aae49461fd49"} err="failed to get container status \"b42d440385b953399bc07de8bdb6974ec06cd9fdc735d457dea6aae49461fd49\": rpc error: code = NotFound desc = could not find container \"b42d440385b953399bc07de8bdb6974ec06cd9fdc735d457dea6aae49461fd49\": container with ID starting with b42d440385b953399bc07de8bdb6974ec06cd9fdc735d457dea6aae49461fd49 not found: ID does not exist" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.322834 4790 scope.go:117] "RemoveContainer" containerID="926712c26d54a03dbf190d2dd69a4c2871a2ac13304b21eb29921e964c369938" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.323138 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"926712c26d54a03dbf190d2dd69a4c2871a2ac13304b21eb29921e964c369938"} err="failed to get container status \"926712c26d54a03dbf190d2dd69a4c2871a2ac13304b21eb29921e964c369938\": rpc error: code = NotFound desc = could not find container \"926712c26d54a03dbf190d2dd69a4c2871a2ac13304b21eb29921e964c369938\": container with ID starting with 926712c26d54a03dbf190d2dd69a4c2871a2ac13304b21eb29921e964c369938 not found: ID does not exist" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.389044 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/debad20a-2518-49f1-a1d6-49f5c39d7316-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.547374 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.562426 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.578200 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:56:38 crc kubenswrapper[4790]: E1124 14:56:38.578755 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="debad20a-2518-49f1-a1d6-49f5c39d7316" containerName="ceilometer-central-agent" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.578773 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="debad20a-2518-49f1-a1d6-49f5c39d7316" containerName="ceilometer-central-agent" Nov 24 14:56:38 crc kubenswrapper[4790]: E1124 14:56:38.578785 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="debad20a-2518-49f1-a1d6-49f5c39d7316" containerName="sg-core" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.578792 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="debad20a-2518-49f1-a1d6-49f5c39d7316" containerName="sg-core" Nov 24 14:56:38 crc kubenswrapper[4790]: E1124 14:56:38.578804 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="debad20a-2518-49f1-a1d6-49f5c39d7316" containerName="ceilometer-notification-agent" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.578811 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="debad20a-2518-49f1-a1d6-49f5c39d7316" containerName="ceilometer-notification-agent" Nov 24 14:56:38 crc kubenswrapper[4790]: E1124 14:56:38.578843 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="debad20a-2518-49f1-a1d6-49f5c39d7316" containerName="proxy-httpd" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.578849 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="debad20a-2518-49f1-a1d6-49f5c39d7316" containerName="proxy-httpd" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.579036 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="debad20a-2518-49f1-a1d6-49f5c39d7316" containerName="sg-core" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.579045 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="debad20a-2518-49f1-a1d6-49f5c39d7316" containerName="ceilometer-notification-agent" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.579058 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="debad20a-2518-49f1-a1d6-49f5c39d7316" containerName="proxy-httpd" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.579070 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="debad20a-2518-49f1-a1d6-49f5c39d7316" containerName="ceilometer-central-agent" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.583243 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.587932 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.588560 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.650279 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.703698 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\") " pod="openstack/ceilometer-0" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.703781 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-scripts\") pod \"ceilometer-0\" (UID: \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\") " pod="openstack/ceilometer-0" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.704017 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-config-data\") pod \"ceilometer-0\" (UID: \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\") " pod="openstack/ceilometer-0" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.704045 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-run-httpd\") pod \"ceilometer-0\" (UID: \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\") " pod="openstack/ceilometer-0" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.704112 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\") " pod="openstack/ceilometer-0" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.704177 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-log-httpd\") pod \"ceilometer-0\" (UID: \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\") " pod="openstack/ceilometer-0" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.704220 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffcvp\" (UniqueName: \"kubernetes.io/projected/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-kube-api-access-ffcvp\") pod \"ceilometer-0\" (UID: \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\") " pod="openstack/ceilometer-0" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.806506 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-config-data\") pod \"ceilometer-0\" (UID: \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\") " pod="openstack/ceilometer-0" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.806551 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-run-httpd\") pod \"ceilometer-0\" (UID: \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\") " pod="openstack/ceilometer-0" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.806595 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\") " pod="openstack/ceilometer-0" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.806634 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-log-httpd\") pod \"ceilometer-0\" (UID: \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\") " pod="openstack/ceilometer-0" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.806672 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffcvp\" (UniqueName: \"kubernetes.io/projected/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-kube-api-access-ffcvp\") pod \"ceilometer-0\" (UID: \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\") " pod="openstack/ceilometer-0" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.806746 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\") " pod="openstack/ceilometer-0" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.806775 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-scripts\") pod \"ceilometer-0\" (UID: \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\") " pod="openstack/ceilometer-0" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.807235 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-run-httpd\") pod \"ceilometer-0\" (UID: \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\") " pod="openstack/ceilometer-0" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.807389 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-log-httpd\") pod \"ceilometer-0\" (UID: \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\") " pod="openstack/ceilometer-0" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.811938 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-config-data\") pod \"ceilometer-0\" (UID: \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\") " pod="openstack/ceilometer-0" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.815580 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\") " pod="openstack/ceilometer-0" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.827313 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\") " pod="openstack/ceilometer-0" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.827801 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-scripts\") pod \"ceilometer-0\" (UID: \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\") " pod="openstack/ceilometer-0" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.830909 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffcvp\" (UniqueName: \"kubernetes.io/projected/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-kube-api-access-ffcvp\") pod \"ceilometer-0\" (UID: \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\") " pod="openstack/ceilometer-0" Nov 24 14:56:38 crc kubenswrapper[4790]: I1124 14:56:38.995104 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:56:39 crc kubenswrapper[4790]: I1124 14:56:39.735328 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:56:40 crc kubenswrapper[4790]: I1124 14:56:40.252168 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14","Type":"ContainerStarted","Data":"0dce2ace66e5388b6769eb529a3cc7b387fc98703807fb8708466301f9137ec9"} Nov 24 14:56:40 crc kubenswrapper[4790]: I1124 14:56:40.254652 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d0c3fa90-a58b-4a63-836a-35744ec234df","Type":"ContainerStarted","Data":"4da98c2b3010005e89f80a9f989cb972a428b3121838753826b1eb20ef14fe82"} Nov 24 14:56:40 crc kubenswrapper[4790]: I1124 14:56:40.356788 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="debad20a-2518-49f1-a1d6-49f5c39d7316" path="/var/lib/kubelet/pods/debad20a-2518-49f1-a1d6-49f5c39d7316/volumes" Nov 24 14:56:41 crc kubenswrapper[4790]: I1124 14:56:41.302114 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d0c3fa90-a58b-4a63-836a-35744ec234df","Type":"ContainerStarted","Data":"bcbdbc4ade29aca79d088e2766a79f9955274de419d3cbea6166a94b9c8deba8"} Nov 24 14:56:41 crc kubenswrapper[4790]: I1124 14:56:41.330342 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14","Type":"ContainerStarted","Data":"e10c8b46ab94c645d1a8865eb9a593912da588c59b9227d9107546a1cd6af700"} Nov 24 14:56:41 crc kubenswrapper[4790]: I1124 14:56:41.357328 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.18205844 podStartE2EDuration="7.357303491s" podCreationTimestamp="2025-11-24 14:56:34 +0000 UTC" firstStartedPulling="2025-11-24 14:56:35.695177908 +0000 UTC m=+6244.075071570" lastFinishedPulling="2025-11-24 14:56:40.870422959 +0000 UTC m=+6249.250316621" observedRunningTime="2025-11-24 14:56:41.354518411 +0000 UTC m=+6249.734412083" watchObservedRunningTime="2025-11-24 14:56:41.357303491 +0000 UTC m=+6249.737197153" Nov 24 14:56:42 crc kubenswrapper[4790]: I1124 14:56:42.325930 4790 scope.go:117] "RemoveContainer" containerID="28308703817a1d4e0e5344ce19268127c600894f2daa6b921d52a9105f014132" Nov 24 14:56:42 crc kubenswrapper[4790]: E1124 14:56:42.326840 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:56:42 crc kubenswrapper[4790]: I1124 14:56:42.355997 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14","Type":"ContainerStarted","Data":"fd0c27b4cee7b039c2b1bb808ab46734e60417816574192b3f8dc2ca2377ad38"} Nov 24 14:56:43 crc kubenswrapper[4790]: I1124 14:56:43.368360 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14","Type":"ContainerStarted","Data":"bd0d4c2dedb6b07f3d2455f84a347407e05895962ff554e3dbc9964f5bb4d8c2"} Nov 24 14:56:45 crc kubenswrapper[4790]: I1124 14:56:45.391381 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14","Type":"ContainerStarted","Data":"f810e3edfc208b53e611205b6c99a0364809d4a1fa049e015dfc0f7de2e5f79f"} Nov 24 14:56:45 crc kubenswrapper[4790]: I1124 14:56:45.391984 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 14:56:45 crc kubenswrapper[4790]: I1124 14:56:45.425500 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.8051051989999998 podStartE2EDuration="7.425477031s" podCreationTimestamp="2025-11-24 14:56:38 +0000 UTC" firstStartedPulling="2025-11-24 14:56:39.73057056 +0000 UTC m=+6248.110464222" lastFinishedPulling="2025-11-24 14:56:44.350942372 +0000 UTC m=+6252.730836054" observedRunningTime="2025-11-24 14:56:45.419116131 +0000 UTC m=+6253.799009793" watchObservedRunningTime="2025-11-24 14:56:45.425477031 +0000 UTC m=+6253.805370693" Nov 24 14:56:47 crc kubenswrapper[4790]: I1124 14:56:47.403932 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-2zln5"] Nov 24 14:56:47 crc kubenswrapper[4790]: I1124 14:56:47.407741 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-2zln5" Nov 24 14:56:47 crc kubenswrapper[4790]: I1124 14:56:47.446871 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-2zln5"] Nov 24 14:56:47 crc kubenswrapper[4790]: I1124 14:56:47.505343 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41b0a8b8-9156-44b3-b7ed-f72675827738-operator-scripts\") pod \"manila-db-create-2zln5\" (UID: \"41b0a8b8-9156-44b3-b7ed-f72675827738\") " pod="openstack/manila-db-create-2zln5" Nov 24 14:56:47 crc kubenswrapper[4790]: I1124 14:56:47.505383 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jx6p2\" (UniqueName: \"kubernetes.io/projected/41b0a8b8-9156-44b3-b7ed-f72675827738-kube-api-access-jx6p2\") pod \"manila-db-create-2zln5\" (UID: \"41b0a8b8-9156-44b3-b7ed-f72675827738\") " pod="openstack/manila-db-create-2zln5" Nov 24 14:56:47 crc kubenswrapper[4790]: I1124 14:56:47.507307 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-fe63-account-create-ppfwz"] Nov 24 14:56:47 crc kubenswrapper[4790]: I1124 14:56:47.508695 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-fe63-account-create-ppfwz" Nov 24 14:56:47 crc kubenswrapper[4790]: I1124 14:56:47.510665 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Nov 24 14:56:47 crc kubenswrapper[4790]: I1124 14:56:47.529680 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-fe63-account-create-ppfwz"] Nov 24 14:56:47 crc kubenswrapper[4790]: I1124 14:56:47.607711 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f95121cf-16ec-4760-9917-84d42ed63728-operator-scripts\") pod \"manila-fe63-account-create-ppfwz\" (UID: \"f95121cf-16ec-4760-9917-84d42ed63728\") " pod="openstack/manila-fe63-account-create-ppfwz" Nov 24 14:56:47 crc kubenswrapper[4790]: I1124 14:56:47.608136 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7htgj\" (UniqueName: \"kubernetes.io/projected/f95121cf-16ec-4760-9917-84d42ed63728-kube-api-access-7htgj\") pod \"manila-fe63-account-create-ppfwz\" (UID: \"f95121cf-16ec-4760-9917-84d42ed63728\") " pod="openstack/manila-fe63-account-create-ppfwz" Nov 24 14:56:47 crc kubenswrapper[4790]: I1124 14:56:47.608187 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41b0a8b8-9156-44b3-b7ed-f72675827738-operator-scripts\") pod \"manila-db-create-2zln5\" (UID: \"41b0a8b8-9156-44b3-b7ed-f72675827738\") " pod="openstack/manila-db-create-2zln5" Nov 24 14:56:47 crc kubenswrapper[4790]: I1124 14:56:47.608211 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jx6p2\" (UniqueName: \"kubernetes.io/projected/41b0a8b8-9156-44b3-b7ed-f72675827738-kube-api-access-jx6p2\") pod \"manila-db-create-2zln5\" (UID: \"41b0a8b8-9156-44b3-b7ed-f72675827738\") " pod="openstack/manila-db-create-2zln5" Nov 24 14:56:47 crc kubenswrapper[4790]: I1124 14:56:47.608982 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41b0a8b8-9156-44b3-b7ed-f72675827738-operator-scripts\") pod \"manila-db-create-2zln5\" (UID: \"41b0a8b8-9156-44b3-b7ed-f72675827738\") " pod="openstack/manila-db-create-2zln5" Nov 24 14:56:47 crc kubenswrapper[4790]: I1124 14:56:47.629386 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jx6p2\" (UniqueName: \"kubernetes.io/projected/41b0a8b8-9156-44b3-b7ed-f72675827738-kube-api-access-jx6p2\") pod \"manila-db-create-2zln5\" (UID: \"41b0a8b8-9156-44b3-b7ed-f72675827738\") " pod="openstack/manila-db-create-2zln5" Nov 24 14:56:47 crc kubenswrapper[4790]: I1124 14:56:47.710127 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7htgj\" (UniqueName: \"kubernetes.io/projected/f95121cf-16ec-4760-9917-84d42ed63728-kube-api-access-7htgj\") pod \"manila-fe63-account-create-ppfwz\" (UID: \"f95121cf-16ec-4760-9917-84d42ed63728\") " pod="openstack/manila-fe63-account-create-ppfwz" Nov 24 14:56:47 crc kubenswrapper[4790]: I1124 14:56:47.710482 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f95121cf-16ec-4760-9917-84d42ed63728-operator-scripts\") pod \"manila-fe63-account-create-ppfwz\" (UID: \"f95121cf-16ec-4760-9917-84d42ed63728\") " pod="openstack/manila-fe63-account-create-ppfwz" Nov 24 14:56:47 crc kubenswrapper[4790]: I1124 14:56:47.711388 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f95121cf-16ec-4760-9917-84d42ed63728-operator-scripts\") pod \"manila-fe63-account-create-ppfwz\" (UID: \"f95121cf-16ec-4760-9917-84d42ed63728\") " pod="openstack/manila-fe63-account-create-ppfwz" Nov 24 14:56:47 crc kubenswrapper[4790]: I1124 14:56:47.728656 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7htgj\" (UniqueName: \"kubernetes.io/projected/f95121cf-16ec-4760-9917-84d42ed63728-kube-api-access-7htgj\") pod \"manila-fe63-account-create-ppfwz\" (UID: \"f95121cf-16ec-4760-9917-84d42ed63728\") " pod="openstack/manila-fe63-account-create-ppfwz" Nov 24 14:56:47 crc kubenswrapper[4790]: I1124 14:56:47.736725 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-2zln5" Nov 24 14:56:47 crc kubenswrapper[4790]: I1124 14:56:47.846535 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-fe63-account-create-ppfwz" Nov 24 14:56:48 crc kubenswrapper[4790]: I1124 14:56:48.313328 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-2zln5"] Nov 24 14:56:48 crc kubenswrapper[4790]: I1124 14:56:48.442550 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-fe63-account-create-ppfwz"] Nov 24 14:56:48 crc kubenswrapper[4790]: I1124 14:56:48.450600 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-2zln5" event={"ID":"41b0a8b8-9156-44b3-b7ed-f72675827738","Type":"ContainerStarted","Data":"d4a38c08c9ce3a92bdae63dbfb24e6014bd560c4486daac360dcd80966013c35"} Nov 24 14:56:48 crc kubenswrapper[4790]: W1124 14:56:48.455798 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf95121cf_16ec_4760_9917_84d42ed63728.slice/crio-98580d1656f701f0d16d01e064ac15c7d9c3ab914dfd605cb2dcdfe61cfec955 WatchSource:0}: Error finding container 98580d1656f701f0d16d01e064ac15c7d9c3ab914dfd605cb2dcdfe61cfec955: Status 404 returned error can't find the container with id 98580d1656f701f0d16d01e064ac15c7d9c3ab914dfd605cb2dcdfe61cfec955 Nov 24 14:56:49 crc kubenswrapper[4790]: I1124 14:56:49.473023 4790 generic.go:334] "Generic (PLEG): container finished" podID="f95121cf-16ec-4760-9917-84d42ed63728" containerID="c94de0a2f3bb9c324c069d299d0ca9f47b924c5d6efe206968c30be485b0f7a2" exitCode=0 Nov 24 14:56:49 crc kubenswrapper[4790]: I1124 14:56:49.473091 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-fe63-account-create-ppfwz" event={"ID":"f95121cf-16ec-4760-9917-84d42ed63728","Type":"ContainerDied","Data":"c94de0a2f3bb9c324c069d299d0ca9f47b924c5d6efe206968c30be485b0f7a2"} Nov 24 14:56:49 crc kubenswrapper[4790]: I1124 14:56:49.475028 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-fe63-account-create-ppfwz" event={"ID":"f95121cf-16ec-4760-9917-84d42ed63728","Type":"ContainerStarted","Data":"98580d1656f701f0d16d01e064ac15c7d9c3ab914dfd605cb2dcdfe61cfec955"} Nov 24 14:56:49 crc kubenswrapper[4790]: I1124 14:56:49.477998 4790 generic.go:334] "Generic (PLEG): container finished" podID="41b0a8b8-9156-44b3-b7ed-f72675827738" containerID="f3469f70e2698fbbf691e9541604d6ce1f343b9a3eb3729dbb6e5dad3eea0788" exitCode=0 Nov 24 14:56:49 crc kubenswrapper[4790]: I1124 14:56:49.478035 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-2zln5" event={"ID":"41b0a8b8-9156-44b3-b7ed-f72675827738","Type":"ContainerDied","Data":"f3469f70e2698fbbf691e9541604d6ce1f343b9a3eb3729dbb6e5dad3eea0788"} Nov 24 14:56:51 crc kubenswrapper[4790]: I1124 14:56:51.059027 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-2zln5" Nov 24 14:56:51 crc kubenswrapper[4790]: I1124 14:56:51.067269 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-fe63-account-create-ppfwz" Nov 24 14:56:51 crc kubenswrapper[4790]: I1124 14:56:51.090813 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f95121cf-16ec-4760-9917-84d42ed63728-operator-scripts\") pod \"f95121cf-16ec-4760-9917-84d42ed63728\" (UID: \"f95121cf-16ec-4760-9917-84d42ed63728\") " Nov 24 14:56:51 crc kubenswrapper[4790]: I1124 14:56:51.090998 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41b0a8b8-9156-44b3-b7ed-f72675827738-operator-scripts\") pod \"41b0a8b8-9156-44b3-b7ed-f72675827738\" (UID: \"41b0a8b8-9156-44b3-b7ed-f72675827738\") " Nov 24 14:56:51 crc kubenswrapper[4790]: I1124 14:56:51.091132 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jx6p2\" (UniqueName: \"kubernetes.io/projected/41b0a8b8-9156-44b3-b7ed-f72675827738-kube-api-access-jx6p2\") pod \"41b0a8b8-9156-44b3-b7ed-f72675827738\" (UID: \"41b0a8b8-9156-44b3-b7ed-f72675827738\") " Nov 24 14:56:51 crc kubenswrapper[4790]: I1124 14:56:51.091230 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7htgj\" (UniqueName: \"kubernetes.io/projected/f95121cf-16ec-4760-9917-84d42ed63728-kube-api-access-7htgj\") pod \"f95121cf-16ec-4760-9917-84d42ed63728\" (UID: \"f95121cf-16ec-4760-9917-84d42ed63728\") " Nov 24 14:56:51 crc kubenswrapper[4790]: I1124 14:56:51.093334 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41b0a8b8-9156-44b3-b7ed-f72675827738-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "41b0a8b8-9156-44b3-b7ed-f72675827738" (UID: "41b0a8b8-9156-44b3-b7ed-f72675827738"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:56:51 crc kubenswrapper[4790]: I1124 14:56:51.093712 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f95121cf-16ec-4760-9917-84d42ed63728-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f95121cf-16ec-4760-9917-84d42ed63728" (UID: "f95121cf-16ec-4760-9917-84d42ed63728"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:56:51 crc kubenswrapper[4790]: I1124 14:56:51.099500 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f95121cf-16ec-4760-9917-84d42ed63728-kube-api-access-7htgj" (OuterVolumeSpecName: "kube-api-access-7htgj") pod "f95121cf-16ec-4760-9917-84d42ed63728" (UID: "f95121cf-16ec-4760-9917-84d42ed63728"). InnerVolumeSpecName "kube-api-access-7htgj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:56:51 crc kubenswrapper[4790]: I1124 14:56:51.101100 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41b0a8b8-9156-44b3-b7ed-f72675827738-kube-api-access-jx6p2" (OuterVolumeSpecName: "kube-api-access-jx6p2") pod "41b0a8b8-9156-44b3-b7ed-f72675827738" (UID: "41b0a8b8-9156-44b3-b7ed-f72675827738"). InnerVolumeSpecName "kube-api-access-jx6p2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:56:51 crc kubenswrapper[4790]: I1124 14:56:51.193391 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f95121cf-16ec-4760-9917-84d42ed63728-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:56:51 crc kubenswrapper[4790]: I1124 14:56:51.193427 4790 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41b0a8b8-9156-44b3-b7ed-f72675827738-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:56:51 crc kubenswrapper[4790]: I1124 14:56:51.193440 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jx6p2\" (UniqueName: \"kubernetes.io/projected/41b0a8b8-9156-44b3-b7ed-f72675827738-kube-api-access-jx6p2\") on node \"crc\" DevicePath \"\"" Nov 24 14:56:51 crc kubenswrapper[4790]: I1124 14:56:51.193454 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7htgj\" (UniqueName: \"kubernetes.io/projected/f95121cf-16ec-4760-9917-84d42ed63728-kube-api-access-7htgj\") on node \"crc\" DevicePath \"\"" Nov 24 14:56:51 crc kubenswrapper[4790]: I1124 14:56:51.506193 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-fe63-account-create-ppfwz" Nov 24 14:56:51 crc kubenswrapper[4790]: I1124 14:56:51.506199 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-fe63-account-create-ppfwz" event={"ID":"f95121cf-16ec-4760-9917-84d42ed63728","Type":"ContainerDied","Data":"98580d1656f701f0d16d01e064ac15c7d9c3ab914dfd605cb2dcdfe61cfec955"} Nov 24 14:56:51 crc kubenswrapper[4790]: I1124 14:56:51.506574 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98580d1656f701f0d16d01e064ac15c7d9c3ab914dfd605cb2dcdfe61cfec955" Nov 24 14:56:51 crc kubenswrapper[4790]: I1124 14:56:51.508708 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-2zln5" event={"ID":"41b0a8b8-9156-44b3-b7ed-f72675827738","Type":"ContainerDied","Data":"d4a38c08c9ce3a92bdae63dbfb24e6014bd560c4486daac360dcd80966013c35"} Nov 24 14:56:51 crc kubenswrapper[4790]: I1124 14:56:51.508778 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d4a38c08c9ce3a92bdae63dbfb24e6014bd560c4486daac360dcd80966013c35" Nov 24 14:56:51 crc kubenswrapper[4790]: I1124 14:56:51.508738 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-2zln5" Nov 24 14:56:52 crc kubenswrapper[4790]: I1124 14:56:52.495967 4790 scope.go:117] "RemoveContainer" containerID="0f08f5749ee72d72ca517437ac8315d8ad0d67680f0eaa3ada7413a972972e1f" Nov 24 14:56:52 crc kubenswrapper[4790]: I1124 14:56:52.581910 4790 scope.go:117] "RemoveContainer" containerID="1aca79fde163fdd085dbb216e6ef776da218068f8d70b5ad9db5568f87c7d80f" Nov 24 14:56:52 crc kubenswrapper[4790]: I1124 14:56:52.613014 4790 scope.go:117] "RemoveContainer" containerID="6e0567e71c58f0b072f917c99cde349dcf07f95554b9b7029917e653f0b4cabf" Nov 24 14:56:52 crc kubenswrapper[4790]: I1124 14:56:52.860389 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-ccjhs"] Nov 24 14:56:52 crc kubenswrapper[4790]: E1124 14:56:52.861114 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f95121cf-16ec-4760-9917-84d42ed63728" containerName="mariadb-account-create" Nov 24 14:56:52 crc kubenswrapper[4790]: I1124 14:56:52.861146 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="f95121cf-16ec-4760-9917-84d42ed63728" containerName="mariadb-account-create" Nov 24 14:56:52 crc kubenswrapper[4790]: E1124 14:56:52.861175 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41b0a8b8-9156-44b3-b7ed-f72675827738" containerName="mariadb-database-create" Nov 24 14:56:52 crc kubenswrapper[4790]: I1124 14:56:52.861188 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="41b0a8b8-9156-44b3-b7ed-f72675827738" containerName="mariadb-database-create" Nov 24 14:56:52 crc kubenswrapper[4790]: I1124 14:56:52.861540 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="f95121cf-16ec-4760-9917-84d42ed63728" containerName="mariadb-account-create" Nov 24 14:56:52 crc kubenswrapper[4790]: I1124 14:56:52.861576 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="41b0a8b8-9156-44b3-b7ed-f72675827738" containerName="mariadb-database-create" Nov 24 14:56:52 crc kubenswrapper[4790]: I1124 14:56:52.862951 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-ccjhs" Nov 24 14:56:52 crc kubenswrapper[4790]: I1124 14:56:52.865523 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-jkwgk" Nov 24 14:56:52 crc kubenswrapper[4790]: I1124 14:56:52.865996 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Nov 24 14:56:52 crc kubenswrapper[4790]: I1124 14:56:52.873186 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-ccjhs"] Nov 24 14:56:52 crc kubenswrapper[4790]: I1124 14:56:52.932287 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/657b7d2e-3626-4c66-a216-e2f2a7993771-config-data\") pod \"manila-db-sync-ccjhs\" (UID: \"657b7d2e-3626-4c66-a216-e2f2a7993771\") " pod="openstack/manila-db-sync-ccjhs" Nov 24 14:56:52 crc kubenswrapper[4790]: I1124 14:56:52.932328 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/657b7d2e-3626-4c66-a216-e2f2a7993771-combined-ca-bundle\") pod \"manila-db-sync-ccjhs\" (UID: \"657b7d2e-3626-4c66-a216-e2f2a7993771\") " pod="openstack/manila-db-sync-ccjhs" Nov 24 14:56:52 crc kubenswrapper[4790]: I1124 14:56:52.932369 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/657b7d2e-3626-4c66-a216-e2f2a7993771-job-config-data\") pod \"manila-db-sync-ccjhs\" (UID: \"657b7d2e-3626-4c66-a216-e2f2a7993771\") " pod="openstack/manila-db-sync-ccjhs" Nov 24 14:56:52 crc kubenswrapper[4790]: I1124 14:56:52.932430 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jt56f\" (UniqueName: \"kubernetes.io/projected/657b7d2e-3626-4c66-a216-e2f2a7993771-kube-api-access-jt56f\") pod \"manila-db-sync-ccjhs\" (UID: \"657b7d2e-3626-4c66-a216-e2f2a7993771\") " pod="openstack/manila-db-sync-ccjhs" Nov 24 14:56:53 crc kubenswrapper[4790]: I1124 14:56:53.034503 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/657b7d2e-3626-4c66-a216-e2f2a7993771-config-data\") pod \"manila-db-sync-ccjhs\" (UID: \"657b7d2e-3626-4c66-a216-e2f2a7993771\") " pod="openstack/manila-db-sync-ccjhs" Nov 24 14:56:53 crc kubenswrapper[4790]: I1124 14:56:53.034561 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/657b7d2e-3626-4c66-a216-e2f2a7993771-combined-ca-bundle\") pod \"manila-db-sync-ccjhs\" (UID: \"657b7d2e-3626-4c66-a216-e2f2a7993771\") " pod="openstack/manila-db-sync-ccjhs" Nov 24 14:56:53 crc kubenswrapper[4790]: I1124 14:56:53.034628 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/657b7d2e-3626-4c66-a216-e2f2a7993771-job-config-data\") pod \"manila-db-sync-ccjhs\" (UID: \"657b7d2e-3626-4c66-a216-e2f2a7993771\") " pod="openstack/manila-db-sync-ccjhs" Nov 24 14:56:53 crc kubenswrapper[4790]: I1124 14:56:53.034724 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jt56f\" (UniqueName: \"kubernetes.io/projected/657b7d2e-3626-4c66-a216-e2f2a7993771-kube-api-access-jt56f\") pod \"manila-db-sync-ccjhs\" (UID: \"657b7d2e-3626-4c66-a216-e2f2a7993771\") " pod="openstack/manila-db-sync-ccjhs" Nov 24 14:56:53 crc kubenswrapper[4790]: I1124 14:56:53.049776 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/657b7d2e-3626-4c66-a216-e2f2a7993771-job-config-data\") pod \"manila-db-sync-ccjhs\" (UID: \"657b7d2e-3626-4c66-a216-e2f2a7993771\") " pod="openstack/manila-db-sync-ccjhs" Nov 24 14:56:53 crc kubenswrapper[4790]: I1124 14:56:53.049945 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/657b7d2e-3626-4c66-a216-e2f2a7993771-combined-ca-bundle\") pod \"manila-db-sync-ccjhs\" (UID: \"657b7d2e-3626-4c66-a216-e2f2a7993771\") " pod="openstack/manila-db-sync-ccjhs" Nov 24 14:56:53 crc kubenswrapper[4790]: I1124 14:56:53.049984 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/657b7d2e-3626-4c66-a216-e2f2a7993771-config-data\") pod \"manila-db-sync-ccjhs\" (UID: \"657b7d2e-3626-4c66-a216-e2f2a7993771\") " pod="openstack/manila-db-sync-ccjhs" Nov 24 14:56:53 crc kubenswrapper[4790]: I1124 14:56:53.061359 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jt56f\" (UniqueName: \"kubernetes.io/projected/657b7d2e-3626-4c66-a216-e2f2a7993771-kube-api-access-jt56f\") pod \"manila-db-sync-ccjhs\" (UID: \"657b7d2e-3626-4c66-a216-e2f2a7993771\") " pod="openstack/manila-db-sync-ccjhs" Nov 24 14:56:53 crc kubenswrapper[4790]: I1124 14:56:53.184469 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-ccjhs" Nov 24 14:56:54 crc kubenswrapper[4790]: W1124 14:56:54.095535 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod657b7d2e_3626_4c66_a216_e2f2a7993771.slice/crio-ff6a6cc96ff8ee43aadbad49c0264e932ecd3f86d851232c7f93f26fb8d0d4a5 WatchSource:0}: Error finding container ff6a6cc96ff8ee43aadbad49c0264e932ecd3f86d851232c7f93f26fb8d0d4a5: Status 404 returned error can't find the container with id ff6a6cc96ff8ee43aadbad49c0264e932ecd3f86d851232c7f93f26fb8d0d4a5 Nov 24 14:56:54 crc kubenswrapper[4790]: I1124 14:56:54.095574 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-ccjhs"] Nov 24 14:56:54 crc kubenswrapper[4790]: I1124 14:56:54.545034 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-ccjhs" event={"ID":"657b7d2e-3626-4c66-a216-e2f2a7993771","Type":"ContainerStarted","Data":"ff6a6cc96ff8ee43aadbad49c0264e932ecd3f86d851232c7f93f26fb8d0d4a5"} Nov 24 14:56:55 crc kubenswrapper[4790]: I1124 14:56:55.315418 4790 scope.go:117] "RemoveContainer" containerID="28308703817a1d4e0e5344ce19268127c600894f2daa6b921d52a9105f014132" Nov 24 14:56:55 crc kubenswrapper[4790]: E1124 14:56:55.317213 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:56:59 crc kubenswrapper[4790]: I1124 14:56:59.623955 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-ccjhs" event={"ID":"657b7d2e-3626-4c66-a216-e2f2a7993771","Type":"ContainerStarted","Data":"4e2c99071bb23fc6540090564ab09dfcab17097ba5d8ec2bd9359866dc926528"} Nov 24 14:56:59 crc kubenswrapper[4790]: I1124 14:56:59.646570 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-ccjhs" podStartSLOduration=2.958236346 podStartE2EDuration="7.646553792s" podCreationTimestamp="2025-11-24 14:56:52 +0000 UTC" firstStartedPulling="2025-11-24 14:56:54.098446439 +0000 UTC m=+6262.478340101" lastFinishedPulling="2025-11-24 14:56:58.786763845 +0000 UTC m=+6267.166657547" observedRunningTime="2025-11-24 14:56:59.64645951 +0000 UTC m=+6268.026353182" watchObservedRunningTime="2025-11-24 14:56:59.646553792 +0000 UTC m=+6268.026447454" Nov 24 14:57:01 crc kubenswrapper[4790]: I1124 14:57:01.656548 4790 generic.go:334] "Generic (PLEG): container finished" podID="657b7d2e-3626-4c66-a216-e2f2a7993771" containerID="4e2c99071bb23fc6540090564ab09dfcab17097ba5d8ec2bd9359866dc926528" exitCode=0 Nov 24 14:57:01 crc kubenswrapper[4790]: I1124 14:57:01.656642 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-ccjhs" event={"ID":"657b7d2e-3626-4c66-a216-e2f2a7993771","Type":"ContainerDied","Data":"4e2c99071bb23fc6540090564ab09dfcab17097ba5d8ec2bd9359866dc926528"} Nov 24 14:57:03 crc kubenswrapper[4790]: I1124 14:57:03.231830 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-ccjhs" Nov 24 14:57:03 crc kubenswrapper[4790]: I1124 14:57:03.275370 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jt56f\" (UniqueName: \"kubernetes.io/projected/657b7d2e-3626-4c66-a216-e2f2a7993771-kube-api-access-jt56f\") pod \"657b7d2e-3626-4c66-a216-e2f2a7993771\" (UID: \"657b7d2e-3626-4c66-a216-e2f2a7993771\") " Nov 24 14:57:03 crc kubenswrapper[4790]: I1124 14:57:03.275488 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/657b7d2e-3626-4c66-a216-e2f2a7993771-combined-ca-bundle\") pod \"657b7d2e-3626-4c66-a216-e2f2a7993771\" (UID: \"657b7d2e-3626-4c66-a216-e2f2a7993771\") " Nov 24 14:57:03 crc kubenswrapper[4790]: I1124 14:57:03.275557 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/657b7d2e-3626-4c66-a216-e2f2a7993771-config-data\") pod \"657b7d2e-3626-4c66-a216-e2f2a7993771\" (UID: \"657b7d2e-3626-4c66-a216-e2f2a7993771\") " Nov 24 14:57:03 crc kubenswrapper[4790]: I1124 14:57:03.275692 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/657b7d2e-3626-4c66-a216-e2f2a7993771-job-config-data\") pod \"657b7d2e-3626-4c66-a216-e2f2a7993771\" (UID: \"657b7d2e-3626-4c66-a216-e2f2a7993771\") " Nov 24 14:57:03 crc kubenswrapper[4790]: I1124 14:57:03.289550 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/657b7d2e-3626-4c66-a216-e2f2a7993771-kube-api-access-jt56f" (OuterVolumeSpecName: "kube-api-access-jt56f") pod "657b7d2e-3626-4c66-a216-e2f2a7993771" (UID: "657b7d2e-3626-4c66-a216-e2f2a7993771"). InnerVolumeSpecName "kube-api-access-jt56f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:57:03 crc kubenswrapper[4790]: I1124 14:57:03.289568 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/657b7d2e-3626-4c66-a216-e2f2a7993771-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "657b7d2e-3626-4c66-a216-e2f2a7993771" (UID: "657b7d2e-3626-4c66-a216-e2f2a7993771"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:57:03 crc kubenswrapper[4790]: I1124 14:57:03.293715 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/657b7d2e-3626-4c66-a216-e2f2a7993771-config-data" (OuterVolumeSpecName: "config-data") pod "657b7d2e-3626-4c66-a216-e2f2a7993771" (UID: "657b7d2e-3626-4c66-a216-e2f2a7993771"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:57:03 crc kubenswrapper[4790]: I1124 14:57:03.313745 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/657b7d2e-3626-4c66-a216-e2f2a7993771-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "657b7d2e-3626-4c66-a216-e2f2a7993771" (UID: "657b7d2e-3626-4c66-a216-e2f2a7993771"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:57:03 crc kubenswrapper[4790]: I1124 14:57:03.378172 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/657b7d2e-3626-4c66-a216-e2f2a7993771-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:57:03 crc kubenswrapper[4790]: I1124 14:57:03.378213 4790 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/657b7d2e-3626-4c66-a216-e2f2a7993771-job-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:57:03 crc kubenswrapper[4790]: I1124 14:57:03.378370 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jt56f\" (UniqueName: \"kubernetes.io/projected/657b7d2e-3626-4c66-a216-e2f2a7993771-kube-api-access-jt56f\") on node \"crc\" DevicePath \"\"" Nov 24 14:57:03 crc kubenswrapper[4790]: I1124 14:57:03.378651 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/657b7d2e-3626-4c66-a216-e2f2a7993771-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:57:03 crc kubenswrapper[4790]: I1124 14:57:03.682516 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-ccjhs" event={"ID":"657b7d2e-3626-4c66-a216-e2f2a7993771","Type":"ContainerDied","Data":"ff6a6cc96ff8ee43aadbad49c0264e932ecd3f86d851232c7f93f26fb8d0d4a5"} Nov 24 14:57:03 crc kubenswrapper[4790]: I1124 14:57:03.682561 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-ccjhs" Nov 24 14:57:03 crc kubenswrapper[4790]: I1124 14:57:03.682569 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff6a6cc96ff8ee43aadbad49c0264e932ecd3f86d851232c7f93f26fb8d0d4a5" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.057706 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Nov 24 14:57:04 crc kubenswrapper[4790]: E1124 14:57:04.058215 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="657b7d2e-3626-4c66-a216-e2f2a7993771" containerName="manila-db-sync" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.058237 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="657b7d2e-3626-4c66-a216-e2f2a7993771" containerName="manila-db-sync" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.058433 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="657b7d2e-3626-4c66-a216-e2f2a7993771" containerName="manila-db-sync" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.059655 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.069222 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.076078 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.076303 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.076424 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-jkwgk" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.076543 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.089024 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.096312 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b75fcca8-34a3-42d4-bfca-43c21b63227b-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"b75fcca8-34a3-42d4-bfca-43c21b63227b\") " pod="openstack/manila-scheduler-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.096574 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7jhf\" (UniqueName: \"kubernetes.io/projected/b75fcca8-34a3-42d4-bfca-43c21b63227b-kube-api-access-n7jhf\") pod \"manila-scheduler-0\" (UID: \"b75fcca8-34a3-42d4-bfca-43c21b63227b\") " pod="openstack/manila-scheduler-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.096602 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b75fcca8-34a3-42d4-bfca-43c21b63227b-scripts\") pod \"manila-scheduler-0\" (UID: \"b75fcca8-34a3-42d4-bfca-43c21b63227b\") " pod="openstack/manila-scheduler-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.096637 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b75fcca8-34a3-42d4-bfca-43c21b63227b-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"b75fcca8-34a3-42d4-bfca-43c21b63227b\") " pod="openstack/manila-scheduler-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.096666 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b75fcca8-34a3-42d4-bfca-43c21b63227b-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"b75fcca8-34a3-42d4-bfca-43c21b63227b\") " pod="openstack/manila-scheduler-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.096780 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b75fcca8-34a3-42d4-bfca-43c21b63227b-config-data\") pod \"manila-scheduler-0\" (UID: \"b75fcca8-34a3-42d4-bfca-43c21b63227b\") " pod="openstack/manila-scheduler-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.098977 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.103509 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.104620 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.159351 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-68bc6cdf65-975r7"] Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.162420 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68bc6cdf65-975r7" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.182523 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68bc6cdf65-975r7"] Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.198800 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b75fcca8-34a3-42d4-bfca-43c21b63227b-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"b75fcca8-34a3-42d4-bfca-43c21b63227b\") " pod="openstack/manila-scheduler-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.198868 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd78a936-5519-480a-b36a-a6ec905cd694-dns-svc\") pod \"dnsmasq-dns-68bc6cdf65-975r7\" (UID: \"cd78a936-5519-480a-b36a-a6ec905cd694\") " pod="openstack/dnsmasq-dns-68bc6cdf65-975r7" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.198914 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7jhf\" (UniqueName: \"kubernetes.io/projected/b75fcca8-34a3-42d4-bfca-43c21b63227b-kube-api-access-n7jhf\") pod \"manila-scheduler-0\" (UID: \"b75fcca8-34a3-42d4-bfca-43c21b63227b\") " pod="openstack/manila-scheduler-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.198934 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xx8g\" (UniqueName: \"kubernetes.io/projected/6cacb59b-e19b-44b9-b4c4-3a12559e64ed-kube-api-access-4xx8g\") pod \"manila-share-share1-0\" (UID: \"6cacb59b-e19b-44b9-b4c4-3a12559e64ed\") " pod="openstack/manila-share-share1-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.198961 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/6cacb59b-e19b-44b9-b4c4-3a12559e64ed-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"6cacb59b-e19b-44b9-b4c4-3a12559e64ed\") " pod="openstack/manila-share-share1-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.198979 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b75fcca8-34a3-42d4-bfca-43c21b63227b-scripts\") pod \"manila-scheduler-0\" (UID: \"b75fcca8-34a3-42d4-bfca-43c21b63227b\") " pod="openstack/manila-scheduler-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.199049 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6cacb59b-e19b-44b9-b4c4-3a12559e64ed-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"6cacb59b-e19b-44b9-b4c4-3a12559e64ed\") " pod="openstack/manila-share-share1-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.199125 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b75fcca8-34a3-42d4-bfca-43c21b63227b-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"b75fcca8-34a3-42d4-bfca-43c21b63227b\") " pod="openstack/manila-scheduler-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.199158 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cacb59b-e19b-44b9-b4c4-3a12559e64ed-config-data\") pod \"manila-share-share1-0\" (UID: \"6cacb59b-e19b-44b9-b4c4-3a12559e64ed\") " pod="openstack/manila-share-share1-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.199197 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b75fcca8-34a3-42d4-bfca-43c21b63227b-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"b75fcca8-34a3-42d4-bfca-43c21b63227b\") " pod="openstack/manila-scheduler-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.199314 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/6cacb59b-e19b-44b9-b4c4-3a12559e64ed-ceph\") pod \"manila-share-share1-0\" (UID: \"6cacb59b-e19b-44b9-b4c4-3a12559e64ed\") " pod="openstack/manila-share-share1-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.199332 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd78a936-5519-480a-b36a-a6ec905cd694-ovsdbserver-nb\") pod \"dnsmasq-dns-68bc6cdf65-975r7\" (UID: \"cd78a936-5519-480a-b36a-a6ec905cd694\") " pod="openstack/dnsmasq-dns-68bc6cdf65-975r7" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.199388 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cacb59b-e19b-44b9-b4c4-3a12559e64ed-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"6cacb59b-e19b-44b9-b4c4-3a12559e64ed\") " pod="openstack/manila-share-share1-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.199455 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd78a936-5519-480a-b36a-a6ec905cd694-config\") pod \"dnsmasq-dns-68bc6cdf65-975r7\" (UID: \"cd78a936-5519-480a-b36a-a6ec905cd694\") " pod="openstack/dnsmasq-dns-68bc6cdf65-975r7" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.199479 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6cacb59b-e19b-44b9-b4c4-3a12559e64ed-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"6cacb59b-e19b-44b9-b4c4-3a12559e64ed\") " pod="openstack/manila-share-share1-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.199508 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd78a936-5519-480a-b36a-a6ec905cd694-ovsdbserver-sb\") pod \"dnsmasq-dns-68bc6cdf65-975r7\" (UID: \"cd78a936-5519-480a-b36a-a6ec905cd694\") " pod="openstack/dnsmasq-dns-68bc6cdf65-975r7" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.199587 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6cacb59b-e19b-44b9-b4c4-3a12559e64ed-scripts\") pod \"manila-share-share1-0\" (UID: \"6cacb59b-e19b-44b9-b4c4-3a12559e64ed\") " pod="openstack/manila-share-share1-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.199628 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b75fcca8-34a3-42d4-bfca-43c21b63227b-config-data\") pod \"manila-scheduler-0\" (UID: \"b75fcca8-34a3-42d4-bfca-43c21b63227b\") " pod="openstack/manila-scheduler-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.199690 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb5hs\" (UniqueName: \"kubernetes.io/projected/cd78a936-5519-480a-b36a-a6ec905cd694-kube-api-access-sb5hs\") pod \"dnsmasq-dns-68bc6cdf65-975r7\" (UID: \"cd78a936-5519-480a-b36a-a6ec905cd694\") " pod="openstack/dnsmasq-dns-68bc6cdf65-975r7" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.198967 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b75fcca8-34a3-42d4-bfca-43c21b63227b-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"b75fcca8-34a3-42d4-bfca-43c21b63227b\") " pod="openstack/manila-scheduler-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.203817 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b75fcca8-34a3-42d4-bfca-43c21b63227b-scripts\") pod \"manila-scheduler-0\" (UID: \"b75fcca8-34a3-42d4-bfca-43c21b63227b\") " pod="openstack/manila-scheduler-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.207588 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b75fcca8-34a3-42d4-bfca-43c21b63227b-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"b75fcca8-34a3-42d4-bfca-43c21b63227b\") " pod="openstack/manila-scheduler-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.211972 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b75fcca8-34a3-42d4-bfca-43c21b63227b-config-data\") pod \"manila-scheduler-0\" (UID: \"b75fcca8-34a3-42d4-bfca-43c21b63227b\") " pod="openstack/manila-scheduler-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.219501 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b75fcca8-34a3-42d4-bfca-43c21b63227b-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"b75fcca8-34a3-42d4-bfca-43c21b63227b\") " pod="openstack/manila-scheduler-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.220459 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7jhf\" (UniqueName: \"kubernetes.io/projected/b75fcca8-34a3-42d4-bfca-43c21b63227b-kube-api-access-n7jhf\") pod \"manila-scheduler-0\" (UID: \"b75fcca8-34a3-42d4-bfca-43c21b63227b\") " pod="openstack/manila-scheduler-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.301909 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd78a936-5519-480a-b36a-a6ec905cd694-dns-svc\") pod \"dnsmasq-dns-68bc6cdf65-975r7\" (UID: \"cd78a936-5519-480a-b36a-a6ec905cd694\") " pod="openstack/dnsmasq-dns-68bc6cdf65-975r7" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.301967 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xx8g\" (UniqueName: \"kubernetes.io/projected/6cacb59b-e19b-44b9-b4c4-3a12559e64ed-kube-api-access-4xx8g\") pod \"manila-share-share1-0\" (UID: \"6cacb59b-e19b-44b9-b4c4-3a12559e64ed\") " pod="openstack/manila-share-share1-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.301992 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/6cacb59b-e19b-44b9-b4c4-3a12559e64ed-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"6cacb59b-e19b-44b9-b4c4-3a12559e64ed\") " pod="openstack/manila-share-share1-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.302018 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6cacb59b-e19b-44b9-b4c4-3a12559e64ed-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"6cacb59b-e19b-44b9-b4c4-3a12559e64ed\") " pod="openstack/manila-share-share1-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.302047 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cacb59b-e19b-44b9-b4c4-3a12559e64ed-config-data\") pod \"manila-share-share1-0\" (UID: \"6cacb59b-e19b-44b9-b4c4-3a12559e64ed\") " pod="openstack/manila-share-share1-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.302099 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/6cacb59b-e19b-44b9-b4c4-3a12559e64ed-ceph\") pod \"manila-share-share1-0\" (UID: \"6cacb59b-e19b-44b9-b4c4-3a12559e64ed\") " pod="openstack/manila-share-share1-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.302115 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd78a936-5519-480a-b36a-a6ec905cd694-ovsdbserver-nb\") pod \"dnsmasq-dns-68bc6cdf65-975r7\" (UID: \"cd78a936-5519-480a-b36a-a6ec905cd694\") " pod="openstack/dnsmasq-dns-68bc6cdf65-975r7" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.302142 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cacb59b-e19b-44b9-b4c4-3a12559e64ed-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"6cacb59b-e19b-44b9-b4c4-3a12559e64ed\") " pod="openstack/manila-share-share1-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.302172 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd78a936-5519-480a-b36a-a6ec905cd694-config\") pod \"dnsmasq-dns-68bc6cdf65-975r7\" (UID: \"cd78a936-5519-480a-b36a-a6ec905cd694\") " pod="openstack/dnsmasq-dns-68bc6cdf65-975r7" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.302188 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6cacb59b-e19b-44b9-b4c4-3a12559e64ed-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"6cacb59b-e19b-44b9-b4c4-3a12559e64ed\") " pod="openstack/manila-share-share1-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.302205 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd78a936-5519-480a-b36a-a6ec905cd694-ovsdbserver-sb\") pod \"dnsmasq-dns-68bc6cdf65-975r7\" (UID: \"cd78a936-5519-480a-b36a-a6ec905cd694\") " pod="openstack/dnsmasq-dns-68bc6cdf65-975r7" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.302252 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6cacb59b-e19b-44b9-b4c4-3a12559e64ed-scripts\") pod \"manila-share-share1-0\" (UID: \"6cacb59b-e19b-44b9-b4c4-3a12559e64ed\") " pod="openstack/manila-share-share1-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.302290 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sb5hs\" (UniqueName: \"kubernetes.io/projected/cd78a936-5519-480a-b36a-a6ec905cd694-kube-api-access-sb5hs\") pod \"dnsmasq-dns-68bc6cdf65-975r7\" (UID: \"cd78a936-5519-480a-b36a-a6ec905cd694\") " pod="openstack/dnsmasq-dns-68bc6cdf65-975r7" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.303405 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd78a936-5519-480a-b36a-a6ec905cd694-dns-svc\") pod \"dnsmasq-dns-68bc6cdf65-975r7\" (UID: \"cd78a936-5519-480a-b36a-a6ec905cd694\") " pod="openstack/dnsmasq-dns-68bc6cdf65-975r7" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.303641 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/6cacb59b-e19b-44b9-b4c4-3a12559e64ed-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"6cacb59b-e19b-44b9-b4c4-3a12559e64ed\") " pod="openstack/manila-share-share1-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.306134 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd78a936-5519-480a-b36a-a6ec905cd694-config\") pod \"dnsmasq-dns-68bc6cdf65-975r7\" (UID: \"cd78a936-5519-480a-b36a-a6ec905cd694\") " pod="openstack/dnsmasq-dns-68bc6cdf65-975r7" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.308106 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6cacb59b-e19b-44b9-b4c4-3a12559e64ed-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"6cacb59b-e19b-44b9-b4c4-3a12559e64ed\") " pod="openstack/manila-share-share1-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.308412 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd78a936-5519-480a-b36a-a6ec905cd694-ovsdbserver-sb\") pod \"dnsmasq-dns-68bc6cdf65-975r7\" (UID: \"cd78a936-5519-480a-b36a-a6ec905cd694\") " pod="openstack/dnsmasq-dns-68bc6cdf65-975r7" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.309658 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd78a936-5519-480a-b36a-a6ec905cd694-ovsdbserver-nb\") pod \"dnsmasq-dns-68bc6cdf65-975r7\" (UID: \"cd78a936-5519-480a-b36a-a6ec905cd694\") " pod="openstack/dnsmasq-dns-68bc6cdf65-975r7" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.314176 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cacb59b-e19b-44b9-b4c4-3a12559e64ed-config-data\") pod \"manila-share-share1-0\" (UID: \"6cacb59b-e19b-44b9-b4c4-3a12559e64ed\") " pod="openstack/manila-share-share1-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.314609 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6cacb59b-e19b-44b9-b4c4-3a12559e64ed-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"6cacb59b-e19b-44b9-b4c4-3a12559e64ed\") " pod="openstack/manila-share-share1-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.316558 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cacb59b-e19b-44b9-b4c4-3a12559e64ed-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"6cacb59b-e19b-44b9-b4c4-3a12559e64ed\") " pod="openstack/manila-share-share1-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.317071 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/6cacb59b-e19b-44b9-b4c4-3a12559e64ed-ceph\") pod \"manila-share-share1-0\" (UID: \"6cacb59b-e19b-44b9-b4c4-3a12559e64ed\") " pod="openstack/manila-share-share1-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.321959 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xx8g\" (UniqueName: \"kubernetes.io/projected/6cacb59b-e19b-44b9-b4c4-3a12559e64ed-kube-api-access-4xx8g\") pod \"manila-share-share1-0\" (UID: \"6cacb59b-e19b-44b9-b4c4-3a12559e64ed\") " pod="openstack/manila-share-share1-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.326674 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6cacb59b-e19b-44b9-b4c4-3a12559e64ed-scripts\") pod \"manila-share-share1-0\" (UID: \"6cacb59b-e19b-44b9-b4c4-3a12559e64ed\") " pod="openstack/manila-share-share1-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.334436 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb5hs\" (UniqueName: \"kubernetes.io/projected/cd78a936-5519-480a-b36a-a6ec905cd694-kube-api-access-sb5hs\") pod \"dnsmasq-dns-68bc6cdf65-975r7\" (UID: \"cd78a936-5519-480a-b36a-a6ec905cd694\") " pod="openstack/dnsmasq-dns-68bc6cdf65-975r7" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.391018 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.393384 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.395296 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.397814 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.422729 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.432400 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.479931 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68bc6cdf65-975r7" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.509691 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d42b6e7d-ef2f-4d91-b9d0-138efb9549eb-config-data-custom\") pod \"manila-api-0\" (UID: \"d42b6e7d-ef2f-4d91-b9d0-138efb9549eb\") " pod="openstack/manila-api-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.509743 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d42b6e7d-ef2f-4d91-b9d0-138efb9549eb-etc-machine-id\") pod \"manila-api-0\" (UID: \"d42b6e7d-ef2f-4d91-b9d0-138efb9549eb\") " pod="openstack/manila-api-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.511843 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ln6tp\" (UniqueName: \"kubernetes.io/projected/d42b6e7d-ef2f-4d91-b9d0-138efb9549eb-kube-api-access-ln6tp\") pod \"manila-api-0\" (UID: \"d42b6e7d-ef2f-4d91-b9d0-138efb9549eb\") " pod="openstack/manila-api-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.511943 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d42b6e7d-ef2f-4d91-b9d0-138efb9549eb-logs\") pod \"manila-api-0\" (UID: \"d42b6e7d-ef2f-4d91-b9d0-138efb9549eb\") " pod="openstack/manila-api-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.512002 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d42b6e7d-ef2f-4d91-b9d0-138efb9549eb-config-data\") pod \"manila-api-0\" (UID: \"d42b6e7d-ef2f-4d91-b9d0-138efb9549eb\") " pod="openstack/manila-api-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.512078 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d42b6e7d-ef2f-4d91-b9d0-138efb9549eb-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"d42b6e7d-ef2f-4d91-b9d0-138efb9549eb\") " pod="openstack/manila-api-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.512413 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d42b6e7d-ef2f-4d91-b9d0-138efb9549eb-scripts\") pod \"manila-api-0\" (UID: \"d42b6e7d-ef2f-4d91-b9d0-138efb9549eb\") " pod="openstack/manila-api-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.618330 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d42b6e7d-ef2f-4d91-b9d0-138efb9549eb-scripts\") pod \"manila-api-0\" (UID: \"d42b6e7d-ef2f-4d91-b9d0-138efb9549eb\") " pod="openstack/manila-api-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.618773 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d42b6e7d-ef2f-4d91-b9d0-138efb9549eb-config-data-custom\") pod \"manila-api-0\" (UID: \"d42b6e7d-ef2f-4d91-b9d0-138efb9549eb\") " pod="openstack/manila-api-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.618827 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d42b6e7d-ef2f-4d91-b9d0-138efb9549eb-etc-machine-id\") pod \"manila-api-0\" (UID: \"d42b6e7d-ef2f-4d91-b9d0-138efb9549eb\") " pod="openstack/manila-api-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.618904 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ln6tp\" (UniqueName: \"kubernetes.io/projected/d42b6e7d-ef2f-4d91-b9d0-138efb9549eb-kube-api-access-ln6tp\") pod \"manila-api-0\" (UID: \"d42b6e7d-ef2f-4d91-b9d0-138efb9549eb\") " pod="openstack/manila-api-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.618951 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d42b6e7d-ef2f-4d91-b9d0-138efb9549eb-logs\") pod \"manila-api-0\" (UID: \"d42b6e7d-ef2f-4d91-b9d0-138efb9549eb\") " pod="openstack/manila-api-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.618988 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d42b6e7d-ef2f-4d91-b9d0-138efb9549eb-config-data\") pod \"manila-api-0\" (UID: \"d42b6e7d-ef2f-4d91-b9d0-138efb9549eb\") " pod="openstack/manila-api-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.619042 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d42b6e7d-ef2f-4d91-b9d0-138efb9549eb-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"d42b6e7d-ef2f-4d91-b9d0-138efb9549eb\") " pod="openstack/manila-api-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.619878 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d42b6e7d-ef2f-4d91-b9d0-138efb9549eb-etc-machine-id\") pod \"manila-api-0\" (UID: \"d42b6e7d-ef2f-4d91-b9d0-138efb9549eb\") " pod="openstack/manila-api-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.625299 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d42b6e7d-ef2f-4d91-b9d0-138efb9549eb-logs\") pod \"manila-api-0\" (UID: \"d42b6e7d-ef2f-4d91-b9d0-138efb9549eb\") " pod="openstack/manila-api-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.629627 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d42b6e7d-ef2f-4d91-b9d0-138efb9549eb-scripts\") pod \"manila-api-0\" (UID: \"d42b6e7d-ef2f-4d91-b9d0-138efb9549eb\") " pod="openstack/manila-api-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.629741 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d42b6e7d-ef2f-4d91-b9d0-138efb9549eb-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"d42b6e7d-ef2f-4d91-b9d0-138efb9549eb\") " pod="openstack/manila-api-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.630831 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d42b6e7d-ef2f-4d91-b9d0-138efb9549eb-config-data-custom\") pod \"manila-api-0\" (UID: \"d42b6e7d-ef2f-4d91-b9d0-138efb9549eb\") " pod="openstack/manila-api-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.632164 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d42b6e7d-ef2f-4d91-b9d0-138efb9549eb-config-data\") pod \"manila-api-0\" (UID: \"d42b6e7d-ef2f-4d91-b9d0-138efb9549eb\") " pod="openstack/manila-api-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.649527 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ln6tp\" (UniqueName: \"kubernetes.io/projected/d42b6e7d-ef2f-4d91-b9d0-138efb9549eb-kube-api-access-ln6tp\") pod \"manila-api-0\" (UID: \"d42b6e7d-ef2f-4d91-b9d0-138efb9549eb\") " pod="openstack/manila-api-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.788115 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 24 14:57:04 crc kubenswrapper[4790]: I1124 14:57:04.936687 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 24 14:57:05 crc kubenswrapper[4790]: I1124 14:57:05.102458 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68bc6cdf65-975r7"] Nov 24 14:57:05 crc kubenswrapper[4790]: I1124 14:57:05.192992 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 24 14:57:05 crc kubenswrapper[4790]: I1124 14:57:05.394337 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Nov 24 14:57:05 crc kubenswrapper[4790]: I1124 14:57:05.726107 4790 generic.go:334] "Generic (PLEG): container finished" podID="cd78a936-5519-480a-b36a-a6ec905cd694" containerID="4e01d22038030468b23c387d57334ceb2e0efaba5a1eda56f841b16ad889663b" exitCode=0 Nov 24 14:57:05 crc kubenswrapper[4790]: I1124 14:57:05.726185 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68bc6cdf65-975r7" event={"ID":"cd78a936-5519-480a-b36a-a6ec905cd694","Type":"ContainerDied","Data":"4e01d22038030468b23c387d57334ceb2e0efaba5a1eda56f841b16ad889663b"} Nov 24 14:57:05 crc kubenswrapper[4790]: I1124 14:57:05.726224 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68bc6cdf65-975r7" event={"ID":"cd78a936-5519-480a-b36a-a6ec905cd694","Type":"ContainerStarted","Data":"db61ba705687b91bcdae96d2ace8d031bf7b289b1f9f06e8a9d2d288af7f7e5c"} Nov 24 14:57:05 crc kubenswrapper[4790]: I1124 14:57:05.728037 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"d42b6e7d-ef2f-4d91-b9d0-138efb9549eb","Type":"ContainerStarted","Data":"ec0af6a1867bd084f378fa6a409d13913126118bccb2ef50380582033ade8326"} Nov 24 14:57:05 crc kubenswrapper[4790]: I1124 14:57:05.731699 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"6cacb59b-e19b-44b9-b4c4-3a12559e64ed","Type":"ContainerStarted","Data":"6f96abba096c02e2c292cf762ca80a0c2eadae071f644d1357460d7569a0d9d7"} Nov 24 14:57:05 crc kubenswrapper[4790]: I1124 14:57:05.734849 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"b75fcca8-34a3-42d4-bfca-43c21b63227b","Type":"ContainerStarted","Data":"1760c5e80c5800750ac77370ad020e3b8342db7dc62cf3b6003a5185e40f564d"} Nov 24 14:57:06 crc kubenswrapper[4790]: I1124 14:57:06.316613 4790 scope.go:117] "RemoveContainer" containerID="28308703817a1d4e0e5344ce19268127c600894f2daa6b921d52a9105f014132" Nov 24 14:57:06 crc kubenswrapper[4790]: E1124 14:57:06.317408 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:57:06 crc kubenswrapper[4790]: I1124 14:57:06.754213 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"b75fcca8-34a3-42d4-bfca-43c21b63227b","Type":"ContainerStarted","Data":"947e9705682bb0e007b4868a480f3805195c0de2efda18a8924876013a95987b"} Nov 24 14:57:06 crc kubenswrapper[4790]: I1124 14:57:06.758415 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68bc6cdf65-975r7" event={"ID":"cd78a936-5519-480a-b36a-a6ec905cd694","Type":"ContainerStarted","Data":"15d6d72023d3e6da20d416deec5568184431107042f2be4b7beaa76570629cdc"} Nov 24 14:57:06 crc kubenswrapper[4790]: I1124 14:57:06.758597 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-68bc6cdf65-975r7" Nov 24 14:57:06 crc kubenswrapper[4790]: I1124 14:57:06.761694 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"d42b6e7d-ef2f-4d91-b9d0-138efb9549eb","Type":"ContainerStarted","Data":"0072ae0c535e2b56ae06692e3e67bb80c60c0089bd3c3a94945674ec9a2e6009"} Nov 24 14:57:06 crc kubenswrapper[4790]: I1124 14:57:06.782679 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-68bc6cdf65-975r7" podStartSLOduration=2.782656079 podStartE2EDuration="2.782656079s" podCreationTimestamp="2025-11-24 14:57:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:57:06.779584261 +0000 UTC m=+6275.159477923" watchObservedRunningTime="2025-11-24 14:57:06.782656079 +0000 UTC m=+6275.162549741" Nov 24 14:57:07 crc kubenswrapper[4790]: I1124 14:57:07.774281 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"b75fcca8-34a3-42d4-bfca-43c21b63227b","Type":"ContainerStarted","Data":"a7b28e8800662d4f19c36c40b013b86ed17db4727e436bbe154d56030f2923d4"} Nov 24 14:57:07 crc kubenswrapper[4790]: I1124 14:57:07.779066 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"d42b6e7d-ef2f-4d91-b9d0-138efb9549eb","Type":"ContainerStarted","Data":"53d19343158f92ebf44c63acd4481bcb21ffc8ecb95a673164049f01b50e9b17"} Nov 24 14:57:07 crc kubenswrapper[4790]: I1124 14:57:07.779191 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Nov 24 14:57:07 crc kubenswrapper[4790]: I1124 14:57:07.802064 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.146014002 podStartE2EDuration="3.802044165s" podCreationTimestamp="2025-11-24 14:57:04 +0000 UTC" firstStartedPulling="2025-11-24 14:57:04.958955605 +0000 UTC m=+6273.338849267" lastFinishedPulling="2025-11-24 14:57:05.614985768 +0000 UTC m=+6273.994879430" observedRunningTime="2025-11-24 14:57:07.79510824 +0000 UTC m=+6276.175001902" watchObservedRunningTime="2025-11-24 14:57:07.802044165 +0000 UTC m=+6276.181937827" Nov 24 14:57:07 crc kubenswrapper[4790]: I1124 14:57:07.830719 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=3.830697878 podStartE2EDuration="3.830697878s" podCreationTimestamp="2025-11-24 14:57:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:57:07.822418569 +0000 UTC m=+6276.202312231" watchObservedRunningTime="2025-11-24 14:57:07.830697878 +0000 UTC m=+6276.210591540" Nov 24 14:57:09 crc kubenswrapper[4790]: I1124 14:57:09.003853 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 14:57:12 crc kubenswrapper[4790]: I1124 14:57:12.837105 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"6cacb59b-e19b-44b9-b4c4-3a12559e64ed","Type":"ContainerStarted","Data":"fb8691e722cdac461a9c5d99bcce5a435c55bcc5dae60164f031c6a25bc1db65"} Nov 24 14:57:12 crc kubenswrapper[4790]: I1124 14:57:12.837535 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"6cacb59b-e19b-44b9-b4c4-3a12559e64ed","Type":"ContainerStarted","Data":"53063430f4c43e7e2fb435fa95e40e5b41d69f8e86111f2885a86b8c212a699d"} Nov 24 14:57:12 crc kubenswrapper[4790]: I1124 14:57:12.876192 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=2.584272521 podStartE2EDuration="8.876130912s" podCreationTimestamp="2025-11-24 14:57:04 +0000 UTC" firstStartedPulling="2025-11-24 14:57:05.195813335 +0000 UTC m=+6273.575706987" lastFinishedPulling="2025-11-24 14:57:11.487671716 +0000 UTC m=+6279.867565378" observedRunningTime="2025-11-24 14:57:12.859709047 +0000 UTC m=+6281.239602729" watchObservedRunningTime="2025-11-24 14:57:12.876130912 +0000 UTC m=+6281.256024584" Nov 24 14:57:14 crc kubenswrapper[4790]: I1124 14:57:14.391628 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Nov 24 14:57:14 crc kubenswrapper[4790]: I1124 14:57:14.433840 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Nov 24 14:57:14 crc kubenswrapper[4790]: I1124 14:57:14.482159 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-68bc6cdf65-975r7" Nov 24 14:57:14 crc kubenswrapper[4790]: I1124 14:57:14.570617 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76c6d786bf-q6t82"] Nov 24 14:57:14 crc kubenswrapper[4790]: I1124 14:57:14.570988 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-76c6d786bf-q6t82" podUID="20f13a25-af88-4161-ae40-d4a73a9a28f5" containerName="dnsmasq-dns" containerID="cri-o://038cdb82e65884f32057a1a3349f4d2a6e48160457f6d2effd5f1379b132367c" gracePeriod=10 Nov 24 14:57:14 crc kubenswrapper[4790]: I1124 14:57:14.887101 4790 generic.go:334] "Generic (PLEG): container finished" podID="20f13a25-af88-4161-ae40-d4a73a9a28f5" containerID="038cdb82e65884f32057a1a3349f4d2a6e48160457f6d2effd5f1379b132367c" exitCode=0 Nov 24 14:57:14 crc kubenswrapper[4790]: I1124 14:57:14.887301 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76c6d786bf-q6t82" event={"ID":"20f13a25-af88-4161-ae40-d4a73a9a28f5","Type":"ContainerDied","Data":"038cdb82e65884f32057a1a3349f4d2a6e48160457f6d2effd5f1379b132367c"} Nov 24 14:57:15 crc kubenswrapper[4790]: I1124 14:57:15.179858 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76c6d786bf-q6t82" Nov 24 14:57:15 crc kubenswrapper[4790]: I1124 14:57:15.204668 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20f13a25-af88-4161-ae40-d4a73a9a28f5-config\") pod \"20f13a25-af88-4161-ae40-d4a73a9a28f5\" (UID: \"20f13a25-af88-4161-ae40-d4a73a9a28f5\") " Nov 24 14:57:15 crc kubenswrapper[4790]: I1124 14:57:15.204895 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20f13a25-af88-4161-ae40-d4a73a9a28f5-ovsdbserver-sb\") pod \"20f13a25-af88-4161-ae40-d4a73a9a28f5\" (UID: \"20f13a25-af88-4161-ae40-d4a73a9a28f5\") " Nov 24 14:57:15 crc kubenswrapper[4790]: I1124 14:57:15.204921 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20f13a25-af88-4161-ae40-d4a73a9a28f5-ovsdbserver-nb\") pod \"20f13a25-af88-4161-ae40-d4a73a9a28f5\" (UID: \"20f13a25-af88-4161-ae40-d4a73a9a28f5\") " Nov 24 14:57:15 crc kubenswrapper[4790]: I1124 14:57:15.204947 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4hws\" (UniqueName: \"kubernetes.io/projected/20f13a25-af88-4161-ae40-d4a73a9a28f5-kube-api-access-b4hws\") pod \"20f13a25-af88-4161-ae40-d4a73a9a28f5\" (UID: \"20f13a25-af88-4161-ae40-d4a73a9a28f5\") " Nov 24 14:57:15 crc kubenswrapper[4790]: I1124 14:57:15.204972 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20f13a25-af88-4161-ae40-d4a73a9a28f5-dns-svc\") pod \"20f13a25-af88-4161-ae40-d4a73a9a28f5\" (UID: \"20f13a25-af88-4161-ae40-d4a73a9a28f5\") " Nov 24 14:57:15 crc kubenswrapper[4790]: I1124 14:57:15.223998 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20f13a25-af88-4161-ae40-d4a73a9a28f5-kube-api-access-b4hws" (OuterVolumeSpecName: "kube-api-access-b4hws") pod "20f13a25-af88-4161-ae40-d4a73a9a28f5" (UID: "20f13a25-af88-4161-ae40-d4a73a9a28f5"). InnerVolumeSpecName "kube-api-access-b4hws". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:57:15 crc kubenswrapper[4790]: I1124 14:57:15.277135 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20f13a25-af88-4161-ae40-d4a73a9a28f5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "20f13a25-af88-4161-ae40-d4a73a9a28f5" (UID: "20f13a25-af88-4161-ae40-d4a73a9a28f5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:57:15 crc kubenswrapper[4790]: I1124 14:57:15.281840 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20f13a25-af88-4161-ae40-d4a73a9a28f5-config" (OuterVolumeSpecName: "config") pod "20f13a25-af88-4161-ae40-d4a73a9a28f5" (UID: "20f13a25-af88-4161-ae40-d4a73a9a28f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:57:15 crc kubenswrapper[4790]: I1124 14:57:15.307114 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20f13a25-af88-4161-ae40-d4a73a9a28f5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "20f13a25-af88-4161-ae40-d4a73a9a28f5" (UID: "20f13a25-af88-4161-ae40-d4a73a9a28f5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:57:15 crc kubenswrapper[4790]: I1124 14:57:15.307663 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20f13a25-af88-4161-ae40-d4a73a9a28f5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:57:15 crc kubenswrapper[4790]: I1124 14:57:15.307692 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4hws\" (UniqueName: \"kubernetes.io/projected/20f13a25-af88-4161-ae40-d4a73a9a28f5-kube-api-access-b4hws\") on node \"crc\" DevicePath \"\"" Nov 24 14:57:15 crc kubenswrapper[4790]: I1124 14:57:15.307704 4790 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20f13a25-af88-4161-ae40-d4a73a9a28f5-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:57:15 crc kubenswrapper[4790]: I1124 14:57:15.307712 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20f13a25-af88-4161-ae40-d4a73a9a28f5-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:57:15 crc kubenswrapper[4790]: I1124 14:57:15.311958 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20f13a25-af88-4161-ae40-d4a73a9a28f5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "20f13a25-af88-4161-ae40-d4a73a9a28f5" (UID: "20f13a25-af88-4161-ae40-d4a73a9a28f5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:57:15 crc kubenswrapper[4790]: I1124 14:57:15.410401 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20f13a25-af88-4161-ae40-d4a73a9a28f5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:57:15 crc kubenswrapper[4790]: I1124 14:57:15.898163 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76c6d786bf-q6t82" event={"ID":"20f13a25-af88-4161-ae40-d4a73a9a28f5","Type":"ContainerDied","Data":"77e613a81a3dc8f8c9ee402a58a37f92813cb995a5478e09f37c8aedb2d19c6a"} Nov 24 14:57:15 crc kubenswrapper[4790]: I1124 14:57:15.898222 4790 scope.go:117] "RemoveContainer" containerID="038cdb82e65884f32057a1a3349f4d2a6e48160457f6d2effd5f1379b132367c" Nov 24 14:57:15 crc kubenswrapper[4790]: I1124 14:57:15.898525 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76c6d786bf-q6t82" Nov 24 14:57:15 crc kubenswrapper[4790]: I1124 14:57:15.932207 4790 scope.go:117] "RemoveContainer" containerID="a5c1f2159c9241a5c830fda662bbbc4261983da3d535ef066afb4531c13910be" Nov 24 14:57:15 crc kubenswrapper[4790]: I1124 14:57:15.934296 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76c6d786bf-q6t82"] Nov 24 14:57:15 crc kubenswrapper[4790]: I1124 14:57:15.944865 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-76c6d786bf-q6t82"] Nov 24 14:57:16 crc kubenswrapper[4790]: I1124 14:57:16.326196 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20f13a25-af88-4161-ae40-d4a73a9a28f5" path="/var/lib/kubelet/pods/20f13a25-af88-4161-ae40-d4a73a9a28f5/volumes" Nov 24 14:57:16 crc kubenswrapper[4790]: I1124 14:57:16.741326 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:57:16 crc kubenswrapper[4790]: I1124 14:57:16.742006 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14" containerName="ceilometer-central-agent" containerID="cri-o://e10c8b46ab94c645d1a8865eb9a593912da588c59b9227d9107546a1cd6af700" gracePeriod=30 Nov 24 14:57:16 crc kubenswrapper[4790]: I1124 14:57:16.742118 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14" containerName="sg-core" containerID="cri-o://bd0d4c2dedb6b07f3d2455f84a347407e05895962ff554e3dbc9964f5bb4d8c2" gracePeriod=30 Nov 24 14:57:16 crc kubenswrapper[4790]: I1124 14:57:16.742174 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14" containerName="ceilometer-notification-agent" containerID="cri-o://fd0c27b4cee7b039c2b1bb808ab46734e60417816574192b3f8dc2ca2377ad38" gracePeriod=30 Nov 24 14:57:16 crc kubenswrapper[4790]: I1124 14:57:16.742119 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14" containerName="proxy-httpd" containerID="cri-o://f810e3edfc208b53e611205b6c99a0364809d4a1fa049e015dfc0f7de2e5f79f" gracePeriod=30 Nov 24 14:57:16 crc kubenswrapper[4790]: I1124 14:57:16.913165 4790 generic.go:334] "Generic (PLEG): container finished" podID="4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14" containerID="bd0d4c2dedb6b07f3d2455f84a347407e05895962ff554e3dbc9964f5bb4d8c2" exitCode=2 Nov 24 14:57:16 crc kubenswrapper[4790]: I1124 14:57:16.913223 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14","Type":"ContainerDied","Data":"bd0d4c2dedb6b07f3d2455f84a347407e05895962ff554e3dbc9964f5bb4d8c2"} Nov 24 14:57:17 crc kubenswrapper[4790]: E1124 14:57:17.760348 4790 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a94207f_fc4f_4f38_8c7e_2b1ebfbfbe14.slice/crio-conmon-fd0c27b4cee7b039c2b1bb808ab46734e60417816574192b3f8dc2ca2377ad38.scope\": RecentStats: unable to find data in memory cache]" Nov 24 14:57:17 crc kubenswrapper[4790]: I1124 14:57:17.854306 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:57:17 crc kubenswrapper[4790]: I1124 14:57:17.931286 4790 generic.go:334] "Generic (PLEG): container finished" podID="4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14" containerID="f810e3edfc208b53e611205b6c99a0364809d4a1fa049e015dfc0f7de2e5f79f" exitCode=0 Nov 24 14:57:17 crc kubenswrapper[4790]: I1124 14:57:17.931849 4790 generic.go:334] "Generic (PLEG): container finished" podID="4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14" containerID="fd0c27b4cee7b039c2b1bb808ab46734e60417816574192b3f8dc2ca2377ad38" exitCode=0 Nov 24 14:57:17 crc kubenswrapper[4790]: I1124 14:57:17.931987 4790 generic.go:334] "Generic (PLEG): container finished" podID="4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14" containerID="e10c8b46ab94c645d1a8865eb9a593912da588c59b9227d9107546a1cd6af700" exitCode=0 Nov 24 14:57:17 crc kubenswrapper[4790]: I1124 14:57:17.932085 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14","Type":"ContainerDied","Data":"f810e3edfc208b53e611205b6c99a0364809d4a1fa049e015dfc0f7de2e5f79f"} Nov 24 14:57:17 crc kubenswrapper[4790]: I1124 14:57:17.932193 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14","Type":"ContainerDied","Data":"fd0c27b4cee7b039c2b1bb808ab46734e60417816574192b3f8dc2ca2377ad38"} Nov 24 14:57:17 crc kubenswrapper[4790]: I1124 14:57:17.932271 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14","Type":"ContainerDied","Data":"e10c8b46ab94c645d1a8865eb9a593912da588c59b9227d9107546a1cd6af700"} Nov 24 14:57:17 crc kubenswrapper[4790]: I1124 14:57:17.932351 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14","Type":"ContainerDied","Data":"0dce2ace66e5388b6769eb529a3cc7b387fc98703807fb8708466301f9137ec9"} Nov 24 14:57:17 crc kubenswrapper[4790]: I1124 14:57:17.932435 4790 scope.go:117] "RemoveContainer" containerID="f810e3edfc208b53e611205b6c99a0364809d4a1fa049e015dfc0f7de2e5f79f" Nov 24 14:57:17 crc kubenswrapper[4790]: I1124 14:57:17.932668 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:57:17 crc kubenswrapper[4790]: I1124 14:57:17.954850 4790 scope.go:117] "RemoveContainer" containerID="bd0d4c2dedb6b07f3d2455f84a347407e05895962ff554e3dbc9964f5bb4d8c2" Nov 24 14:57:17 crc kubenswrapper[4790]: I1124 14:57:17.979294 4790 scope.go:117] "RemoveContainer" containerID="fd0c27b4cee7b039c2b1bb808ab46734e60417816574192b3f8dc2ca2377ad38" Nov 24 14:57:17 crc kubenswrapper[4790]: I1124 14:57:17.987495 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-combined-ca-bundle\") pod \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\" (UID: \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\") " Nov 24 14:57:17 crc kubenswrapper[4790]: I1124 14:57:17.987581 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-config-data\") pod \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\" (UID: \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\") " Nov 24 14:57:17 crc kubenswrapper[4790]: I1124 14:57:17.987662 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-run-httpd\") pod \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\" (UID: \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\") " Nov 24 14:57:17 crc kubenswrapper[4790]: I1124 14:57:17.987705 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-scripts\") pod \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\" (UID: \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\") " Nov 24 14:57:17 crc kubenswrapper[4790]: I1124 14:57:17.987761 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-sg-core-conf-yaml\") pod \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\" (UID: \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\") " Nov 24 14:57:17 crc kubenswrapper[4790]: I1124 14:57:17.987798 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffcvp\" (UniqueName: \"kubernetes.io/projected/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-kube-api-access-ffcvp\") pod \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\" (UID: \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\") " Nov 24 14:57:17 crc kubenswrapper[4790]: I1124 14:57:17.987845 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-log-httpd\") pod \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\" (UID: \"4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14\") " Nov 24 14:57:17 crc kubenswrapper[4790]: I1124 14:57:17.988066 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14" (UID: "4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:57:17 crc kubenswrapper[4790]: I1124 14:57:17.988385 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14" (UID: "4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:57:17 crc kubenswrapper[4790]: I1124 14:57:17.988473 4790 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 14:57:17 crc kubenswrapper[4790]: I1124 14:57:17.992526 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-scripts" (OuterVolumeSpecName: "scripts") pod "4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14" (UID: "4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:57:17 crc kubenswrapper[4790]: I1124 14:57:17.992847 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-kube-api-access-ffcvp" (OuterVolumeSpecName: "kube-api-access-ffcvp") pod "4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14" (UID: "4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14"). InnerVolumeSpecName "kube-api-access-ffcvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.018303 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14" (UID: "4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.033964 4790 scope.go:117] "RemoveContainer" containerID="e10c8b46ab94c645d1a8865eb9a593912da588c59b9227d9107546a1cd6af700" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.075487 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14" (UID: "4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.095642 4790 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.095683 4790 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.095697 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffcvp\" (UniqueName: \"kubernetes.io/projected/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-kube-api-access-ffcvp\") on node \"crc\" DevicePath \"\"" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.095712 4790 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.095725 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.106984 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-config-data" (OuterVolumeSpecName: "config-data") pod "4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14" (UID: "4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.197993 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.205458 4790 scope.go:117] "RemoveContainer" containerID="f810e3edfc208b53e611205b6c99a0364809d4a1fa049e015dfc0f7de2e5f79f" Nov 24 14:57:18 crc kubenswrapper[4790]: E1124 14:57:18.205864 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f810e3edfc208b53e611205b6c99a0364809d4a1fa049e015dfc0f7de2e5f79f\": container with ID starting with f810e3edfc208b53e611205b6c99a0364809d4a1fa049e015dfc0f7de2e5f79f not found: ID does not exist" containerID="f810e3edfc208b53e611205b6c99a0364809d4a1fa049e015dfc0f7de2e5f79f" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.205910 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f810e3edfc208b53e611205b6c99a0364809d4a1fa049e015dfc0f7de2e5f79f"} err="failed to get container status \"f810e3edfc208b53e611205b6c99a0364809d4a1fa049e015dfc0f7de2e5f79f\": rpc error: code = NotFound desc = could not find container \"f810e3edfc208b53e611205b6c99a0364809d4a1fa049e015dfc0f7de2e5f79f\": container with ID starting with f810e3edfc208b53e611205b6c99a0364809d4a1fa049e015dfc0f7de2e5f79f not found: ID does not exist" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.205936 4790 scope.go:117] "RemoveContainer" containerID="bd0d4c2dedb6b07f3d2455f84a347407e05895962ff554e3dbc9964f5bb4d8c2" Nov 24 14:57:18 crc kubenswrapper[4790]: E1124 14:57:18.206349 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd0d4c2dedb6b07f3d2455f84a347407e05895962ff554e3dbc9964f5bb4d8c2\": container with ID starting with bd0d4c2dedb6b07f3d2455f84a347407e05895962ff554e3dbc9964f5bb4d8c2 not found: ID does not exist" containerID="bd0d4c2dedb6b07f3d2455f84a347407e05895962ff554e3dbc9964f5bb4d8c2" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.206381 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd0d4c2dedb6b07f3d2455f84a347407e05895962ff554e3dbc9964f5bb4d8c2"} err="failed to get container status \"bd0d4c2dedb6b07f3d2455f84a347407e05895962ff554e3dbc9964f5bb4d8c2\": rpc error: code = NotFound desc = could not find container \"bd0d4c2dedb6b07f3d2455f84a347407e05895962ff554e3dbc9964f5bb4d8c2\": container with ID starting with bd0d4c2dedb6b07f3d2455f84a347407e05895962ff554e3dbc9964f5bb4d8c2 not found: ID does not exist" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.206404 4790 scope.go:117] "RemoveContainer" containerID="fd0c27b4cee7b039c2b1bb808ab46734e60417816574192b3f8dc2ca2377ad38" Nov 24 14:57:18 crc kubenswrapper[4790]: E1124 14:57:18.206943 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd0c27b4cee7b039c2b1bb808ab46734e60417816574192b3f8dc2ca2377ad38\": container with ID starting with fd0c27b4cee7b039c2b1bb808ab46734e60417816574192b3f8dc2ca2377ad38 not found: ID does not exist" containerID="fd0c27b4cee7b039c2b1bb808ab46734e60417816574192b3f8dc2ca2377ad38" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.206970 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd0c27b4cee7b039c2b1bb808ab46734e60417816574192b3f8dc2ca2377ad38"} err="failed to get container status \"fd0c27b4cee7b039c2b1bb808ab46734e60417816574192b3f8dc2ca2377ad38\": rpc error: code = NotFound desc = could not find container \"fd0c27b4cee7b039c2b1bb808ab46734e60417816574192b3f8dc2ca2377ad38\": container with ID starting with fd0c27b4cee7b039c2b1bb808ab46734e60417816574192b3f8dc2ca2377ad38 not found: ID does not exist" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.206987 4790 scope.go:117] "RemoveContainer" containerID="e10c8b46ab94c645d1a8865eb9a593912da588c59b9227d9107546a1cd6af700" Nov 24 14:57:18 crc kubenswrapper[4790]: E1124 14:57:18.207444 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e10c8b46ab94c645d1a8865eb9a593912da588c59b9227d9107546a1cd6af700\": container with ID starting with e10c8b46ab94c645d1a8865eb9a593912da588c59b9227d9107546a1cd6af700 not found: ID does not exist" containerID="e10c8b46ab94c645d1a8865eb9a593912da588c59b9227d9107546a1cd6af700" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.207517 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e10c8b46ab94c645d1a8865eb9a593912da588c59b9227d9107546a1cd6af700"} err="failed to get container status \"e10c8b46ab94c645d1a8865eb9a593912da588c59b9227d9107546a1cd6af700\": rpc error: code = NotFound desc = could not find container \"e10c8b46ab94c645d1a8865eb9a593912da588c59b9227d9107546a1cd6af700\": container with ID starting with e10c8b46ab94c645d1a8865eb9a593912da588c59b9227d9107546a1cd6af700 not found: ID does not exist" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.207546 4790 scope.go:117] "RemoveContainer" containerID="f810e3edfc208b53e611205b6c99a0364809d4a1fa049e015dfc0f7de2e5f79f" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.207872 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f810e3edfc208b53e611205b6c99a0364809d4a1fa049e015dfc0f7de2e5f79f"} err="failed to get container status \"f810e3edfc208b53e611205b6c99a0364809d4a1fa049e015dfc0f7de2e5f79f\": rpc error: code = NotFound desc = could not find container \"f810e3edfc208b53e611205b6c99a0364809d4a1fa049e015dfc0f7de2e5f79f\": container with ID starting with f810e3edfc208b53e611205b6c99a0364809d4a1fa049e015dfc0f7de2e5f79f not found: ID does not exist" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.207910 4790 scope.go:117] "RemoveContainer" containerID="bd0d4c2dedb6b07f3d2455f84a347407e05895962ff554e3dbc9964f5bb4d8c2" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.208212 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd0d4c2dedb6b07f3d2455f84a347407e05895962ff554e3dbc9964f5bb4d8c2"} err="failed to get container status \"bd0d4c2dedb6b07f3d2455f84a347407e05895962ff554e3dbc9964f5bb4d8c2\": rpc error: code = NotFound desc = could not find container \"bd0d4c2dedb6b07f3d2455f84a347407e05895962ff554e3dbc9964f5bb4d8c2\": container with ID starting with bd0d4c2dedb6b07f3d2455f84a347407e05895962ff554e3dbc9964f5bb4d8c2 not found: ID does not exist" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.208254 4790 scope.go:117] "RemoveContainer" containerID="fd0c27b4cee7b039c2b1bb808ab46734e60417816574192b3f8dc2ca2377ad38" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.208541 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd0c27b4cee7b039c2b1bb808ab46734e60417816574192b3f8dc2ca2377ad38"} err="failed to get container status \"fd0c27b4cee7b039c2b1bb808ab46734e60417816574192b3f8dc2ca2377ad38\": rpc error: code = NotFound desc = could not find container \"fd0c27b4cee7b039c2b1bb808ab46734e60417816574192b3f8dc2ca2377ad38\": container with ID starting with fd0c27b4cee7b039c2b1bb808ab46734e60417816574192b3f8dc2ca2377ad38 not found: ID does not exist" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.208566 4790 scope.go:117] "RemoveContainer" containerID="e10c8b46ab94c645d1a8865eb9a593912da588c59b9227d9107546a1cd6af700" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.208829 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e10c8b46ab94c645d1a8865eb9a593912da588c59b9227d9107546a1cd6af700"} err="failed to get container status \"e10c8b46ab94c645d1a8865eb9a593912da588c59b9227d9107546a1cd6af700\": rpc error: code = NotFound desc = could not find container \"e10c8b46ab94c645d1a8865eb9a593912da588c59b9227d9107546a1cd6af700\": container with ID starting with e10c8b46ab94c645d1a8865eb9a593912da588c59b9227d9107546a1cd6af700 not found: ID does not exist" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.208851 4790 scope.go:117] "RemoveContainer" containerID="f810e3edfc208b53e611205b6c99a0364809d4a1fa049e015dfc0f7de2e5f79f" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.209188 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f810e3edfc208b53e611205b6c99a0364809d4a1fa049e015dfc0f7de2e5f79f"} err="failed to get container status \"f810e3edfc208b53e611205b6c99a0364809d4a1fa049e015dfc0f7de2e5f79f\": rpc error: code = NotFound desc = could not find container \"f810e3edfc208b53e611205b6c99a0364809d4a1fa049e015dfc0f7de2e5f79f\": container with ID starting with f810e3edfc208b53e611205b6c99a0364809d4a1fa049e015dfc0f7de2e5f79f not found: ID does not exist" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.209207 4790 scope.go:117] "RemoveContainer" containerID="bd0d4c2dedb6b07f3d2455f84a347407e05895962ff554e3dbc9964f5bb4d8c2" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.209462 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd0d4c2dedb6b07f3d2455f84a347407e05895962ff554e3dbc9964f5bb4d8c2"} err="failed to get container status \"bd0d4c2dedb6b07f3d2455f84a347407e05895962ff554e3dbc9964f5bb4d8c2\": rpc error: code = NotFound desc = could not find container \"bd0d4c2dedb6b07f3d2455f84a347407e05895962ff554e3dbc9964f5bb4d8c2\": container with ID starting with bd0d4c2dedb6b07f3d2455f84a347407e05895962ff554e3dbc9964f5bb4d8c2 not found: ID does not exist" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.209482 4790 scope.go:117] "RemoveContainer" containerID="fd0c27b4cee7b039c2b1bb808ab46734e60417816574192b3f8dc2ca2377ad38" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.209693 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd0c27b4cee7b039c2b1bb808ab46734e60417816574192b3f8dc2ca2377ad38"} err="failed to get container status \"fd0c27b4cee7b039c2b1bb808ab46734e60417816574192b3f8dc2ca2377ad38\": rpc error: code = NotFound desc = could not find container \"fd0c27b4cee7b039c2b1bb808ab46734e60417816574192b3f8dc2ca2377ad38\": container with ID starting with fd0c27b4cee7b039c2b1bb808ab46734e60417816574192b3f8dc2ca2377ad38 not found: ID does not exist" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.209715 4790 scope.go:117] "RemoveContainer" containerID="e10c8b46ab94c645d1a8865eb9a593912da588c59b9227d9107546a1cd6af700" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.209924 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e10c8b46ab94c645d1a8865eb9a593912da588c59b9227d9107546a1cd6af700"} err="failed to get container status \"e10c8b46ab94c645d1a8865eb9a593912da588c59b9227d9107546a1cd6af700\": rpc error: code = NotFound desc = could not find container \"e10c8b46ab94c645d1a8865eb9a593912da588c59b9227d9107546a1cd6af700\": container with ID starting with e10c8b46ab94c645d1a8865eb9a593912da588c59b9227d9107546a1cd6af700 not found: ID does not exist" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.266815 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.274287 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.296948 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:57:18 crc kubenswrapper[4790]: E1124 14:57:18.297353 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14" containerName="ceilometer-notification-agent" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.297369 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14" containerName="ceilometer-notification-agent" Nov 24 14:57:18 crc kubenswrapper[4790]: E1124 14:57:18.297393 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14" containerName="ceilometer-central-agent" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.297399 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14" containerName="ceilometer-central-agent" Nov 24 14:57:18 crc kubenswrapper[4790]: E1124 14:57:18.297417 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20f13a25-af88-4161-ae40-d4a73a9a28f5" containerName="init" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.297423 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="20f13a25-af88-4161-ae40-d4a73a9a28f5" containerName="init" Nov 24 14:57:18 crc kubenswrapper[4790]: E1124 14:57:18.297430 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20f13a25-af88-4161-ae40-d4a73a9a28f5" containerName="dnsmasq-dns" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.297435 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="20f13a25-af88-4161-ae40-d4a73a9a28f5" containerName="dnsmasq-dns" Nov 24 14:57:18 crc kubenswrapper[4790]: E1124 14:57:18.297459 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14" containerName="sg-core" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.297464 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14" containerName="sg-core" Nov 24 14:57:18 crc kubenswrapper[4790]: E1124 14:57:18.297475 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14" containerName="proxy-httpd" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.297480 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14" containerName="proxy-httpd" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.297661 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14" containerName="sg-core" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.297682 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="20f13a25-af88-4161-ae40-d4a73a9a28f5" containerName="dnsmasq-dns" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.297695 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14" containerName="ceilometer-central-agent" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.297711 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14" containerName="ceilometer-notification-agent" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.297723 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14" containerName="proxy-httpd" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.299661 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.305375 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.305979 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.331251 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14" path="/var/lib/kubelet/pods/4a94207f-fc4f-4f38-8c7e-2b1ebfbfbe14/volumes" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.332253 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.401772 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/01048e56-f62e-4007-b717-ac76629de1ed-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"01048e56-f62e-4007-b717-ac76629de1ed\") " pod="openstack/ceilometer-0" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.401830 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01048e56-f62e-4007-b717-ac76629de1ed-scripts\") pod \"ceilometer-0\" (UID: \"01048e56-f62e-4007-b717-ac76629de1ed\") " pod="openstack/ceilometer-0" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.402242 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bng9p\" (UniqueName: \"kubernetes.io/projected/01048e56-f62e-4007-b717-ac76629de1ed-kube-api-access-bng9p\") pod \"ceilometer-0\" (UID: \"01048e56-f62e-4007-b717-ac76629de1ed\") " pod="openstack/ceilometer-0" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.402298 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01048e56-f62e-4007-b717-ac76629de1ed-config-data\") pod \"ceilometer-0\" (UID: \"01048e56-f62e-4007-b717-ac76629de1ed\") " pod="openstack/ceilometer-0" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.402335 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01048e56-f62e-4007-b717-ac76629de1ed-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"01048e56-f62e-4007-b717-ac76629de1ed\") " pod="openstack/ceilometer-0" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.402436 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01048e56-f62e-4007-b717-ac76629de1ed-run-httpd\") pod \"ceilometer-0\" (UID: \"01048e56-f62e-4007-b717-ac76629de1ed\") " pod="openstack/ceilometer-0" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.402480 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01048e56-f62e-4007-b717-ac76629de1ed-log-httpd\") pod \"ceilometer-0\" (UID: \"01048e56-f62e-4007-b717-ac76629de1ed\") " pod="openstack/ceilometer-0" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.504155 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/01048e56-f62e-4007-b717-ac76629de1ed-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"01048e56-f62e-4007-b717-ac76629de1ed\") " pod="openstack/ceilometer-0" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.504218 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01048e56-f62e-4007-b717-ac76629de1ed-scripts\") pod \"ceilometer-0\" (UID: \"01048e56-f62e-4007-b717-ac76629de1ed\") " pod="openstack/ceilometer-0" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.504400 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bng9p\" (UniqueName: \"kubernetes.io/projected/01048e56-f62e-4007-b717-ac76629de1ed-kube-api-access-bng9p\") pod \"ceilometer-0\" (UID: \"01048e56-f62e-4007-b717-ac76629de1ed\") " pod="openstack/ceilometer-0" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.504435 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01048e56-f62e-4007-b717-ac76629de1ed-config-data\") pod \"ceilometer-0\" (UID: \"01048e56-f62e-4007-b717-ac76629de1ed\") " pod="openstack/ceilometer-0" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.504464 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01048e56-f62e-4007-b717-ac76629de1ed-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"01048e56-f62e-4007-b717-ac76629de1ed\") " pod="openstack/ceilometer-0" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.504507 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01048e56-f62e-4007-b717-ac76629de1ed-run-httpd\") pod \"ceilometer-0\" (UID: \"01048e56-f62e-4007-b717-ac76629de1ed\") " pod="openstack/ceilometer-0" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.504532 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01048e56-f62e-4007-b717-ac76629de1ed-log-httpd\") pod \"ceilometer-0\" (UID: \"01048e56-f62e-4007-b717-ac76629de1ed\") " pod="openstack/ceilometer-0" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.505097 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01048e56-f62e-4007-b717-ac76629de1ed-log-httpd\") pod \"ceilometer-0\" (UID: \"01048e56-f62e-4007-b717-ac76629de1ed\") " pod="openstack/ceilometer-0" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.505136 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01048e56-f62e-4007-b717-ac76629de1ed-run-httpd\") pod \"ceilometer-0\" (UID: \"01048e56-f62e-4007-b717-ac76629de1ed\") " pod="openstack/ceilometer-0" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.515803 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01048e56-f62e-4007-b717-ac76629de1ed-scripts\") pod \"ceilometer-0\" (UID: \"01048e56-f62e-4007-b717-ac76629de1ed\") " pod="openstack/ceilometer-0" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.517603 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01048e56-f62e-4007-b717-ac76629de1ed-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"01048e56-f62e-4007-b717-ac76629de1ed\") " pod="openstack/ceilometer-0" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.519480 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/01048e56-f62e-4007-b717-ac76629de1ed-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"01048e56-f62e-4007-b717-ac76629de1ed\") " pod="openstack/ceilometer-0" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.519781 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01048e56-f62e-4007-b717-ac76629de1ed-config-data\") pod \"ceilometer-0\" (UID: \"01048e56-f62e-4007-b717-ac76629de1ed\") " pod="openstack/ceilometer-0" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.527336 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bng9p\" (UniqueName: \"kubernetes.io/projected/01048e56-f62e-4007-b717-ac76629de1ed-kube-api-access-bng9p\") pod \"ceilometer-0\" (UID: \"01048e56-f62e-4007-b717-ac76629de1ed\") " pod="openstack/ceilometer-0" Nov 24 14:57:18 crc kubenswrapper[4790]: I1124 14:57:18.630662 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:57:19 crc kubenswrapper[4790]: I1124 14:57:19.080068 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:57:19 crc kubenswrapper[4790]: W1124 14:57:19.084684 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01048e56_f62e_4007_b717_ac76629de1ed.slice/crio-31811f31f164324c3aa9f99d15199e1141057c0d63f9647b42513cb36df16688 WatchSource:0}: Error finding container 31811f31f164324c3aa9f99d15199e1141057c0d63f9647b42513cb36df16688: Status 404 returned error can't find the container with id 31811f31f164324c3aa9f99d15199e1141057c0d63f9647b42513cb36df16688 Nov 24 14:57:19 crc kubenswrapper[4790]: I1124 14:57:19.958329 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01048e56-f62e-4007-b717-ac76629de1ed","Type":"ContainerStarted","Data":"eec2209e3cc13fcfb1fb9015db6b3a5e65dcd79db3c2b3ecb70eb8456b4709a2"} Nov 24 14:57:19 crc kubenswrapper[4790]: I1124 14:57:19.958907 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01048e56-f62e-4007-b717-ac76629de1ed","Type":"ContainerStarted","Data":"31811f31f164324c3aa9f99d15199e1141057c0d63f9647b42513cb36df16688"} Nov 24 14:57:20 crc kubenswrapper[4790]: I1124 14:57:20.999164 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01048e56-f62e-4007-b717-ac76629de1ed","Type":"ContainerStarted","Data":"74be2c3f2137f0114f3d7cc25b50f7777fc0b7d9e635eb54ede8596d8e857d89"} Nov 24 14:57:21 crc kubenswrapper[4790]: I1124 14:57:21.316774 4790 scope.go:117] "RemoveContainer" containerID="28308703817a1d4e0e5344ce19268127c600894f2daa6b921d52a9105f014132" Nov 24 14:57:21 crc kubenswrapper[4790]: E1124 14:57:21.317510 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:57:22 crc kubenswrapper[4790]: I1124 14:57:22.014696 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01048e56-f62e-4007-b717-ac76629de1ed","Type":"ContainerStarted","Data":"6dc413909e55c8e2244e5a836fc37f7826a144a5c3fc8f7f7fe95a92414a428d"} Nov 24 14:57:24 crc kubenswrapper[4790]: I1124 14:57:24.037772 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01048e56-f62e-4007-b717-ac76629de1ed","Type":"ContainerStarted","Data":"2c2614aa6a7644c4e8491858988f47af68b2259032fbfaf20ad546093c616b66"} Nov 24 14:57:24 crc kubenswrapper[4790]: I1124 14:57:24.038525 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 14:57:24 crc kubenswrapper[4790]: I1124 14:57:24.073285 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.9445853199999998 podStartE2EDuration="6.073264938s" podCreationTimestamp="2025-11-24 14:57:18 +0000 UTC" firstStartedPulling="2025-11-24 14:57:19.087584063 +0000 UTC m=+6287.467477745" lastFinishedPulling="2025-11-24 14:57:23.216263691 +0000 UTC m=+6291.596157363" observedRunningTime="2025-11-24 14:57:24.063667306 +0000 UTC m=+6292.443561008" watchObservedRunningTime="2025-11-24 14:57:24.073264938 +0000 UTC m=+6292.453158600" Nov 24 14:57:25 crc kubenswrapper[4790]: I1124 14:57:25.909211 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Nov 24 14:57:25 crc kubenswrapper[4790]: I1124 14:57:25.913329 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Nov 24 14:57:26 crc kubenswrapper[4790]: I1124 14:57:26.069634 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Nov 24 14:57:33 crc kubenswrapper[4790]: I1124 14:57:33.315979 4790 scope.go:117] "RemoveContainer" containerID="28308703817a1d4e0e5344ce19268127c600894f2daa6b921d52a9105f014132" Nov 24 14:57:33 crc kubenswrapper[4790]: E1124 14:57:33.316827 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 14:57:40 crc kubenswrapper[4790]: I1124 14:57:40.929347 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-87dpj"] Nov 24 14:57:40 crc kubenswrapper[4790]: I1124 14:57:40.934982 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-87dpj" Nov 24 14:57:40 crc kubenswrapper[4790]: I1124 14:57:40.943410 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-87dpj"] Nov 24 14:57:41 crc kubenswrapper[4790]: I1124 14:57:41.058351 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km4v2\" (UniqueName: \"kubernetes.io/projected/b9864022-64b2-41a9-bdc3-55beb82adb57-kube-api-access-km4v2\") pod \"community-operators-87dpj\" (UID: \"b9864022-64b2-41a9-bdc3-55beb82adb57\") " pod="openshift-marketplace/community-operators-87dpj" Nov 24 14:57:41 crc kubenswrapper[4790]: I1124 14:57:41.058754 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9864022-64b2-41a9-bdc3-55beb82adb57-utilities\") pod \"community-operators-87dpj\" (UID: \"b9864022-64b2-41a9-bdc3-55beb82adb57\") " pod="openshift-marketplace/community-operators-87dpj" Nov 24 14:57:41 crc kubenswrapper[4790]: I1124 14:57:41.058970 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9864022-64b2-41a9-bdc3-55beb82adb57-catalog-content\") pod \"community-operators-87dpj\" (UID: \"b9864022-64b2-41a9-bdc3-55beb82adb57\") " pod="openshift-marketplace/community-operators-87dpj" Nov 24 14:57:41 crc kubenswrapper[4790]: I1124 14:57:41.161587 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km4v2\" (UniqueName: \"kubernetes.io/projected/b9864022-64b2-41a9-bdc3-55beb82adb57-kube-api-access-km4v2\") pod \"community-operators-87dpj\" (UID: \"b9864022-64b2-41a9-bdc3-55beb82adb57\") " pod="openshift-marketplace/community-operators-87dpj" Nov 24 14:57:41 crc kubenswrapper[4790]: I1124 14:57:41.161661 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9864022-64b2-41a9-bdc3-55beb82adb57-utilities\") pod \"community-operators-87dpj\" (UID: \"b9864022-64b2-41a9-bdc3-55beb82adb57\") " pod="openshift-marketplace/community-operators-87dpj" Nov 24 14:57:41 crc kubenswrapper[4790]: I1124 14:57:41.161702 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9864022-64b2-41a9-bdc3-55beb82adb57-catalog-content\") pod \"community-operators-87dpj\" (UID: \"b9864022-64b2-41a9-bdc3-55beb82adb57\") " pod="openshift-marketplace/community-operators-87dpj" Nov 24 14:57:41 crc kubenswrapper[4790]: I1124 14:57:41.162233 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9864022-64b2-41a9-bdc3-55beb82adb57-catalog-content\") pod \"community-operators-87dpj\" (UID: \"b9864022-64b2-41a9-bdc3-55beb82adb57\") " pod="openshift-marketplace/community-operators-87dpj" Nov 24 14:57:41 crc kubenswrapper[4790]: I1124 14:57:41.162450 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9864022-64b2-41a9-bdc3-55beb82adb57-utilities\") pod \"community-operators-87dpj\" (UID: \"b9864022-64b2-41a9-bdc3-55beb82adb57\") " pod="openshift-marketplace/community-operators-87dpj" Nov 24 14:57:41 crc kubenswrapper[4790]: I1124 14:57:41.187210 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km4v2\" (UniqueName: \"kubernetes.io/projected/b9864022-64b2-41a9-bdc3-55beb82adb57-kube-api-access-km4v2\") pod \"community-operators-87dpj\" (UID: \"b9864022-64b2-41a9-bdc3-55beb82adb57\") " pod="openshift-marketplace/community-operators-87dpj" Nov 24 14:57:41 crc kubenswrapper[4790]: I1124 14:57:41.265595 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-87dpj" Nov 24 14:57:41 crc kubenswrapper[4790]: I1124 14:57:41.742063 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-87dpj"] Nov 24 14:57:41 crc kubenswrapper[4790]: W1124 14:57:41.747308 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9864022_64b2_41a9_bdc3_55beb82adb57.slice/crio-4024ade319f118db900efd85f389017ef544024cfde09abaa84def674ea1cf99 WatchSource:0}: Error finding container 4024ade319f118db900efd85f389017ef544024cfde09abaa84def674ea1cf99: Status 404 returned error can't find the container with id 4024ade319f118db900efd85f389017ef544024cfde09abaa84def674ea1cf99 Nov 24 14:57:42 crc kubenswrapper[4790]: I1124 14:57:42.255559 4790 generic.go:334] "Generic (PLEG): container finished" podID="b9864022-64b2-41a9-bdc3-55beb82adb57" containerID="5d2b195a9d75078ce95eaf1bdc6c31966858340f20a6b4b6d301c461b39e2045" exitCode=0 Nov 24 14:57:42 crc kubenswrapper[4790]: I1124 14:57:42.255828 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-87dpj" event={"ID":"b9864022-64b2-41a9-bdc3-55beb82adb57","Type":"ContainerDied","Data":"5d2b195a9d75078ce95eaf1bdc6c31966858340f20a6b4b6d301c461b39e2045"} Nov 24 14:57:42 crc kubenswrapper[4790]: I1124 14:57:42.255858 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-87dpj" event={"ID":"b9864022-64b2-41a9-bdc3-55beb82adb57","Type":"ContainerStarted","Data":"4024ade319f118db900efd85f389017ef544024cfde09abaa84def674ea1cf99"} Nov 24 14:57:47 crc kubenswrapper[4790]: I1124 14:57:47.317133 4790 scope.go:117] "RemoveContainer" containerID="28308703817a1d4e0e5344ce19268127c600894f2daa6b921d52a9105f014132" Nov 24 14:57:47 crc kubenswrapper[4790]: I1124 14:57:47.318506 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-87dpj" event={"ID":"b9864022-64b2-41a9-bdc3-55beb82adb57","Type":"ContainerStarted","Data":"a8b7d9b7936b18c47586475f9b81c4543aeff3843d1bc835160517dd57c44375"} Nov 24 14:57:48 crc kubenswrapper[4790]: I1124 14:57:48.333581 4790 generic.go:334] "Generic (PLEG): container finished" podID="b9864022-64b2-41a9-bdc3-55beb82adb57" containerID="a8b7d9b7936b18c47586475f9b81c4543aeff3843d1bc835160517dd57c44375" exitCode=0 Nov 24 14:57:48 crc kubenswrapper[4790]: I1124 14:57:48.336637 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-87dpj" event={"ID":"b9864022-64b2-41a9-bdc3-55beb82adb57","Type":"ContainerDied","Data":"a8b7d9b7936b18c47586475f9b81c4543aeff3843d1bc835160517dd57c44375"} Nov 24 14:57:48 crc kubenswrapper[4790]: I1124 14:57:48.338179 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"848bb411b4e62828968580971f8aaefd4e3c2cc50ce3c086c482e5c081b37c96"} Nov 24 14:57:48 crc kubenswrapper[4790]: I1124 14:57:48.638598 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 14:57:49 crc kubenswrapper[4790]: I1124 14:57:49.356164 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-87dpj" event={"ID":"b9864022-64b2-41a9-bdc3-55beb82adb57","Type":"ContainerStarted","Data":"7214f327b5c535944bee9ab487c874d0dc452a618cff958d081997893e299e1f"} Nov 24 14:57:49 crc kubenswrapper[4790]: I1124 14:57:49.383795 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-87dpj" podStartSLOduration=2.777327923 podStartE2EDuration="9.383773567s" podCreationTimestamp="2025-11-24 14:57:40 +0000 UTC" firstStartedPulling="2025-11-24 14:57:42.258454292 +0000 UTC m=+6310.638347964" lastFinishedPulling="2025-11-24 14:57:48.864899946 +0000 UTC m=+6317.244793608" observedRunningTime="2025-11-24 14:57:49.376082182 +0000 UTC m=+6317.755975844" watchObservedRunningTime="2025-11-24 14:57:49.383773567 +0000 UTC m=+6317.763667229" Nov 24 14:57:51 crc kubenswrapper[4790]: I1124 14:57:51.266638 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-87dpj" Nov 24 14:57:51 crc kubenswrapper[4790]: I1124 14:57:51.267096 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-87dpj" Nov 24 14:57:52 crc kubenswrapper[4790]: I1124 14:57:52.334686 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-87dpj" podUID="b9864022-64b2-41a9-bdc3-55beb82adb57" containerName="registry-server" probeResult="failure" output=< Nov 24 14:57:52 crc kubenswrapper[4790]: timeout: failed to connect service ":50051" within 1s Nov 24 14:57:52 crc kubenswrapper[4790]: > Nov 24 14:58:01 crc kubenswrapper[4790]: I1124 14:58:01.322780 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-87dpj" Nov 24 14:58:01 crc kubenswrapper[4790]: I1124 14:58:01.397686 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-87dpj" Nov 24 14:58:02 crc kubenswrapper[4790]: I1124 14:58:02.568821 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-87dpj"] Nov 24 14:58:02 crc kubenswrapper[4790]: I1124 14:58:02.964824 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pdbm7"] Nov 24 14:58:02 crc kubenswrapper[4790]: I1124 14:58:02.965960 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pdbm7" podUID="1a415403-8f42-4dd2-9387-30b9c2af07b3" containerName="registry-server" containerID="cri-o://bb4a95744b54bbbdb68dbd8cd4a650b7a66aa6135408708c5cfc9b0014b04512" gracePeriod=2 Nov 24 14:58:03 crc kubenswrapper[4790]: I1124 14:58:03.510864 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pdbm7" Nov 24 14:58:03 crc kubenswrapper[4790]: I1124 14:58:03.537475 4790 generic.go:334] "Generic (PLEG): container finished" podID="1a415403-8f42-4dd2-9387-30b9c2af07b3" containerID="bb4a95744b54bbbdb68dbd8cd4a650b7a66aa6135408708c5cfc9b0014b04512" exitCode=0 Nov 24 14:58:03 crc kubenswrapper[4790]: I1124 14:58:03.537514 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pdbm7" event={"ID":"1a415403-8f42-4dd2-9387-30b9c2af07b3","Type":"ContainerDied","Data":"bb4a95744b54bbbdb68dbd8cd4a650b7a66aa6135408708c5cfc9b0014b04512"} Nov 24 14:58:03 crc kubenswrapper[4790]: I1124 14:58:03.537538 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pdbm7" event={"ID":"1a415403-8f42-4dd2-9387-30b9c2af07b3","Type":"ContainerDied","Data":"79b36d8dc97d19c7bec48f96ab057cc114dec4e3c1b2506301f1ed2fb1b73246"} Nov 24 14:58:03 crc kubenswrapper[4790]: I1124 14:58:03.537555 4790 scope.go:117] "RemoveContainer" containerID="bb4a95744b54bbbdb68dbd8cd4a650b7a66aa6135408708c5cfc9b0014b04512" Nov 24 14:58:03 crc kubenswrapper[4790]: I1124 14:58:03.537637 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pdbm7" Nov 24 14:58:03 crc kubenswrapper[4790]: I1124 14:58:03.567477 4790 scope.go:117] "RemoveContainer" containerID="546b8f3b04dd9052f1c204a7eb7dd6b20cd306ecb9342e506d76e403d6f73bb8" Nov 24 14:58:03 crc kubenswrapper[4790]: I1124 14:58:03.594509 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a415403-8f42-4dd2-9387-30b9c2af07b3-catalog-content\") pod \"1a415403-8f42-4dd2-9387-30b9c2af07b3\" (UID: \"1a415403-8f42-4dd2-9387-30b9c2af07b3\") " Nov 24 14:58:03 crc kubenswrapper[4790]: I1124 14:58:03.594551 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4lnb\" (UniqueName: \"kubernetes.io/projected/1a415403-8f42-4dd2-9387-30b9c2af07b3-kube-api-access-q4lnb\") pod \"1a415403-8f42-4dd2-9387-30b9c2af07b3\" (UID: \"1a415403-8f42-4dd2-9387-30b9c2af07b3\") " Nov 24 14:58:03 crc kubenswrapper[4790]: I1124 14:58:03.594635 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a415403-8f42-4dd2-9387-30b9c2af07b3-utilities\") pod \"1a415403-8f42-4dd2-9387-30b9c2af07b3\" (UID: \"1a415403-8f42-4dd2-9387-30b9c2af07b3\") " Nov 24 14:58:03 crc kubenswrapper[4790]: I1124 14:58:03.595114 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a415403-8f42-4dd2-9387-30b9c2af07b3-utilities" (OuterVolumeSpecName: "utilities") pod "1a415403-8f42-4dd2-9387-30b9c2af07b3" (UID: "1a415403-8f42-4dd2-9387-30b9c2af07b3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:58:03 crc kubenswrapper[4790]: I1124 14:58:03.600462 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a415403-8f42-4dd2-9387-30b9c2af07b3-kube-api-access-q4lnb" (OuterVolumeSpecName: "kube-api-access-q4lnb") pod "1a415403-8f42-4dd2-9387-30b9c2af07b3" (UID: "1a415403-8f42-4dd2-9387-30b9c2af07b3"). InnerVolumeSpecName "kube-api-access-q4lnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:58:03 crc kubenswrapper[4790]: I1124 14:58:03.600928 4790 scope.go:117] "RemoveContainer" containerID="8d5064d024f41a27cd92a7b60a8d0f8704660ecde1ea290c63f128a03b91c586" Nov 24 14:58:03 crc kubenswrapper[4790]: I1124 14:58:03.677102 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a415403-8f42-4dd2-9387-30b9c2af07b3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1a415403-8f42-4dd2-9387-30b9c2af07b3" (UID: "1a415403-8f42-4dd2-9387-30b9c2af07b3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:58:03 crc kubenswrapper[4790]: I1124 14:58:03.679365 4790 scope.go:117] "RemoveContainer" containerID="bb4a95744b54bbbdb68dbd8cd4a650b7a66aa6135408708c5cfc9b0014b04512" Nov 24 14:58:03 crc kubenswrapper[4790]: E1124 14:58:03.679817 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb4a95744b54bbbdb68dbd8cd4a650b7a66aa6135408708c5cfc9b0014b04512\": container with ID starting with bb4a95744b54bbbdb68dbd8cd4a650b7a66aa6135408708c5cfc9b0014b04512 not found: ID does not exist" containerID="bb4a95744b54bbbdb68dbd8cd4a650b7a66aa6135408708c5cfc9b0014b04512" Nov 24 14:58:03 crc kubenswrapper[4790]: I1124 14:58:03.679865 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb4a95744b54bbbdb68dbd8cd4a650b7a66aa6135408708c5cfc9b0014b04512"} err="failed to get container status \"bb4a95744b54bbbdb68dbd8cd4a650b7a66aa6135408708c5cfc9b0014b04512\": rpc error: code = NotFound desc = could not find container \"bb4a95744b54bbbdb68dbd8cd4a650b7a66aa6135408708c5cfc9b0014b04512\": container with ID starting with bb4a95744b54bbbdb68dbd8cd4a650b7a66aa6135408708c5cfc9b0014b04512 not found: ID does not exist" Nov 24 14:58:03 crc kubenswrapper[4790]: I1124 14:58:03.679908 4790 scope.go:117] "RemoveContainer" containerID="546b8f3b04dd9052f1c204a7eb7dd6b20cd306ecb9342e506d76e403d6f73bb8" Nov 24 14:58:03 crc kubenswrapper[4790]: E1124 14:58:03.680216 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"546b8f3b04dd9052f1c204a7eb7dd6b20cd306ecb9342e506d76e403d6f73bb8\": container with ID starting with 546b8f3b04dd9052f1c204a7eb7dd6b20cd306ecb9342e506d76e403d6f73bb8 not found: ID does not exist" containerID="546b8f3b04dd9052f1c204a7eb7dd6b20cd306ecb9342e506d76e403d6f73bb8" Nov 24 14:58:03 crc kubenswrapper[4790]: I1124 14:58:03.680255 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"546b8f3b04dd9052f1c204a7eb7dd6b20cd306ecb9342e506d76e403d6f73bb8"} err="failed to get container status \"546b8f3b04dd9052f1c204a7eb7dd6b20cd306ecb9342e506d76e403d6f73bb8\": rpc error: code = NotFound desc = could not find container \"546b8f3b04dd9052f1c204a7eb7dd6b20cd306ecb9342e506d76e403d6f73bb8\": container with ID starting with 546b8f3b04dd9052f1c204a7eb7dd6b20cd306ecb9342e506d76e403d6f73bb8 not found: ID does not exist" Nov 24 14:58:03 crc kubenswrapper[4790]: I1124 14:58:03.680270 4790 scope.go:117] "RemoveContainer" containerID="8d5064d024f41a27cd92a7b60a8d0f8704660ecde1ea290c63f128a03b91c586" Nov 24 14:58:03 crc kubenswrapper[4790]: E1124 14:58:03.680587 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d5064d024f41a27cd92a7b60a8d0f8704660ecde1ea290c63f128a03b91c586\": container with ID starting with 8d5064d024f41a27cd92a7b60a8d0f8704660ecde1ea290c63f128a03b91c586 not found: ID does not exist" containerID="8d5064d024f41a27cd92a7b60a8d0f8704660ecde1ea290c63f128a03b91c586" Nov 24 14:58:03 crc kubenswrapper[4790]: I1124 14:58:03.680608 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d5064d024f41a27cd92a7b60a8d0f8704660ecde1ea290c63f128a03b91c586"} err="failed to get container status \"8d5064d024f41a27cd92a7b60a8d0f8704660ecde1ea290c63f128a03b91c586\": rpc error: code = NotFound desc = could not find container \"8d5064d024f41a27cd92a7b60a8d0f8704660ecde1ea290c63f128a03b91c586\": container with ID starting with 8d5064d024f41a27cd92a7b60a8d0f8704660ecde1ea290c63f128a03b91c586 not found: ID does not exist" Nov 24 14:58:03 crc kubenswrapper[4790]: I1124 14:58:03.700612 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a415403-8f42-4dd2-9387-30b9c2af07b3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:58:03 crc kubenswrapper[4790]: I1124 14:58:03.700650 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4lnb\" (UniqueName: \"kubernetes.io/projected/1a415403-8f42-4dd2-9387-30b9c2af07b3-kube-api-access-q4lnb\") on node \"crc\" DevicePath \"\"" Nov 24 14:58:03 crc kubenswrapper[4790]: I1124 14:58:03.700663 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a415403-8f42-4dd2-9387-30b9c2af07b3-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:58:03 crc kubenswrapper[4790]: I1124 14:58:03.869414 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pdbm7"] Nov 24 14:58:03 crc kubenswrapper[4790]: I1124 14:58:03.882441 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pdbm7"] Nov 24 14:58:04 crc kubenswrapper[4790]: I1124 14:58:04.327323 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a415403-8f42-4dd2-9387-30b9c2af07b3" path="/var/lib/kubelet/pods/1a415403-8f42-4dd2-9387-30b9c2af07b3/volumes" Nov 24 14:58:12 crc kubenswrapper[4790]: I1124 14:58:12.793062 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d447559f9-hl956"] Nov 24 14:58:12 crc kubenswrapper[4790]: E1124 14:58:12.795014 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a415403-8f42-4dd2-9387-30b9c2af07b3" containerName="registry-server" Nov 24 14:58:12 crc kubenswrapper[4790]: I1124 14:58:12.795104 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a415403-8f42-4dd2-9387-30b9c2af07b3" containerName="registry-server" Nov 24 14:58:12 crc kubenswrapper[4790]: E1124 14:58:12.795178 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a415403-8f42-4dd2-9387-30b9c2af07b3" containerName="extract-utilities" Nov 24 14:58:12 crc kubenswrapper[4790]: I1124 14:58:12.795233 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a415403-8f42-4dd2-9387-30b9c2af07b3" containerName="extract-utilities" Nov 24 14:58:12 crc kubenswrapper[4790]: E1124 14:58:12.795315 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a415403-8f42-4dd2-9387-30b9c2af07b3" containerName="extract-content" Nov 24 14:58:12 crc kubenswrapper[4790]: I1124 14:58:12.795370 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a415403-8f42-4dd2-9387-30b9c2af07b3" containerName="extract-content" Nov 24 14:58:12 crc kubenswrapper[4790]: I1124 14:58:12.795661 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a415403-8f42-4dd2-9387-30b9c2af07b3" containerName="registry-server" Nov 24 14:58:12 crc kubenswrapper[4790]: I1124 14:58:12.796872 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d447559f9-hl956" Nov 24 14:58:12 crc kubenswrapper[4790]: I1124 14:58:12.802778 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1" Nov 24 14:58:12 crc kubenswrapper[4790]: I1124 14:58:12.827362 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d447559f9-hl956"] Nov 24 14:58:12 crc kubenswrapper[4790]: I1124 14:58:12.892411 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9452ec1e-f1d0-4190-868c-3919b7caa55d-dns-svc\") pod \"dnsmasq-dns-5d447559f9-hl956\" (UID: \"9452ec1e-f1d0-4190-868c-3919b7caa55d\") " pod="openstack/dnsmasq-dns-5d447559f9-hl956" Nov 24 14:58:12 crc kubenswrapper[4790]: I1124 14:58:12.892493 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9452ec1e-f1d0-4190-868c-3919b7caa55d-ovsdbserver-nb\") pod \"dnsmasq-dns-5d447559f9-hl956\" (UID: \"9452ec1e-f1d0-4190-868c-3919b7caa55d\") " pod="openstack/dnsmasq-dns-5d447559f9-hl956" Nov 24 14:58:12 crc kubenswrapper[4790]: I1124 14:58:12.892573 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9452ec1e-f1d0-4190-868c-3919b7caa55d-ovsdbserver-sb\") pod \"dnsmasq-dns-5d447559f9-hl956\" (UID: \"9452ec1e-f1d0-4190-868c-3919b7caa55d\") " pod="openstack/dnsmasq-dns-5d447559f9-hl956" Nov 24 14:58:12 crc kubenswrapper[4790]: I1124 14:58:12.892621 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/9452ec1e-f1d0-4190-868c-3919b7caa55d-openstack-cell1\") pod \"dnsmasq-dns-5d447559f9-hl956\" (UID: \"9452ec1e-f1d0-4190-868c-3919b7caa55d\") " pod="openstack/dnsmasq-dns-5d447559f9-hl956" Nov 24 14:58:12 crc kubenswrapper[4790]: I1124 14:58:12.892708 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sszjz\" (UniqueName: \"kubernetes.io/projected/9452ec1e-f1d0-4190-868c-3919b7caa55d-kube-api-access-sszjz\") pod \"dnsmasq-dns-5d447559f9-hl956\" (UID: \"9452ec1e-f1d0-4190-868c-3919b7caa55d\") " pod="openstack/dnsmasq-dns-5d447559f9-hl956" Nov 24 14:58:12 crc kubenswrapper[4790]: I1124 14:58:12.892842 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9452ec1e-f1d0-4190-868c-3919b7caa55d-config\") pod \"dnsmasq-dns-5d447559f9-hl956\" (UID: \"9452ec1e-f1d0-4190-868c-3919b7caa55d\") " pod="openstack/dnsmasq-dns-5d447559f9-hl956" Nov 24 14:58:12 crc kubenswrapper[4790]: I1124 14:58:12.995236 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9452ec1e-f1d0-4190-868c-3919b7caa55d-config\") pod \"dnsmasq-dns-5d447559f9-hl956\" (UID: \"9452ec1e-f1d0-4190-868c-3919b7caa55d\") " pod="openstack/dnsmasq-dns-5d447559f9-hl956" Nov 24 14:58:12 crc kubenswrapper[4790]: I1124 14:58:12.995303 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9452ec1e-f1d0-4190-868c-3919b7caa55d-dns-svc\") pod \"dnsmasq-dns-5d447559f9-hl956\" (UID: \"9452ec1e-f1d0-4190-868c-3919b7caa55d\") " pod="openstack/dnsmasq-dns-5d447559f9-hl956" Nov 24 14:58:12 crc kubenswrapper[4790]: I1124 14:58:12.995332 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9452ec1e-f1d0-4190-868c-3919b7caa55d-ovsdbserver-nb\") pod \"dnsmasq-dns-5d447559f9-hl956\" (UID: \"9452ec1e-f1d0-4190-868c-3919b7caa55d\") " pod="openstack/dnsmasq-dns-5d447559f9-hl956" Nov 24 14:58:12 crc kubenswrapper[4790]: I1124 14:58:12.995371 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9452ec1e-f1d0-4190-868c-3919b7caa55d-ovsdbserver-sb\") pod \"dnsmasq-dns-5d447559f9-hl956\" (UID: \"9452ec1e-f1d0-4190-868c-3919b7caa55d\") " pod="openstack/dnsmasq-dns-5d447559f9-hl956" Nov 24 14:58:12 crc kubenswrapper[4790]: I1124 14:58:12.995404 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/9452ec1e-f1d0-4190-868c-3919b7caa55d-openstack-cell1\") pod \"dnsmasq-dns-5d447559f9-hl956\" (UID: \"9452ec1e-f1d0-4190-868c-3919b7caa55d\") " pod="openstack/dnsmasq-dns-5d447559f9-hl956" Nov 24 14:58:12 crc kubenswrapper[4790]: I1124 14:58:12.995452 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sszjz\" (UniqueName: \"kubernetes.io/projected/9452ec1e-f1d0-4190-868c-3919b7caa55d-kube-api-access-sszjz\") pod \"dnsmasq-dns-5d447559f9-hl956\" (UID: \"9452ec1e-f1d0-4190-868c-3919b7caa55d\") " pod="openstack/dnsmasq-dns-5d447559f9-hl956" Nov 24 14:58:12 crc kubenswrapper[4790]: I1124 14:58:12.996518 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9452ec1e-f1d0-4190-868c-3919b7caa55d-config\") pod \"dnsmasq-dns-5d447559f9-hl956\" (UID: \"9452ec1e-f1d0-4190-868c-3919b7caa55d\") " pod="openstack/dnsmasq-dns-5d447559f9-hl956" Nov 24 14:58:12 crc kubenswrapper[4790]: I1124 14:58:12.997049 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9452ec1e-f1d0-4190-868c-3919b7caa55d-dns-svc\") pod \"dnsmasq-dns-5d447559f9-hl956\" (UID: \"9452ec1e-f1d0-4190-868c-3919b7caa55d\") " pod="openstack/dnsmasq-dns-5d447559f9-hl956" Nov 24 14:58:12 crc kubenswrapper[4790]: I1124 14:58:12.997537 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9452ec1e-f1d0-4190-868c-3919b7caa55d-ovsdbserver-nb\") pod \"dnsmasq-dns-5d447559f9-hl956\" (UID: \"9452ec1e-f1d0-4190-868c-3919b7caa55d\") " pod="openstack/dnsmasq-dns-5d447559f9-hl956" Nov 24 14:58:12 crc kubenswrapper[4790]: I1124 14:58:12.998423 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/9452ec1e-f1d0-4190-868c-3919b7caa55d-openstack-cell1\") pod \"dnsmasq-dns-5d447559f9-hl956\" (UID: \"9452ec1e-f1d0-4190-868c-3919b7caa55d\") " pod="openstack/dnsmasq-dns-5d447559f9-hl956" Nov 24 14:58:12 crc kubenswrapper[4790]: I1124 14:58:12.998667 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9452ec1e-f1d0-4190-868c-3919b7caa55d-ovsdbserver-sb\") pod \"dnsmasq-dns-5d447559f9-hl956\" (UID: \"9452ec1e-f1d0-4190-868c-3919b7caa55d\") " pod="openstack/dnsmasq-dns-5d447559f9-hl956" Nov 24 14:58:13 crc kubenswrapper[4790]: I1124 14:58:13.025853 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sszjz\" (UniqueName: \"kubernetes.io/projected/9452ec1e-f1d0-4190-868c-3919b7caa55d-kube-api-access-sszjz\") pod \"dnsmasq-dns-5d447559f9-hl956\" (UID: \"9452ec1e-f1d0-4190-868c-3919b7caa55d\") " pod="openstack/dnsmasq-dns-5d447559f9-hl956" Nov 24 14:58:13 crc kubenswrapper[4790]: I1124 14:58:13.125903 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d447559f9-hl956" Nov 24 14:58:13 crc kubenswrapper[4790]: I1124 14:58:13.624228 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d447559f9-hl956"] Nov 24 14:58:13 crc kubenswrapper[4790]: I1124 14:58:13.725999 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d447559f9-hl956" event={"ID":"9452ec1e-f1d0-4190-868c-3919b7caa55d","Type":"ContainerStarted","Data":"db88fd0c0a7c34ea1973b223396c911cc77c2bac46658c0215a03729b7588d8f"} Nov 24 14:58:14 crc kubenswrapper[4790]: I1124 14:58:14.738801 4790 generic.go:334] "Generic (PLEG): container finished" podID="9452ec1e-f1d0-4190-868c-3919b7caa55d" containerID="5cdf1deb251436ecabde2d77bd70a5f22bed414f843fd2f8534a947f800121dd" exitCode=0 Nov 24 14:58:14 crc kubenswrapper[4790]: I1124 14:58:14.738943 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d447559f9-hl956" event={"ID":"9452ec1e-f1d0-4190-868c-3919b7caa55d","Type":"ContainerDied","Data":"5cdf1deb251436ecabde2d77bd70a5f22bed414f843fd2f8534a947f800121dd"} Nov 24 14:58:15 crc kubenswrapper[4790]: I1124 14:58:15.754618 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d447559f9-hl956" event={"ID":"9452ec1e-f1d0-4190-868c-3919b7caa55d","Type":"ContainerStarted","Data":"2422650d7bc0a1f76cc50950beb536141ebac551cf6f094fd72a1c3f2ab1b688"} Nov 24 14:58:15 crc kubenswrapper[4790]: I1124 14:58:15.755526 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d447559f9-hl956" Nov 24 14:58:15 crc kubenswrapper[4790]: I1124 14:58:15.785611 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d447559f9-hl956" podStartSLOduration=3.785577305 podStartE2EDuration="3.785577305s" podCreationTimestamp="2025-11-24 14:58:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:58:15.773713096 +0000 UTC m=+6344.153606758" watchObservedRunningTime="2025-11-24 14:58:15.785577305 +0000 UTC m=+6344.165471007" Nov 24 14:58:23 crc kubenswrapper[4790]: I1124 14:58:23.127104 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5d447559f9-hl956" Nov 24 14:58:23 crc kubenswrapper[4790]: I1124 14:58:23.283346 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68bc6cdf65-975r7"] Nov 24 14:58:23 crc kubenswrapper[4790]: I1124 14:58:23.283642 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-68bc6cdf65-975r7" podUID="cd78a936-5519-480a-b36a-a6ec905cd694" containerName="dnsmasq-dns" containerID="cri-o://15d6d72023d3e6da20d416deec5568184431107042f2be4b7beaa76570629cdc" gracePeriod=10 Nov 24 14:58:23 crc kubenswrapper[4790]: I1124 14:58:23.479818 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cb6b9bbc5-vrcjr"] Nov 24 14:58:23 crc kubenswrapper[4790]: I1124 14:58:23.481985 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6b9bbc5-vrcjr" Nov 24 14:58:23 crc kubenswrapper[4790]: I1124 14:58:23.504517 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6b9bbc5-vrcjr"] Nov 24 14:58:23 crc kubenswrapper[4790]: I1124 14:58:23.506603 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35ee9564-c213-4eba-b011-467edd892cd0-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6b9bbc5-vrcjr\" (UID: \"35ee9564-c213-4eba-b011-467edd892cd0\") " pod="openstack/dnsmasq-dns-cb6b9bbc5-vrcjr" Nov 24 14:58:23 crc kubenswrapper[4790]: I1124 14:58:23.506672 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35ee9564-c213-4eba-b011-467edd892cd0-dns-svc\") pod \"dnsmasq-dns-cb6b9bbc5-vrcjr\" (UID: \"35ee9564-c213-4eba-b011-467edd892cd0\") " pod="openstack/dnsmasq-dns-cb6b9bbc5-vrcjr" Nov 24 14:58:23 crc kubenswrapper[4790]: I1124 14:58:23.506705 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/35ee9564-c213-4eba-b011-467edd892cd0-openstack-cell1\") pod \"dnsmasq-dns-cb6b9bbc5-vrcjr\" (UID: \"35ee9564-c213-4eba-b011-467edd892cd0\") " pod="openstack/dnsmasq-dns-cb6b9bbc5-vrcjr" Nov 24 14:58:23 crc kubenswrapper[4790]: I1124 14:58:23.506826 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35ee9564-c213-4eba-b011-467edd892cd0-config\") pod \"dnsmasq-dns-cb6b9bbc5-vrcjr\" (UID: \"35ee9564-c213-4eba-b011-467edd892cd0\") " pod="openstack/dnsmasq-dns-cb6b9bbc5-vrcjr" Nov 24 14:58:23 crc kubenswrapper[4790]: I1124 14:58:23.507114 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5j4js\" (UniqueName: \"kubernetes.io/projected/35ee9564-c213-4eba-b011-467edd892cd0-kube-api-access-5j4js\") pod \"dnsmasq-dns-cb6b9bbc5-vrcjr\" (UID: \"35ee9564-c213-4eba-b011-467edd892cd0\") " pod="openstack/dnsmasq-dns-cb6b9bbc5-vrcjr" Nov 24 14:58:23 crc kubenswrapper[4790]: I1124 14:58:23.507169 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35ee9564-c213-4eba-b011-467edd892cd0-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6b9bbc5-vrcjr\" (UID: \"35ee9564-c213-4eba-b011-467edd892cd0\") " pod="openstack/dnsmasq-dns-cb6b9bbc5-vrcjr" Nov 24 14:58:23 crc kubenswrapper[4790]: I1124 14:58:23.609383 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5j4js\" (UniqueName: \"kubernetes.io/projected/35ee9564-c213-4eba-b011-467edd892cd0-kube-api-access-5j4js\") pod \"dnsmasq-dns-cb6b9bbc5-vrcjr\" (UID: \"35ee9564-c213-4eba-b011-467edd892cd0\") " pod="openstack/dnsmasq-dns-cb6b9bbc5-vrcjr" Nov 24 14:58:23 crc kubenswrapper[4790]: I1124 14:58:23.609440 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35ee9564-c213-4eba-b011-467edd892cd0-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6b9bbc5-vrcjr\" (UID: \"35ee9564-c213-4eba-b011-467edd892cd0\") " pod="openstack/dnsmasq-dns-cb6b9bbc5-vrcjr" Nov 24 14:58:23 crc kubenswrapper[4790]: I1124 14:58:23.609513 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35ee9564-c213-4eba-b011-467edd892cd0-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6b9bbc5-vrcjr\" (UID: \"35ee9564-c213-4eba-b011-467edd892cd0\") " pod="openstack/dnsmasq-dns-cb6b9bbc5-vrcjr" Nov 24 14:58:23 crc kubenswrapper[4790]: I1124 14:58:23.609581 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35ee9564-c213-4eba-b011-467edd892cd0-dns-svc\") pod \"dnsmasq-dns-cb6b9bbc5-vrcjr\" (UID: \"35ee9564-c213-4eba-b011-467edd892cd0\") " pod="openstack/dnsmasq-dns-cb6b9bbc5-vrcjr" Nov 24 14:58:23 crc kubenswrapper[4790]: I1124 14:58:23.609607 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/35ee9564-c213-4eba-b011-467edd892cd0-openstack-cell1\") pod \"dnsmasq-dns-cb6b9bbc5-vrcjr\" (UID: \"35ee9564-c213-4eba-b011-467edd892cd0\") " pod="openstack/dnsmasq-dns-cb6b9bbc5-vrcjr" Nov 24 14:58:23 crc kubenswrapper[4790]: I1124 14:58:23.609690 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35ee9564-c213-4eba-b011-467edd892cd0-config\") pod \"dnsmasq-dns-cb6b9bbc5-vrcjr\" (UID: \"35ee9564-c213-4eba-b011-467edd892cd0\") " pod="openstack/dnsmasq-dns-cb6b9bbc5-vrcjr" Nov 24 14:58:23 crc kubenswrapper[4790]: I1124 14:58:23.610660 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35ee9564-c213-4eba-b011-467edd892cd0-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6b9bbc5-vrcjr\" (UID: \"35ee9564-c213-4eba-b011-467edd892cd0\") " pod="openstack/dnsmasq-dns-cb6b9bbc5-vrcjr" Nov 24 14:58:23 crc kubenswrapper[4790]: I1124 14:58:23.610905 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35ee9564-c213-4eba-b011-467edd892cd0-dns-svc\") pod \"dnsmasq-dns-cb6b9bbc5-vrcjr\" (UID: \"35ee9564-c213-4eba-b011-467edd892cd0\") " pod="openstack/dnsmasq-dns-cb6b9bbc5-vrcjr" Nov 24 14:58:23 crc kubenswrapper[4790]: I1124 14:58:23.610985 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/35ee9564-c213-4eba-b011-467edd892cd0-openstack-cell1\") pod \"dnsmasq-dns-cb6b9bbc5-vrcjr\" (UID: \"35ee9564-c213-4eba-b011-467edd892cd0\") " pod="openstack/dnsmasq-dns-cb6b9bbc5-vrcjr" Nov 24 14:58:23 crc kubenswrapper[4790]: I1124 14:58:23.611444 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35ee9564-c213-4eba-b011-467edd892cd0-config\") pod \"dnsmasq-dns-cb6b9bbc5-vrcjr\" (UID: \"35ee9564-c213-4eba-b011-467edd892cd0\") " pod="openstack/dnsmasq-dns-cb6b9bbc5-vrcjr" Nov 24 14:58:23 crc kubenswrapper[4790]: I1124 14:58:23.611499 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35ee9564-c213-4eba-b011-467edd892cd0-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6b9bbc5-vrcjr\" (UID: \"35ee9564-c213-4eba-b011-467edd892cd0\") " pod="openstack/dnsmasq-dns-cb6b9bbc5-vrcjr" Nov 24 14:58:23 crc kubenswrapper[4790]: I1124 14:58:23.634576 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5j4js\" (UniqueName: \"kubernetes.io/projected/35ee9564-c213-4eba-b011-467edd892cd0-kube-api-access-5j4js\") pod \"dnsmasq-dns-cb6b9bbc5-vrcjr\" (UID: \"35ee9564-c213-4eba-b011-467edd892cd0\") " pod="openstack/dnsmasq-dns-cb6b9bbc5-vrcjr" Nov 24 14:58:23 crc kubenswrapper[4790]: I1124 14:58:23.831156 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6b9bbc5-vrcjr" Nov 24 14:58:23 crc kubenswrapper[4790]: I1124 14:58:23.854712 4790 generic.go:334] "Generic (PLEG): container finished" podID="cd78a936-5519-480a-b36a-a6ec905cd694" containerID="15d6d72023d3e6da20d416deec5568184431107042f2be4b7beaa76570629cdc" exitCode=0 Nov 24 14:58:23 crc kubenswrapper[4790]: I1124 14:58:23.854776 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68bc6cdf65-975r7" event={"ID":"cd78a936-5519-480a-b36a-a6ec905cd694","Type":"ContainerDied","Data":"15d6d72023d3e6da20d416deec5568184431107042f2be4b7beaa76570629cdc"} Nov 24 14:58:24 crc kubenswrapper[4790]: W1124 14:58:24.329179 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35ee9564_c213_4eba_b011_467edd892cd0.slice/crio-b7a4ed69ded236bc01b08d587a21fc71d16246b285f517421aa2b3dea4a467b6 WatchSource:0}: Error finding container b7a4ed69ded236bc01b08d587a21fc71d16246b285f517421aa2b3dea4a467b6: Status 404 returned error can't find the container with id b7a4ed69ded236bc01b08d587a21fc71d16246b285f517421aa2b3dea4a467b6 Nov 24 14:58:24 crc kubenswrapper[4790]: I1124 14:58:24.352624 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6b9bbc5-vrcjr"] Nov 24 14:58:24 crc kubenswrapper[4790]: I1124 14:58:24.422244 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68bc6cdf65-975r7" Nov 24 14:58:24 crc kubenswrapper[4790]: I1124 14:58:24.529385 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd78a936-5519-480a-b36a-a6ec905cd694-ovsdbserver-nb\") pod \"cd78a936-5519-480a-b36a-a6ec905cd694\" (UID: \"cd78a936-5519-480a-b36a-a6ec905cd694\") " Nov 24 14:58:24 crc kubenswrapper[4790]: I1124 14:58:24.529470 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd78a936-5519-480a-b36a-a6ec905cd694-dns-svc\") pod \"cd78a936-5519-480a-b36a-a6ec905cd694\" (UID: \"cd78a936-5519-480a-b36a-a6ec905cd694\") " Nov 24 14:58:24 crc kubenswrapper[4790]: I1124 14:58:24.529501 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd78a936-5519-480a-b36a-a6ec905cd694-config\") pod \"cd78a936-5519-480a-b36a-a6ec905cd694\" (UID: \"cd78a936-5519-480a-b36a-a6ec905cd694\") " Nov 24 14:58:24 crc kubenswrapper[4790]: I1124 14:58:24.532713 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb5hs\" (UniqueName: \"kubernetes.io/projected/cd78a936-5519-480a-b36a-a6ec905cd694-kube-api-access-sb5hs\") pod \"cd78a936-5519-480a-b36a-a6ec905cd694\" (UID: \"cd78a936-5519-480a-b36a-a6ec905cd694\") " Nov 24 14:58:24 crc kubenswrapper[4790]: I1124 14:58:24.532788 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd78a936-5519-480a-b36a-a6ec905cd694-ovsdbserver-sb\") pod \"cd78a936-5519-480a-b36a-a6ec905cd694\" (UID: \"cd78a936-5519-480a-b36a-a6ec905cd694\") " Nov 24 14:58:24 crc kubenswrapper[4790]: I1124 14:58:24.540372 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd78a936-5519-480a-b36a-a6ec905cd694-kube-api-access-sb5hs" (OuterVolumeSpecName: "kube-api-access-sb5hs") pod "cd78a936-5519-480a-b36a-a6ec905cd694" (UID: "cd78a936-5519-480a-b36a-a6ec905cd694"). InnerVolumeSpecName "kube-api-access-sb5hs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:58:24 crc kubenswrapper[4790]: I1124 14:58:24.601230 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd78a936-5519-480a-b36a-a6ec905cd694-config" (OuterVolumeSpecName: "config") pod "cd78a936-5519-480a-b36a-a6ec905cd694" (UID: "cd78a936-5519-480a-b36a-a6ec905cd694"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:58:24 crc kubenswrapper[4790]: I1124 14:58:24.602429 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd78a936-5519-480a-b36a-a6ec905cd694-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cd78a936-5519-480a-b36a-a6ec905cd694" (UID: "cd78a936-5519-480a-b36a-a6ec905cd694"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:58:24 crc kubenswrapper[4790]: I1124 14:58:24.602424 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd78a936-5519-480a-b36a-a6ec905cd694-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cd78a936-5519-480a-b36a-a6ec905cd694" (UID: "cd78a936-5519-480a-b36a-a6ec905cd694"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:58:24 crc kubenswrapper[4790]: I1124 14:58:24.636034 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb5hs\" (UniqueName: \"kubernetes.io/projected/cd78a936-5519-480a-b36a-a6ec905cd694-kube-api-access-sb5hs\") on node \"crc\" DevicePath \"\"" Nov 24 14:58:24 crc kubenswrapper[4790]: I1124 14:58:24.636286 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd78a936-5519-480a-b36a-a6ec905cd694-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:58:24 crc kubenswrapper[4790]: I1124 14:58:24.636297 4790 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd78a936-5519-480a-b36a-a6ec905cd694-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:58:24 crc kubenswrapper[4790]: I1124 14:58:24.636305 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd78a936-5519-480a-b36a-a6ec905cd694-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:58:24 crc kubenswrapper[4790]: I1124 14:58:24.646224 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd78a936-5519-480a-b36a-a6ec905cd694-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cd78a936-5519-480a-b36a-a6ec905cd694" (UID: "cd78a936-5519-480a-b36a-a6ec905cd694"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:58:24 crc kubenswrapper[4790]: I1124 14:58:24.737898 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd78a936-5519-480a-b36a-a6ec905cd694-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:58:24 crc kubenswrapper[4790]: I1124 14:58:24.866218 4790 generic.go:334] "Generic (PLEG): container finished" podID="35ee9564-c213-4eba-b011-467edd892cd0" containerID="119040e32ec8c128782b7a74e4b5371b6c48c53565ddbcd098f5ddf2f12b2ef1" exitCode=0 Nov 24 14:58:24 crc kubenswrapper[4790]: I1124 14:58:24.866278 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6b9bbc5-vrcjr" event={"ID":"35ee9564-c213-4eba-b011-467edd892cd0","Type":"ContainerDied","Data":"119040e32ec8c128782b7a74e4b5371b6c48c53565ddbcd098f5ddf2f12b2ef1"} Nov 24 14:58:24 crc kubenswrapper[4790]: I1124 14:58:24.866348 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6b9bbc5-vrcjr" event={"ID":"35ee9564-c213-4eba-b011-467edd892cd0","Type":"ContainerStarted","Data":"b7a4ed69ded236bc01b08d587a21fc71d16246b285f517421aa2b3dea4a467b6"} Nov 24 14:58:24 crc kubenswrapper[4790]: I1124 14:58:24.870393 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68bc6cdf65-975r7" event={"ID":"cd78a936-5519-480a-b36a-a6ec905cd694","Type":"ContainerDied","Data":"db61ba705687b91bcdae96d2ace8d031bf7b289b1f9f06e8a9d2d288af7f7e5c"} Nov 24 14:58:24 crc kubenswrapper[4790]: I1124 14:58:24.870498 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68bc6cdf65-975r7" Nov 24 14:58:24 crc kubenswrapper[4790]: I1124 14:58:24.874364 4790 scope.go:117] "RemoveContainer" containerID="15d6d72023d3e6da20d416deec5568184431107042f2be4b7beaa76570629cdc" Nov 24 14:58:24 crc kubenswrapper[4790]: I1124 14:58:24.906757 4790 scope.go:117] "RemoveContainer" containerID="4e01d22038030468b23c387d57334ceb2e0efaba5a1eda56f841b16ad889663b" Nov 24 14:58:24 crc kubenswrapper[4790]: I1124 14:58:24.923925 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68bc6cdf65-975r7"] Nov 24 14:58:24 crc kubenswrapper[4790]: I1124 14:58:24.933600 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-68bc6cdf65-975r7"] Nov 24 14:58:25 crc kubenswrapper[4790]: I1124 14:58:25.883086 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6b9bbc5-vrcjr" event={"ID":"35ee9564-c213-4eba-b011-467edd892cd0","Type":"ContainerStarted","Data":"f7d368902f90c3fcfcedde95b418dea5dd3607e4ea296c247b13876b1d157fb8"} Nov 24 14:58:25 crc kubenswrapper[4790]: I1124 14:58:25.883623 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cb6b9bbc5-vrcjr" Nov 24 14:58:25 crc kubenswrapper[4790]: I1124 14:58:25.904541 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cb6b9bbc5-vrcjr" podStartSLOduration=2.9045243960000002 podStartE2EDuration="2.904524396s" podCreationTimestamp="2025-11-24 14:58:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:58:25.90224784 +0000 UTC m=+6354.282141502" watchObservedRunningTime="2025-11-24 14:58:25.904524396 +0000 UTC m=+6354.284418058" Nov 24 14:58:26 crc kubenswrapper[4790]: I1124 14:58:26.325191 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd78a936-5519-480a-b36a-a6ec905cd694" path="/var/lib/kubelet/pods/cd78a936-5519-480a-b36a-a6ec905cd694/volumes" Nov 24 14:58:33 crc kubenswrapper[4790]: I1124 14:58:33.833323 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cb6b9bbc5-vrcjr" Nov 24 14:58:33 crc kubenswrapper[4790]: I1124 14:58:33.914698 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d447559f9-hl956"] Nov 24 14:58:33 crc kubenswrapper[4790]: I1124 14:58:33.914996 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5d447559f9-hl956" podUID="9452ec1e-f1d0-4190-868c-3919b7caa55d" containerName="dnsmasq-dns" containerID="cri-o://2422650d7bc0a1f76cc50950beb536141ebac551cf6f094fd72a1c3f2ab1b688" gracePeriod=10 Nov 24 14:58:34 crc kubenswrapper[4790]: I1124 14:58:34.412667 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d447559f9-hl956" Nov 24 14:58:34 crc kubenswrapper[4790]: I1124 14:58:34.487298 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9452ec1e-f1d0-4190-868c-3919b7caa55d-ovsdbserver-nb\") pod \"9452ec1e-f1d0-4190-868c-3919b7caa55d\" (UID: \"9452ec1e-f1d0-4190-868c-3919b7caa55d\") " Nov 24 14:58:34 crc kubenswrapper[4790]: I1124 14:58:34.487436 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sszjz\" (UniqueName: \"kubernetes.io/projected/9452ec1e-f1d0-4190-868c-3919b7caa55d-kube-api-access-sszjz\") pod \"9452ec1e-f1d0-4190-868c-3919b7caa55d\" (UID: \"9452ec1e-f1d0-4190-868c-3919b7caa55d\") " Nov 24 14:58:34 crc kubenswrapper[4790]: I1124 14:58:34.487643 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9452ec1e-f1d0-4190-868c-3919b7caa55d-ovsdbserver-sb\") pod \"9452ec1e-f1d0-4190-868c-3919b7caa55d\" (UID: \"9452ec1e-f1d0-4190-868c-3919b7caa55d\") " Nov 24 14:58:34 crc kubenswrapper[4790]: I1124 14:58:34.487817 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9452ec1e-f1d0-4190-868c-3919b7caa55d-dns-svc\") pod \"9452ec1e-f1d0-4190-868c-3919b7caa55d\" (UID: \"9452ec1e-f1d0-4190-868c-3919b7caa55d\") " Nov 24 14:58:34 crc kubenswrapper[4790]: I1124 14:58:34.487951 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9452ec1e-f1d0-4190-868c-3919b7caa55d-config\") pod \"9452ec1e-f1d0-4190-868c-3919b7caa55d\" (UID: \"9452ec1e-f1d0-4190-868c-3919b7caa55d\") " Nov 24 14:58:34 crc kubenswrapper[4790]: I1124 14:58:34.488024 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/9452ec1e-f1d0-4190-868c-3919b7caa55d-openstack-cell1\") pod \"9452ec1e-f1d0-4190-868c-3919b7caa55d\" (UID: \"9452ec1e-f1d0-4190-868c-3919b7caa55d\") " Nov 24 14:58:34 crc kubenswrapper[4790]: I1124 14:58:34.493122 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9452ec1e-f1d0-4190-868c-3919b7caa55d-kube-api-access-sszjz" (OuterVolumeSpecName: "kube-api-access-sszjz") pod "9452ec1e-f1d0-4190-868c-3919b7caa55d" (UID: "9452ec1e-f1d0-4190-868c-3919b7caa55d"). InnerVolumeSpecName "kube-api-access-sszjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:58:34 crc kubenswrapper[4790]: I1124 14:58:34.549011 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9452ec1e-f1d0-4190-868c-3919b7caa55d-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "9452ec1e-f1d0-4190-868c-3919b7caa55d" (UID: "9452ec1e-f1d0-4190-868c-3919b7caa55d"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:58:34 crc kubenswrapper[4790]: I1124 14:58:34.554590 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9452ec1e-f1d0-4190-868c-3919b7caa55d-config" (OuterVolumeSpecName: "config") pod "9452ec1e-f1d0-4190-868c-3919b7caa55d" (UID: "9452ec1e-f1d0-4190-868c-3919b7caa55d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:58:34 crc kubenswrapper[4790]: I1124 14:58:34.558716 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9452ec1e-f1d0-4190-868c-3919b7caa55d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9452ec1e-f1d0-4190-868c-3919b7caa55d" (UID: "9452ec1e-f1d0-4190-868c-3919b7caa55d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:58:34 crc kubenswrapper[4790]: I1124 14:58:34.565661 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9452ec1e-f1d0-4190-868c-3919b7caa55d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9452ec1e-f1d0-4190-868c-3919b7caa55d" (UID: "9452ec1e-f1d0-4190-868c-3919b7caa55d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:58:34 crc kubenswrapper[4790]: I1124 14:58:34.567680 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9452ec1e-f1d0-4190-868c-3919b7caa55d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9452ec1e-f1d0-4190-868c-3919b7caa55d" (UID: "9452ec1e-f1d0-4190-868c-3919b7caa55d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:58:34 crc kubenswrapper[4790]: I1124 14:58:34.591152 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9452ec1e-f1d0-4190-868c-3919b7caa55d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:58:34 crc kubenswrapper[4790]: I1124 14:58:34.591192 4790 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9452ec1e-f1d0-4190-868c-3919b7caa55d-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:58:34 crc kubenswrapper[4790]: I1124 14:58:34.591202 4790 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9452ec1e-f1d0-4190-868c-3919b7caa55d-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:58:34 crc kubenswrapper[4790]: I1124 14:58:34.591210 4790 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/9452ec1e-f1d0-4190-868c-3919b7caa55d-openstack-cell1\") on node \"crc\" DevicePath \"\"" Nov 24 14:58:34 crc kubenswrapper[4790]: I1124 14:58:34.591217 4790 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9452ec1e-f1d0-4190-868c-3919b7caa55d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:58:34 crc kubenswrapper[4790]: I1124 14:58:34.591227 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sszjz\" (UniqueName: \"kubernetes.io/projected/9452ec1e-f1d0-4190-868c-3919b7caa55d-kube-api-access-sszjz\") on node \"crc\" DevicePath \"\"" Nov 24 14:58:34 crc kubenswrapper[4790]: I1124 14:58:34.994722 4790 generic.go:334] "Generic (PLEG): container finished" podID="9452ec1e-f1d0-4190-868c-3919b7caa55d" containerID="2422650d7bc0a1f76cc50950beb536141ebac551cf6f094fd72a1c3f2ab1b688" exitCode=0 Nov 24 14:58:34 crc kubenswrapper[4790]: I1124 14:58:34.994811 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d447559f9-hl956" event={"ID":"9452ec1e-f1d0-4190-868c-3919b7caa55d","Type":"ContainerDied","Data":"2422650d7bc0a1f76cc50950beb536141ebac551cf6f094fd72a1c3f2ab1b688"} Nov 24 14:58:34 crc kubenswrapper[4790]: I1124 14:58:34.996009 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d447559f9-hl956" event={"ID":"9452ec1e-f1d0-4190-868c-3919b7caa55d","Type":"ContainerDied","Data":"db88fd0c0a7c34ea1973b223396c911cc77c2bac46658c0215a03729b7588d8f"} Nov 24 14:58:34 crc kubenswrapper[4790]: I1124 14:58:34.994868 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d447559f9-hl956" Nov 24 14:58:34 crc kubenswrapper[4790]: I1124 14:58:34.996077 4790 scope.go:117] "RemoveContainer" containerID="2422650d7bc0a1f76cc50950beb536141ebac551cf6f094fd72a1c3f2ab1b688" Nov 24 14:58:35 crc kubenswrapper[4790]: I1124 14:58:35.015137 4790 scope.go:117] "RemoveContainer" containerID="5cdf1deb251436ecabde2d77bd70a5f22bed414f843fd2f8534a947f800121dd" Nov 24 14:58:35 crc kubenswrapper[4790]: I1124 14:58:35.042341 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d447559f9-hl956"] Nov 24 14:58:35 crc kubenswrapper[4790]: I1124 14:58:35.043825 4790 scope.go:117] "RemoveContainer" containerID="2422650d7bc0a1f76cc50950beb536141ebac551cf6f094fd72a1c3f2ab1b688" Nov 24 14:58:35 crc kubenswrapper[4790]: E1124 14:58:35.044821 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2422650d7bc0a1f76cc50950beb536141ebac551cf6f094fd72a1c3f2ab1b688\": container with ID starting with 2422650d7bc0a1f76cc50950beb536141ebac551cf6f094fd72a1c3f2ab1b688 not found: ID does not exist" containerID="2422650d7bc0a1f76cc50950beb536141ebac551cf6f094fd72a1c3f2ab1b688" Nov 24 14:58:35 crc kubenswrapper[4790]: I1124 14:58:35.044871 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2422650d7bc0a1f76cc50950beb536141ebac551cf6f094fd72a1c3f2ab1b688"} err="failed to get container status \"2422650d7bc0a1f76cc50950beb536141ebac551cf6f094fd72a1c3f2ab1b688\": rpc error: code = NotFound desc = could not find container \"2422650d7bc0a1f76cc50950beb536141ebac551cf6f094fd72a1c3f2ab1b688\": container with ID starting with 2422650d7bc0a1f76cc50950beb536141ebac551cf6f094fd72a1c3f2ab1b688 not found: ID does not exist" Nov 24 14:58:35 crc kubenswrapper[4790]: I1124 14:58:35.044912 4790 scope.go:117] "RemoveContainer" containerID="5cdf1deb251436ecabde2d77bd70a5f22bed414f843fd2f8534a947f800121dd" Nov 24 14:58:35 crc kubenswrapper[4790]: E1124 14:58:35.045386 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5cdf1deb251436ecabde2d77bd70a5f22bed414f843fd2f8534a947f800121dd\": container with ID starting with 5cdf1deb251436ecabde2d77bd70a5f22bed414f843fd2f8534a947f800121dd not found: ID does not exist" containerID="5cdf1deb251436ecabde2d77bd70a5f22bed414f843fd2f8534a947f800121dd" Nov 24 14:58:35 crc kubenswrapper[4790]: I1124 14:58:35.045407 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cdf1deb251436ecabde2d77bd70a5f22bed414f843fd2f8534a947f800121dd"} err="failed to get container status \"5cdf1deb251436ecabde2d77bd70a5f22bed414f843fd2f8534a947f800121dd\": rpc error: code = NotFound desc = could not find container \"5cdf1deb251436ecabde2d77bd70a5f22bed414f843fd2f8534a947f800121dd\": container with ID starting with 5cdf1deb251436ecabde2d77bd70a5f22bed414f843fd2f8534a947f800121dd not found: ID does not exist" Nov 24 14:58:35 crc kubenswrapper[4790]: I1124 14:58:35.051134 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d447559f9-hl956"] Nov 24 14:58:36 crc kubenswrapper[4790]: I1124 14:58:36.338422 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9452ec1e-f1d0-4190-868c-3919b7caa55d" path="/var/lib/kubelet/pods/9452ec1e-f1d0-4190-868c-3919b7caa55d/volumes" Nov 24 14:58:44 crc kubenswrapper[4790]: I1124 14:58:44.777506 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk"] Nov 24 14:58:44 crc kubenswrapper[4790]: E1124 14:58:44.779089 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9452ec1e-f1d0-4190-868c-3919b7caa55d" containerName="dnsmasq-dns" Nov 24 14:58:44 crc kubenswrapper[4790]: I1124 14:58:44.779115 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="9452ec1e-f1d0-4190-868c-3919b7caa55d" containerName="dnsmasq-dns" Nov 24 14:58:44 crc kubenswrapper[4790]: E1124 14:58:44.779145 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9452ec1e-f1d0-4190-868c-3919b7caa55d" containerName="init" Nov 24 14:58:44 crc kubenswrapper[4790]: I1124 14:58:44.779159 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="9452ec1e-f1d0-4190-868c-3919b7caa55d" containerName="init" Nov 24 14:58:44 crc kubenswrapper[4790]: E1124 14:58:44.779180 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd78a936-5519-480a-b36a-a6ec905cd694" containerName="dnsmasq-dns" Nov 24 14:58:44 crc kubenswrapper[4790]: I1124 14:58:44.779195 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd78a936-5519-480a-b36a-a6ec905cd694" containerName="dnsmasq-dns" Nov 24 14:58:44 crc kubenswrapper[4790]: E1124 14:58:44.779231 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd78a936-5519-480a-b36a-a6ec905cd694" containerName="init" Nov 24 14:58:44 crc kubenswrapper[4790]: I1124 14:58:44.779249 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd78a936-5519-480a-b36a-a6ec905cd694" containerName="init" Nov 24 14:58:44 crc kubenswrapper[4790]: I1124 14:58:44.779703 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd78a936-5519-480a-b36a-a6ec905cd694" containerName="dnsmasq-dns" Nov 24 14:58:44 crc kubenswrapper[4790]: I1124 14:58:44.779751 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="9452ec1e-f1d0-4190-868c-3919b7caa55d" containerName="dnsmasq-dns" Nov 24 14:58:44 crc kubenswrapper[4790]: I1124 14:58:44.781173 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk" Nov 24 14:58:44 crc kubenswrapper[4790]: I1124 14:58:44.784053 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 14:58:44 crc kubenswrapper[4790]: I1124 14:58:44.784068 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fc497" Nov 24 14:58:44 crc kubenswrapper[4790]: I1124 14:58:44.786005 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 14:58:44 crc kubenswrapper[4790]: I1124 14:58:44.786320 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 14:58:44 crc kubenswrapper[4790]: I1124 14:58:44.804992 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk"] Nov 24 14:58:44 crc kubenswrapper[4790]: I1124 14:58:44.855217 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba0f2f40-136f-402f-b25c-4fdf2236b7d3-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk\" (UID: \"ba0f2f40-136f-402f-b25c-4fdf2236b7d3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk" Nov 24 14:58:44 crc kubenswrapper[4790]: I1124 14:58:44.855845 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgbg2\" (UniqueName: \"kubernetes.io/projected/ba0f2f40-136f-402f-b25c-4fdf2236b7d3-kube-api-access-dgbg2\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk\" (UID: \"ba0f2f40-136f-402f-b25c-4fdf2236b7d3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk" Nov 24 14:58:44 crc kubenswrapper[4790]: I1124 14:58:44.856109 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ba0f2f40-136f-402f-b25c-4fdf2236b7d3-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk\" (UID: \"ba0f2f40-136f-402f-b25c-4fdf2236b7d3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk" Nov 24 14:58:44 crc kubenswrapper[4790]: I1124 14:58:44.856242 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba0f2f40-136f-402f-b25c-4fdf2236b7d3-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk\" (UID: \"ba0f2f40-136f-402f-b25c-4fdf2236b7d3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk" Nov 24 14:58:44 crc kubenswrapper[4790]: I1124 14:58:44.856380 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba0f2f40-136f-402f-b25c-4fdf2236b7d3-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk\" (UID: \"ba0f2f40-136f-402f-b25c-4fdf2236b7d3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk" Nov 24 14:58:44 crc kubenswrapper[4790]: I1124 14:58:44.959279 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba0f2f40-136f-402f-b25c-4fdf2236b7d3-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk\" (UID: \"ba0f2f40-136f-402f-b25c-4fdf2236b7d3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk" Nov 24 14:58:44 crc kubenswrapper[4790]: I1124 14:58:44.959426 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgbg2\" (UniqueName: \"kubernetes.io/projected/ba0f2f40-136f-402f-b25c-4fdf2236b7d3-kube-api-access-dgbg2\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk\" (UID: \"ba0f2f40-136f-402f-b25c-4fdf2236b7d3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk" Nov 24 14:58:44 crc kubenswrapper[4790]: I1124 14:58:44.959538 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ba0f2f40-136f-402f-b25c-4fdf2236b7d3-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk\" (UID: \"ba0f2f40-136f-402f-b25c-4fdf2236b7d3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk" Nov 24 14:58:44 crc kubenswrapper[4790]: I1124 14:58:44.959569 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba0f2f40-136f-402f-b25c-4fdf2236b7d3-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk\" (UID: \"ba0f2f40-136f-402f-b25c-4fdf2236b7d3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk" Nov 24 14:58:44 crc kubenswrapper[4790]: I1124 14:58:44.959604 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba0f2f40-136f-402f-b25c-4fdf2236b7d3-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk\" (UID: \"ba0f2f40-136f-402f-b25c-4fdf2236b7d3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk" Nov 24 14:58:44 crc kubenswrapper[4790]: I1124 14:58:44.965797 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ba0f2f40-136f-402f-b25c-4fdf2236b7d3-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk\" (UID: \"ba0f2f40-136f-402f-b25c-4fdf2236b7d3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk" Nov 24 14:58:44 crc kubenswrapper[4790]: I1124 14:58:44.968396 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba0f2f40-136f-402f-b25c-4fdf2236b7d3-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk\" (UID: \"ba0f2f40-136f-402f-b25c-4fdf2236b7d3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk" Nov 24 14:58:44 crc kubenswrapper[4790]: I1124 14:58:44.973821 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba0f2f40-136f-402f-b25c-4fdf2236b7d3-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk\" (UID: \"ba0f2f40-136f-402f-b25c-4fdf2236b7d3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk" Nov 24 14:58:44 crc kubenswrapper[4790]: I1124 14:58:44.978697 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba0f2f40-136f-402f-b25c-4fdf2236b7d3-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk\" (UID: \"ba0f2f40-136f-402f-b25c-4fdf2236b7d3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk" Nov 24 14:58:44 crc kubenswrapper[4790]: I1124 14:58:44.985820 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgbg2\" (UniqueName: \"kubernetes.io/projected/ba0f2f40-136f-402f-b25c-4fdf2236b7d3-kube-api-access-dgbg2\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk\" (UID: \"ba0f2f40-136f-402f-b25c-4fdf2236b7d3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk" Nov 24 14:58:45 crc kubenswrapper[4790]: I1124 14:58:45.118316 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk" Nov 24 14:58:45 crc kubenswrapper[4790]: I1124 14:58:45.745843 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk"] Nov 24 14:58:46 crc kubenswrapper[4790]: I1124 14:58:46.139052 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk" event={"ID":"ba0f2f40-136f-402f-b25c-4fdf2236b7d3","Type":"ContainerStarted","Data":"ab67cdedb715aef1a32ab7efb1b0a5f28794b73cde19939d48ff9a63dc654839"} Nov 24 14:59:01 crc kubenswrapper[4790]: E1124 14:59:01.476361 4790 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/openstack-k8s-operators/openstack-ansibleee-runner:3b0bb6b72928679854fd8df2180c23d5c3286974" Nov 24 14:59:01 crc kubenswrapper[4790]: E1124 14:59:01.476875 4790 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/openstack-k8s-operators/openstack-ansibleee-runner:3b0bb6b72928679854fd8df2180c23d5c3286974" Nov 24 14:59:01 crc kubenswrapper[4790]: E1124 14:59:01.477130 4790 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 24 14:59:01 crc kubenswrapper[4790]: container &Container{Name:pre-adoption-validation-openstack-pre-adoption-openstack-cell1,Image:quay.rdoproject.org/openstack-k8s-operators/openstack-ansibleee-runner:3b0bb6b72928679854fd8df2180c23d5c3286974,Command:[],Args:[ansible-runner run /runner -p osp.edpm.pre_adoption_validation -i pre-adoption-validation-openstack-pre-adoption-openstack-cell1],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ANSIBLE_CALLBACKS_ENABLED,Value:ansible.posix.profile_tasks,ValueFrom:nil,},EnvVar{Name:ANSIBLE_CALLBACK_RESULT_FORMAT,Value:yaml,ValueFrom:nil,},EnvVar{Name:ANSIBLE_FORCE_COLOR,Value:True,ValueFrom:nil,},EnvVar{Name:ANSIBLE_DISPLAY_ARGS_TO_STDOUT,Value:True,ValueFrom:nil,},EnvVar{Name:ANSIBLE_SSH_ARGS,Value:-C -o ControlMaster=auto -o ControlPersist=80s,ValueFrom:nil,},EnvVar{Name:ANSIBLE_VERBOSITY,Value:1,ValueFrom:nil,},EnvVar{Name:RUNNER_PLAYBOOK,Value: Nov 24 14:59:01 crc kubenswrapper[4790]: osp.edpm.pre_adoption_validation Nov 24 14:59:01 crc kubenswrapper[4790]: Nov 24 14:59:01 crc kubenswrapper[4790]: ,ValueFrom:nil,},EnvVar{Name:RUNNER_EXTRA_VARS,Value: Nov 24 14:59:01 crc kubenswrapper[4790]: edpm_override_hosts: openstack-cell1 Nov 24 14:59:01 crc kubenswrapper[4790]: edpm_service_type: pre-adoption-validation Nov 24 14:59:01 crc kubenswrapper[4790]: edpm_services_override: [pre-adoption-validation] Nov 24 14:59:01 crc kubenswrapper[4790]: Nov 24 14:59:01 crc kubenswrapper[4790]: Nov 24 14:59:01 crc kubenswrapper[4790]: ,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ceph,ReadOnly:true,MountPath:/etc/ceph,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:pre-adoption-validation-combined-ca-bundle,ReadOnly:false,MountPath:/var/lib/openstack/cacerts/pre-adoption-validation,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/runner/env/ssh_key,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:inventory,ReadOnly:false,MountPath:/runner/inventory/hosts,SubPath:inventory,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dgbg2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:openstack-aee-default-env,},Optional:*true,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk_openstack(ba0f2f40-136f-402f-b25c-4fdf2236b7d3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled Nov 24 14:59:01 crc kubenswrapper[4790]: > logger="UnhandledError" Nov 24 14:59:01 crc kubenswrapper[4790]: E1124 14:59:01.478587 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"pre-adoption-validation-openstack-pre-adoption-openstack-cell1\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk" podUID="ba0f2f40-136f-402f-b25c-4fdf2236b7d3" Nov 24 14:59:02 crc kubenswrapper[4790]: E1124 14:59:02.340286 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"pre-adoption-validation-openstack-pre-adoption-openstack-cell1\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/openstack-k8s-operators/openstack-ansibleee-runner:3b0bb6b72928679854fd8df2180c23d5c3286974\\\"\"" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk" podUID="ba0f2f40-136f-402f-b25c-4fdf2236b7d3" Nov 24 14:59:14 crc kubenswrapper[4790]: I1124 14:59:14.056122 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-create-bzlld"] Nov 24 14:59:14 crc kubenswrapper[4790]: I1124 14:59:14.067228 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-create-bzlld"] Nov 24 14:59:14 crc kubenswrapper[4790]: I1124 14:59:14.332045 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9038e28a-efcc-48e9-97ac-5c1b6417fb0f" path="/var/lib/kubelet/pods/9038e28a-efcc-48e9-97ac-5c1b6417fb0f/volumes" Nov 24 14:59:15 crc kubenswrapper[4790]: I1124 14:59:15.051739 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-bccc-account-create-j8zg9"] Nov 24 14:59:15 crc kubenswrapper[4790]: I1124 14:59:15.072575 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-bccc-account-create-j8zg9"] Nov 24 14:59:16 crc kubenswrapper[4790]: I1124 14:59:16.336145 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85093f1b-8746-41e3-b112-58dde9f6d54f" path="/var/lib/kubelet/pods/85093f1b-8746-41e3-b112-58dde9f6d54f/volumes" Nov 24 14:59:18 crc kubenswrapper[4790]: I1124 14:59:18.539649 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk" event={"ID":"ba0f2f40-136f-402f-b25c-4fdf2236b7d3","Type":"ContainerStarted","Data":"7e1200334a0171650ff372bf525e82546ef52cf7519fba2ea474917162360c2e"} Nov 24 14:59:18 crc kubenswrapper[4790]: I1124 14:59:18.558945 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk" podStartSLOduration=2.7330332889999998 podStartE2EDuration="34.558920947s" podCreationTimestamp="2025-11-24 14:58:44 +0000 UTC" firstStartedPulling="2025-11-24 14:58:45.744683449 +0000 UTC m=+6374.124577111" lastFinishedPulling="2025-11-24 14:59:17.570571107 +0000 UTC m=+6405.950464769" observedRunningTime="2025-11-24 14:59:18.557714093 +0000 UTC m=+6406.937607755" watchObservedRunningTime="2025-11-24 14:59:18.558920947 +0000 UTC m=+6406.938814619" Nov 24 14:59:21 crc kubenswrapper[4790]: I1124 14:59:21.036227 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-persistence-db-create-sdmzn"] Nov 24 14:59:21 crc kubenswrapper[4790]: I1124 14:59:21.048175 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-persistence-db-create-sdmzn"] Nov 24 14:59:22 crc kubenswrapper[4790]: I1124 14:59:22.034161 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-1102-account-create-hrp2v"] Nov 24 14:59:22 crc kubenswrapper[4790]: I1124 14:59:22.049499 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-1102-account-create-hrp2v"] Nov 24 14:59:22 crc kubenswrapper[4790]: I1124 14:59:22.339738 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57b1601f-40f6-49cd-ab92-85efc8fb18e1" path="/var/lib/kubelet/pods/57b1601f-40f6-49cd-ab92-85efc8fb18e1/volumes" Nov 24 14:59:22 crc kubenswrapper[4790]: I1124 14:59:22.342056 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1e0c3ca-accf-4636-82c8-bb73ad8304e4" path="/var/lib/kubelet/pods/a1e0c3ca-accf-4636-82c8-bb73ad8304e4/volumes" Nov 24 14:59:31 crc kubenswrapper[4790]: I1124 14:59:31.728616 4790 generic.go:334] "Generic (PLEG): container finished" podID="ba0f2f40-136f-402f-b25c-4fdf2236b7d3" containerID="7e1200334a0171650ff372bf525e82546ef52cf7519fba2ea474917162360c2e" exitCode=0 Nov 24 14:59:31 crc kubenswrapper[4790]: I1124 14:59:31.728737 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk" event={"ID":"ba0f2f40-136f-402f-b25c-4fdf2236b7d3","Type":"ContainerDied","Data":"7e1200334a0171650ff372bf525e82546ef52cf7519fba2ea474917162360c2e"} Nov 24 14:59:33 crc kubenswrapper[4790]: I1124 14:59:33.332098 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk" Nov 24 14:59:33 crc kubenswrapper[4790]: I1124 14:59:33.504966 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba0f2f40-136f-402f-b25c-4fdf2236b7d3-pre-adoption-validation-combined-ca-bundle\") pod \"ba0f2f40-136f-402f-b25c-4fdf2236b7d3\" (UID: \"ba0f2f40-136f-402f-b25c-4fdf2236b7d3\") " Nov 24 14:59:33 crc kubenswrapper[4790]: I1124 14:59:33.505003 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba0f2f40-136f-402f-b25c-4fdf2236b7d3-ssh-key\") pod \"ba0f2f40-136f-402f-b25c-4fdf2236b7d3\" (UID: \"ba0f2f40-136f-402f-b25c-4fdf2236b7d3\") " Nov 24 14:59:33 crc kubenswrapper[4790]: I1124 14:59:33.505091 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgbg2\" (UniqueName: \"kubernetes.io/projected/ba0f2f40-136f-402f-b25c-4fdf2236b7d3-kube-api-access-dgbg2\") pod \"ba0f2f40-136f-402f-b25c-4fdf2236b7d3\" (UID: \"ba0f2f40-136f-402f-b25c-4fdf2236b7d3\") " Nov 24 14:59:33 crc kubenswrapper[4790]: I1124 14:59:33.505117 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba0f2f40-136f-402f-b25c-4fdf2236b7d3-inventory\") pod \"ba0f2f40-136f-402f-b25c-4fdf2236b7d3\" (UID: \"ba0f2f40-136f-402f-b25c-4fdf2236b7d3\") " Nov 24 14:59:33 crc kubenswrapper[4790]: I1124 14:59:33.505369 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ba0f2f40-136f-402f-b25c-4fdf2236b7d3-ceph\") pod \"ba0f2f40-136f-402f-b25c-4fdf2236b7d3\" (UID: \"ba0f2f40-136f-402f-b25c-4fdf2236b7d3\") " Nov 24 14:59:33 crc kubenswrapper[4790]: I1124 14:59:33.510926 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba0f2f40-136f-402f-b25c-4fdf2236b7d3-kube-api-access-dgbg2" (OuterVolumeSpecName: "kube-api-access-dgbg2") pod "ba0f2f40-136f-402f-b25c-4fdf2236b7d3" (UID: "ba0f2f40-136f-402f-b25c-4fdf2236b7d3"). InnerVolumeSpecName "kube-api-access-dgbg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:59:33 crc kubenswrapper[4790]: I1124 14:59:33.511683 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba0f2f40-136f-402f-b25c-4fdf2236b7d3-pre-adoption-validation-combined-ca-bundle" (OuterVolumeSpecName: "pre-adoption-validation-combined-ca-bundle") pod "ba0f2f40-136f-402f-b25c-4fdf2236b7d3" (UID: "ba0f2f40-136f-402f-b25c-4fdf2236b7d3"). InnerVolumeSpecName "pre-adoption-validation-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:59:33 crc kubenswrapper[4790]: I1124 14:59:33.512437 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba0f2f40-136f-402f-b25c-4fdf2236b7d3-ceph" (OuterVolumeSpecName: "ceph") pod "ba0f2f40-136f-402f-b25c-4fdf2236b7d3" (UID: "ba0f2f40-136f-402f-b25c-4fdf2236b7d3"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:59:33 crc kubenswrapper[4790]: I1124 14:59:33.544518 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba0f2f40-136f-402f-b25c-4fdf2236b7d3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ba0f2f40-136f-402f-b25c-4fdf2236b7d3" (UID: "ba0f2f40-136f-402f-b25c-4fdf2236b7d3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:59:33 crc kubenswrapper[4790]: I1124 14:59:33.552424 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba0f2f40-136f-402f-b25c-4fdf2236b7d3-inventory" (OuterVolumeSpecName: "inventory") pod "ba0f2f40-136f-402f-b25c-4fdf2236b7d3" (UID: "ba0f2f40-136f-402f-b25c-4fdf2236b7d3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:59:33 crc kubenswrapper[4790]: I1124 14:59:33.608632 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgbg2\" (UniqueName: \"kubernetes.io/projected/ba0f2f40-136f-402f-b25c-4fdf2236b7d3-kube-api-access-dgbg2\") on node \"crc\" DevicePath \"\"" Nov 24 14:59:33 crc kubenswrapper[4790]: I1124 14:59:33.608694 4790 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba0f2f40-136f-402f-b25c-4fdf2236b7d3-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 14:59:33 crc kubenswrapper[4790]: I1124 14:59:33.608714 4790 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ba0f2f40-136f-402f-b25c-4fdf2236b7d3-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 14:59:33 crc kubenswrapper[4790]: I1124 14:59:33.608734 4790 reconciler_common.go:293] "Volume detached for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba0f2f40-136f-402f-b25c-4fdf2236b7d3-pre-adoption-validation-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:59:33 crc kubenswrapper[4790]: I1124 14:59:33.608754 4790 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba0f2f40-136f-402f-b25c-4fdf2236b7d3-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:59:33 crc kubenswrapper[4790]: I1124 14:59:33.768976 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk" event={"ID":"ba0f2f40-136f-402f-b25c-4fdf2236b7d3","Type":"ContainerDied","Data":"ab67cdedb715aef1a32ab7efb1b0a5f28794b73cde19939d48ff9a63dc654839"} Nov 24 14:59:33 crc kubenswrapper[4790]: I1124 14:59:33.769031 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab67cdedb715aef1a32ab7efb1b0a5f28794b73cde19939d48ff9a63dc654839" Nov 24 14:59:33 crc kubenswrapper[4790]: I1124 14:59:33.769078 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk" Nov 24 14:59:38 crc kubenswrapper[4790]: I1124 14:59:38.352553 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6"] Nov 24 14:59:38 crc kubenswrapper[4790]: E1124 14:59:38.353983 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba0f2f40-136f-402f-b25c-4fdf2236b7d3" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Nov 24 14:59:38 crc kubenswrapper[4790]: I1124 14:59:38.354016 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba0f2f40-136f-402f-b25c-4fdf2236b7d3" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Nov 24 14:59:38 crc kubenswrapper[4790]: I1124 14:59:38.354536 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba0f2f40-136f-402f-b25c-4fdf2236b7d3" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Nov 24 14:59:38 crc kubenswrapper[4790]: I1124 14:59:38.356062 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6" Nov 24 14:59:38 crc kubenswrapper[4790]: I1124 14:59:38.364393 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 14:59:38 crc kubenswrapper[4790]: I1124 14:59:38.365875 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 14:59:38 crc kubenswrapper[4790]: I1124 14:59:38.365966 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fc497" Nov 24 14:59:38 crc kubenswrapper[4790]: I1124 14:59:38.366271 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 14:59:38 crc kubenswrapper[4790]: I1124 14:59:38.385548 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6"] Nov 24 14:59:38 crc kubenswrapper[4790]: I1124 14:59:38.542144 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0effec18-dddb-4967-a5f8-1a98979e3338-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6\" (UID: \"0effec18-dddb-4967-a5f8-1a98979e3338\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6" Nov 24 14:59:38 crc kubenswrapper[4790]: I1124 14:59:38.542246 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0effec18-dddb-4967-a5f8-1a98979e3338-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6\" (UID: \"0effec18-dddb-4967-a5f8-1a98979e3338\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6" Nov 24 14:59:38 crc kubenswrapper[4790]: I1124 14:59:38.542351 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv2bj\" (UniqueName: \"kubernetes.io/projected/0effec18-dddb-4967-a5f8-1a98979e3338-kube-api-access-bv2bj\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6\" (UID: \"0effec18-dddb-4967-a5f8-1a98979e3338\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6" Nov 24 14:59:38 crc kubenswrapper[4790]: I1124 14:59:38.542388 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0effec18-dddb-4967-a5f8-1a98979e3338-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6\" (UID: \"0effec18-dddb-4967-a5f8-1a98979e3338\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6" Nov 24 14:59:38 crc kubenswrapper[4790]: I1124 14:59:38.542413 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0effec18-dddb-4967-a5f8-1a98979e3338-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6\" (UID: \"0effec18-dddb-4967-a5f8-1a98979e3338\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6" Nov 24 14:59:38 crc kubenswrapper[4790]: I1124 14:59:38.644920 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0effec18-dddb-4967-a5f8-1a98979e3338-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6\" (UID: \"0effec18-dddb-4967-a5f8-1a98979e3338\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6" Nov 24 14:59:38 crc kubenswrapper[4790]: I1124 14:59:38.645110 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0effec18-dddb-4967-a5f8-1a98979e3338-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6\" (UID: \"0effec18-dddb-4967-a5f8-1a98979e3338\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6" Nov 24 14:59:38 crc kubenswrapper[4790]: I1124 14:59:38.645300 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv2bj\" (UniqueName: \"kubernetes.io/projected/0effec18-dddb-4967-a5f8-1a98979e3338-kube-api-access-bv2bj\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6\" (UID: \"0effec18-dddb-4967-a5f8-1a98979e3338\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6" Nov 24 14:59:38 crc kubenswrapper[4790]: I1124 14:59:38.645374 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0effec18-dddb-4967-a5f8-1a98979e3338-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6\" (UID: \"0effec18-dddb-4967-a5f8-1a98979e3338\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6" Nov 24 14:59:38 crc kubenswrapper[4790]: I1124 14:59:38.645453 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0effec18-dddb-4967-a5f8-1a98979e3338-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6\" (UID: \"0effec18-dddb-4967-a5f8-1a98979e3338\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6" Nov 24 14:59:38 crc kubenswrapper[4790]: I1124 14:59:38.653195 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0effec18-dddb-4967-a5f8-1a98979e3338-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6\" (UID: \"0effec18-dddb-4967-a5f8-1a98979e3338\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6" Nov 24 14:59:38 crc kubenswrapper[4790]: I1124 14:59:38.653198 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0effec18-dddb-4967-a5f8-1a98979e3338-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6\" (UID: \"0effec18-dddb-4967-a5f8-1a98979e3338\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6" Nov 24 14:59:38 crc kubenswrapper[4790]: I1124 14:59:38.654089 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0effec18-dddb-4967-a5f8-1a98979e3338-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6\" (UID: \"0effec18-dddb-4967-a5f8-1a98979e3338\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6" Nov 24 14:59:38 crc kubenswrapper[4790]: I1124 14:59:38.664495 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0effec18-dddb-4967-a5f8-1a98979e3338-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6\" (UID: \"0effec18-dddb-4967-a5f8-1a98979e3338\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6" Nov 24 14:59:38 crc kubenswrapper[4790]: I1124 14:59:38.682802 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv2bj\" (UniqueName: \"kubernetes.io/projected/0effec18-dddb-4967-a5f8-1a98979e3338-kube-api-access-bv2bj\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6\" (UID: \"0effec18-dddb-4967-a5f8-1a98979e3338\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6" Nov 24 14:59:38 crc kubenswrapper[4790]: I1124 14:59:38.684651 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6" Nov 24 14:59:39 crc kubenswrapper[4790]: I1124 14:59:39.349450 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6"] Nov 24 14:59:39 crc kubenswrapper[4790]: I1124 14:59:39.842994 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6" event={"ID":"0effec18-dddb-4967-a5f8-1a98979e3338","Type":"ContainerStarted","Data":"c5689da16dc1b47d24271e836bba6bad763721756a42083787ba79987a81a1d0"} Nov 24 14:59:40 crc kubenswrapper[4790]: I1124 14:59:40.857002 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6" event={"ID":"0effec18-dddb-4967-a5f8-1a98979e3338","Type":"ContainerStarted","Data":"b9e707df255ff9aaba0fc3e34b5c9508408bf92671598883395b7a29beec6fb0"} Nov 24 14:59:40 crc kubenswrapper[4790]: I1124 14:59:40.895652 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6" podStartSLOduration=2.6941377749999997 podStartE2EDuration="2.895625003s" podCreationTimestamp="2025-11-24 14:59:38 +0000 UTC" firstStartedPulling="2025-11-24 14:59:39.352087231 +0000 UTC m=+6427.731980903" lastFinishedPulling="2025-11-24 14:59:39.553574459 +0000 UTC m=+6427.933468131" observedRunningTime="2025-11-24 14:59:40.882188578 +0000 UTC m=+6429.262082250" watchObservedRunningTime="2025-11-24 14:59:40.895625003 +0000 UTC m=+6429.275518705" Nov 24 15:00:00 crc kubenswrapper[4790]: I1124 15:00:00.164044 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399940-48kdm"] Nov 24 15:00:00 crc kubenswrapper[4790]: I1124 15:00:00.169056 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-48kdm" Nov 24 15:00:00 crc kubenswrapper[4790]: I1124 15:00:00.173637 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 15:00:00 crc kubenswrapper[4790]: I1124 15:00:00.173914 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 15:00:00 crc kubenswrapper[4790]: I1124 15:00:00.181689 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399940-48kdm"] Nov 24 15:00:00 crc kubenswrapper[4790]: I1124 15:00:00.264815 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fzkl\" (UniqueName: \"kubernetes.io/projected/2e688f6b-ed59-42f5-b5fb-3199f6ac7570-kube-api-access-7fzkl\") pod \"collect-profiles-29399940-48kdm\" (UID: \"2e688f6b-ed59-42f5-b5fb-3199f6ac7570\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-48kdm" Nov 24 15:00:00 crc kubenswrapper[4790]: I1124 15:00:00.265401 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2e688f6b-ed59-42f5-b5fb-3199f6ac7570-config-volume\") pod \"collect-profiles-29399940-48kdm\" (UID: \"2e688f6b-ed59-42f5-b5fb-3199f6ac7570\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-48kdm" Nov 24 15:00:00 crc kubenswrapper[4790]: I1124 15:00:00.265517 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2e688f6b-ed59-42f5-b5fb-3199f6ac7570-secret-volume\") pod \"collect-profiles-29399940-48kdm\" (UID: \"2e688f6b-ed59-42f5-b5fb-3199f6ac7570\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-48kdm" Nov 24 15:00:00 crc kubenswrapper[4790]: I1124 15:00:00.368200 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2e688f6b-ed59-42f5-b5fb-3199f6ac7570-config-volume\") pod \"collect-profiles-29399940-48kdm\" (UID: \"2e688f6b-ed59-42f5-b5fb-3199f6ac7570\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-48kdm" Nov 24 15:00:00 crc kubenswrapper[4790]: I1124 15:00:00.368407 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2e688f6b-ed59-42f5-b5fb-3199f6ac7570-secret-volume\") pod \"collect-profiles-29399940-48kdm\" (UID: \"2e688f6b-ed59-42f5-b5fb-3199f6ac7570\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-48kdm" Nov 24 15:00:00 crc kubenswrapper[4790]: I1124 15:00:00.368639 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fzkl\" (UniqueName: \"kubernetes.io/projected/2e688f6b-ed59-42f5-b5fb-3199f6ac7570-kube-api-access-7fzkl\") pod \"collect-profiles-29399940-48kdm\" (UID: \"2e688f6b-ed59-42f5-b5fb-3199f6ac7570\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-48kdm" Nov 24 15:00:00 crc kubenswrapper[4790]: I1124 15:00:00.369433 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2e688f6b-ed59-42f5-b5fb-3199f6ac7570-config-volume\") pod \"collect-profiles-29399940-48kdm\" (UID: \"2e688f6b-ed59-42f5-b5fb-3199f6ac7570\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-48kdm" Nov 24 15:00:00 crc kubenswrapper[4790]: I1124 15:00:00.376124 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2e688f6b-ed59-42f5-b5fb-3199f6ac7570-secret-volume\") pod \"collect-profiles-29399940-48kdm\" (UID: \"2e688f6b-ed59-42f5-b5fb-3199f6ac7570\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-48kdm" Nov 24 15:00:00 crc kubenswrapper[4790]: I1124 15:00:00.395134 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fzkl\" (UniqueName: \"kubernetes.io/projected/2e688f6b-ed59-42f5-b5fb-3199f6ac7570-kube-api-access-7fzkl\") pod \"collect-profiles-29399940-48kdm\" (UID: \"2e688f6b-ed59-42f5-b5fb-3199f6ac7570\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-48kdm" Nov 24 15:00:00 crc kubenswrapper[4790]: I1124 15:00:00.507635 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-48kdm" Nov 24 15:00:01 crc kubenswrapper[4790]: I1124 15:00:01.037108 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399940-48kdm"] Nov 24 15:00:01 crc kubenswrapper[4790]: W1124 15:00:01.040783 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e688f6b_ed59_42f5_b5fb_3199f6ac7570.slice/crio-f5fa1fe4dab1b1493e0c632edac30d761fb0a1bd19dd97275ce9f2d2cf0a6c17 WatchSource:0}: Error finding container f5fa1fe4dab1b1493e0c632edac30d761fb0a1bd19dd97275ce9f2d2cf0a6c17: Status 404 returned error can't find the container with id f5fa1fe4dab1b1493e0c632edac30d761fb0a1bd19dd97275ce9f2d2cf0a6c17 Nov 24 15:00:01 crc kubenswrapper[4790]: I1124 15:00:01.207180 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-48kdm" event={"ID":"2e688f6b-ed59-42f5-b5fb-3199f6ac7570","Type":"ContainerStarted","Data":"f5fa1fe4dab1b1493e0c632edac30d761fb0a1bd19dd97275ce9f2d2cf0a6c17"} Nov 24 15:00:01 crc kubenswrapper[4790]: I1124 15:00:01.459077 4790 scope.go:117] "RemoveContainer" containerID="a49ae406ba85a7abbfc33e2a4270e26ad1729908b9d461a811cfe8f4dbf153be" Nov 24 15:00:01 crc kubenswrapper[4790]: I1124 15:00:01.490613 4790 scope.go:117] "RemoveContainer" containerID="6ed027e62487b0ee9f3027bdeffb9da69254fa8f826f89c13563c3527f0a6634" Nov 24 15:00:01 crc kubenswrapper[4790]: I1124 15:00:01.543409 4790 scope.go:117] "RemoveContainer" containerID="88ef69ff632a65bbcec6feb14ea1545eba4f781f2083a2d3b071757ad3e80949" Nov 24 15:00:01 crc kubenswrapper[4790]: I1124 15:00:01.587096 4790 scope.go:117] "RemoveContainer" containerID="0a4bb43787f72515fb17ba0b5afce80e2325d47ac13d50a956015c340f3ea575" Nov 24 15:00:02 crc kubenswrapper[4790]: I1124 15:00:02.225127 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-48kdm" event={"ID":"2e688f6b-ed59-42f5-b5fb-3199f6ac7570","Type":"ContainerStarted","Data":"bca98e9f7a1e74274b9a4a717d4ca591f5ad30f24c0084ec3b64204d853803ab"} Nov 24 15:00:02 crc kubenswrapper[4790]: I1124 15:00:02.247998 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-48kdm" podStartSLOduration=2.247978223 podStartE2EDuration="2.247978223s" podCreationTimestamp="2025-11-24 15:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 15:00:02.24544067 +0000 UTC m=+6450.625334372" watchObservedRunningTime="2025-11-24 15:00:02.247978223 +0000 UTC m=+6450.627871885" Nov 24 15:00:03 crc kubenswrapper[4790]: E1124 15:00:03.185125 4790 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e688f6b_ed59_42f5_b5fb_3199f6ac7570.slice/crio-bca98e9f7a1e74274b9a4a717d4ca591f5ad30f24c0084ec3b64204d853803ab.scope\": RecentStats: unable to find data in memory cache]" Nov 24 15:00:03 crc kubenswrapper[4790]: I1124 15:00:03.244787 4790 generic.go:334] "Generic (PLEG): container finished" podID="2e688f6b-ed59-42f5-b5fb-3199f6ac7570" containerID="bca98e9f7a1e74274b9a4a717d4ca591f5ad30f24c0084ec3b64204d853803ab" exitCode=0 Nov 24 15:00:03 crc kubenswrapper[4790]: I1124 15:00:03.244830 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-48kdm" event={"ID":"2e688f6b-ed59-42f5-b5fb-3199f6ac7570","Type":"ContainerDied","Data":"bca98e9f7a1e74274b9a4a717d4ca591f5ad30f24c0084ec3b64204d853803ab"} Nov 24 15:00:04 crc kubenswrapper[4790]: I1124 15:00:04.059799 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-sync-qgdgj"] Nov 24 15:00:04 crc kubenswrapper[4790]: I1124 15:00:04.073695 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-sync-qgdgj"] Nov 24 15:00:04 crc kubenswrapper[4790]: I1124 15:00:04.343219 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5ef870d-41ea-48e7-9a44-3f4a610f0a62" path="/var/lib/kubelet/pods/c5ef870d-41ea-48e7-9a44-3f4a610f0a62/volumes" Nov 24 15:00:04 crc kubenswrapper[4790]: I1124 15:00:04.754025 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-48kdm" Nov 24 15:00:04 crc kubenswrapper[4790]: I1124 15:00:04.881195 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2e688f6b-ed59-42f5-b5fb-3199f6ac7570-secret-volume\") pod \"2e688f6b-ed59-42f5-b5fb-3199f6ac7570\" (UID: \"2e688f6b-ed59-42f5-b5fb-3199f6ac7570\") " Nov 24 15:00:04 crc kubenswrapper[4790]: I1124 15:00:04.881317 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fzkl\" (UniqueName: \"kubernetes.io/projected/2e688f6b-ed59-42f5-b5fb-3199f6ac7570-kube-api-access-7fzkl\") pod \"2e688f6b-ed59-42f5-b5fb-3199f6ac7570\" (UID: \"2e688f6b-ed59-42f5-b5fb-3199f6ac7570\") " Nov 24 15:00:04 crc kubenswrapper[4790]: I1124 15:00:04.881448 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2e688f6b-ed59-42f5-b5fb-3199f6ac7570-config-volume\") pod \"2e688f6b-ed59-42f5-b5fb-3199f6ac7570\" (UID: \"2e688f6b-ed59-42f5-b5fb-3199f6ac7570\") " Nov 24 15:00:04 crc kubenswrapper[4790]: I1124 15:00:04.882316 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e688f6b-ed59-42f5-b5fb-3199f6ac7570-config-volume" (OuterVolumeSpecName: "config-volume") pod "2e688f6b-ed59-42f5-b5fb-3199f6ac7570" (UID: "2e688f6b-ed59-42f5-b5fb-3199f6ac7570"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 15:00:04 crc kubenswrapper[4790]: I1124 15:00:04.888630 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e688f6b-ed59-42f5-b5fb-3199f6ac7570-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2e688f6b-ed59-42f5-b5fb-3199f6ac7570" (UID: "2e688f6b-ed59-42f5-b5fb-3199f6ac7570"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:00:04 crc kubenswrapper[4790]: I1124 15:00:04.889984 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e688f6b-ed59-42f5-b5fb-3199f6ac7570-kube-api-access-7fzkl" (OuterVolumeSpecName: "kube-api-access-7fzkl") pod "2e688f6b-ed59-42f5-b5fb-3199f6ac7570" (UID: "2e688f6b-ed59-42f5-b5fb-3199f6ac7570"). InnerVolumeSpecName "kube-api-access-7fzkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:00:04 crc kubenswrapper[4790]: I1124 15:00:04.984254 4790 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2e688f6b-ed59-42f5-b5fb-3199f6ac7570-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 15:00:04 crc kubenswrapper[4790]: I1124 15:00:04.984296 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fzkl\" (UniqueName: \"kubernetes.io/projected/2e688f6b-ed59-42f5-b5fb-3199f6ac7570-kube-api-access-7fzkl\") on node \"crc\" DevicePath \"\"" Nov 24 15:00:04 crc kubenswrapper[4790]: I1124 15:00:04.984310 4790 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2e688f6b-ed59-42f5-b5fb-3199f6ac7570-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 15:00:05 crc kubenswrapper[4790]: I1124 15:00:05.274065 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-48kdm" event={"ID":"2e688f6b-ed59-42f5-b5fb-3199f6ac7570","Type":"ContainerDied","Data":"f5fa1fe4dab1b1493e0c632edac30d761fb0a1bd19dd97275ce9f2d2cf0a6c17"} Nov 24 15:00:05 crc kubenswrapper[4790]: I1124 15:00:05.274706 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f5fa1fe4dab1b1493e0c632edac30d761fb0a1bd19dd97275ce9f2d2cf0a6c17" Nov 24 15:00:05 crc kubenswrapper[4790]: I1124 15:00:05.274132 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399940-48kdm" Nov 24 15:00:05 crc kubenswrapper[4790]: I1124 15:00:05.345244 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399895-t6g7z"] Nov 24 15:00:05 crc kubenswrapper[4790]: I1124 15:00:05.355635 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399895-t6g7z"] Nov 24 15:00:06 crc kubenswrapper[4790]: I1124 15:00:06.330082 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bd45d25-7193-44bd-9a32-06329ca3fac1" path="/var/lib/kubelet/pods/2bd45d25-7193-44bd-9a32-06329ca3fac1/volumes" Nov 24 15:00:13 crc kubenswrapper[4790]: I1124 15:00:13.939282 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:00:13 crc kubenswrapper[4790]: I1124 15:00:13.940194 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:00:43 crc kubenswrapper[4790]: I1124 15:00:43.938682 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:00:43 crc kubenswrapper[4790]: I1124 15:00:43.939640 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:01:00 crc kubenswrapper[4790]: I1124 15:01:00.166929 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29399941-28pp2"] Nov 24 15:01:00 crc kubenswrapper[4790]: E1124 15:01:00.168455 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e688f6b-ed59-42f5-b5fb-3199f6ac7570" containerName="collect-profiles" Nov 24 15:01:00 crc kubenswrapper[4790]: I1124 15:01:00.168491 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e688f6b-ed59-42f5-b5fb-3199f6ac7570" containerName="collect-profiles" Nov 24 15:01:00 crc kubenswrapper[4790]: I1124 15:01:00.168949 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e688f6b-ed59-42f5-b5fb-3199f6ac7570" containerName="collect-profiles" Nov 24 15:01:00 crc kubenswrapper[4790]: I1124 15:01:00.170627 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399941-28pp2" Nov 24 15:01:00 crc kubenswrapper[4790]: I1124 15:01:00.180211 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29399941-28pp2"] Nov 24 15:01:00 crc kubenswrapper[4790]: I1124 15:01:00.372245 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/292b9a6b-b644-46a2-b4ba-2ba30d46e345-fernet-keys\") pod \"keystone-cron-29399941-28pp2\" (UID: \"292b9a6b-b644-46a2-b4ba-2ba30d46e345\") " pod="openstack/keystone-cron-29399941-28pp2" Nov 24 15:01:00 crc kubenswrapper[4790]: I1124 15:01:00.372499 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/292b9a6b-b644-46a2-b4ba-2ba30d46e345-config-data\") pod \"keystone-cron-29399941-28pp2\" (UID: \"292b9a6b-b644-46a2-b4ba-2ba30d46e345\") " pod="openstack/keystone-cron-29399941-28pp2" Nov 24 15:01:00 crc kubenswrapper[4790]: I1124 15:01:00.372586 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8w4b\" (UniqueName: \"kubernetes.io/projected/292b9a6b-b644-46a2-b4ba-2ba30d46e345-kube-api-access-x8w4b\") pod \"keystone-cron-29399941-28pp2\" (UID: \"292b9a6b-b644-46a2-b4ba-2ba30d46e345\") " pod="openstack/keystone-cron-29399941-28pp2" Nov 24 15:01:00 crc kubenswrapper[4790]: I1124 15:01:00.372726 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/292b9a6b-b644-46a2-b4ba-2ba30d46e345-combined-ca-bundle\") pod \"keystone-cron-29399941-28pp2\" (UID: \"292b9a6b-b644-46a2-b4ba-2ba30d46e345\") " pod="openstack/keystone-cron-29399941-28pp2" Nov 24 15:01:00 crc kubenswrapper[4790]: I1124 15:01:00.475531 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/292b9a6b-b644-46a2-b4ba-2ba30d46e345-config-data\") pod \"keystone-cron-29399941-28pp2\" (UID: \"292b9a6b-b644-46a2-b4ba-2ba30d46e345\") " pod="openstack/keystone-cron-29399941-28pp2" Nov 24 15:01:00 crc kubenswrapper[4790]: I1124 15:01:00.475673 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8w4b\" (UniqueName: \"kubernetes.io/projected/292b9a6b-b644-46a2-b4ba-2ba30d46e345-kube-api-access-x8w4b\") pod \"keystone-cron-29399941-28pp2\" (UID: \"292b9a6b-b644-46a2-b4ba-2ba30d46e345\") " pod="openstack/keystone-cron-29399941-28pp2" Nov 24 15:01:00 crc kubenswrapper[4790]: I1124 15:01:00.476426 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/292b9a6b-b644-46a2-b4ba-2ba30d46e345-combined-ca-bundle\") pod \"keystone-cron-29399941-28pp2\" (UID: \"292b9a6b-b644-46a2-b4ba-2ba30d46e345\") " pod="openstack/keystone-cron-29399941-28pp2" Nov 24 15:01:00 crc kubenswrapper[4790]: I1124 15:01:00.476755 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/292b9a6b-b644-46a2-b4ba-2ba30d46e345-fernet-keys\") pod \"keystone-cron-29399941-28pp2\" (UID: \"292b9a6b-b644-46a2-b4ba-2ba30d46e345\") " pod="openstack/keystone-cron-29399941-28pp2" Nov 24 15:01:00 crc kubenswrapper[4790]: I1124 15:01:00.483835 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/292b9a6b-b644-46a2-b4ba-2ba30d46e345-config-data\") pod \"keystone-cron-29399941-28pp2\" (UID: \"292b9a6b-b644-46a2-b4ba-2ba30d46e345\") " pod="openstack/keystone-cron-29399941-28pp2" Nov 24 15:01:00 crc kubenswrapper[4790]: I1124 15:01:00.485249 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/292b9a6b-b644-46a2-b4ba-2ba30d46e345-combined-ca-bundle\") pod \"keystone-cron-29399941-28pp2\" (UID: \"292b9a6b-b644-46a2-b4ba-2ba30d46e345\") " pod="openstack/keystone-cron-29399941-28pp2" Nov 24 15:01:00 crc kubenswrapper[4790]: I1124 15:01:00.496251 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/292b9a6b-b644-46a2-b4ba-2ba30d46e345-fernet-keys\") pod \"keystone-cron-29399941-28pp2\" (UID: \"292b9a6b-b644-46a2-b4ba-2ba30d46e345\") " pod="openstack/keystone-cron-29399941-28pp2" Nov 24 15:01:00 crc kubenswrapper[4790]: I1124 15:01:00.499223 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8w4b\" (UniqueName: \"kubernetes.io/projected/292b9a6b-b644-46a2-b4ba-2ba30d46e345-kube-api-access-x8w4b\") pod \"keystone-cron-29399941-28pp2\" (UID: \"292b9a6b-b644-46a2-b4ba-2ba30d46e345\") " pod="openstack/keystone-cron-29399941-28pp2" Nov 24 15:01:00 crc kubenswrapper[4790]: I1124 15:01:00.506181 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399941-28pp2" Nov 24 15:01:01 crc kubenswrapper[4790]: I1124 15:01:01.027797 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29399941-28pp2"] Nov 24 15:01:01 crc kubenswrapper[4790]: W1124 15:01:01.031621 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod292b9a6b_b644_46a2_b4ba_2ba30d46e345.slice/crio-786a48d668d4259194aee9eaccbb453ee05e38b2ff266096e87e09810b0c089b WatchSource:0}: Error finding container 786a48d668d4259194aee9eaccbb453ee05e38b2ff266096e87e09810b0c089b: Status 404 returned error can't find the container with id 786a48d668d4259194aee9eaccbb453ee05e38b2ff266096e87e09810b0c089b Nov 24 15:01:01 crc kubenswrapper[4790]: I1124 15:01:01.733009 4790 scope.go:117] "RemoveContainer" containerID="ce841e968dcaaf518b1267509e0f378bb950ec20d1f40926209f390055d76e84" Nov 24 15:01:01 crc kubenswrapper[4790]: I1124 15:01:01.763270 4790 scope.go:117] "RemoveContainer" containerID="9adb971e143f83f59505b53ba65f7f8b951a7edfbb0affb96c95b125c67c46a2" Nov 24 15:01:01 crc kubenswrapper[4790]: I1124 15:01:01.831682 4790 scope.go:117] "RemoveContainer" containerID="f836dfded5e80709de981a9ca934d8ffdd578779daa2106a025e0792c6a7b839" Nov 24 15:01:01 crc kubenswrapper[4790]: I1124 15:01:01.969521 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399941-28pp2" event={"ID":"292b9a6b-b644-46a2-b4ba-2ba30d46e345","Type":"ContainerStarted","Data":"231b7a780e0acd6da493ee34d76d75b9f668b0db283d3c3bd10b9a152f45e785"} Nov 24 15:01:01 crc kubenswrapper[4790]: I1124 15:01:01.969561 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399941-28pp2" event={"ID":"292b9a6b-b644-46a2-b4ba-2ba30d46e345","Type":"ContainerStarted","Data":"786a48d668d4259194aee9eaccbb453ee05e38b2ff266096e87e09810b0c089b"} Nov 24 15:01:01 crc kubenswrapper[4790]: I1124 15:01:01.988005 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29399941-28pp2" podStartSLOduration=1.987987908 podStartE2EDuration="1.987987908s" podCreationTimestamp="2025-11-24 15:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 15:01:01.981837722 +0000 UTC m=+6510.361731384" watchObservedRunningTime="2025-11-24 15:01:01.987987908 +0000 UTC m=+6510.367881560" Nov 24 15:01:05 crc kubenswrapper[4790]: I1124 15:01:05.009315 4790 generic.go:334] "Generic (PLEG): container finished" podID="292b9a6b-b644-46a2-b4ba-2ba30d46e345" containerID="231b7a780e0acd6da493ee34d76d75b9f668b0db283d3c3bd10b9a152f45e785" exitCode=0 Nov 24 15:01:05 crc kubenswrapper[4790]: I1124 15:01:05.009401 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399941-28pp2" event={"ID":"292b9a6b-b644-46a2-b4ba-2ba30d46e345","Type":"ContainerDied","Data":"231b7a780e0acd6da493ee34d76d75b9f668b0db283d3c3bd10b9a152f45e785"} Nov 24 15:01:06 crc kubenswrapper[4790]: I1124 15:01:06.470803 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399941-28pp2" Nov 24 15:01:06 crc kubenswrapper[4790]: I1124 15:01:06.546963 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/292b9a6b-b644-46a2-b4ba-2ba30d46e345-config-data\") pod \"292b9a6b-b644-46a2-b4ba-2ba30d46e345\" (UID: \"292b9a6b-b644-46a2-b4ba-2ba30d46e345\") " Nov 24 15:01:06 crc kubenswrapper[4790]: I1124 15:01:06.547051 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/292b9a6b-b644-46a2-b4ba-2ba30d46e345-fernet-keys\") pod \"292b9a6b-b644-46a2-b4ba-2ba30d46e345\" (UID: \"292b9a6b-b644-46a2-b4ba-2ba30d46e345\") " Nov 24 15:01:06 crc kubenswrapper[4790]: I1124 15:01:06.547141 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8w4b\" (UniqueName: \"kubernetes.io/projected/292b9a6b-b644-46a2-b4ba-2ba30d46e345-kube-api-access-x8w4b\") pod \"292b9a6b-b644-46a2-b4ba-2ba30d46e345\" (UID: \"292b9a6b-b644-46a2-b4ba-2ba30d46e345\") " Nov 24 15:01:06 crc kubenswrapper[4790]: I1124 15:01:06.547216 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/292b9a6b-b644-46a2-b4ba-2ba30d46e345-combined-ca-bundle\") pod \"292b9a6b-b644-46a2-b4ba-2ba30d46e345\" (UID: \"292b9a6b-b644-46a2-b4ba-2ba30d46e345\") " Nov 24 15:01:06 crc kubenswrapper[4790]: I1124 15:01:06.554198 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/292b9a6b-b644-46a2-b4ba-2ba30d46e345-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "292b9a6b-b644-46a2-b4ba-2ba30d46e345" (UID: "292b9a6b-b644-46a2-b4ba-2ba30d46e345"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:01:06 crc kubenswrapper[4790]: I1124 15:01:06.554604 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/292b9a6b-b644-46a2-b4ba-2ba30d46e345-kube-api-access-x8w4b" (OuterVolumeSpecName: "kube-api-access-x8w4b") pod "292b9a6b-b644-46a2-b4ba-2ba30d46e345" (UID: "292b9a6b-b644-46a2-b4ba-2ba30d46e345"). InnerVolumeSpecName "kube-api-access-x8w4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:01:06 crc kubenswrapper[4790]: I1124 15:01:06.584024 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/292b9a6b-b644-46a2-b4ba-2ba30d46e345-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "292b9a6b-b644-46a2-b4ba-2ba30d46e345" (UID: "292b9a6b-b644-46a2-b4ba-2ba30d46e345"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:01:06 crc kubenswrapper[4790]: I1124 15:01:06.612830 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/292b9a6b-b644-46a2-b4ba-2ba30d46e345-config-data" (OuterVolumeSpecName: "config-data") pod "292b9a6b-b644-46a2-b4ba-2ba30d46e345" (UID: "292b9a6b-b644-46a2-b4ba-2ba30d46e345"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:01:06 crc kubenswrapper[4790]: I1124 15:01:06.649474 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/292b9a6b-b644-46a2-b4ba-2ba30d46e345-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:01:06 crc kubenswrapper[4790]: I1124 15:01:06.649509 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/292b9a6b-b644-46a2-b4ba-2ba30d46e345-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 15:01:06 crc kubenswrapper[4790]: I1124 15:01:06.649517 4790 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/292b9a6b-b644-46a2-b4ba-2ba30d46e345-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 15:01:06 crc kubenswrapper[4790]: I1124 15:01:06.649527 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8w4b\" (UniqueName: \"kubernetes.io/projected/292b9a6b-b644-46a2-b4ba-2ba30d46e345-kube-api-access-x8w4b\") on node \"crc\" DevicePath \"\"" Nov 24 15:01:07 crc kubenswrapper[4790]: I1124 15:01:07.036045 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399941-28pp2" event={"ID":"292b9a6b-b644-46a2-b4ba-2ba30d46e345","Type":"ContainerDied","Data":"786a48d668d4259194aee9eaccbb453ee05e38b2ff266096e87e09810b0c089b"} Nov 24 15:01:07 crc kubenswrapper[4790]: I1124 15:01:07.036410 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="786a48d668d4259194aee9eaccbb453ee05e38b2ff266096e87e09810b0c089b" Nov 24 15:01:07 crc kubenswrapper[4790]: I1124 15:01:07.036136 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399941-28pp2" Nov 24 15:01:13 crc kubenswrapper[4790]: I1124 15:01:13.939500 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:01:13 crc kubenswrapper[4790]: I1124 15:01:13.940392 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:01:13 crc kubenswrapper[4790]: I1124 15:01:13.940482 4790 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 15:01:13 crc kubenswrapper[4790]: I1124 15:01:13.941866 4790 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"848bb411b4e62828968580971f8aaefd4e3c2cc50ce3c086c482e5c081b37c96"} pod="openshift-machine-config-operator/machine-config-daemon-xz49t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 15:01:13 crc kubenswrapper[4790]: I1124 15:01:13.942039 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" containerID="cri-o://848bb411b4e62828968580971f8aaefd4e3c2cc50ce3c086c482e5c081b37c96" gracePeriod=600 Nov 24 15:01:14 crc kubenswrapper[4790]: I1124 15:01:14.147544 4790 generic.go:334] "Generic (PLEG): container finished" podID="0d73b133-48f0-455f-8f6a-742e633f631a" containerID="848bb411b4e62828968580971f8aaefd4e3c2cc50ce3c086c482e5c081b37c96" exitCode=0 Nov 24 15:01:14 crc kubenswrapper[4790]: I1124 15:01:14.147608 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerDied","Data":"848bb411b4e62828968580971f8aaefd4e3c2cc50ce3c086c482e5c081b37c96"} Nov 24 15:01:14 crc kubenswrapper[4790]: I1124 15:01:14.147657 4790 scope.go:117] "RemoveContainer" containerID="28308703817a1d4e0e5344ce19268127c600894f2daa6b921d52a9105f014132" Nov 24 15:01:15 crc kubenswrapper[4790]: I1124 15:01:15.165528 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"a4432de01408933d930865fc3e585b6bb44fe87c6f5674553d2039220c20c92e"} Nov 24 15:02:01 crc kubenswrapper[4790]: I1124 15:02:01.947927 4790 scope.go:117] "RemoveContainer" containerID="3f15a721b100c376d986283df9271dd98af54c99d32f88b55b9e5916d76d5511" Nov 24 15:02:01 crc kubenswrapper[4790]: I1124 15:02:01.989389 4790 scope.go:117] "RemoveContainer" containerID="2017ab431c0705f7a464fc6a3033bb6312f0d9fb4b202198caed6be1ba577b36" Nov 24 15:02:02 crc kubenswrapper[4790]: I1124 15:02:02.039340 4790 scope.go:117] "RemoveContainer" containerID="d5cb2bb1b93c40424bb48a6a0e7f76518b73206f6e2011df2fd1ab5f2673ff95" Nov 24 15:03:35 crc kubenswrapper[4790]: I1124 15:03:35.071000 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-pd5mv"] Nov 24 15:03:35 crc kubenswrapper[4790]: I1124 15:03:35.088188 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-ff8b-account-create-fw2ks"] Nov 24 15:03:35 crc kubenswrapper[4790]: I1124 15:03:35.103332 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-pd5mv"] Nov 24 15:03:35 crc kubenswrapper[4790]: I1124 15:03:35.115567 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-ff8b-account-create-fw2ks"] Nov 24 15:03:36 crc kubenswrapper[4790]: I1124 15:03:36.336944 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6501aed1-881c-481d-abd1-92da3c7d0121" path="/var/lib/kubelet/pods/6501aed1-881c-481d-abd1-92da3c7d0121/volumes" Nov 24 15:03:36 crc kubenswrapper[4790]: I1124 15:03:36.338241 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f152e218-4ab5-45a5-a87b-c8d4178c0f20" path="/var/lib/kubelet/pods/f152e218-4ab5-45a5-a87b-c8d4178c0f20/volumes" Nov 24 15:03:43 crc kubenswrapper[4790]: I1124 15:03:43.939043 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:03:43 crc kubenswrapper[4790]: I1124 15:03:43.939773 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:03:48 crc kubenswrapper[4790]: I1124 15:03:48.053293 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-j42q5"] Nov 24 15:03:48 crc kubenswrapper[4790]: I1124 15:03:48.069981 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-j42q5"] Nov 24 15:03:48 crc kubenswrapper[4790]: I1124 15:03:48.337351 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d55c02d-16d4-41aa-9ac5-98236ea35b71" path="/var/lib/kubelet/pods/9d55c02d-16d4-41aa-9ac5-98236ea35b71/volumes" Nov 24 15:04:02 crc kubenswrapper[4790]: I1124 15:04:02.163190 4790 scope.go:117] "RemoveContainer" containerID="f85f79c2bf1bdb2d3e9824fb8e78a4de6856485471122600b3b2c511fa64f873" Nov 24 15:04:02 crc kubenswrapper[4790]: I1124 15:04:02.192278 4790 scope.go:117] "RemoveContainer" containerID="ee8bb4fd17e401091383fb7fb4fa7bf0909ed573b8dde432e905058966c643b3" Nov 24 15:04:02 crc kubenswrapper[4790]: I1124 15:04:02.263610 4790 scope.go:117] "RemoveContainer" containerID="ec9bedbc7b636fd77da187df1c0e895ea495bc019f0ff22ce501ab052a36d412" Nov 24 15:04:13 crc kubenswrapper[4790]: I1124 15:04:13.938402 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:04:13 crc kubenswrapper[4790]: I1124 15:04:13.939053 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:04:28 crc kubenswrapper[4790]: I1124 15:04:28.821188 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wqbf7"] Nov 24 15:04:28 crc kubenswrapper[4790]: E1124 15:04:28.822239 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="292b9a6b-b644-46a2-b4ba-2ba30d46e345" containerName="keystone-cron" Nov 24 15:04:28 crc kubenswrapper[4790]: I1124 15:04:28.822251 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="292b9a6b-b644-46a2-b4ba-2ba30d46e345" containerName="keystone-cron" Nov 24 15:04:28 crc kubenswrapper[4790]: I1124 15:04:28.822508 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="292b9a6b-b644-46a2-b4ba-2ba30d46e345" containerName="keystone-cron" Nov 24 15:04:28 crc kubenswrapper[4790]: I1124 15:04:28.824164 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wqbf7" Nov 24 15:04:28 crc kubenswrapper[4790]: I1124 15:04:28.859173 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wqbf7"] Nov 24 15:04:28 crc kubenswrapper[4790]: I1124 15:04:28.965594 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a23cdfa-e59f-4a03-b368-f3cfc8964a0b-catalog-content\") pod \"certified-operators-wqbf7\" (UID: \"3a23cdfa-e59f-4a03-b368-f3cfc8964a0b\") " pod="openshift-marketplace/certified-operators-wqbf7" Nov 24 15:04:28 crc kubenswrapper[4790]: I1124 15:04:28.965794 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a23cdfa-e59f-4a03-b368-f3cfc8964a0b-utilities\") pod \"certified-operators-wqbf7\" (UID: \"3a23cdfa-e59f-4a03-b368-f3cfc8964a0b\") " pod="openshift-marketplace/certified-operators-wqbf7" Nov 24 15:04:28 crc kubenswrapper[4790]: I1124 15:04:28.965844 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ltcc\" (UniqueName: \"kubernetes.io/projected/3a23cdfa-e59f-4a03-b368-f3cfc8964a0b-kube-api-access-5ltcc\") pod \"certified-operators-wqbf7\" (UID: \"3a23cdfa-e59f-4a03-b368-f3cfc8964a0b\") " pod="openshift-marketplace/certified-operators-wqbf7" Nov 24 15:04:29 crc kubenswrapper[4790]: I1124 15:04:29.067782 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a23cdfa-e59f-4a03-b368-f3cfc8964a0b-utilities\") pod \"certified-operators-wqbf7\" (UID: \"3a23cdfa-e59f-4a03-b368-f3cfc8964a0b\") " pod="openshift-marketplace/certified-operators-wqbf7" Nov 24 15:04:29 crc kubenswrapper[4790]: I1124 15:04:29.068283 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ltcc\" (UniqueName: \"kubernetes.io/projected/3a23cdfa-e59f-4a03-b368-f3cfc8964a0b-kube-api-access-5ltcc\") pod \"certified-operators-wqbf7\" (UID: \"3a23cdfa-e59f-4a03-b368-f3cfc8964a0b\") " pod="openshift-marketplace/certified-operators-wqbf7" Nov 24 15:04:29 crc kubenswrapper[4790]: I1124 15:04:29.068515 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a23cdfa-e59f-4a03-b368-f3cfc8964a0b-utilities\") pod \"certified-operators-wqbf7\" (UID: \"3a23cdfa-e59f-4a03-b368-f3cfc8964a0b\") " pod="openshift-marketplace/certified-operators-wqbf7" Nov 24 15:04:29 crc kubenswrapper[4790]: I1124 15:04:29.068702 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a23cdfa-e59f-4a03-b368-f3cfc8964a0b-catalog-content\") pod \"certified-operators-wqbf7\" (UID: \"3a23cdfa-e59f-4a03-b368-f3cfc8964a0b\") " pod="openshift-marketplace/certified-operators-wqbf7" Nov 24 15:04:29 crc kubenswrapper[4790]: I1124 15:04:29.069152 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a23cdfa-e59f-4a03-b368-f3cfc8964a0b-catalog-content\") pod \"certified-operators-wqbf7\" (UID: \"3a23cdfa-e59f-4a03-b368-f3cfc8964a0b\") " pod="openshift-marketplace/certified-operators-wqbf7" Nov 24 15:04:29 crc kubenswrapper[4790]: I1124 15:04:29.089834 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ltcc\" (UniqueName: \"kubernetes.io/projected/3a23cdfa-e59f-4a03-b368-f3cfc8964a0b-kube-api-access-5ltcc\") pod \"certified-operators-wqbf7\" (UID: \"3a23cdfa-e59f-4a03-b368-f3cfc8964a0b\") " pod="openshift-marketplace/certified-operators-wqbf7" Nov 24 15:04:29 crc kubenswrapper[4790]: I1124 15:04:29.145275 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wqbf7" Nov 24 15:04:29 crc kubenswrapper[4790]: I1124 15:04:29.673275 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wqbf7"] Nov 24 15:04:29 crc kubenswrapper[4790]: I1124 15:04:29.689478 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wqbf7" event={"ID":"3a23cdfa-e59f-4a03-b368-f3cfc8964a0b","Type":"ContainerStarted","Data":"37ece91c99cd2b84486a18f511633e7f3030fa4e3382deec91ccc88df8a74d39"} Nov 24 15:04:30 crc kubenswrapper[4790]: I1124 15:04:30.703410 4790 generic.go:334] "Generic (PLEG): container finished" podID="3a23cdfa-e59f-4a03-b368-f3cfc8964a0b" containerID="6cf2c9526249e1eeea4f76b9ac9abaeed474c5926ad4ae8d7c3783784b2f9c6f" exitCode=0 Nov 24 15:04:30 crc kubenswrapper[4790]: I1124 15:04:30.703536 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wqbf7" event={"ID":"3a23cdfa-e59f-4a03-b368-f3cfc8964a0b","Type":"ContainerDied","Data":"6cf2c9526249e1eeea4f76b9ac9abaeed474c5926ad4ae8d7c3783784b2f9c6f"} Nov 24 15:04:30 crc kubenswrapper[4790]: I1124 15:04:30.706539 4790 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 15:04:31 crc kubenswrapper[4790]: I1124 15:04:31.714005 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wqbf7" event={"ID":"3a23cdfa-e59f-4a03-b368-f3cfc8964a0b","Type":"ContainerStarted","Data":"1fbe2a3d5d80bebe2cfc1abb0bc5629c3ce45983ecb2681322376df0898bd3cf"} Nov 24 15:04:32 crc kubenswrapper[4790]: I1124 15:04:32.727271 4790 generic.go:334] "Generic (PLEG): container finished" podID="3a23cdfa-e59f-4a03-b368-f3cfc8964a0b" containerID="1fbe2a3d5d80bebe2cfc1abb0bc5629c3ce45983ecb2681322376df0898bd3cf" exitCode=0 Nov 24 15:04:32 crc kubenswrapper[4790]: I1124 15:04:32.727339 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wqbf7" event={"ID":"3a23cdfa-e59f-4a03-b368-f3cfc8964a0b","Type":"ContainerDied","Data":"1fbe2a3d5d80bebe2cfc1abb0bc5629c3ce45983ecb2681322376df0898bd3cf"} Nov 24 15:04:33 crc kubenswrapper[4790]: I1124 15:04:33.740308 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wqbf7" event={"ID":"3a23cdfa-e59f-4a03-b368-f3cfc8964a0b","Type":"ContainerStarted","Data":"f78230ee3ddfca1845dd4468b6635fc445fce7df27c19c380943554ba58a7592"} Nov 24 15:04:33 crc kubenswrapper[4790]: I1124 15:04:33.775574 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wqbf7" podStartSLOduration=3.277093868 podStartE2EDuration="5.77554757s" podCreationTimestamp="2025-11-24 15:04:28 +0000 UTC" firstStartedPulling="2025-11-24 15:04:30.706288313 +0000 UTC m=+6719.086181975" lastFinishedPulling="2025-11-24 15:04:33.204741985 +0000 UTC m=+6721.584635677" observedRunningTime="2025-11-24 15:04:33.762033883 +0000 UTC m=+6722.141927545" watchObservedRunningTime="2025-11-24 15:04:33.77554757 +0000 UTC m=+6722.155441242" Nov 24 15:04:39 crc kubenswrapper[4790]: I1124 15:04:39.145413 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wqbf7" Nov 24 15:04:39 crc kubenswrapper[4790]: I1124 15:04:39.146037 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wqbf7" Nov 24 15:04:39 crc kubenswrapper[4790]: I1124 15:04:39.206402 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wqbf7" Nov 24 15:04:39 crc kubenswrapper[4790]: I1124 15:04:39.907687 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wqbf7" Nov 24 15:04:39 crc kubenswrapper[4790]: I1124 15:04:39.986436 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wqbf7"] Nov 24 15:04:41 crc kubenswrapper[4790]: I1124 15:04:41.845978 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wqbf7" podUID="3a23cdfa-e59f-4a03-b368-f3cfc8964a0b" containerName="registry-server" containerID="cri-o://f78230ee3ddfca1845dd4468b6635fc445fce7df27c19c380943554ba58a7592" gracePeriod=2 Nov 24 15:04:42 crc kubenswrapper[4790]: E1124 15:04:42.121873 4790 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a23cdfa_e59f_4a03_b368_f3cfc8964a0b.slice/crio-f78230ee3ddfca1845dd4468b6635fc445fce7df27c19c380943554ba58a7592.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a23cdfa_e59f_4a03_b368_f3cfc8964a0b.slice/crio-conmon-f78230ee3ddfca1845dd4468b6635fc445fce7df27c19c380943554ba58a7592.scope\": RecentStats: unable to find data in memory cache]" Nov 24 15:04:42 crc kubenswrapper[4790]: I1124 15:04:42.352925 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wqbf7" Nov 24 15:04:42 crc kubenswrapper[4790]: I1124 15:04:42.541741 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a23cdfa-e59f-4a03-b368-f3cfc8964a0b-catalog-content\") pod \"3a23cdfa-e59f-4a03-b368-f3cfc8964a0b\" (UID: \"3a23cdfa-e59f-4a03-b368-f3cfc8964a0b\") " Nov 24 15:04:42 crc kubenswrapper[4790]: I1124 15:04:42.541824 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a23cdfa-e59f-4a03-b368-f3cfc8964a0b-utilities\") pod \"3a23cdfa-e59f-4a03-b368-f3cfc8964a0b\" (UID: \"3a23cdfa-e59f-4a03-b368-f3cfc8964a0b\") " Nov 24 15:04:42 crc kubenswrapper[4790]: I1124 15:04:42.541959 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ltcc\" (UniqueName: \"kubernetes.io/projected/3a23cdfa-e59f-4a03-b368-f3cfc8964a0b-kube-api-access-5ltcc\") pod \"3a23cdfa-e59f-4a03-b368-f3cfc8964a0b\" (UID: \"3a23cdfa-e59f-4a03-b368-f3cfc8964a0b\") " Nov 24 15:04:42 crc kubenswrapper[4790]: I1124 15:04:42.543380 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a23cdfa-e59f-4a03-b368-f3cfc8964a0b-utilities" (OuterVolumeSpecName: "utilities") pod "3a23cdfa-e59f-4a03-b368-f3cfc8964a0b" (UID: "3a23cdfa-e59f-4a03-b368-f3cfc8964a0b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:04:42 crc kubenswrapper[4790]: I1124 15:04:42.544414 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a23cdfa-e59f-4a03-b368-f3cfc8964a0b-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 15:04:42 crc kubenswrapper[4790]: I1124 15:04:42.552128 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a23cdfa-e59f-4a03-b368-f3cfc8964a0b-kube-api-access-5ltcc" (OuterVolumeSpecName: "kube-api-access-5ltcc") pod "3a23cdfa-e59f-4a03-b368-f3cfc8964a0b" (UID: "3a23cdfa-e59f-4a03-b368-f3cfc8964a0b"). InnerVolumeSpecName "kube-api-access-5ltcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:04:42 crc kubenswrapper[4790]: I1124 15:04:42.646791 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ltcc\" (UniqueName: \"kubernetes.io/projected/3a23cdfa-e59f-4a03-b368-f3cfc8964a0b-kube-api-access-5ltcc\") on node \"crc\" DevicePath \"\"" Nov 24 15:04:42 crc kubenswrapper[4790]: I1124 15:04:42.739250 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a23cdfa-e59f-4a03-b368-f3cfc8964a0b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3a23cdfa-e59f-4a03-b368-f3cfc8964a0b" (UID: "3a23cdfa-e59f-4a03-b368-f3cfc8964a0b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:04:42 crc kubenswrapper[4790]: I1124 15:04:42.749259 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a23cdfa-e59f-4a03-b368-f3cfc8964a0b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 15:04:42 crc kubenswrapper[4790]: I1124 15:04:42.866392 4790 generic.go:334] "Generic (PLEG): container finished" podID="3a23cdfa-e59f-4a03-b368-f3cfc8964a0b" containerID="f78230ee3ddfca1845dd4468b6635fc445fce7df27c19c380943554ba58a7592" exitCode=0 Nov 24 15:04:42 crc kubenswrapper[4790]: I1124 15:04:42.866473 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wqbf7" event={"ID":"3a23cdfa-e59f-4a03-b368-f3cfc8964a0b","Type":"ContainerDied","Data":"f78230ee3ddfca1845dd4468b6635fc445fce7df27c19c380943554ba58a7592"} Nov 24 15:04:42 crc kubenswrapper[4790]: I1124 15:04:42.866605 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wqbf7" event={"ID":"3a23cdfa-e59f-4a03-b368-f3cfc8964a0b","Type":"ContainerDied","Data":"37ece91c99cd2b84486a18f511633e7f3030fa4e3382deec91ccc88df8a74d39"} Nov 24 15:04:42 crc kubenswrapper[4790]: I1124 15:04:42.866640 4790 scope.go:117] "RemoveContainer" containerID="f78230ee3ddfca1845dd4468b6635fc445fce7df27c19c380943554ba58a7592" Nov 24 15:04:42 crc kubenswrapper[4790]: I1124 15:04:42.866507 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wqbf7" Nov 24 15:04:42 crc kubenswrapper[4790]: I1124 15:04:42.898684 4790 scope.go:117] "RemoveContainer" containerID="1fbe2a3d5d80bebe2cfc1abb0bc5629c3ce45983ecb2681322376df0898bd3cf" Nov 24 15:04:42 crc kubenswrapper[4790]: I1124 15:04:42.913958 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wqbf7"] Nov 24 15:04:42 crc kubenswrapper[4790]: I1124 15:04:42.921709 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wqbf7"] Nov 24 15:04:42 crc kubenswrapper[4790]: I1124 15:04:42.937165 4790 scope.go:117] "RemoveContainer" containerID="6cf2c9526249e1eeea4f76b9ac9abaeed474c5926ad4ae8d7c3783784b2f9c6f" Nov 24 15:04:42 crc kubenswrapper[4790]: I1124 15:04:42.984749 4790 scope.go:117] "RemoveContainer" containerID="f78230ee3ddfca1845dd4468b6635fc445fce7df27c19c380943554ba58a7592" Nov 24 15:04:42 crc kubenswrapper[4790]: E1124 15:04:42.985399 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f78230ee3ddfca1845dd4468b6635fc445fce7df27c19c380943554ba58a7592\": container with ID starting with f78230ee3ddfca1845dd4468b6635fc445fce7df27c19c380943554ba58a7592 not found: ID does not exist" containerID="f78230ee3ddfca1845dd4468b6635fc445fce7df27c19c380943554ba58a7592" Nov 24 15:04:42 crc kubenswrapper[4790]: I1124 15:04:42.985452 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f78230ee3ddfca1845dd4468b6635fc445fce7df27c19c380943554ba58a7592"} err="failed to get container status \"f78230ee3ddfca1845dd4468b6635fc445fce7df27c19c380943554ba58a7592\": rpc error: code = NotFound desc = could not find container \"f78230ee3ddfca1845dd4468b6635fc445fce7df27c19c380943554ba58a7592\": container with ID starting with f78230ee3ddfca1845dd4468b6635fc445fce7df27c19c380943554ba58a7592 not found: ID does not exist" Nov 24 15:04:42 crc kubenswrapper[4790]: I1124 15:04:42.985486 4790 scope.go:117] "RemoveContainer" containerID="1fbe2a3d5d80bebe2cfc1abb0bc5629c3ce45983ecb2681322376df0898bd3cf" Nov 24 15:04:42 crc kubenswrapper[4790]: E1124 15:04:42.986280 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fbe2a3d5d80bebe2cfc1abb0bc5629c3ce45983ecb2681322376df0898bd3cf\": container with ID starting with 1fbe2a3d5d80bebe2cfc1abb0bc5629c3ce45983ecb2681322376df0898bd3cf not found: ID does not exist" containerID="1fbe2a3d5d80bebe2cfc1abb0bc5629c3ce45983ecb2681322376df0898bd3cf" Nov 24 15:04:42 crc kubenswrapper[4790]: I1124 15:04:42.986328 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fbe2a3d5d80bebe2cfc1abb0bc5629c3ce45983ecb2681322376df0898bd3cf"} err="failed to get container status \"1fbe2a3d5d80bebe2cfc1abb0bc5629c3ce45983ecb2681322376df0898bd3cf\": rpc error: code = NotFound desc = could not find container \"1fbe2a3d5d80bebe2cfc1abb0bc5629c3ce45983ecb2681322376df0898bd3cf\": container with ID starting with 1fbe2a3d5d80bebe2cfc1abb0bc5629c3ce45983ecb2681322376df0898bd3cf not found: ID does not exist" Nov 24 15:04:42 crc kubenswrapper[4790]: I1124 15:04:42.986359 4790 scope.go:117] "RemoveContainer" containerID="6cf2c9526249e1eeea4f76b9ac9abaeed474c5926ad4ae8d7c3783784b2f9c6f" Nov 24 15:04:42 crc kubenswrapper[4790]: E1124 15:04:42.986715 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6cf2c9526249e1eeea4f76b9ac9abaeed474c5926ad4ae8d7c3783784b2f9c6f\": container with ID starting with 6cf2c9526249e1eeea4f76b9ac9abaeed474c5926ad4ae8d7c3783784b2f9c6f not found: ID does not exist" containerID="6cf2c9526249e1eeea4f76b9ac9abaeed474c5926ad4ae8d7c3783784b2f9c6f" Nov 24 15:04:42 crc kubenswrapper[4790]: I1124 15:04:42.986749 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6cf2c9526249e1eeea4f76b9ac9abaeed474c5926ad4ae8d7c3783784b2f9c6f"} err="failed to get container status \"6cf2c9526249e1eeea4f76b9ac9abaeed474c5926ad4ae8d7c3783784b2f9c6f\": rpc error: code = NotFound desc = could not find container \"6cf2c9526249e1eeea4f76b9ac9abaeed474c5926ad4ae8d7c3783784b2f9c6f\": container with ID starting with 6cf2c9526249e1eeea4f76b9ac9abaeed474c5926ad4ae8d7c3783784b2f9c6f not found: ID does not exist" Nov 24 15:04:43 crc kubenswrapper[4790]: I1124 15:04:43.938955 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:04:43 crc kubenswrapper[4790]: I1124 15:04:43.939283 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:04:43 crc kubenswrapper[4790]: I1124 15:04:43.939352 4790 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 15:04:43 crc kubenswrapper[4790]: I1124 15:04:43.940462 4790 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a4432de01408933d930865fc3e585b6bb44fe87c6f5674553d2039220c20c92e"} pod="openshift-machine-config-operator/machine-config-daemon-xz49t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 15:04:43 crc kubenswrapper[4790]: I1124 15:04:43.940551 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" containerID="cri-o://a4432de01408933d930865fc3e585b6bb44fe87c6f5674553d2039220c20c92e" gracePeriod=600 Nov 24 15:04:44 crc kubenswrapper[4790]: E1124 15:04:44.077601 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:04:44 crc kubenswrapper[4790]: I1124 15:04:44.334633 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a23cdfa-e59f-4a03-b368-f3cfc8964a0b" path="/var/lib/kubelet/pods/3a23cdfa-e59f-4a03-b368-f3cfc8964a0b/volumes" Nov 24 15:04:44 crc kubenswrapper[4790]: I1124 15:04:44.906603 4790 generic.go:334] "Generic (PLEG): container finished" podID="0d73b133-48f0-455f-8f6a-742e633f631a" containerID="a4432de01408933d930865fc3e585b6bb44fe87c6f5674553d2039220c20c92e" exitCode=0 Nov 24 15:04:44 crc kubenswrapper[4790]: I1124 15:04:44.906687 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerDied","Data":"a4432de01408933d930865fc3e585b6bb44fe87c6f5674553d2039220c20c92e"} Nov 24 15:04:44 crc kubenswrapper[4790]: I1124 15:04:44.906744 4790 scope.go:117] "RemoveContainer" containerID="848bb411b4e62828968580971f8aaefd4e3c2cc50ce3c086c482e5c081b37c96" Nov 24 15:04:44 crc kubenswrapper[4790]: I1124 15:04:44.907653 4790 scope.go:117] "RemoveContainer" containerID="a4432de01408933d930865fc3e585b6bb44fe87c6f5674553d2039220c20c92e" Nov 24 15:04:44 crc kubenswrapper[4790]: E1124 15:04:44.908374 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:04:56 crc kubenswrapper[4790]: I1124 15:04:56.341306 4790 scope.go:117] "RemoveContainer" containerID="a4432de01408933d930865fc3e585b6bb44fe87c6f5674553d2039220c20c92e" Nov 24 15:04:56 crc kubenswrapper[4790]: E1124 15:04:56.343771 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:05:07 crc kubenswrapper[4790]: I1124 15:05:07.314901 4790 scope.go:117] "RemoveContainer" containerID="a4432de01408933d930865fc3e585b6bb44fe87c6f5674553d2039220c20c92e" Nov 24 15:05:07 crc kubenswrapper[4790]: E1124 15:05:07.315729 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:05:21 crc kubenswrapper[4790]: I1124 15:05:21.317435 4790 scope.go:117] "RemoveContainer" containerID="a4432de01408933d930865fc3e585b6bb44fe87c6f5674553d2039220c20c92e" Nov 24 15:05:21 crc kubenswrapper[4790]: E1124 15:05:21.320822 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:05:33 crc kubenswrapper[4790]: I1124 15:05:33.316762 4790 scope.go:117] "RemoveContainer" containerID="a4432de01408933d930865fc3e585b6bb44fe87c6f5674553d2039220c20c92e" Nov 24 15:05:33 crc kubenswrapper[4790]: E1124 15:05:33.317989 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:05:37 crc kubenswrapper[4790]: I1124 15:05:37.595459 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tzbl4"] Nov 24 15:05:37 crc kubenswrapper[4790]: E1124 15:05:37.598751 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a23cdfa-e59f-4a03-b368-f3cfc8964a0b" containerName="extract-utilities" Nov 24 15:05:37 crc kubenswrapper[4790]: I1124 15:05:37.598856 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a23cdfa-e59f-4a03-b368-f3cfc8964a0b" containerName="extract-utilities" Nov 24 15:05:37 crc kubenswrapper[4790]: E1124 15:05:37.599106 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a23cdfa-e59f-4a03-b368-f3cfc8964a0b" containerName="extract-content" Nov 24 15:05:37 crc kubenswrapper[4790]: I1124 15:05:37.599192 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a23cdfa-e59f-4a03-b368-f3cfc8964a0b" containerName="extract-content" Nov 24 15:05:37 crc kubenswrapper[4790]: E1124 15:05:37.599317 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a23cdfa-e59f-4a03-b368-f3cfc8964a0b" containerName="registry-server" Nov 24 15:05:37 crc kubenswrapper[4790]: I1124 15:05:37.599407 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a23cdfa-e59f-4a03-b368-f3cfc8964a0b" containerName="registry-server" Nov 24 15:05:37 crc kubenswrapper[4790]: I1124 15:05:37.599754 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a23cdfa-e59f-4a03-b368-f3cfc8964a0b" containerName="registry-server" Nov 24 15:05:37 crc kubenswrapper[4790]: I1124 15:05:37.613118 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tzbl4"] Nov 24 15:05:37 crc kubenswrapper[4790]: I1124 15:05:37.613295 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tzbl4" Nov 24 15:05:37 crc kubenswrapper[4790]: I1124 15:05:37.717626 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00582e63-67c8-48ab-be45-26a5188708f9-utilities\") pod \"redhat-operators-tzbl4\" (UID: \"00582e63-67c8-48ab-be45-26a5188708f9\") " pod="openshift-marketplace/redhat-operators-tzbl4" Nov 24 15:05:37 crc kubenswrapper[4790]: I1124 15:05:37.717739 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzrmp\" (UniqueName: \"kubernetes.io/projected/00582e63-67c8-48ab-be45-26a5188708f9-kube-api-access-nzrmp\") pod \"redhat-operators-tzbl4\" (UID: \"00582e63-67c8-48ab-be45-26a5188708f9\") " pod="openshift-marketplace/redhat-operators-tzbl4" Nov 24 15:05:37 crc kubenswrapper[4790]: I1124 15:05:37.717780 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00582e63-67c8-48ab-be45-26a5188708f9-catalog-content\") pod \"redhat-operators-tzbl4\" (UID: \"00582e63-67c8-48ab-be45-26a5188708f9\") " pod="openshift-marketplace/redhat-operators-tzbl4" Nov 24 15:05:37 crc kubenswrapper[4790]: I1124 15:05:37.821019 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00582e63-67c8-48ab-be45-26a5188708f9-utilities\") pod \"redhat-operators-tzbl4\" (UID: \"00582e63-67c8-48ab-be45-26a5188708f9\") " pod="openshift-marketplace/redhat-operators-tzbl4" Nov 24 15:05:37 crc kubenswrapper[4790]: I1124 15:05:37.821220 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzrmp\" (UniqueName: \"kubernetes.io/projected/00582e63-67c8-48ab-be45-26a5188708f9-kube-api-access-nzrmp\") pod \"redhat-operators-tzbl4\" (UID: \"00582e63-67c8-48ab-be45-26a5188708f9\") " pod="openshift-marketplace/redhat-operators-tzbl4" Nov 24 15:05:37 crc kubenswrapper[4790]: I1124 15:05:37.821311 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00582e63-67c8-48ab-be45-26a5188708f9-catalog-content\") pod \"redhat-operators-tzbl4\" (UID: \"00582e63-67c8-48ab-be45-26a5188708f9\") " pod="openshift-marketplace/redhat-operators-tzbl4" Nov 24 15:05:37 crc kubenswrapper[4790]: I1124 15:05:37.821675 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00582e63-67c8-48ab-be45-26a5188708f9-utilities\") pod \"redhat-operators-tzbl4\" (UID: \"00582e63-67c8-48ab-be45-26a5188708f9\") " pod="openshift-marketplace/redhat-operators-tzbl4" Nov 24 15:05:37 crc kubenswrapper[4790]: I1124 15:05:37.822268 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00582e63-67c8-48ab-be45-26a5188708f9-catalog-content\") pod \"redhat-operators-tzbl4\" (UID: \"00582e63-67c8-48ab-be45-26a5188708f9\") " pod="openshift-marketplace/redhat-operators-tzbl4" Nov 24 15:05:37 crc kubenswrapper[4790]: I1124 15:05:37.846077 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzrmp\" (UniqueName: \"kubernetes.io/projected/00582e63-67c8-48ab-be45-26a5188708f9-kube-api-access-nzrmp\") pod \"redhat-operators-tzbl4\" (UID: \"00582e63-67c8-48ab-be45-26a5188708f9\") " pod="openshift-marketplace/redhat-operators-tzbl4" Nov 24 15:05:37 crc kubenswrapper[4790]: I1124 15:05:37.945857 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tzbl4" Nov 24 15:05:38 crc kubenswrapper[4790]: I1124 15:05:38.486980 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tzbl4"] Nov 24 15:05:38 crc kubenswrapper[4790]: I1124 15:05:38.663262 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tzbl4" event={"ID":"00582e63-67c8-48ab-be45-26a5188708f9","Type":"ContainerStarted","Data":"2dc74167df32a0844f45f3e3cb7f9cfa0a1942aed3fc2efce0b10e5134f351a5"} Nov 24 15:05:39 crc kubenswrapper[4790]: I1124 15:05:39.678317 4790 generic.go:334] "Generic (PLEG): container finished" podID="00582e63-67c8-48ab-be45-26a5188708f9" containerID="06ff89a8e208650b683834206a9fe722c90d88acaf6c0aa26f7245a0764c1f4a" exitCode=0 Nov 24 15:05:39 crc kubenswrapper[4790]: I1124 15:05:39.678439 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tzbl4" event={"ID":"00582e63-67c8-48ab-be45-26a5188708f9","Type":"ContainerDied","Data":"06ff89a8e208650b683834206a9fe722c90d88acaf6c0aa26f7245a0764c1f4a"} Nov 24 15:05:41 crc kubenswrapper[4790]: I1124 15:05:41.706829 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tzbl4" event={"ID":"00582e63-67c8-48ab-be45-26a5188708f9","Type":"ContainerStarted","Data":"95b3801017f3f6e9be48712c085473b88a2aa18f3e90ded63fd0edd274ae2a89"} Nov 24 15:05:45 crc kubenswrapper[4790]: I1124 15:05:45.766779 4790 generic.go:334] "Generic (PLEG): container finished" podID="00582e63-67c8-48ab-be45-26a5188708f9" containerID="95b3801017f3f6e9be48712c085473b88a2aa18f3e90ded63fd0edd274ae2a89" exitCode=0 Nov 24 15:05:45 crc kubenswrapper[4790]: I1124 15:05:45.767471 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tzbl4" event={"ID":"00582e63-67c8-48ab-be45-26a5188708f9","Type":"ContainerDied","Data":"95b3801017f3f6e9be48712c085473b88a2aa18f3e90ded63fd0edd274ae2a89"} Nov 24 15:05:46 crc kubenswrapper[4790]: I1124 15:05:46.783830 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tzbl4" event={"ID":"00582e63-67c8-48ab-be45-26a5188708f9","Type":"ContainerStarted","Data":"8e766da7268d9706f267a70ca3b4d7727cd293aba8420f79b57c57f5c40bcdf5"} Nov 24 15:05:46 crc kubenswrapper[4790]: I1124 15:05:46.822552 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tzbl4" podStartSLOduration=3.301730103 podStartE2EDuration="9.822523586s" podCreationTimestamp="2025-11-24 15:05:37 +0000 UTC" firstStartedPulling="2025-11-24 15:05:39.684116439 +0000 UTC m=+6788.064010131" lastFinishedPulling="2025-11-24 15:05:46.204909922 +0000 UTC m=+6794.584803614" observedRunningTime="2025-11-24 15:05:46.814360272 +0000 UTC m=+6795.194253964" watchObservedRunningTime="2025-11-24 15:05:46.822523586 +0000 UTC m=+6795.202417248" Nov 24 15:05:47 crc kubenswrapper[4790]: I1124 15:05:47.315507 4790 scope.go:117] "RemoveContainer" containerID="a4432de01408933d930865fc3e585b6bb44fe87c6f5674553d2039220c20c92e" Nov 24 15:05:47 crc kubenswrapper[4790]: E1124 15:05:47.316207 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:05:47 crc kubenswrapper[4790]: I1124 15:05:47.946091 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tzbl4" Nov 24 15:05:47 crc kubenswrapper[4790]: I1124 15:05:47.946157 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tzbl4" Nov 24 15:05:49 crc kubenswrapper[4790]: I1124 15:05:49.038572 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tzbl4" podUID="00582e63-67c8-48ab-be45-26a5188708f9" containerName="registry-server" probeResult="failure" output=< Nov 24 15:05:49 crc kubenswrapper[4790]: timeout: failed to connect service ":50051" within 1s Nov 24 15:05:49 crc kubenswrapper[4790]: > Nov 24 15:05:58 crc kubenswrapper[4790]: I1124 15:05:58.028307 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tzbl4" Nov 24 15:05:58 crc kubenswrapper[4790]: I1124 15:05:58.101312 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tzbl4" Nov 24 15:05:58 crc kubenswrapper[4790]: I1124 15:05:58.276366 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tzbl4"] Nov 24 15:05:59 crc kubenswrapper[4790]: I1124 15:05:59.315514 4790 scope.go:117] "RemoveContainer" containerID="a4432de01408933d930865fc3e585b6bb44fe87c6f5674553d2039220c20c92e" Nov 24 15:05:59 crc kubenswrapper[4790]: E1124 15:05:59.316556 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:05:59 crc kubenswrapper[4790]: I1124 15:05:59.940103 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tzbl4" podUID="00582e63-67c8-48ab-be45-26a5188708f9" containerName="registry-server" containerID="cri-o://8e766da7268d9706f267a70ca3b4d7727cd293aba8420f79b57c57f5c40bcdf5" gracePeriod=2 Nov 24 15:06:00 crc kubenswrapper[4790]: I1124 15:06:00.459734 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tzbl4" Nov 24 15:06:00 crc kubenswrapper[4790]: I1124 15:06:00.558088 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00582e63-67c8-48ab-be45-26a5188708f9-utilities\") pod \"00582e63-67c8-48ab-be45-26a5188708f9\" (UID: \"00582e63-67c8-48ab-be45-26a5188708f9\") " Nov 24 15:06:00 crc kubenswrapper[4790]: I1124 15:06:00.558299 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00582e63-67c8-48ab-be45-26a5188708f9-catalog-content\") pod \"00582e63-67c8-48ab-be45-26a5188708f9\" (UID: \"00582e63-67c8-48ab-be45-26a5188708f9\") " Nov 24 15:06:00 crc kubenswrapper[4790]: I1124 15:06:00.558380 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzrmp\" (UniqueName: \"kubernetes.io/projected/00582e63-67c8-48ab-be45-26a5188708f9-kube-api-access-nzrmp\") pod \"00582e63-67c8-48ab-be45-26a5188708f9\" (UID: \"00582e63-67c8-48ab-be45-26a5188708f9\") " Nov 24 15:06:00 crc kubenswrapper[4790]: I1124 15:06:00.559029 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00582e63-67c8-48ab-be45-26a5188708f9-utilities" (OuterVolumeSpecName: "utilities") pod "00582e63-67c8-48ab-be45-26a5188708f9" (UID: "00582e63-67c8-48ab-be45-26a5188708f9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:06:00 crc kubenswrapper[4790]: I1124 15:06:00.563462 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00582e63-67c8-48ab-be45-26a5188708f9-kube-api-access-nzrmp" (OuterVolumeSpecName: "kube-api-access-nzrmp") pod "00582e63-67c8-48ab-be45-26a5188708f9" (UID: "00582e63-67c8-48ab-be45-26a5188708f9"). InnerVolumeSpecName "kube-api-access-nzrmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:06:00 crc kubenswrapper[4790]: I1124 15:06:00.661629 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00582e63-67c8-48ab-be45-26a5188708f9-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 15:06:00 crc kubenswrapper[4790]: I1124 15:06:00.661672 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzrmp\" (UniqueName: \"kubernetes.io/projected/00582e63-67c8-48ab-be45-26a5188708f9-kube-api-access-nzrmp\") on node \"crc\" DevicePath \"\"" Nov 24 15:06:00 crc kubenswrapper[4790]: I1124 15:06:00.690390 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00582e63-67c8-48ab-be45-26a5188708f9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "00582e63-67c8-48ab-be45-26a5188708f9" (UID: "00582e63-67c8-48ab-be45-26a5188708f9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:06:00 crc kubenswrapper[4790]: I1124 15:06:00.763507 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00582e63-67c8-48ab-be45-26a5188708f9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 15:06:00 crc kubenswrapper[4790]: I1124 15:06:00.953377 4790 generic.go:334] "Generic (PLEG): container finished" podID="00582e63-67c8-48ab-be45-26a5188708f9" containerID="8e766da7268d9706f267a70ca3b4d7727cd293aba8420f79b57c57f5c40bcdf5" exitCode=0 Nov 24 15:06:00 crc kubenswrapper[4790]: I1124 15:06:00.953432 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tzbl4" event={"ID":"00582e63-67c8-48ab-be45-26a5188708f9","Type":"ContainerDied","Data":"8e766da7268d9706f267a70ca3b4d7727cd293aba8420f79b57c57f5c40bcdf5"} Nov 24 15:06:00 crc kubenswrapper[4790]: I1124 15:06:00.953469 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tzbl4" event={"ID":"00582e63-67c8-48ab-be45-26a5188708f9","Type":"ContainerDied","Data":"2dc74167df32a0844f45f3e3cb7f9cfa0a1942aed3fc2efce0b10e5134f351a5"} Nov 24 15:06:00 crc kubenswrapper[4790]: I1124 15:06:00.953488 4790 scope.go:117] "RemoveContainer" containerID="8e766da7268d9706f267a70ca3b4d7727cd293aba8420f79b57c57f5c40bcdf5" Nov 24 15:06:00 crc kubenswrapper[4790]: I1124 15:06:00.953491 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tzbl4" Nov 24 15:06:00 crc kubenswrapper[4790]: I1124 15:06:00.982214 4790 scope.go:117] "RemoveContainer" containerID="95b3801017f3f6e9be48712c085473b88a2aa18f3e90ded63fd0edd274ae2a89" Nov 24 15:06:00 crc kubenswrapper[4790]: I1124 15:06:00.994096 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tzbl4"] Nov 24 15:06:01 crc kubenswrapper[4790]: I1124 15:06:01.001677 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tzbl4"] Nov 24 15:06:01 crc kubenswrapper[4790]: I1124 15:06:01.021269 4790 scope.go:117] "RemoveContainer" containerID="06ff89a8e208650b683834206a9fe722c90d88acaf6c0aa26f7245a0764c1f4a" Nov 24 15:06:01 crc kubenswrapper[4790]: I1124 15:06:01.059673 4790 scope.go:117] "RemoveContainer" containerID="8e766da7268d9706f267a70ca3b4d7727cd293aba8420f79b57c57f5c40bcdf5" Nov 24 15:06:01 crc kubenswrapper[4790]: E1124 15:06:01.060825 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e766da7268d9706f267a70ca3b4d7727cd293aba8420f79b57c57f5c40bcdf5\": container with ID starting with 8e766da7268d9706f267a70ca3b4d7727cd293aba8420f79b57c57f5c40bcdf5 not found: ID does not exist" containerID="8e766da7268d9706f267a70ca3b4d7727cd293aba8420f79b57c57f5c40bcdf5" Nov 24 15:06:01 crc kubenswrapper[4790]: I1124 15:06:01.061118 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e766da7268d9706f267a70ca3b4d7727cd293aba8420f79b57c57f5c40bcdf5"} err="failed to get container status \"8e766da7268d9706f267a70ca3b4d7727cd293aba8420f79b57c57f5c40bcdf5\": rpc error: code = NotFound desc = could not find container \"8e766da7268d9706f267a70ca3b4d7727cd293aba8420f79b57c57f5c40bcdf5\": container with ID starting with 8e766da7268d9706f267a70ca3b4d7727cd293aba8420f79b57c57f5c40bcdf5 not found: ID does not exist" Nov 24 15:06:01 crc kubenswrapper[4790]: I1124 15:06:01.061149 4790 scope.go:117] "RemoveContainer" containerID="95b3801017f3f6e9be48712c085473b88a2aa18f3e90ded63fd0edd274ae2a89" Nov 24 15:06:01 crc kubenswrapper[4790]: E1124 15:06:01.061562 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95b3801017f3f6e9be48712c085473b88a2aa18f3e90ded63fd0edd274ae2a89\": container with ID starting with 95b3801017f3f6e9be48712c085473b88a2aa18f3e90ded63fd0edd274ae2a89 not found: ID does not exist" containerID="95b3801017f3f6e9be48712c085473b88a2aa18f3e90ded63fd0edd274ae2a89" Nov 24 15:06:01 crc kubenswrapper[4790]: I1124 15:06:01.061624 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95b3801017f3f6e9be48712c085473b88a2aa18f3e90ded63fd0edd274ae2a89"} err="failed to get container status \"95b3801017f3f6e9be48712c085473b88a2aa18f3e90ded63fd0edd274ae2a89\": rpc error: code = NotFound desc = could not find container \"95b3801017f3f6e9be48712c085473b88a2aa18f3e90ded63fd0edd274ae2a89\": container with ID starting with 95b3801017f3f6e9be48712c085473b88a2aa18f3e90ded63fd0edd274ae2a89 not found: ID does not exist" Nov 24 15:06:01 crc kubenswrapper[4790]: I1124 15:06:01.061661 4790 scope.go:117] "RemoveContainer" containerID="06ff89a8e208650b683834206a9fe722c90d88acaf6c0aa26f7245a0764c1f4a" Nov 24 15:06:01 crc kubenswrapper[4790]: E1124 15:06:01.062444 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06ff89a8e208650b683834206a9fe722c90d88acaf6c0aa26f7245a0764c1f4a\": container with ID starting with 06ff89a8e208650b683834206a9fe722c90d88acaf6c0aa26f7245a0764c1f4a not found: ID does not exist" containerID="06ff89a8e208650b683834206a9fe722c90d88acaf6c0aa26f7245a0764c1f4a" Nov 24 15:06:01 crc kubenswrapper[4790]: I1124 15:06:01.062529 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06ff89a8e208650b683834206a9fe722c90d88acaf6c0aa26f7245a0764c1f4a"} err="failed to get container status \"06ff89a8e208650b683834206a9fe722c90d88acaf6c0aa26f7245a0764c1f4a\": rpc error: code = NotFound desc = could not find container \"06ff89a8e208650b683834206a9fe722c90d88acaf6c0aa26f7245a0764c1f4a\": container with ID starting with 06ff89a8e208650b683834206a9fe722c90d88acaf6c0aa26f7245a0764c1f4a not found: ID does not exist" Nov 24 15:06:02 crc kubenswrapper[4790]: I1124 15:06:02.334256 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00582e63-67c8-48ab-be45-26a5188708f9" path="/var/lib/kubelet/pods/00582e63-67c8-48ab-be45-26a5188708f9/volumes" Nov 24 15:06:13 crc kubenswrapper[4790]: I1124 15:06:13.316269 4790 scope.go:117] "RemoveContainer" containerID="a4432de01408933d930865fc3e585b6bb44fe87c6f5674553d2039220c20c92e" Nov 24 15:06:13 crc kubenswrapper[4790]: E1124 15:06:13.317568 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:06:23 crc kubenswrapper[4790]: I1124 15:06:23.072098 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-jtsf4"] Nov 24 15:06:23 crc kubenswrapper[4790]: I1124 15:06:23.088665 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-bcf5-account-create-ktm8m"] Nov 24 15:06:23 crc kubenswrapper[4790]: I1124 15:06:23.099153 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-bcf5-account-create-ktm8m"] Nov 24 15:06:23 crc kubenswrapper[4790]: I1124 15:06:23.108793 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-jtsf4"] Nov 24 15:06:24 crc kubenswrapper[4790]: I1124 15:06:24.338366 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="236160ca-0716-4cfe-9ae2-bd706d30cc3c" path="/var/lib/kubelet/pods/236160ca-0716-4cfe-9ae2-bd706d30cc3c/volumes" Nov 24 15:06:24 crc kubenswrapper[4790]: I1124 15:06:24.340477 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8b24900-d7bb-4af6-9279-4b93f117e15e" path="/var/lib/kubelet/pods/f8b24900-d7bb-4af6-9279-4b93f117e15e/volumes" Nov 24 15:06:27 crc kubenswrapper[4790]: I1124 15:06:27.316016 4790 scope.go:117] "RemoveContainer" containerID="a4432de01408933d930865fc3e585b6bb44fe87c6f5674553d2039220c20c92e" Nov 24 15:06:27 crc kubenswrapper[4790]: E1124 15:06:27.316827 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:06:34 crc kubenswrapper[4790]: I1124 15:06:34.064041 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-qjttx"] Nov 24 15:06:34 crc kubenswrapper[4790]: I1124 15:06:34.076534 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-qjttx"] Nov 24 15:06:34 crc kubenswrapper[4790]: I1124 15:06:34.331843 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df033470-00fb-45ac-920a-fa5d75d69d73" path="/var/lib/kubelet/pods/df033470-00fb-45ac-920a-fa5d75d69d73/volumes" Nov 24 15:06:36 crc kubenswrapper[4790]: I1124 15:06:36.771471 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-h4gjb"] Nov 24 15:06:36 crc kubenswrapper[4790]: E1124 15:06:36.772731 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00582e63-67c8-48ab-be45-26a5188708f9" containerName="registry-server" Nov 24 15:06:36 crc kubenswrapper[4790]: I1124 15:06:36.772758 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="00582e63-67c8-48ab-be45-26a5188708f9" containerName="registry-server" Nov 24 15:06:36 crc kubenswrapper[4790]: E1124 15:06:36.772880 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00582e63-67c8-48ab-be45-26a5188708f9" containerName="extract-content" Nov 24 15:06:36 crc kubenswrapper[4790]: I1124 15:06:36.772932 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="00582e63-67c8-48ab-be45-26a5188708f9" containerName="extract-content" Nov 24 15:06:36 crc kubenswrapper[4790]: E1124 15:06:36.772973 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00582e63-67c8-48ab-be45-26a5188708f9" containerName="extract-utilities" Nov 24 15:06:36 crc kubenswrapper[4790]: I1124 15:06:36.772991 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="00582e63-67c8-48ab-be45-26a5188708f9" containerName="extract-utilities" Nov 24 15:06:36 crc kubenswrapper[4790]: I1124 15:06:36.773435 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="00582e63-67c8-48ab-be45-26a5188708f9" containerName="registry-server" Nov 24 15:06:36 crc kubenswrapper[4790]: I1124 15:06:36.776574 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h4gjb" Nov 24 15:06:36 crc kubenswrapper[4790]: I1124 15:06:36.783704 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h4gjb"] Nov 24 15:06:36 crc kubenswrapper[4790]: I1124 15:06:36.794289 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtzh4\" (UniqueName: \"kubernetes.io/projected/8d5e1a8f-d234-4a89-8035-d559f00365bf-kube-api-access-wtzh4\") pod \"redhat-marketplace-h4gjb\" (UID: \"8d5e1a8f-d234-4a89-8035-d559f00365bf\") " pod="openshift-marketplace/redhat-marketplace-h4gjb" Nov 24 15:06:36 crc kubenswrapper[4790]: I1124 15:06:36.794383 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d5e1a8f-d234-4a89-8035-d559f00365bf-utilities\") pod \"redhat-marketplace-h4gjb\" (UID: \"8d5e1a8f-d234-4a89-8035-d559f00365bf\") " pod="openshift-marketplace/redhat-marketplace-h4gjb" Nov 24 15:06:36 crc kubenswrapper[4790]: I1124 15:06:36.794434 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d5e1a8f-d234-4a89-8035-d559f00365bf-catalog-content\") pod \"redhat-marketplace-h4gjb\" (UID: \"8d5e1a8f-d234-4a89-8035-d559f00365bf\") " pod="openshift-marketplace/redhat-marketplace-h4gjb" Nov 24 15:06:36 crc kubenswrapper[4790]: I1124 15:06:36.897381 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtzh4\" (UniqueName: \"kubernetes.io/projected/8d5e1a8f-d234-4a89-8035-d559f00365bf-kube-api-access-wtzh4\") pod \"redhat-marketplace-h4gjb\" (UID: \"8d5e1a8f-d234-4a89-8035-d559f00365bf\") " pod="openshift-marketplace/redhat-marketplace-h4gjb" Nov 24 15:06:36 crc kubenswrapper[4790]: I1124 15:06:36.897484 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d5e1a8f-d234-4a89-8035-d559f00365bf-utilities\") pod \"redhat-marketplace-h4gjb\" (UID: \"8d5e1a8f-d234-4a89-8035-d559f00365bf\") " pod="openshift-marketplace/redhat-marketplace-h4gjb" Nov 24 15:06:36 crc kubenswrapper[4790]: I1124 15:06:36.897533 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d5e1a8f-d234-4a89-8035-d559f00365bf-catalog-content\") pod \"redhat-marketplace-h4gjb\" (UID: \"8d5e1a8f-d234-4a89-8035-d559f00365bf\") " pod="openshift-marketplace/redhat-marketplace-h4gjb" Nov 24 15:06:36 crc kubenswrapper[4790]: I1124 15:06:36.898631 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d5e1a8f-d234-4a89-8035-d559f00365bf-utilities\") pod \"redhat-marketplace-h4gjb\" (UID: \"8d5e1a8f-d234-4a89-8035-d559f00365bf\") " pod="openshift-marketplace/redhat-marketplace-h4gjb" Nov 24 15:06:36 crc kubenswrapper[4790]: I1124 15:06:36.898664 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d5e1a8f-d234-4a89-8035-d559f00365bf-catalog-content\") pod \"redhat-marketplace-h4gjb\" (UID: \"8d5e1a8f-d234-4a89-8035-d559f00365bf\") " pod="openshift-marketplace/redhat-marketplace-h4gjb" Nov 24 15:06:36 crc kubenswrapper[4790]: I1124 15:06:36.920701 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtzh4\" (UniqueName: \"kubernetes.io/projected/8d5e1a8f-d234-4a89-8035-d559f00365bf-kube-api-access-wtzh4\") pod \"redhat-marketplace-h4gjb\" (UID: \"8d5e1a8f-d234-4a89-8035-d559f00365bf\") " pod="openshift-marketplace/redhat-marketplace-h4gjb" Nov 24 15:06:37 crc kubenswrapper[4790]: I1124 15:06:37.104670 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h4gjb" Nov 24 15:06:37 crc kubenswrapper[4790]: I1124 15:06:37.636579 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h4gjb"] Nov 24 15:06:38 crc kubenswrapper[4790]: I1124 15:06:38.457476 4790 generic.go:334] "Generic (PLEG): container finished" podID="8d5e1a8f-d234-4a89-8035-d559f00365bf" containerID="4ed4615aba2f6f985973ff46116e96ed3ca8c67091b07c8fd5a98bfa9cf76d49" exitCode=0 Nov 24 15:06:38 crc kubenswrapper[4790]: I1124 15:06:38.457909 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4gjb" event={"ID":"8d5e1a8f-d234-4a89-8035-d559f00365bf","Type":"ContainerDied","Data":"4ed4615aba2f6f985973ff46116e96ed3ca8c67091b07c8fd5a98bfa9cf76d49"} Nov 24 15:06:38 crc kubenswrapper[4790]: I1124 15:06:38.457966 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4gjb" event={"ID":"8d5e1a8f-d234-4a89-8035-d559f00365bf","Type":"ContainerStarted","Data":"bb10fbf8136baf4a218723173d877191cb704f132bbff8e6012ff0d2cc596c48"} Nov 24 15:06:39 crc kubenswrapper[4790]: I1124 15:06:39.316183 4790 scope.go:117] "RemoveContainer" containerID="a4432de01408933d930865fc3e585b6bb44fe87c6f5674553d2039220c20c92e" Nov 24 15:06:39 crc kubenswrapper[4790]: E1124 15:06:39.316821 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:06:40 crc kubenswrapper[4790]: I1124 15:06:40.491320 4790 generic.go:334] "Generic (PLEG): container finished" podID="8d5e1a8f-d234-4a89-8035-d559f00365bf" containerID="78fe3c6d93de589eb0643e65217aef987935eb4819fbdaad7480aa4e68e55192" exitCode=0 Nov 24 15:06:40 crc kubenswrapper[4790]: I1124 15:06:40.491391 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4gjb" event={"ID":"8d5e1a8f-d234-4a89-8035-d559f00365bf","Type":"ContainerDied","Data":"78fe3c6d93de589eb0643e65217aef987935eb4819fbdaad7480aa4e68e55192"} Nov 24 15:06:41 crc kubenswrapper[4790]: I1124 15:06:41.507761 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4gjb" event={"ID":"8d5e1a8f-d234-4a89-8035-d559f00365bf","Type":"ContainerStarted","Data":"60259804df751cfaed8f2b457aeb75ce35f7bd0add75f6aafb17ac68f0ad5295"} Nov 24 15:06:41 crc kubenswrapper[4790]: I1124 15:06:41.535732 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-h4gjb" podStartSLOduration=3.096134842 podStartE2EDuration="5.535712648s" podCreationTimestamp="2025-11-24 15:06:36 +0000 UTC" firstStartedPulling="2025-11-24 15:06:38.461057387 +0000 UTC m=+6846.840951089" lastFinishedPulling="2025-11-24 15:06:40.900635193 +0000 UTC m=+6849.280528895" observedRunningTime="2025-11-24 15:06:41.532141076 +0000 UTC m=+6849.912034768" watchObservedRunningTime="2025-11-24 15:06:41.535712648 +0000 UTC m=+6849.915606300" Nov 24 15:06:47 crc kubenswrapper[4790]: I1124 15:06:47.105811 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-h4gjb" Nov 24 15:06:47 crc kubenswrapper[4790]: I1124 15:06:47.106470 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-h4gjb" Nov 24 15:06:47 crc kubenswrapper[4790]: I1124 15:06:47.196092 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-h4gjb" Nov 24 15:06:47 crc kubenswrapper[4790]: I1124 15:06:47.628695 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-h4gjb" Nov 24 15:06:47 crc kubenswrapper[4790]: I1124 15:06:47.684324 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h4gjb"] Nov 24 15:06:49 crc kubenswrapper[4790]: I1124 15:06:49.600201 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-h4gjb" podUID="8d5e1a8f-d234-4a89-8035-d559f00365bf" containerName="registry-server" containerID="cri-o://60259804df751cfaed8f2b457aeb75ce35f7bd0add75f6aafb17ac68f0ad5295" gracePeriod=2 Nov 24 15:06:50 crc kubenswrapper[4790]: I1124 15:06:50.206318 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h4gjb" Nov 24 15:06:50 crc kubenswrapper[4790]: I1124 15:06:50.242241 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtzh4\" (UniqueName: \"kubernetes.io/projected/8d5e1a8f-d234-4a89-8035-d559f00365bf-kube-api-access-wtzh4\") pod \"8d5e1a8f-d234-4a89-8035-d559f00365bf\" (UID: \"8d5e1a8f-d234-4a89-8035-d559f00365bf\") " Nov 24 15:06:50 crc kubenswrapper[4790]: I1124 15:06:50.242372 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d5e1a8f-d234-4a89-8035-d559f00365bf-utilities\") pod \"8d5e1a8f-d234-4a89-8035-d559f00365bf\" (UID: \"8d5e1a8f-d234-4a89-8035-d559f00365bf\") " Nov 24 15:06:50 crc kubenswrapper[4790]: I1124 15:06:50.242408 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d5e1a8f-d234-4a89-8035-d559f00365bf-catalog-content\") pod \"8d5e1a8f-d234-4a89-8035-d559f00365bf\" (UID: \"8d5e1a8f-d234-4a89-8035-d559f00365bf\") " Nov 24 15:06:50 crc kubenswrapper[4790]: I1124 15:06:50.243997 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d5e1a8f-d234-4a89-8035-d559f00365bf-utilities" (OuterVolumeSpecName: "utilities") pod "8d5e1a8f-d234-4a89-8035-d559f00365bf" (UID: "8d5e1a8f-d234-4a89-8035-d559f00365bf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:06:50 crc kubenswrapper[4790]: I1124 15:06:50.256112 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d5e1a8f-d234-4a89-8035-d559f00365bf-kube-api-access-wtzh4" (OuterVolumeSpecName: "kube-api-access-wtzh4") pod "8d5e1a8f-d234-4a89-8035-d559f00365bf" (UID: "8d5e1a8f-d234-4a89-8035-d559f00365bf"). InnerVolumeSpecName "kube-api-access-wtzh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:06:50 crc kubenswrapper[4790]: I1124 15:06:50.272310 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d5e1a8f-d234-4a89-8035-d559f00365bf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8d5e1a8f-d234-4a89-8035-d559f00365bf" (UID: "8d5e1a8f-d234-4a89-8035-d559f00365bf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:06:50 crc kubenswrapper[4790]: I1124 15:06:50.350578 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtzh4\" (UniqueName: \"kubernetes.io/projected/8d5e1a8f-d234-4a89-8035-d559f00365bf-kube-api-access-wtzh4\") on node \"crc\" DevicePath \"\"" Nov 24 15:06:50 crc kubenswrapper[4790]: I1124 15:06:50.350621 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d5e1a8f-d234-4a89-8035-d559f00365bf-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 15:06:50 crc kubenswrapper[4790]: I1124 15:06:50.350636 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d5e1a8f-d234-4a89-8035-d559f00365bf-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 15:06:50 crc kubenswrapper[4790]: I1124 15:06:50.618186 4790 generic.go:334] "Generic (PLEG): container finished" podID="8d5e1a8f-d234-4a89-8035-d559f00365bf" containerID="60259804df751cfaed8f2b457aeb75ce35f7bd0add75f6aafb17ac68f0ad5295" exitCode=0 Nov 24 15:06:50 crc kubenswrapper[4790]: I1124 15:06:50.618339 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h4gjb" Nov 24 15:06:50 crc kubenswrapper[4790]: I1124 15:06:50.618383 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4gjb" event={"ID":"8d5e1a8f-d234-4a89-8035-d559f00365bf","Type":"ContainerDied","Data":"60259804df751cfaed8f2b457aeb75ce35f7bd0add75f6aafb17ac68f0ad5295"} Nov 24 15:06:50 crc kubenswrapper[4790]: I1124 15:06:50.619784 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4gjb" event={"ID":"8d5e1a8f-d234-4a89-8035-d559f00365bf","Type":"ContainerDied","Data":"bb10fbf8136baf4a218723173d877191cb704f132bbff8e6012ff0d2cc596c48"} Nov 24 15:06:50 crc kubenswrapper[4790]: I1124 15:06:50.619839 4790 scope.go:117] "RemoveContainer" containerID="60259804df751cfaed8f2b457aeb75ce35f7bd0add75f6aafb17ac68f0ad5295" Nov 24 15:06:50 crc kubenswrapper[4790]: I1124 15:06:50.657871 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h4gjb"] Nov 24 15:06:50 crc kubenswrapper[4790]: I1124 15:06:50.667770 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-h4gjb"] Nov 24 15:06:50 crc kubenswrapper[4790]: I1124 15:06:50.671206 4790 scope.go:117] "RemoveContainer" containerID="78fe3c6d93de589eb0643e65217aef987935eb4819fbdaad7480aa4e68e55192" Nov 24 15:06:50 crc kubenswrapper[4790]: I1124 15:06:50.703049 4790 scope.go:117] "RemoveContainer" containerID="4ed4615aba2f6f985973ff46116e96ed3ca8c67091b07c8fd5a98bfa9cf76d49" Nov 24 15:06:50 crc kubenswrapper[4790]: I1124 15:06:50.787322 4790 scope.go:117] "RemoveContainer" containerID="60259804df751cfaed8f2b457aeb75ce35f7bd0add75f6aafb17ac68f0ad5295" Nov 24 15:06:50 crc kubenswrapper[4790]: E1124 15:06:50.787775 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60259804df751cfaed8f2b457aeb75ce35f7bd0add75f6aafb17ac68f0ad5295\": container with ID starting with 60259804df751cfaed8f2b457aeb75ce35f7bd0add75f6aafb17ac68f0ad5295 not found: ID does not exist" containerID="60259804df751cfaed8f2b457aeb75ce35f7bd0add75f6aafb17ac68f0ad5295" Nov 24 15:06:50 crc kubenswrapper[4790]: I1124 15:06:50.787813 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60259804df751cfaed8f2b457aeb75ce35f7bd0add75f6aafb17ac68f0ad5295"} err="failed to get container status \"60259804df751cfaed8f2b457aeb75ce35f7bd0add75f6aafb17ac68f0ad5295\": rpc error: code = NotFound desc = could not find container \"60259804df751cfaed8f2b457aeb75ce35f7bd0add75f6aafb17ac68f0ad5295\": container with ID starting with 60259804df751cfaed8f2b457aeb75ce35f7bd0add75f6aafb17ac68f0ad5295 not found: ID does not exist" Nov 24 15:06:50 crc kubenswrapper[4790]: I1124 15:06:50.787907 4790 scope.go:117] "RemoveContainer" containerID="78fe3c6d93de589eb0643e65217aef987935eb4819fbdaad7480aa4e68e55192" Nov 24 15:06:50 crc kubenswrapper[4790]: E1124 15:06:50.788166 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78fe3c6d93de589eb0643e65217aef987935eb4819fbdaad7480aa4e68e55192\": container with ID starting with 78fe3c6d93de589eb0643e65217aef987935eb4819fbdaad7480aa4e68e55192 not found: ID does not exist" containerID="78fe3c6d93de589eb0643e65217aef987935eb4819fbdaad7480aa4e68e55192" Nov 24 15:06:50 crc kubenswrapper[4790]: I1124 15:06:50.788191 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78fe3c6d93de589eb0643e65217aef987935eb4819fbdaad7480aa4e68e55192"} err="failed to get container status \"78fe3c6d93de589eb0643e65217aef987935eb4819fbdaad7480aa4e68e55192\": rpc error: code = NotFound desc = could not find container \"78fe3c6d93de589eb0643e65217aef987935eb4819fbdaad7480aa4e68e55192\": container with ID starting with 78fe3c6d93de589eb0643e65217aef987935eb4819fbdaad7480aa4e68e55192 not found: ID does not exist" Nov 24 15:06:50 crc kubenswrapper[4790]: I1124 15:06:50.788210 4790 scope.go:117] "RemoveContainer" containerID="4ed4615aba2f6f985973ff46116e96ed3ca8c67091b07c8fd5a98bfa9cf76d49" Nov 24 15:06:50 crc kubenswrapper[4790]: E1124 15:06:50.788574 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ed4615aba2f6f985973ff46116e96ed3ca8c67091b07c8fd5a98bfa9cf76d49\": container with ID starting with 4ed4615aba2f6f985973ff46116e96ed3ca8c67091b07c8fd5a98bfa9cf76d49 not found: ID does not exist" containerID="4ed4615aba2f6f985973ff46116e96ed3ca8c67091b07c8fd5a98bfa9cf76d49" Nov 24 15:06:50 crc kubenswrapper[4790]: I1124 15:06:50.788618 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ed4615aba2f6f985973ff46116e96ed3ca8c67091b07c8fd5a98bfa9cf76d49"} err="failed to get container status \"4ed4615aba2f6f985973ff46116e96ed3ca8c67091b07c8fd5a98bfa9cf76d49\": rpc error: code = NotFound desc = could not find container \"4ed4615aba2f6f985973ff46116e96ed3ca8c67091b07c8fd5a98bfa9cf76d49\": container with ID starting with 4ed4615aba2f6f985973ff46116e96ed3ca8c67091b07c8fd5a98bfa9cf76d49 not found: ID does not exist" Nov 24 15:06:52 crc kubenswrapper[4790]: I1124 15:06:52.050976 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-2zln5"] Nov 24 15:06:52 crc kubenswrapper[4790]: I1124 15:06:52.067767 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-fe63-account-create-ppfwz"] Nov 24 15:06:52 crc kubenswrapper[4790]: I1124 15:06:52.078106 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-2zln5"] Nov 24 15:06:52 crc kubenswrapper[4790]: I1124 15:06:52.086873 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-fe63-account-create-ppfwz"] Nov 24 15:06:52 crc kubenswrapper[4790]: I1124 15:06:52.335169 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41b0a8b8-9156-44b3-b7ed-f72675827738" path="/var/lib/kubelet/pods/41b0a8b8-9156-44b3-b7ed-f72675827738/volumes" Nov 24 15:06:52 crc kubenswrapper[4790]: I1124 15:06:52.336367 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d5e1a8f-d234-4a89-8035-d559f00365bf" path="/var/lib/kubelet/pods/8d5e1a8f-d234-4a89-8035-d559f00365bf/volumes" Nov 24 15:06:52 crc kubenswrapper[4790]: I1124 15:06:52.338121 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f95121cf-16ec-4760-9917-84d42ed63728" path="/var/lib/kubelet/pods/f95121cf-16ec-4760-9917-84d42ed63728/volumes" Nov 24 15:06:53 crc kubenswrapper[4790]: I1124 15:06:53.314932 4790 scope.go:117] "RemoveContainer" containerID="a4432de01408933d930865fc3e585b6bb44fe87c6f5674553d2039220c20c92e" Nov 24 15:06:53 crc kubenswrapper[4790]: E1124 15:06:53.315321 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:07:02 crc kubenswrapper[4790]: I1124 15:07:02.532352 4790 scope.go:117] "RemoveContainer" containerID="0f55a96bfe303125f6f43218805e6ec19ad7633c4a6b92a1d332475a7f046501" Nov 24 15:07:02 crc kubenswrapper[4790]: I1124 15:07:02.588279 4790 scope.go:117] "RemoveContainer" containerID="d3bb2790edd355c76c55d6b5837eaa5ce4ab7eea196da627ba2299094fa002e2" Nov 24 15:07:02 crc kubenswrapper[4790]: I1124 15:07:02.636901 4790 scope.go:117] "RemoveContainer" containerID="6feeb517f1e72804474985e9df49115d56dc0d1c4c5125f8d6c86fff4b9dddab" Nov 24 15:07:02 crc kubenswrapper[4790]: I1124 15:07:02.690216 4790 scope.go:117] "RemoveContainer" containerID="f3469f70e2698fbbf691e9541604d6ce1f343b9a3eb3729dbb6e5dad3eea0788" Nov 24 15:07:02 crc kubenswrapper[4790]: I1124 15:07:02.745020 4790 scope.go:117] "RemoveContainer" containerID="c94de0a2f3bb9c324c069d299d0ca9f47b924c5d6efe206968c30be485b0f7a2" Nov 24 15:07:03 crc kubenswrapper[4790]: I1124 15:07:03.073938 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-ccjhs"] Nov 24 15:07:03 crc kubenswrapper[4790]: I1124 15:07:03.089526 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-ccjhs"] Nov 24 15:07:04 crc kubenswrapper[4790]: I1124 15:07:04.333737 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="657b7d2e-3626-4c66-a216-e2f2a7993771" path="/var/lib/kubelet/pods/657b7d2e-3626-4c66-a216-e2f2a7993771/volumes" Nov 24 15:07:08 crc kubenswrapper[4790]: I1124 15:07:08.315632 4790 scope.go:117] "RemoveContainer" containerID="a4432de01408933d930865fc3e585b6bb44fe87c6f5674553d2039220c20c92e" Nov 24 15:07:08 crc kubenswrapper[4790]: E1124 15:07:08.316565 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:07:19 crc kubenswrapper[4790]: I1124 15:07:19.315424 4790 scope.go:117] "RemoveContainer" containerID="a4432de01408933d930865fc3e585b6bb44fe87c6f5674553d2039220c20c92e" Nov 24 15:07:19 crc kubenswrapper[4790]: E1124 15:07:19.316668 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:07:32 crc kubenswrapper[4790]: I1124 15:07:32.320951 4790 scope.go:117] "RemoveContainer" containerID="a4432de01408933d930865fc3e585b6bb44fe87c6f5674553d2039220c20c92e" Nov 24 15:07:32 crc kubenswrapper[4790]: E1124 15:07:32.322624 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:07:45 crc kubenswrapper[4790]: I1124 15:07:45.315053 4790 scope.go:117] "RemoveContainer" containerID="a4432de01408933d930865fc3e585b6bb44fe87c6f5674553d2039220c20c92e" Nov 24 15:07:45 crc kubenswrapper[4790]: E1124 15:07:45.315912 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:07:57 crc kubenswrapper[4790]: I1124 15:07:57.315001 4790 scope.go:117] "RemoveContainer" containerID="a4432de01408933d930865fc3e585b6bb44fe87c6f5674553d2039220c20c92e" Nov 24 15:07:57 crc kubenswrapper[4790]: E1124 15:07:57.316256 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:08:02 crc kubenswrapper[4790]: I1124 15:08:02.960519 4790 scope.go:117] "RemoveContainer" containerID="4e2c99071bb23fc6540090564ab09dfcab17097ba5d8ec2bd9359866dc926528" Nov 24 15:08:11 crc kubenswrapper[4790]: I1124 15:08:11.316977 4790 scope.go:117] "RemoveContainer" containerID="a4432de01408933d930865fc3e585b6bb44fe87c6f5674553d2039220c20c92e" Nov 24 15:08:11 crc kubenswrapper[4790]: E1124 15:08:11.318231 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:08:20 crc kubenswrapper[4790]: I1124 15:08:20.884837 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jz2kk"] Nov 24 15:08:20 crc kubenswrapper[4790]: E1124 15:08:20.886537 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d5e1a8f-d234-4a89-8035-d559f00365bf" containerName="extract-utilities" Nov 24 15:08:20 crc kubenswrapper[4790]: I1124 15:08:20.886571 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d5e1a8f-d234-4a89-8035-d559f00365bf" containerName="extract-utilities" Nov 24 15:08:20 crc kubenswrapper[4790]: E1124 15:08:20.886607 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d5e1a8f-d234-4a89-8035-d559f00365bf" containerName="extract-content" Nov 24 15:08:20 crc kubenswrapper[4790]: I1124 15:08:20.886625 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d5e1a8f-d234-4a89-8035-d559f00365bf" containerName="extract-content" Nov 24 15:08:20 crc kubenswrapper[4790]: E1124 15:08:20.886655 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d5e1a8f-d234-4a89-8035-d559f00365bf" containerName="registry-server" Nov 24 15:08:20 crc kubenswrapper[4790]: I1124 15:08:20.886671 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d5e1a8f-d234-4a89-8035-d559f00365bf" containerName="registry-server" Nov 24 15:08:20 crc kubenswrapper[4790]: I1124 15:08:20.887314 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d5e1a8f-d234-4a89-8035-d559f00365bf" containerName="registry-server" Nov 24 15:08:20 crc kubenswrapper[4790]: I1124 15:08:20.896222 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jz2kk" Nov 24 15:08:20 crc kubenswrapper[4790]: I1124 15:08:20.915908 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jz2kk"] Nov 24 15:08:21 crc kubenswrapper[4790]: I1124 15:08:21.073007 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a2cd0fa-f98f-47ad-8f50-80375d5a5630-utilities\") pod \"community-operators-jz2kk\" (UID: \"3a2cd0fa-f98f-47ad-8f50-80375d5a5630\") " pod="openshift-marketplace/community-operators-jz2kk" Nov 24 15:08:21 crc kubenswrapper[4790]: I1124 15:08:21.073085 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a2cd0fa-f98f-47ad-8f50-80375d5a5630-catalog-content\") pod \"community-operators-jz2kk\" (UID: \"3a2cd0fa-f98f-47ad-8f50-80375d5a5630\") " pod="openshift-marketplace/community-operators-jz2kk" Nov 24 15:08:21 crc kubenswrapper[4790]: I1124 15:08:21.073294 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mj9j8\" (UniqueName: \"kubernetes.io/projected/3a2cd0fa-f98f-47ad-8f50-80375d5a5630-kube-api-access-mj9j8\") pod \"community-operators-jz2kk\" (UID: \"3a2cd0fa-f98f-47ad-8f50-80375d5a5630\") " pod="openshift-marketplace/community-operators-jz2kk" Nov 24 15:08:21 crc kubenswrapper[4790]: I1124 15:08:21.175427 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a2cd0fa-f98f-47ad-8f50-80375d5a5630-utilities\") pod \"community-operators-jz2kk\" (UID: \"3a2cd0fa-f98f-47ad-8f50-80375d5a5630\") " pod="openshift-marketplace/community-operators-jz2kk" Nov 24 15:08:21 crc kubenswrapper[4790]: I1124 15:08:21.176257 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a2cd0fa-f98f-47ad-8f50-80375d5a5630-utilities\") pod \"community-operators-jz2kk\" (UID: \"3a2cd0fa-f98f-47ad-8f50-80375d5a5630\") " pod="openshift-marketplace/community-operators-jz2kk" Nov 24 15:08:21 crc kubenswrapper[4790]: I1124 15:08:21.176311 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a2cd0fa-f98f-47ad-8f50-80375d5a5630-catalog-content\") pod \"community-operators-jz2kk\" (UID: \"3a2cd0fa-f98f-47ad-8f50-80375d5a5630\") " pod="openshift-marketplace/community-operators-jz2kk" Nov 24 15:08:21 crc kubenswrapper[4790]: I1124 15:08:21.176620 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mj9j8\" (UniqueName: \"kubernetes.io/projected/3a2cd0fa-f98f-47ad-8f50-80375d5a5630-kube-api-access-mj9j8\") pod \"community-operators-jz2kk\" (UID: \"3a2cd0fa-f98f-47ad-8f50-80375d5a5630\") " pod="openshift-marketplace/community-operators-jz2kk" Nov 24 15:08:21 crc kubenswrapper[4790]: I1124 15:08:21.176656 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a2cd0fa-f98f-47ad-8f50-80375d5a5630-catalog-content\") pod \"community-operators-jz2kk\" (UID: \"3a2cd0fa-f98f-47ad-8f50-80375d5a5630\") " pod="openshift-marketplace/community-operators-jz2kk" Nov 24 15:08:21 crc kubenswrapper[4790]: I1124 15:08:21.201262 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mj9j8\" (UniqueName: \"kubernetes.io/projected/3a2cd0fa-f98f-47ad-8f50-80375d5a5630-kube-api-access-mj9j8\") pod \"community-operators-jz2kk\" (UID: \"3a2cd0fa-f98f-47ad-8f50-80375d5a5630\") " pod="openshift-marketplace/community-operators-jz2kk" Nov 24 15:08:21 crc kubenswrapper[4790]: I1124 15:08:21.228831 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jz2kk" Nov 24 15:08:21 crc kubenswrapper[4790]: I1124 15:08:21.774429 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jz2kk"] Nov 24 15:08:21 crc kubenswrapper[4790]: W1124 15:08:21.775985 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a2cd0fa_f98f_47ad_8f50_80375d5a5630.slice/crio-82550d1b27ed623c11590730600a8835906d84281ef1835809efa7ba03e0542d WatchSource:0}: Error finding container 82550d1b27ed623c11590730600a8835906d84281ef1835809efa7ba03e0542d: Status 404 returned error can't find the container with id 82550d1b27ed623c11590730600a8835906d84281ef1835809efa7ba03e0542d Nov 24 15:08:21 crc kubenswrapper[4790]: I1124 15:08:21.854337 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jz2kk" event={"ID":"3a2cd0fa-f98f-47ad-8f50-80375d5a5630","Type":"ContainerStarted","Data":"82550d1b27ed623c11590730600a8835906d84281ef1835809efa7ba03e0542d"} Nov 24 15:08:22 crc kubenswrapper[4790]: I1124 15:08:22.871158 4790 generic.go:334] "Generic (PLEG): container finished" podID="3a2cd0fa-f98f-47ad-8f50-80375d5a5630" containerID="6eb90b412a54117eb3d63aa5a658e500aa17233b605c74066a3557ac999c3812" exitCode=0 Nov 24 15:08:22 crc kubenswrapper[4790]: I1124 15:08:22.871316 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jz2kk" event={"ID":"3a2cd0fa-f98f-47ad-8f50-80375d5a5630","Type":"ContainerDied","Data":"6eb90b412a54117eb3d63aa5a658e500aa17233b605c74066a3557ac999c3812"} Nov 24 15:08:23 crc kubenswrapper[4790]: I1124 15:08:23.884174 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jz2kk" event={"ID":"3a2cd0fa-f98f-47ad-8f50-80375d5a5630","Type":"ContainerStarted","Data":"52ad60005a96109a9ac2cdde3c80aa06cca61c0f3e14b6d07f67c0ee409c86b1"} Nov 24 15:08:24 crc kubenswrapper[4790]: I1124 15:08:24.897648 4790 generic.go:334] "Generic (PLEG): container finished" podID="3a2cd0fa-f98f-47ad-8f50-80375d5a5630" containerID="52ad60005a96109a9ac2cdde3c80aa06cca61c0f3e14b6d07f67c0ee409c86b1" exitCode=0 Nov 24 15:08:24 crc kubenswrapper[4790]: I1124 15:08:24.897859 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jz2kk" event={"ID":"3a2cd0fa-f98f-47ad-8f50-80375d5a5630","Type":"ContainerDied","Data":"52ad60005a96109a9ac2cdde3c80aa06cca61c0f3e14b6d07f67c0ee409c86b1"} Nov 24 15:08:26 crc kubenswrapper[4790]: I1124 15:08:26.315698 4790 scope.go:117] "RemoveContainer" containerID="a4432de01408933d930865fc3e585b6bb44fe87c6f5674553d2039220c20c92e" Nov 24 15:08:26 crc kubenswrapper[4790]: E1124 15:08:26.316680 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:08:26 crc kubenswrapper[4790]: I1124 15:08:26.931841 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jz2kk" event={"ID":"3a2cd0fa-f98f-47ad-8f50-80375d5a5630","Type":"ContainerStarted","Data":"1d63d96d582f772de6a45908f3fcb1cbd5fd92e4381342a96a07313c3f48ccac"} Nov 24 15:08:26 crc kubenswrapper[4790]: I1124 15:08:26.967484 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jz2kk" podStartSLOduration=4.466572066 podStartE2EDuration="6.967453108s" podCreationTimestamp="2025-11-24 15:08:20 +0000 UTC" firstStartedPulling="2025-11-24 15:08:22.874811266 +0000 UTC m=+6951.254704968" lastFinishedPulling="2025-11-24 15:08:25.375692328 +0000 UTC m=+6953.755586010" observedRunningTime="2025-11-24 15:08:26.955233688 +0000 UTC m=+6955.335127410" watchObservedRunningTime="2025-11-24 15:08:26.967453108 +0000 UTC m=+6955.347346810" Nov 24 15:08:31 crc kubenswrapper[4790]: I1124 15:08:31.229985 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jz2kk" Nov 24 15:08:31 crc kubenswrapper[4790]: I1124 15:08:31.230595 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jz2kk" Nov 24 15:08:31 crc kubenswrapper[4790]: I1124 15:08:31.309980 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jz2kk" Nov 24 15:08:32 crc kubenswrapper[4790]: I1124 15:08:32.070329 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jz2kk" Nov 24 15:08:32 crc kubenswrapper[4790]: I1124 15:08:32.138667 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jz2kk"] Nov 24 15:08:34 crc kubenswrapper[4790]: I1124 15:08:34.019797 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jz2kk" podUID="3a2cd0fa-f98f-47ad-8f50-80375d5a5630" containerName="registry-server" containerID="cri-o://1d63d96d582f772de6a45908f3fcb1cbd5fd92e4381342a96a07313c3f48ccac" gracePeriod=2 Nov 24 15:08:34 crc kubenswrapper[4790]: I1124 15:08:34.525869 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jz2kk" Nov 24 15:08:34 crc kubenswrapper[4790]: I1124 15:08:34.632476 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a2cd0fa-f98f-47ad-8f50-80375d5a5630-utilities\") pod \"3a2cd0fa-f98f-47ad-8f50-80375d5a5630\" (UID: \"3a2cd0fa-f98f-47ad-8f50-80375d5a5630\") " Nov 24 15:08:34 crc kubenswrapper[4790]: I1124 15:08:34.632852 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a2cd0fa-f98f-47ad-8f50-80375d5a5630-catalog-content\") pod \"3a2cd0fa-f98f-47ad-8f50-80375d5a5630\" (UID: \"3a2cd0fa-f98f-47ad-8f50-80375d5a5630\") " Nov 24 15:08:34 crc kubenswrapper[4790]: I1124 15:08:34.633087 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mj9j8\" (UniqueName: \"kubernetes.io/projected/3a2cd0fa-f98f-47ad-8f50-80375d5a5630-kube-api-access-mj9j8\") pod \"3a2cd0fa-f98f-47ad-8f50-80375d5a5630\" (UID: \"3a2cd0fa-f98f-47ad-8f50-80375d5a5630\") " Nov 24 15:08:34 crc kubenswrapper[4790]: I1124 15:08:34.633530 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a2cd0fa-f98f-47ad-8f50-80375d5a5630-utilities" (OuterVolumeSpecName: "utilities") pod "3a2cd0fa-f98f-47ad-8f50-80375d5a5630" (UID: "3a2cd0fa-f98f-47ad-8f50-80375d5a5630"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:08:34 crc kubenswrapper[4790]: I1124 15:08:34.639714 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a2cd0fa-f98f-47ad-8f50-80375d5a5630-kube-api-access-mj9j8" (OuterVolumeSpecName: "kube-api-access-mj9j8") pod "3a2cd0fa-f98f-47ad-8f50-80375d5a5630" (UID: "3a2cd0fa-f98f-47ad-8f50-80375d5a5630"). InnerVolumeSpecName "kube-api-access-mj9j8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:08:34 crc kubenswrapper[4790]: I1124 15:08:34.690872 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a2cd0fa-f98f-47ad-8f50-80375d5a5630-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3a2cd0fa-f98f-47ad-8f50-80375d5a5630" (UID: "3a2cd0fa-f98f-47ad-8f50-80375d5a5630"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:08:34 crc kubenswrapper[4790]: I1124 15:08:34.737077 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mj9j8\" (UniqueName: \"kubernetes.io/projected/3a2cd0fa-f98f-47ad-8f50-80375d5a5630-kube-api-access-mj9j8\") on node \"crc\" DevicePath \"\"" Nov 24 15:08:34 crc kubenswrapper[4790]: I1124 15:08:34.737121 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a2cd0fa-f98f-47ad-8f50-80375d5a5630-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 15:08:34 crc kubenswrapper[4790]: I1124 15:08:34.737134 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a2cd0fa-f98f-47ad-8f50-80375d5a5630-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 15:08:35 crc kubenswrapper[4790]: I1124 15:08:35.035097 4790 generic.go:334] "Generic (PLEG): container finished" podID="3a2cd0fa-f98f-47ad-8f50-80375d5a5630" containerID="1d63d96d582f772de6a45908f3fcb1cbd5fd92e4381342a96a07313c3f48ccac" exitCode=0 Nov 24 15:08:35 crc kubenswrapper[4790]: I1124 15:08:35.035178 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jz2kk" event={"ID":"3a2cd0fa-f98f-47ad-8f50-80375d5a5630","Type":"ContainerDied","Data":"1d63d96d582f772de6a45908f3fcb1cbd5fd92e4381342a96a07313c3f48ccac"} Nov 24 15:08:35 crc kubenswrapper[4790]: I1124 15:08:35.035226 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jz2kk" event={"ID":"3a2cd0fa-f98f-47ad-8f50-80375d5a5630","Type":"ContainerDied","Data":"82550d1b27ed623c11590730600a8835906d84281ef1835809efa7ba03e0542d"} Nov 24 15:08:35 crc kubenswrapper[4790]: I1124 15:08:35.035262 4790 scope.go:117] "RemoveContainer" containerID="1d63d96d582f772de6a45908f3fcb1cbd5fd92e4381342a96a07313c3f48ccac" Nov 24 15:08:35 crc kubenswrapper[4790]: I1124 15:08:35.035335 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jz2kk" Nov 24 15:08:35 crc kubenswrapper[4790]: I1124 15:08:35.070995 4790 scope.go:117] "RemoveContainer" containerID="52ad60005a96109a9ac2cdde3c80aa06cca61c0f3e14b6d07f67c0ee409c86b1" Nov 24 15:08:35 crc kubenswrapper[4790]: I1124 15:08:35.094459 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jz2kk"] Nov 24 15:08:35 crc kubenswrapper[4790]: I1124 15:08:35.111723 4790 scope.go:117] "RemoveContainer" containerID="6eb90b412a54117eb3d63aa5a658e500aa17233b605c74066a3557ac999c3812" Nov 24 15:08:35 crc kubenswrapper[4790]: I1124 15:08:35.114964 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jz2kk"] Nov 24 15:08:35 crc kubenswrapper[4790]: I1124 15:08:35.160480 4790 scope.go:117] "RemoveContainer" containerID="1d63d96d582f772de6a45908f3fcb1cbd5fd92e4381342a96a07313c3f48ccac" Nov 24 15:08:35 crc kubenswrapper[4790]: E1124 15:08:35.160984 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d63d96d582f772de6a45908f3fcb1cbd5fd92e4381342a96a07313c3f48ccac\": container with ID starting with 1d63d96d582f772de6a45908f3fcb1cbd5fd92e4381342a96a07313c3f48ccac not found: ID does not exist" containerID="1d63d96d582f772de6a45908f3fcb1cbd5fd92e4381342a96a07313c3f48ccac" Nov 24 15:08:35 crc kubenswrapper[4790]: I1124 15:08:35.161041 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d63d96d582f772de6a45908f3fcb1cbd5fd92e4381342a96a07313c3f48ccac"} err="failed to get container status \"1d63d96d582f772de6a45908f3fcb1cbd5fd92e4381342a96a07313c3f48ccac\": rpc error: code = NotFound desc = could not find container \"1d63d96d582f772de6a45908f3fcb1cbd5fd92e4381342a96a07313c3f48ccac\": container with ID starting with 1d63d96d582f772de6a45908f3fcb1cbd5fd92e4381342a96a07313c3f48ccac not found: ID does not exist" Nov 24 15:08:35 crc kubenswrapper[4790]: I1124 15:08:35.161075 4790 scope.go:117] "RemoveContainer" containerID="52ad60005a96109a9ac2cdde3c80aa06cca61c0f3e14b6d07f67c0ee409c86b1" Nov 24 15:08:35 crc kubenswrapper[4790]: E1124 15:08:35.161483 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52ad60005a96109a9ac2cdde3c80aa06cca61c0f3e14b6d07f67c0ee409c86b1\": container with ID starting with 52ad60005a96109a9ac2cdde3c80aa06cca61c0f3e14b6d07f67c0ee409c86b1 not found: ID does not exist" containerID="52ad60005a96109a9ac2cdde3c80aa06cca61c0f3e14b6d07f67c0ee409c86b1" Nov 24 15:08:35 crc kubenswrapper[4790]: I1124 15:08:35.161548 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52ad60005a96109a9ac2cdde3c80aa06cca61c0f3e14b6d07f67c0ee409c86b1"} err="failed to get container status \"52ad60005a96109a9ac2cdde3c80aa06cca61c0f3e14b6d07f67c0ee409c86b1\": rpc error: code = NotFound desc = could not find container \"52ad60005a96109a9ac2cdde3c80aa06cca61c0f3e14b6d07f67c0ee409c86b1\": container with ID starting with 52ad60005a96109a9ac2cdde3c80aa06cca61c0f3e14b6d07f67c0ee409c86b1 not found: ID does not exist" Nov 24 15:08:35 crc kubenswrapper[4790]: I1124 15:08:35.161588 4790 scope.go:117] "RemoveContainer" containerID="6eb90b412a54117eb3d63aa5a658e500aa17233b605c74066a3557ac999c3812" Nov 24 15:08:35 crc kubenswrapper[4790]: E1124 15:08:35.161990 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6eb90b412a54117eb3d63aa5a658e500aa17233b605c74066a3557ac999c3812\": container with ID starting with 6eb90b412a54117eb3d63aa5a658e500aa17233b605c74066a3557ac999c3812 not found: ID does not exist" containerID="6eb90b412a54117eb3d63aa5a658e500aa17233b605c74066a3557ac999c3812" Nov 24 15:08:35 crc kubenswrapper[4790]: I1124 15:08:35.162042 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6eb90b412a54117eb3d63aa5a658e500aa17233b605c74066a3557ac999c3812"} err="failed to get container status \"6eb90b412a54117eb3d63aa5a658e500aa17233b605c74066a3557ac999c3812\": rpc error: code = NotFound desc = could not find container \"6eb90b412a54117eb3d63aa5a658e500aa17233b605c74066a3557ac999c3812\": container with ID starting with 6eb90b412a54117eb3d63aa5a658e500aa17233b605c74066a3557ac999c3812 not found: ID does not exist" Nov 24 15:08:36 crc kubenswrapper[4790]: I1124 15:08:36.336933 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a2cd0fa-f98f-47ad-8f50-80375d5a5630" path="/var/lib/kubelet/pods/3a2cd0fa-f98f-47ad-8f50-80375d5a5630/volumes" Nov 24 15:08:37 crc kubenswrapper[4790]: I1124 15:08:37.314396 4790 scope.go:117] "RemoveContainer" containerID="a4432de01408933d930865fc3e585b6bb44fe87c6f5674553d2039220c20c92e" Nov 24 15:08:37 crc kubenswrapper[4790]: E1124 15:08:37.315098 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:08:48 crc kubenswrapper[4790]: I1124 15:08:48.315388 4790 scope.go:117] "RemoveContainer" containerID="a4432de01408933d930865fc3e585b6bb44fe87c6f5674553d2039220c20c92e" Nov 24 15:08:48 crc kubenswrapper[4790]: E1124 15:08:48.316352 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:09:00 crc kubenswrapper[4790]: I1124 15:09:00.315071 4790 scope.go:117] "RemoveContainer" containerID="a4432de01408933d930865fc3e585b6bb44fe87c6f5674553d2039220c20c92e" Nov 24 15:09:00 crc kubenswrapper[4790]: E1124 15:09:00.316096 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:09:12 crc kubenswrapper[4790]: I1124 15:09:12.323155 4790 scope.go:117] "RemoveContainer" containerID="a4432de01408933d930865fc3e585b6bb44fe87c6f5674553d2039220c20c92e" Nov 24 15:09:12 crc kubenswrapper[4790]: E1124 15:09:12.323994 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:09:27 crc kubenswrapper[4790]: I1124 15:09:27.315309 4790 scope.go:117] "RemoveContainer" containerID="a4432de01408933d930865fc3e585b6bb44fe87c6f5674553d2039220c20c92e" Nov 24 15:09:27 crc kubenswrapper[4790]: E1124 15:09:27.316082 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:09:41 crc kubenswrapper[4790]: I1124 15:09:41.314676 4790 scope.go:117] "RemoveContainer" containerID="a4432de01408933d930865fc3e585b6bb44fe87c6f5674553d2039220c20c92e" Nov 24 15:09:41 crc kubenswrapper[4790]: E1124 15:09:41.315629 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:09:56 crc kubenswrapper[4790]: I1124 15:09:56.316211 4790 scope.go:117] "RemoveContainer" containerID="a4432de01408933d930865fc3e585b6bb44fe87c6f5674553d2039220c20c92e" Nov 24 15:09:57 crc kubenswrapper[4790]: I1124 15:09:57.054711 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"f34c5fa54378b1b7f2049ef63829861f39b73c8e72d607198fb71c136fcce70c"} Nov 24 15:10:20 crc kubenswrapper[4790]: I1124 15:10:20.362671 4790 generic.go:334] "Generic (PLEG): container finished" podID="0effec18-dddb-4967-a5f8-1a98979e3338" containerID="b9e707df255ff9aaba0fc3e34b5c9508408bf92671598883395b7a29beec6fb0" exitCode=0 Nov 24 15:10:20 crc kubenswrapper[4790]: I1124 15:10:20.362778 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6" event={"ID":"0effec18-dddb-4967-a5f8-1a98979e3338","Type":"ContainerDied","Data":"b9e707df255ff9aaba0fc3e34b5c9508408bf92671598883395b7a29beec6fb0"} Nov 24 15:10:21 crc kubenswrapper[4790]: I1124 15:10:21.937762 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6" Nov 24 15:10:22 crc kubenswrapper[4790]: I1124 15:10:22.003153 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bv2bj\" (UniqueName: \"kubernetes.io/projected/0effec18-dddb-4967-a5f8-1a98979e3338-kube-api-access-bv2bj\") pod \"0effec18-dddb-4967-a5f8-1a98979e3338\" (UID: \"0effec18-dddb-4967-a5f8-1a98979e3338\") " Nov 24 15:10:22 crc kubenswrapper[4790]: I1124 15:10:22.003360 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0effec18-dddb-4967-a5f8-1a98979e3338-tripleo-cleanup-combined-ca-bundle\") pod \"0effec18-dddb-4967-a5f8-1a98979e3338\" (UID: \"0effec18-dddb-4967-a5f8-1a98979e3338\") " Nov 24 15:10:22 crc kubenswrapper[4790]: I1124 15:10:22.003457 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0effec18-dddb-4967-a5f8-1a98979e3338-inventory\") pod \"0effec18-dddb-4967-a5f8-1a98979e3338\" (UID: \"0effec18-dddb-4967-a5f8-1a98979e3338\") " Nov 24 15:10:22 crc kubenswrapper[4790]: I1124 15:10:22.003524 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0effec18-dddb-4967-a5f8-1a98979e3338-ssh-key\") pod \"0effec18-dddb-4967-a5f8-1a98979e3338\" (UID: \"0effec18-dddb-4967-a5f8-1a98979e3338\") " Nov 24 15:10:22 crc kubenswrapper[4790]: I1124 15:10:22.003620 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0effec18-dddb-4967-a5f8-1a98979e3338-ceph\") pod \"0effec18-dddb-4967-a5f8-1a98979e3338\" (UID: \"0effec18-dddb-4967-a5f8-1a98979e3338\") " Nov 24 15:10:22 crc kubenswrapper[4790]: I1124 15:10:22.009286 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0effec18-dddb-4967-a5f8-1a98979e3338-ceph" (OuterVolumeSpecName: "ceph") pod "0effec18-dddb-4967-a5f8-1a98979e3338" (UID: "0effec18-dddb-4967-a5f8-1a98979e3338"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:10:22 crc kubenswrapper[4790]: I1124 15:10:22.009400 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0effec18-dddb-4967-a5f8-1a98979e3338-kube-api-access-bv2bj" (OuterVolumeSpecName: "kube-api-access-bv2bj") pod "0effec18-dddb-4967-a5f8-1a98979e3338" (UID: "0effec18-dddb-4967-a5f8-1a98979e3338"). InnerVolumeSpecName "kube-api-access-bv2bj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:10:22 crc kubenswrapper[4790]: I1124 15:10:22.010214 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0effec18-dddb-4967-a5f8-1a98979e3338-tripleo-cleanup-combined-ca-bundle" (OuterVolumeSpecName: "tripleo-cleanup-combined-ca-bundle") pod "0effec18-dddb-4967-a5f8-1a98979e3338" (UID: "0effec18-dddb-4967-a5f8-1a98979e3338"). InnerVolumeSpecName "tripleo-cleanup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:10:22 crc kubenswrapper[4790]: I1124 15:10:22.032714 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0effec18-dddb-4967-a5f8-1a98979e3338-inventory" (OuterVolumeSpecName: "inventory") pod "0effec18-dddb-4967-a5f8-1a98979e3338" (UID: "0effec18-dddb-4967-a5f8-1a98979e3338"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:10:22 crc kubenswrapper[4790]: I1124 15:10:22.042980 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0effec18-dddb-4967-a5f8-1a98979e3338-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0effec18-dddb-4967-a5f8-1a98979e3338" (UID: "0effec18-dddb-4967-a5f8-1a98979e3338"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:10:22 crc kubenswrapper[4790]: I1124 15:10:22.106051 4790 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0effec18-dddb-4967-a5f8-1a98979e3338-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 15:10:22 crc kubenswrapper[4790]: I1124 15:10:22.106085 4790 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0effec18-dddb-4967-a5f8-1a98979e3338-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 15:10:22 crc kubenswrapper[4790]: I1124 15:10:22.106094 4790 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0effec18-dddb-4967-a5f8-1a98979e3338-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 15:10:22 crc kubenswrapper[4790]: I1124 15:10:22.106105 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bv2bj\" (UniqueName: \"kubernetes.io/projected/0effec18-dddb-4967-a5f8-1a98979e3338-kube-api-access-bv2bj\") on node \"crc\" DevicePath \"\"" Nov 24 15:10:22 crc kubenswrapper[4790]: I1124 15:10:22.106116 4790 reconciler_common.go:293] "Volume detached for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0effec18-dddb-4967-a5f8-1a98979e3338-tripleo-cleanup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:10:22 crc kubenswrapper[4790]: I1124 15:10:22.399205 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6" event={"ID":"0effec18-dddb-4967-a5f8-1a98979e3338","Type":"ContainerDied","Data":"c5689da16dc1b47d24271e836bba6bad763721756a42083787ba79987a81a1d0"} Nov 24 15:10:22 crc kubenswrapper[4790]: I1124 15:10:22.399270 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c5689da16dc1b47d24271e836bba6bad763721756a42083787ba79987a81a1d0" Nov 24 15:10:22 crc kubenswrapper[4790]: I1124 15:10:22.399457 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6" Nov 24 15:10:22 crc kubenswrapper[4790]: E1124 15:10:22.626985 4790 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0effec18_dddb_4967_a5f8_1a98979e3338.slice/crio-c5689da16dc1b47d24271e836bba6bad763721756a42083787ba79987a81a1d0\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0effec18_dddb_4967_a5f8_1a98979e3338.slice\": RecentStats: unable to find data in memory cache]" Nov 24 15:10:27 crc kubenswrapper[4790]: I1124 15:10:27.218373 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-9cmwr"] Nov 24 15:10:27 crc kubenswrapper[4790]: E1124 15:10:27.219535 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a2cd0fa-f98f-47ad-8f50-80375d5a5630" containerName="extract-utilities" Nov 24 15:10:27 crc kubenswrapper[4790]: I1124 15:10:27.219552 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a2cd0fa-f98f-47ad-8f50-80375d5a5630" containerName="extract-utilities" Nov 24 15:10:27 crc kubenswrapper[4790]: E1124 15:10:27.219572 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0effec18-dddb-4967-a5f8-1a98979e3338" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Nov 24 15:10:27 crc kubenswrapper[4790]: I1124 15:10:27.219582 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="0effec18-dddb-4967-a5f8-1a98979e3338" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Nov 24 15:10:27 crc kubenswrapper[4790]: E1124 15:10:27.219627 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a2cd0fa-f98f-47ad-8f50-80375d5a5630" containerName="registry-server" Nov 24 15:10:27 crc kubenswrapper[4790]: I1124 15:10:27.219635 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a2cd0fa-f98f-47ad-8f50-80375d5a5630" containerName="registry-server" Nov 24 15:10:27 crc kubenswrapper[4790]: E1124 15:10:27.219656 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a2cd0fa-f98f-47ad-8f50-80375d5a5630" containerName="extract-content" Nov 24 15:10:27 crc kubenswrapper[4790]: I1124 15:10:27.219664 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a2cd0fa-f98f-47ad-8f50-80375d5a5630" containerName="extract-content" Nov 24 15:10:27 crc kubenswrapper[4790]: I1124 15:10:27.219923 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="0effec18-dddb-4967-a5f8-1a98979e3338" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Nov 24 15:10:27 crc kubenswrapper[4790]: I1124 15:10:27.219936 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a2cd0fa-f98f-47ad-8f50-80375d5a5630" containerName="registry-server" Nov 24 15:10:27 crc kubenswrapper[4790]: I1124 15:10:27.220805 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-9cmwr" Nov 24 15:10:27 crc kubenswrapper[4790]: I1124 15:10:27.228531 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 15:10:27 crc kubenswrapper[4790]: I1124 15:10:27.228854 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 15:10:27 crc kubenswrapper[4790]: I1124 15:10:27.229189 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 15:10:27 crc kubenswrapper[4790]: I1124 15:10:27.229244 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fc497" Nov 24 15:10:27 crc kubenswrapper[4790]: I1124 15:10:27.242414 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-9cmwr"] Nov 24 15:10:27 crc kubenswrapper[4790]: I1124 15:10:27.339227 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5991caca-bb27-439c-b5f8-4137b17639d9-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-9cmwr\" (UID: \"5991caca-bb27-439c-b5f8-4137b17639d9\") " pod="openstack/bootstrap-openstack-openstack-cell1-9cmwr" Nov 24 15:10:27 crc kubenswrapper[4790]: I1124 15:10:27.339424 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sn22\" (UniqueName: \"kubernetes.io/projected/5991caca-bb27-439c-b5f8-4137b17639d9-kube-api-access-4sn22\") pod \"bootstrap-openstack-openstack-cell1-9cmwr\" (UID: \"5991caca-bb27-439c-b5f8-4137b17639d9\") " pod="openstack/bootstrap-openstack-openstack-cell1-9cmwr" Nov 24 15:10:27 crc kubenswrapper[4790]: I1124 15:10:27.339488 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5991caca-bb27-439c-b5f8-4137b17639d9-inventory\") pod \"bootstrap-openstack-openstack-cell1-9cmwr\" (UID: \"5991caca-bb27-439c-b5f8-4137b17639d9\") " pod="openstack/bootstrap-openstack-openstack-cell1-9cmwr" Nov 24 15:10:27 crc kubenswrapper[4790]: I1124 15:10:27.339523 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5991caca-bb27-439c-b5f8-4137b17639d9-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-9cmwr\" (UID: \"5991caca-bb27-439c-b5f8-4137b17639d9\") " pod="openstack/bootstrap-openstack-openstack-cell1-9cmwr" Nov 24 15:10:27 crc kubenswrapper[4790]: I1124 15:10:27.339645 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5991caca-bb27-439c-b5f8-4137b17639d9-ceph\") pod \"bootstrap-openstack-openstack-cell1-9cmwr\" (UID: \"5991caca-bb27-439c-b5f8-4137b17639d9\") " pod="openstack/bootstrap-openstack-openstack-cell1-9cmwr" Nov 24 15:10:27 crc kubenswrapper[4790]: I1124 15:10:27.441978 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5991caca-bb27-439c-b5f8-4137b17639d9-ceph\") pod \"bootstrap-openstack-openstack-cell1-9cmwr\" (UID: \"5991caca-bb27-439c-b5f8-4137b17639d9\") " pod="openstack/bootstrap-openstack-openstack-cell1-9cmwr" Nov 24 15:10:27 crc kubenswrapper[4790]: I1124 15:10:27.442291 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5991caca-bb27-439c-b5f8-4137b17639d9-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-9cmwr\" (UID: \"5991caca-bb27-439c-b5f8-4137b17639d9\") " pod="openstack/bootstrap-openstack-openstack-cell1-9cmwr" Nov 24 15:10:27 crc kubenswrapper[4790]: I1124 15:10:27.442607 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sn22\" (UniqueName: \"kubernetes.io/projected/5991caca-bb27-439c-b5f8-4137b17639d9-kube-api-access-4sn22\") pod \"bootstrap-openstack-openstack-cell1-9cmwr\" (UID: \"5991caca-bb27-439c-b5f8-4137b17639d9\") " pod="openstack/bootstrap-openstack-openstack-cell1-9cmwr" Nov 24 15:10:27 crc kubenswrapper[4790]: I1124 15:10:27.442663 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5991caca-bb27-439c-b5f8-4137b17639d9-inventory\") pod \"bootstrap-openstack-openstack-cell1-9cmwr\" (UID: \"5991caca-bb27-439c-b5f8-4137b17639d9\") " pod="openstack/bootstrap-openstack-openstack-cell1-9cmwr" Nov 24 15:10:27 crc kubenswrapper[4790]: I1124 15:10:27.442702 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5991caca-bb27-439c-b5f8-4137b17639d9-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-9cmwr\" (UID: \"5991caca-bb27-439c-b5f8-4137b17639d9\") " pod="openstack/bootstrap-openstack-openstack-cell1-9cmwr" Nov 24 15:10:27 crc kubenswrapper[4790]: I1124 15:10:27.452109 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5991caca-bb27-439c-b5f8-4137b17639d9-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-9cmwr\" (UID: \"5991caca-bb27-439c-b5f8-4137b17639d9\") " pod="openstack/bootstrap-openstack-openstack-cell1-9cmwr" Nov 24 15:10:27 crc kubenswrapper[4790]: I1124 15:10:27.452785 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5991caca-bb27-439c-b5f8-4137b17639d9-ceph\") pod \"bootstrap-openstack-openstack-cell1-9cmwr\" (UID: \"5991caca-bb27-439c-b5f8-4137b17639d9\") " pod="openstack/bootstrap-openstack-openstack-cell1-9cmwr" Nov 24 15:10:27 crc kubenswrapper[4790]: I1124 15:10:27.453539 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5991caca-bb27-439c-b5f8-4137b17639d9-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-9cmwr\" (UID: \"5991caca-bb27-439c-b5f8-4137b17639d9\") " pod="openstack/bootstrap-openstack-openstack-cell1-9cmwr" Nov 24 15:10:27 crc kubenswrapper[4790]: I1124 15:10:27.466601 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5991caca-bb27-439c-b5f8-4137b17639d9-inventory\") pod \"bootstrap-openstack-openstack-cell1-9cmwr\" (UID: \"5991caca-bb27-439c-b5f8-4137b17639d9\") " pod="openstack/bootstrap-openstack-openstack-cell1-9cmwr" Nov 24 15:10:27 crc kubenswrapper[4790]: I1124 15:10:27.471480 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sn22\" (UniqueName: \"kubernetes.io/projected/5991caca-bb27-439c-b5f8-4137b17639d9-kube-api-access-4sn22\") pod \"bootstrap-openstack-openstack-cell1-9cmwr\" (UID: \"5991caca-bb27-439c-b5f8-4137b17639d9\") " pod="openstack/bootstrap-openstack-openstack-cell1-9cmwr" Nov 24 15:10:27 crc kubenswrapper[4790]: I1124 15:10:27.548377 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-9cmwr" Nov 24 15:10:28 crc kubenswrapper[4790]: I1124 15:10:28.144871 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-9cmwr"] Nov 24 15:10:28 crc kubenswrapper[4790]: W1124 15:10:28.151809 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5991caca_bb27_439c_b5f8_4137b17639d9.slice/crio-ab2aef447a6838d4bf1fd2c9d664c8b3e040efa57653d2900c1736caf497dbe3 WatchSource:0}: Error finding container ab2aef447a6838d4bf1fd2c9d664c8b3e040efa57653d2900c1736caf497dbe3: Status 404 returned error can't find the container with id ab2aef447a6838d4bf1fd2c9d664c8b3e040efa57653d2900c1736caf497dbe3 Nov 24 15:10:28 crc kubenswrapper[4790]: I1124 15:10:28.155597 4790 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 15:10:28 crc kubenswrapper[4790]: I1124 15:10:28.482282 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-9cmwr" event={"ID":"5991caca-bb27-439c-b5f8-4137b17639d9","Type":"ContainerStarted","Data":"ab2aef447a6838d4bf1fd2c9d664c8b3e040efa57653d2900c1736caf497dbe3"} Nov 24 15:10:29 crc kubenswrapper[4790]: I1124 15:10:29.495211 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-9cmwr" event={"ID":"5991caca-bb27-439c-b5f8-4137b17639d9","Type":"ContainerStarted","Data":"8eae605fb3294a97505460dac9d55175e3ced2606ec18aa64828b2df680e53f5"} Nov 24 15:10:29 crc kubenswrapper[4790]: I1124 15:10:29.523458 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-openstack-openstack-cell1-9cmwr" podStartSLOduration=2.319393962 podStartE2EDuration="2.523246809s" podCreationTimestamp="2025-11-24 15:10:27 +0000 UTC" firstStartedPulling="2025-11-24 15:10:28.155313449 +0000 UTC m=+7076.535207121" lastFinishedPulling="2025-11-24 15:10:28.359166296 +0000 UTC m=+7076.739059968" observedRunningTime="2025-11-24 15:10:29.516435564 +0000 UTC m=+7077.896329256" watchObservedRunningTime="2025-11-24 15:10:29.523246809 +0000 UTC m=+7077.903140501" Nov 24 15:12:13 crc kubenswrapper[4790]: I1124 15:12:13.938437 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:12:13 crc kubenswrapper[4790]: I1124 15:12:13.939842 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:12:43 crc kubenswrapper[4790]: I1124 15:12:43.938849 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:12:43 crc kubenswrapper[4790]: I1124 15:12:43.939771 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:13:13 crc kubenswrapper[4790]: I1124 15:13:13.938242 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:13:13 crc kubenswrapper[4790]: I1124 15:13:13.939452 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:13:13 crc kubenswrapper[4790]: I1124 15:13:13.939548 4790 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 15:13:13 crc kubenswrapper[4790]: I1124 15:13:13.940990 4790 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f34c5fa54378b1b7f2049ef63829861f39b73c8e72d607198fb71c136fcce70c"} pod="openshift-machine-config-operator/machine-config-daemon-xz49t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 15:13:13 crc kubenswrapper[4790]: I1124 15:13:13.941122 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" containerID="cri-o://f34c5fa54378b1b7f2049ef63829861f39b73c8e72d607198fb71c136fcce70c" gracePeriod=600 Nov 24 15:13:14 crc kubenswrapper[4790]: I1124 15:13:14.656038 4790 generic.go:334] "Generic (PLEG): container finished" podID="0d73b133-48f0-455f-8f6a-742e633f631a" containerID="f34c5fa54378b1b7f2049ef63829861f39b73c8e72d607198fb71c136fcce70c" exitCode=0 Nov 24 15:13:14 crc kubenswrapper[4790]: I1124 15:13:14.656097 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerDied","Data":"f34c5fa54378b1b7f2049ef63829861f39b73c8e72d607198fb71c136fcce70c"} Nov 24 15:13:14 crc kubenswrapper[4790]: I1124 15:13:14.657134 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"ed032daa21005f5603fc87ffbe2ed53274be2bbefd5cbb4153e64e56dd530b81"} Nov 24 15:13:14 crc kubenswrapper[4790]: I1124 15:13:14.657179 4790 scope.go:117] "RemoveContainer" containerID="a4432de01408933d930865fc3e585b6bb44fe87c6f5674553d2039220c20c92e" Nov 24 15:13:42 crc kubenswrapper[4790]: I1124 15:13:42.972048 4790 generic.go:334] "Generic (PLEG): container finished" podID="5991caca-bb27-439c-b5f8-4137b17639d9" containerID="8eae605fb3294a97505460dac9d55175e3ced2606ec18aa64828b2df680e53f5" exitCode=0 Nov 24 15:13:42 crc kubenswrapper[4790]: I1124 15:13:42.972153 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-9cmwr" event={"ID":"5991caca-bb27-439c-b5f8-4137b17639d9","Type":"ContainerDied","Data":"8eae605fb3294a97505460dac9d55175e3ced2606ec18aa64828b2df680e53f5"} Nov 24 15:13:44 crc kubenswrapper[4790]: I1124 15:13:44.633281 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-9cmwr" Nov 24 15:13:44 crc kubenswrapper[4790]: I1124 15:13:44.807232 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5991caca-bb27-439c-b5f8-4137b17639d9-ssh-key\") pod \"5991caca-bb27-439c-b5f8-4137b17639d9\" (UID: \"5991caca-bb27-439c-b5f8-4137b17639d9\") " Nov 24 15:13:44 crc kubenswrapper[4790]: I1124 15:13:44.807319 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4sn22\" (UniqueName: \"kubernetes.io/projected/5991caca-bb27-439c-b5f8-4137b17639d9-kube-api-access-4sn22\") pod \"5991caca-bb27-439c-b5f8-4137b17639d9\" (UID: \"5991caca-bb27-439c-b5f8-4137b17639d9\") " Nov 24 15:13:44 crc kubenswrapper[4790]: I1124 15:13:44.807349 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5991caca-bb27-439c-b5f8-4137b17639d9-bootstrap-combined-ca-bundle\") pod \"5991caca-bb27-439c-b5f8-4137b17639d9\" (UID: \"5991caca-bb27-439c-b5f8-4137b17639d9\") " Nov 24 15:13:44 crc kubenswrapper[4790]: I1124 15:13:44.807385 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5991caca-bb27-439c-b5f8-4137b17639d9-ceph\") pod \"5991caca-bb27-439c-b5f8-4137b17639d9\" (UID: \"5991caca-bb27-439c-b5f8-4137b17639d9\") " Nov 24 15:13:44 crc kubenswrapper[4790]: I1124 15:13:44.807476 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5991caca-bb27-439c-b5f8-4137b17639d9-inventory\") pod \"5991caca-bb27-439c-b5f8-4137b17639d9\" (UID: \"5991caca-bb27-439c-b5f8-4137b17639d9\") " Nov 24 15:13:44 crc kubenswrapper[4790]: I1124 15:13:44.817692 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5991caca-bb27-439c-b5f8-4137b17639d9-kube-api-access-4sn22" (OuterVolumeSpecName: "kube-api-access-4sn22") pod "5991caca-bb27-439c-b5f8-4137b17639d9" (UID: "5991caca-bb27-439c-b5f8-4137b17639d9"). InnerVolumeSpecName "kube-api-access-4sn22". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:13:44 crc kubenswrapper[4790]: I1124 15:13:44.818262 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5991caca-bb27-439c-b5f8-4137b17639d9-ceph" (OuterVolumeSpecName: "ceph") pod "5991caca-bb27-439c-b5f8-4137b17639d9" (UID: "5991caca-bb27-439c-b5f8-4137b17639d9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:13:44 crc kubenswrapper[4790]: I1124 15:13:44.820339 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5991caca-bb27-439c-b5f8-4137b17639d9-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "5991caca-bb27-439c-b5f8-4137b17639d9" (UID: "5991caca-bb27-439c-b5f8-4137b17639d9"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:13:44 crc kubenswrapper[4790]: I1124 15:13:44.841221 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5991caca-bb27-439c-b5f8-4137b17639d9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5991caca-bb27-439c-b5f8-4137b17639d9" (UID: "5991caca-bb27-439c-b5f8-4137b17639d9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:13:44 crc kubenswrapper[4790]: I1124 15:13:44.883201 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5991caca-bb27-439c-b5f8-4137b17639d9-inventory" (OuterVolumeSpecName: "inventory") pod "5991caca-bb27-439c-b5f8-4137b17639d9" (UID: "5991caca-bb27-439c-b5f8-4137b17639d9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:13:44 crc kubenswrapper[4790]: I1124 15:13:44.910624 4790 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5991caca-bb27-439c-b5f8-4137b17639d9-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 15:13:44 crc kubenswrapper[4790]: I1124 15:13:44.910693 4790 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5991caca-bb27-439c-b5f8-4137b17639d9-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 15:13:44 crc kubenswrapper[4790]: I1124 15:13:44.910725 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4sn22\" (UniqueName: \"kubernetes.io/projected/5991caca-bb27-439c-b5f8-4137b17639d9-kube-api-access-4sn22\") on node \"crc\" DevicePath \"\"" Nov 24 15:13:44 crc kubenswrapper[4790]: I1124 15:13:44.910754 4790 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5991caca-bb27-439c-b5f8-4137b17639d9-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:13:44 crc kubenswrapper[4790]: I1124 15:13:44.910780 4790 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5991caca-bb27-439c-b5f8-4137b17639d9-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 15:13:44 crc kubenswrapper[4790]: I1124 15:13:44.994667 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-9cmwr" event={"ID":"5991caca-bb27-439c-b5f8-4137b17639d9","Type":"ContainerDied","Data":"ab2aef447a6838d4bf1fd2c9d664c8b3e040efa57653d2900c1736caf497dbe3"} Nov 24 15:13:44 crc kubenswrapper[4790]: I1124 15:13:44.994713 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab2aef447a6838d4bf1fd2c9d664c8b3e040efa57653d2900c1736caf497dbe3" Nov 24 15:13:44 crc kubenswrapper[4790]: I1124 15:13:44.994718 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-9cmwr" Nov 24 15:13:45 crc kubenswrapper[4790]: I1124 15:13:45.097004 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-d8www"] Nov 24 15:13:45 crc kubenswrapper[4790]: E1124 15:13:45.097542 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5991caca-bb27-439c-b5f8-4137b17639d9" containerName="bootstrap-openstack-openstack-cell1" Nov 24 15:13:45 crc kubenswrapper[4790]: I1124 15:13:45.097563 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="5991caca-bb27-439c-b5f8-4137b17639d9" containerName="bootstrap-openstack-openstack-cell1" Nov 24 15:13:45 crc kubenswrapper[4790]: I1124 15:13:45.097856 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="5991caca-bb27-439c-b5f8-4137b17639d9" containerName="bootstrap-openstack-openstack-cell1" Nov 24 15:13:45 crc kubenswrapper[4790]: I1124 15:13:45.098837 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-d8www" Nov 24 15:13:45 crc kubenswrapper[4790]: I1124 15:13:45.100979 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fc497" Nov 24 15:13:45 crc kubenswrapper[4790]: I1124 15:13:45.101563 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 15:13:45 crc kubenswrapper[4790]: I1124 15:13:45.101761 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 15:13:45 crc kubenswrapper[4790]: I1124 15:13:45.101918 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 15:13:45 crc kubenswrapper[4790]: I1124 15:13:45.116305 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-d8www"] Nov 24 15:13:45 crc kubenswrapper[4790]: I1124 15:13:45.123630 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9qv4\" (UniqueName: \"kubernetes.io/projected/e3c33642-ddec-46a5-b86d-220f086ed2cf-kube-api-access-v9qv4\") pod \"download-cache-openstack-openstack-cell1-d8www\" (UID: \"e3c33642-ddec-46a5-b86d-220f086ed2cf\") " pod="openstack/download-cache-openstack-openstack-cell1-d8www" Nov 24 15:13:45 crc kubenswrapper[4790]: I1124 15:13:45.123795 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3c33642-ddec-46a5-b86d-220f086ed2cf-inventory\") pod \"download-cache-openstack-openstack-cell1-d8www\" (UID: \"e3c33642-ddec-46a5-b86d-220f086ed2cf\") " pod="openstack/download-cache-openstack-openstack-cell1-d8www" Nov 24 15:13:45 crc kubenswrapper[4790]: I1124 15:13:45.124280 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3c33642-ddec-46a5-b86d-220f086ed2cf-ssh-key\") pod \"download-cache-openstack-openstack-cell1-d8www\" (UID: \"e3c33642-ddec-46a5-b86d-220f086ed2cf\") " pod="openstack/download-cache-openstack-openstack-cell1-d8www" Nov 24 15:13:45 crc kubenswrapper[4790]: I1124 15:13:45.124445 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e3c33642-ddec-46a5-b86d-220f086ed2cf-ceph\") pod \"download-cache-openstack-openstack-cell1-d8www\" (UID: \"e3c33642-ddec-46a5-b86d-220f086ed2cf\") " pod="openstack/download-cache-openstack-openstack-cell1-d8www" Nov 24 15:13:45 crc kubenswrapper[4790]: I1124 15:13:45.226470 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9qv4\" (UniqueName: \"kubernetes.io/projected/e3c33642-ddec-46a5-b86d-220f086ed2cf-kube-api-access-v9qv4\") pod \"download-cache-openstack-openstack-cell1-d8www\" (UID: \"e3c33642-ddec-46a5-b86d-220f086ed2cf\") " pod="openstack/download-cache-openstack-openstack-cell1-d8www" Nov 24 15:13:45 crc kubenswrapper[4790]: I1124 15:13:45.226563 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3c33642-ddec-46a5-b86d-220f086ed2cf-inventory\") pod \"download-cache-openstack-openstack-cell1-d8www\" (UID: \"e3c33642-ddec-46a5-b86d-220f086ed2cf\") " pod="openstack/download-cache-openstack-openstack-cell1-d8www" Nov 24 15:13:45 crc kubenswrapper[4790]: I1124 15:13:45.226658 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3c33642-ddec-46a5-b86d-220f086ed2cf-ssh-key\") pod \"download-cache-openstack-openstack-cell1-d8www\" (UID: \"e3c33642-ddec-46a5-b86d-220f086ed2cf\") " pod="openstack/download-cache-openstack-openstack-cell1-d8www" Nov 24 15:13:45 crc kubenswrapper[4790]: I1124 15:13:45.226724 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e3c33642-ddec-46a5-b86d-220f086ed2cf-ceph\") pod \"download-cache-openstack-openstack-cell1-d8www\" (UID: \"e3c33642-ddec-46a5-b86d-220f086ed2cf\") " pod="openstack/download-cache-openstack-openstack-cell1-d8www" Nov 24 15:13:45 crc kubenswrapper[4790]: I1124 15:13:45.230561 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3c33642-ddec-46a5-b86d-220f086ed2cf-inventory\") pod \"download-cache-openstack-openstack-cell1-d8www\" (UID: \"e3c33642-ddec-46a5-b86d-220f086ed2cf\") " pod="openstack/download-cache-openstack-openstack-cell1-d8www" Nov 24 15:13:45 crc kubenswrapper[4790]: I1124 15:13:45.230623 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e3c33642-ddec-46a5-b86d-220f086ed2cf-ceph\") pod \"download-cache-openstack-openstack-cell1-d8www\" (UID: \"e3c33642-ddec-46a5-b86d-220f086ed2cf\") " pod="openstack/download-cache-openstack-openstack-cell1-d8www" Nov 24 15:13:45 crc kubenswrapper[4790]: I1124 15:13:45.231403 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3c33642-ddec-46a5-b86d-220f086ed2cf-ssh-key\") pod \"download-cache-openstack-openstack-cell1-d8www\" (UID: \"e3c33642-ddec-46a5-b86d-220f086ed2cf\") " pod="openstack/download-cache-openstack-openstack-cell1-d8www" Nov 24 15:13:45 crc kubenswrapper[4790]: I1124 15:13:45.246541 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9qv4\" (UniqueName: \"kubernetes.io/projected/e3c33642-ddec-46a5-b86d-220f086ed2cf-kube-api-access-v9qv4\") pod \"download-cache-openstack-openstack-cell1-d8www\" (UID: \"e3c33642-ddec-46a5-b86d-220f086ed2cf\") " pod="openstack/download-cache-openstack-openstack-cell1-d8www" Nov 24 15:13:45 crc kubenswrapper[4790]: I1124 15:13:45.424615 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-d8www" Nov 24 15:13:46 crc kubenswrapper[4790]: I1124 15:13:46.045210 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-d8www"] Nov 24 15:13:47 crc kubenswrapper[4790]: I1124 15:13:47.017655 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-d8www" event={"ID":"e3c33642-ddec-46a5-b86d-220f086ed2cf","Type":"ContainerStarted","Data":"8cb6058225ebb65edd2d1200d22aa629aa5d3fc915bac75c11db4a7d9c3b47e4"} Nov 24 15:13:47 crc kubenswrapper[4790]: I1124 15:13:47.018050 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-d8www" event={"ID":"e3c33642-ddec-46a5-b86d-220f086ed2cf","Type":"ContainerStarted","Data":"414801491e66d0e1da71c5262c63c1d8632f0bb0d2b71975423c736522cdc3b8"} Nov 24 15:13:47 crc kubenswrapper[4790]: I1124 15:13:47.054149 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-openstack-openstack-cell1-d8www" podStartSLOduration=1.87049734 podStartE2EDuration="2.054107503s" podCreationTimestamp="2025-11-24 15:13:45 +0000 UTC" firstStartedPulling="2025-11-24 15:13:46.048195484 +0000 UTC m=+7274.428089146" lastFinishedPulling="2025-11-24 15:13:46.231805647 +0000 UTC m=+7274.611699309" observedRunningTime="2025-11-24 15:13:47.039294314 +0000 UTC m=+7275.419188046" watchObservedRunningTime="2025-11-24 15:13:47.054107503 +0000 UTC m=+7275.434001205" Nov 24 15:15:00 crc kubenswrapper[4790]: I1124 15:15:00.197445 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399955-nvl5w"] Nov 24 15:15:00 crc kubenswrapper[4790]: I1124 15:15:00.199195 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-nvl5w" Nov 24 15:15:00 crc kubenswrapper[4790]: I1124 15:15:00.200837 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 15:15:00 crc kubenswrapper[4790]: I1124 15:15:00.201092 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 15:15:00 crc kubenswrapper[4790]: I1124 15:15:00.208646 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399955-nvl5w"] Nov 24 15:15:00 crc kubenswrapper[4790]: I1124 15:15:00.289520 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/833d9b1e-3d1a-4269-8dbb-26091e6479be-secret-volume\") pod \"collect-profiles-29399955-nvl5w\" (UID: \"833d9b1e-3d1a-4269-8dbb-26091e6479be\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-nvl5w" Nov 24 15:15:00 crc kubenswrapper[4790]: I1124 15:15:00.289604 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/833d9b1e-3d1a-4269-8dbb-26091e6479be-config-volume\") pod \"collect-profiles-29399955-nvl5w\" (UID: \"833d9b1e-3d1a-4269-8dbb-26091e6479be\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-nvl5w" Nov 24 15:15:00 crc kubenswrapper[4790]: I1124 15:15:00.289765 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rc8dj\" (UniqueName: \"kubernetes.io/projected/833d9b1e-3d1a-4269-8dbb-26091e6479be-kube-api-access-rc8dj\") pod \"collect-profiles-29399955-nvl5w\" (UID: \"833d9b1e-3d1a-4269-8dbb-26091e6479be\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-nvl5w" Nov 24 15:15:00 crc kubenswrapper[4790]: I1124 15:15:00.391188 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/833d9b1e-3d1a-4269-8dbb-26091e6479be-secret-volume\") pod \"collect-profiles-29399955-nvl5w\" (UID: \"833d9b1e-3d1a-4269-8dbb-26091e6479be\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-nvl5w" Nov 24 15:15:00 crc kubenswrapper[4790]: I1124 15:15:00.391261 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/833d9b1e-3d1a-4269-8dbb-26091e6479be-config-volume\") pod \"collect-profiles-29399955-nvl5w\" (UID: \"833d9b1e-3d1a-4269-8dbb-26091e6479be\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-nvl5w" Nov 24 15:15:00 crc kubenswrapper[4790]: I1124 15:15:00.391310 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rc8dj\" (UniqueName: \"kubernetes.io/projected/833d9b1e-3d1a-4269-8dbb-26091e6479be-kube-api-access-rc8dj\") pod \"collect-profiles-29399955-nvl5w\" (UID: \"833d9b1e-3d1a-4269-8dbb-26091e6479be\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-nvl5w" Nov 24 15:15:00 crc kubenswrapper[4790]: I1124 15:15:00.392621 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/833d9b1e-3d1a-4269-8dbb-26091e6479be-config-volume\") pod \"collect-profiles-29399955-nvl5w\" (UID: \"833d9b1e-3d1a-4269-8dbb-26091e6479be\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-nvl5w" Nov 24 15:15:00 crc kubenswrapper[4790]: I1124 15:15:00.400241 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/833d9b1e-3d1a-4269-8dbb-26091e6479be-secret-volume\") pod \"collect-profiles-29399955-nvl5w\" (UID: \"833d9b1e-3d1a-4269-8dbb-26091e6479be\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-nvl5w" Nov 24 15:15:00 crc kubenswrapper[4790]: I1124 15:15:00.419838 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rc8dj\" (UniqueName: \"kubernetes.io/projected/833d9b1e-3d1a-4269-8dbb-26091e6479be-kube-api-access-rc8dj\") pod \"collect-profiles-29399955-nvl5w\" (UID: \"833d9b1e-3d1a-4269-8dbb-26091e6479be\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-nvl5w" Nov 24 15:15:00 crc kubenswrapper[4790]: I1124 15:15:00.454846 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-59gl2" Nov 24 15:15:00 crc kubenswrapper[4790]: I1124 15:15:00.519569 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-nvl5w" Nov 24 15:15:01 crc kubenswrapper[4790]: I1124 15:15:01.694341 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399955-nvl5w"] Nov 24 15:15:01 crc kubenswrapper[4790]: I1124 15:15:01.918830 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-nvl5w" event={"ID":"833d9b1e-3d1a-4269-8dbb-26091e6479be","Type":"ContainerStarted","Data":"f6629e9e00132a19cdb1e8cb2f35e186041658cfbc9bb94d2c71f1b5f442ca0d"} Nov 24 15:15:01 crc kubenswrapper[4790]: I1124 15:15:01.918930 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-nvl5w" event={"ID":"833d9b1e-3d1a-4269-8dbb-26091e6479be","Type":"ContainerStarted","Data":"2278f428c9ef9387006cf4a22f8ac1df15af569178177a0fb894096c125bc40c"} Nov 24 15:15:01 crc kubenswrapper[4790]: I1124 15:15:01.951846 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-nvl5w" podStartSLOduration=1.951818351 podStartE2EDuration="1.951818351s" podCreationTimestamp="2025-11-24 15:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 15:15:01.949703641 +0000 UTC m=+7350.329597313" watchObservedRunningTime="2025-11-24 15:15:01.951818351 +0000 UTC m=+7350.331712023" Nov 24 15:15:02 crc kubenswrapper[4790]: I1124 15:15:02.932345 4790 generic.go:334] "Generic (PLEG): container finished" podID="833d9b1e-3d1a-4269-8dbb-26091e6479be" containerID="f6629e9e00132a19cdb1e8cb2f35e186041658cfbc9bb94d2c71f1b5f442ca0d" exitCode=0 Nov 24 15:15:02 crc kubenswrapper[4790]: I1124 15:15:02.932427 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-nvl5w" event={"ID":"833d9b1e-3d1a-4269-8dbb-26091e6479be","Type":"ContainerDied","Data":"f6629e9e00132a19cdb1e8cb2f35e186041658cfbc9bb94d2c71f1b5f442ca0d"} Nov 24 15:15:03 crc kubenswrapper[4790]: I1124 15:15:03.255308 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gsmqf"] Nov 24 15:15:03 crc kubenswrapper[4790]: I1124 15:15:03.259744 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gsmqf" Nov 24 15:15:03 crc kubenswrapper[4790]: I1124 15:15:03.277361 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gsmqf"] Nov 24 15:15:03 crc kubenswrapper[4790]: I1124 15:15:03.362304 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c67690a9-9819-45bb-9d46-24ebc69923b9-catalog-content\") pod \"certified-operators-gsmqf\" (UID: \"c67690a9-9819-45bb-9d46-24ebc69923b9\") " pod="openshift-marketplace/certified-operators-gsmqf" Nov 24 15:15:03 crc kubenswrapper[4790]: I1124 15:15:03.362351 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c67690a9-9819-45bb-9d46-24ebc69923b9-utilities\") pod \"certified-operators-gsmqf\" (UID: \"c67690a9-9819-45bb-9d46-24ebc69923b9\") " pod="openshift-marketplace/certified-operators-gsmqf" Nov 24 15:15:03 crc kubenswrapper[4790]: I1124 15:15:03.362433 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnxbv\" (UniqueName: \"kubernetes.io/projected/c67690a9-9819-45bb-9d46-24ebc69923b9-kube-api-access-fnxbv\") pod \"certified-operators-gsmqf\" (UID: \"c67690a9-9819-45bb-9d46-24ebc69923b9\") " pod="openshift-marketplace/certified-operators-gsmqf" Nov 24 15:15:03 crc kubenswrapper[4790]: I1124 15:15:03.464212 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c67690a9-9819-45bb-9d46-24ebc69923b9-catalog-content\") pod \"certified-operators-gsmqf\" (UID: \"c67690a9-9819-45bb-9d46-24ebc69923b9\") " pod="openshift-marketplace/certified-operators-gsmqf" Nov 24 15:15:03 crc kubenswrapper[4790]: I1124 15:15:03.464257 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c67690a9-9819-45bb-9d46-24ebc69923b9-utilities\") pod \"certified-operators-gsmqf\" (UID: \"c67690a9-9819-45bb-9d46-24ebc69923b9\") " pod="openshift-marketplace/certified-operators-gsmqf" Nov 24 15:15:03 crc kubenswrapper[4790]: I1124 15:15:03.464330 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnxbv\" (UniqueName: \"kubernetes.io/projected/c67690a9-9819-45bb-9d46-24ebc69923b9-kube-api-access-fnxbv\") pod \"certified-operators-gsmqf\" (UID: \"c67690a9-9819-45bb-9d46-24ebc69923b9\") " pod="openshift-marketplace/certified-operators-gsmqf" Nov 24 15:15:03 crc kubenswrapper[4790]: I1124 15:15:03.464927 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c67690a9-9819-45bb-9d46-24ebc69923b9-catalog-content\") pod \"certified-operators-gsmqf\" (UID: \"c67690a9-9819-45bb-9d46-24ebc69923b9\") " pod="openshift-marketplace/certified-operators-gsmqf" Nov 24 15:15:03 crc kubenswrapper[4790]: I1124 15:15:03.465027 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c67690a9-9819-45bb-9d46-24ebc69923b9-utilities\") pod \"certified-operators-gsmqf\" (UID: \"c67690a9-9819-45bb-9d46-24ebc69923b9\") " pod="openshift-marketplace/certified-operators-gsmqf" Nov 24 15:15:03 crc kubenswrapper[4790]: I1124 15:15:03.489599 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnxbv\" (UniqueName: \"kubernetes.io/projected/c67690a9-9819-45bb-9d46-24ebc69923b9-kube-api-access-fnxbv\") pod \"certified-operators-gsmqf\" (UID: \"c67690a9-9819-45bb-9d46-24ebc69923b9\") " pod="openshift-marketplace/certified-operators-gsmqf" Nov 24 15:15:03 crc kubenswrapper[4790]: I1124 15:15:03.580441 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gsmqf" Nov 24 15:15:04 crc kubenswrapper[4790]: I1124 15:15:04.136709 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gsmqf"] Nov 24 15:15:04 crc kubenswrapper[4790]: W1124 15:15:04.164016 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc67690a9_9819_45bb_9d46_24ebc69923b9.slice/crio-fb925f4df3f3bb70cba2ae7de415717eff58cff769310c94b9f0cb9432715312 WatchSource:0}: Error finding container fb925f4df3f3bb70cba2ae7de415717eff58cff769310c94b9f0cb9432715312: Status 404 returned error can't find the container with id fb925f4df3f3bb70cba2ae7de415717eff58cff769310c94b9f0cb9432715312 Nov 24 15:15:04 crc kubenswrapper[4790]: I1124 15:15:04.297258 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-nvl5w" Nov 24 15:15:04 crc kubenswrapper[4790]: I1124 15:15:04.380956 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/833d9b1e-3d1a-4269-8dbb-26091e6479be-secret-volume\") pod \"833d9b1e-3d1a-4269-8dbb-26091e6479be\" (UID: \"833d9b1e-3d1a-4269-8dbb-26091e6479be\") " Nov 24 15:15:04 crc kubenswrapper[4790]: I1124 15:15:04.381015 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/833d9b1e-3d1a-4269-8dbb-26091e6479be-config-volume\") pod \"833d9b1e-3d1a-4269-8dbb-26091e6479be\" (UID: \"833d9b1e-3d1a-4269-8dbb-26091e6479be\") " Nov 24 15:15:04 crc kubenswrapper[4790]: I1124 15:15:04.381121 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rc8dj\" (UniqueName: \"kubernetes.io/projected/833d9b1e-3d1a-4269-8dbb-26091e6479be-kube-api-access-rc8dj\") pod \"833d9b1e-3d1a-4269-8dbb-26091e6479be\" (UID: \"833d9b1e-3d1a-4269-8dbb-26091e6479be\") " Nov 24 15:15:04 crc kubenswrapper[4790]: I1124 15:15:04.382028 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/833d9b1e-3d1a-4269-8dbb-26091e6479be-config-volume" (OuterVolumeSpecName: "config-volume") pod "833d9b1e-3d1a-4269-8dbb-26091e6479be" (UID: "833d9b1e-3d1a-4269-8dbb-26091e6479be"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 15:15:04 crc kubenswrapper[4790]: I1124 15:15:04.383123 4790 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/833d9b1e-3d1a-4269-8dbb-26091e6479be-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 15:15:04 crc kubenswrapper[4790]: I1124 15:15:04.386300 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/833d9b1e-3d1a-4269-8dbb-26091e6479be-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "833d9b1e-3d1a-4269-8dbb-26091e6479be" (UID: "833d9b1e-3d1a-4269-8dbb-26091e6479be"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:15:04 crc kubenswrapper[4790]: I1124 15:15:04.386766 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/833d9b1e-3d1a-4269-8dbb-26091e6479be-kube-api-access-rc8dj" (OuterVolumeSpecName: "kube-api-access-rc8dj") pod "833d9b1e-3d1a-4269-8dbb-26091e6479be" (UID: "833d9b1e-3d1a-4269-8dbb-26091e6479be"). InnerVolumeSpecName "kube-api-access-rc8dj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:15:04 crc kubenswrapper[4790]: I1124 15:15:04.484258 4790 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/833d9b1e-3d1a-4269-8dbb-26091e6479be-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 15:15:04 crc kubenswrapper[4790]: I1124 15:15:04.484287 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rc8dj\" (UniqueName: \"kubernetes.io/projected/833d9b1e-3d1a-4269-8dbb-26091e6479be-kube-api-access-rc8dj\") on node \"crc\" DevicePath \"\"" Nov 24 15:15:04 crc kubenswrapper[4790]: I1124 15:15:04.768670 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399910-j9csk"] Nov 24 15:15:04 crc kubenswrapper[4790]: I1124 15:15:04.777209 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399910-j9csk"] Nov 24 15:15:04 crc kubenswrapper[4790]: I1124 15:15:04.968872 4790 generic.go:334] "Generic (PLEG): container finished" podID="c67690a9-9819-45bb-9d46-24ebc69923b9" containerID="b4840e6956582ab07d2454cca0393537977a54cdf029dffd63bd63ef9b44947b" exitCode=0 Nov 24 15:15:04 crc kubenswrapper[4790]: I1124 15:15:04.969021 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gsmqf" event={"ID":"c67690a9-9819-45bb-9d46-24ebc69923b9","Type":"ContainerDied","Data":"b4840e6956582ab07d2454cca0393537977a54cdf029dffd63bd63ef9b44947b"} Nov 24 15:15:04 crc kubenswrapper[4790]: I1124 15:15:04.969369 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gsmqf" event={"ID":"c67690a9-9819-45bb-9d46-24ebc69923b9","Type":"ContainerStarted","Data":"fb925f4df3f3bb70cba2ae7de415717eff58cff769310c94b9f0cb9432715312"} Nov 24 15:15:04 crc kubenswrapper[4790]: I1124 15:15:04.980387 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-nvl5w" event={"ID":"833d9b1e-3d1a-4269-8dbb-26091e6479be","Type":"ContainerDied","Data":"2278f428c9ef9387006cf4a22f8ac1df15af569178177a0fb894096c125bc40c"} Nov 24 15:15:04 crc kubenswrapper[4790]: I1124 15:15:04.980436 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2278f428c9ef9387006cf4a22f8ac1df15af569178177a0fb894096c125bc40c" Nov 24 15:15:04 crc kubenswrapper[4790]: I1124 15:15:04.980497 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399955-nvl5w" Nov 24 15:15:06 crc kubenswrapper[4790]: I1124 15:15:06.332294 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39c253b2-56c2-46c1-bafd-587d40437a2d" path="/var/lib/kubelet/pods/39c253b2-56c2-46c1-bafd-587d40437a2d/volumes" Nov 24 15:15:11 crc kubenswrapper[4790]: I1124 15:15:11.044946 4790 generic.go:334] "Generic (PLEG): container finished" podID="c67690a9-9819-45bb-9d46-24ebc69923b9" containerID="780c3879c6a76743ace62386cfc8a9f3f4d0f8dfb10d031d70b4b67cd0018879" exitCode=0 Nov 24 15:15:11 crc kubenswrapper[4790]: I1124 15:15:11.045044 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gsmqf" event={"ID":"c67690a9-9819-45bb-9d46-24ebc69923b9","Type":"ContainerDied","Data":"780c3879c6a76743ace62386cfc8a9f3f4d0f8dfb10d031d70b4b67cd0018879"} Nov 24 15:15:12 crc kubenswrapper[4790]: I1124 15:15:12.059234 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gsmqf" event={"ID":"c67690a9-9819-45bb-9d46-24ebc69923b9","Type":"ContainerStarted","Data":"3d1b9e88f278c45442ae4f8ebecb72fd96a721ca6c3dbdf42bfa6413ea6a2df8"} Nov 24 15:15:12 crc kubenswrapper[4790]: I1124 15:15:12.089952 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gsmqf" podStartSLOduration=2.455867318 podStartE2EDuration="9.089923571s" podCreationTimestamp="2025-11-24 15:15:03 +0000 UTC" firstStartedPulling="2025-11-24 15:15:04.972148049 +0000 UTC m=+7353.352041731" lastFinishedPulling="2025-11-24 15:15:11.606204322 +0000 UTC m=+7359.986097984" observedRunningTime="2025-11-24 15:15:12.077915172 +0000 UTC m=+7360.457808864" watchObservedRunningTime="2025-11-24 15:15:12.089923571 +0000 UTC m=+7360.469817253" Nov 24 15:15:13 crc kubenswrapper[4790]: I1124 15:15:13.581610 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gsmqf" Nov 24 15:15:13 crc kubenswrapper[4790]: I1124 15:15:13.582136 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gsmqf" Nov 24 15:15:13 crc kubenswrapper[4790]: I1124 15:15:13.626610 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gsmqf" Nov 24 15:15:23 crc kubenswrapper[4790]: I1124 15:15:23.640511 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gsmqf" Nov 24 15:15:23 crc kubenswrapper[4790]: I1124 15:15:23.723367 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gsmqf"] Nov 24 15:15:23 crc kubenswrapper[4790]: I1124 15:15:23.771962 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pn2zj"] Nov 24 15:15:23 crc kubenswrapper[4790]: I1124 15:15:23.772605 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pn2zj" podUID="730d2a27-6ef4-4a97-8e6b-5272e26aa247" containerName="registry-server" containerID="cri-o://d39e79890018b3cf73d68d6bed504a60b38b6539c4d31478378a16e563642b19" gracePeriod=2 Nov 24 15:15:24 crc kubenswrapper[4790]: I1124 15:15:24.211493 4790 generic.go:334] "Generic (PLEG): container finished" podID="e3c33642-ddec-46a5-b86d-220f086ed2cf" containerID="8cb6058225ebb65edd2d1200d22aa629aa5d3fc915bac75c11db4a7d9c3b47e4" exitCode=0 Nov 24 15:15:24 crc kubenswrapper[4790]: I1124 15:15:24.211547 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-d8www" event={"ID":"e3c33642-ddec-46a5-b86d-220f086ed2cf","Type":"ContainerDied","Data":"8cb6058225ebb65edd2d1200d22aa629aa5d3fc915bac75c11db4a7d9c3b47e4"} Nov 24 15:15:24 crc kubenswrapper[4790]: I1124 15:15:24.214977 4790 generic.go:334] "Generic (PLEG): container finished" podID="730d2a27-6ef4-4a97-8e6b-5272e26aa247" containerID="d39e79890018b3cf73d68d6bed504a60b38b6539c4d31478378a16e563642b19" exitCode=0 Nov 24 15:15:24 crc kubenswrapper[4790]: I1124 15:15:24.215989 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pn2zj" event={"ID":"730d2a27-6ef4-4a97-8e6b-5272e26aa247","Type":"ContainerDied","Data":"d39e79890018b3cf73d68d6bed504a60b38b6539c4d31478378a16e563642b19"} Nov 24 15:15:24 crc kubenswrapper[4790]: I1124 15:15:24.327594 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pn2zj" Nov 24 15:15:24 crc kubenswrapper[4790]: I1124 15:15:24.448616 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/730d2a27-6ef4-4a97-8e6b-5272e26aa247-utilities\") pod \"730d2a27-6ef4-4a97-8e6b-5272e26aa247\" (UID: \"730d2a27-6ef4-4a97-8e6b-5272e26aa247\") " Nov 24 15:15:24 crc kubenswrapper[4790]: I1124 15:15:24.449374 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/730d2a27-6ef4-4a97-8e6b-5272e26aa247-utilities" (OuterVolumeSpecName: "utilities") pod "730d2a27-6ef4-4a97-8e6b-5272e26aa247" (UID: "730d2a27-6ef4-4a97-8e6b-5272e26aa247"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:15:24 crc kubenswrapper[4790]: I1124 15:15:24.449406 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnnxf\" (UniqueName: \"kubernetes.io/projected/730d2a27-6ef4-4a97-8e6b-5272e26aa247-kube-api-access-mnnxf\") pod \"730d2a27-6ef4-4a97-8e6b-5272e26aa247\" (UID: \"730d2a27-6ef4-4a97-8e6b-5272e26aa247\") " Nov 24 15:15:24 crc kubenswrapper[4790]: I1124 15:15:24.449567 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/730d2a27-6ef4-4a97-8e6b-5272e26aa247-catalog-content\") pod \"730d2a27-6ef4-4a97-8e6b-5272e26aa247\" (UID: \"730d2a27-6ef4-4a97-8e6b-5272e26aa247\") " Nov 24 15:15:24 crc kubenswrapper[4790]: I1124 15:15:24.453869 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/730d2a27-6ef4-4a97-8e6b-5272e26aa247-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 15:15:24 crc kubenswrapper[4790]: I1124 15:15:24.462223 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/730d2a27-6ef4-4a97-8e6b-5272e26aa247-kube-api-access-mnnxf" (OuterVolumeSpecName: "kube-api-access-mnnxf") pod "730d2a27-6ef4-4a97-8e6b-5272e26aa247" (UID: "730d2a27-6ef4-4a97-8e6b-5272e26aa247"). InnerVolumeSpecName "kube-api-access-mnnxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:15:24 crc kubenswrapper[4790]: I1124 15:15:24.503332 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/730d2a27-6ef4-4a97-8e6b-5272e26aa247-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "730d2a27-6ef4-4a97-8e6b-5272e26aa247" (UID: "730d2a27-6ef4-4a97-8e6b-5272e26aa247"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:15:24 crc kubenswrapper[4790]: I1124 15:15:24.555688 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnnxf\" (UniqueName: \"kubernetes.io/projected/730d2a27-6ef4-4a97-8e6b-5272e26aa247-kube-api-access-mnnxf\") on node \"crc\" DevicePath \"\"" Nov 24 15:15:24 crc kubenswrapper[4790]: I1124 15:15:24.555721 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/730d2a27-6ef4-4a97-8e6b-5272e26aa247-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 15:15:25 crc kubenswrapper[4790]: I1124 15:15:25.229910 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pn2zj" Nov 24 15:15:25 crc kubenswrapper[4790]: I1124 15:15:25.229904 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pn2zj" event={"ID":"730d2a27-6ef4-4a97-8e6b-5272e26aa247","Type":"ContainerDied","Data":"93faa936195d38d35036222f7f9c1107654d55dc421d50c7af36ff41d35fea26"} Nov 24 15:15:25 crc kubenswrapper[4790]: I1124 15:15:25.232941 4790 scope.go:117] "RemoveContainer" containerID="d39e79890018b3cf73d68d6bed504a60b38b6539c4d31478378a16e563642b19" Nov 24 15:15:25 crc kubenswrapper[4790]: I1124 15:15:25.294087 4790 scope.go:117] "RemoveContainer" containerID="67b32023432d3f56c81185e30864fa6f05fe6e07d006872288ce88ee7913fb20" Nov 24 15:15:25 crc kubenswrapper[4790]: I1124 15:15:25.295818 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pn2zj"] Nov 24 15:15:25 crc kubenswrapper[4790]: I1124 15:15:25.308157 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pn2zj"] Nov 24 15:15:25 crc kubenswrapper[4790]: I1124 15:15:25.335152 4790 scope.go:117] "RemoveContainer" containerID="4acc542cd92b0a895489568ad066e7ad98b6159a6423d2738df93d81adce0f80" Nov 24 15:15:25 crc kubenswrapper[4790]: I1124 15:15:25.808244 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-d8www" Nov 24 15:15:25 crc kubenswrapper[4790]: I1124 15:15:25.997085 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9qv4\" (UniqueName: \"kubernetes.io/projected/e3c33642-ddec-46a5-b86d-220f086ed2cf-kube-api-access-v9qv4\") pod \"e3c33642-ddec-46a5-b86d-220f086ed2cf\" (UID: \"e3c33642-ddec-46a5-b86d-220f086ed2cf\") " Nov 24 15:15:25 crc kubenswrapper[4790]: I1124 15:15:25.997154 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e3c33642-ddec-46a5-b86d-220f086ed2cf-ceph\") pod \"e3c33642-ddec-46a5-b86d-220f086ed2cf\" (UID: \"e3c33642-ddec-46a5-b86d-220f086ed2cf\") " Nov 24 15:15:25 crc kubenswrapper[4790]: I1124 15:15:25.997249 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3c33642-ddec-46a5-b86d-220f086ed2cf-inventory\") pod \"e3c33642-ddec-46a5-b86d-220f086ed2cf\" (UID: \"e3c33642-ddec-46a5-b86d-220f086ed2cf\") " Nov 24 15:15:25 crc kubenswrapper[4790]: I1124 15:15:25.997375 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3c33642-ddec-46a5-b86d-220f086ed2cf-ssh-key\") pod \"e3c33642-ddec-46a5-b86d-220f086ed2cf\" (UID: \"e3c33642-ddec-46a5-b86d-220f086ed2cf\") " Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.004670 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3c33642-ddec-46a5-b86d-220f086ed2cf-kube-api-access-v9qv4" (OuterVolumeSpecName: "kube-api-access-v9qv4") pod "e3c33642-ddec-46a5-b86d-220f086ed2cf" (UID: "e3c33642-ddec-46a5-b86d-220f086ed2cf"). InnerVolumeSpecName "kube-api-access-v9qv4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.006013 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3c33642-ddec-46a5-b86d-220f086ed2cf-ceph" (OuterVolumeSpecName: "ceph") pod "e3c33642-ddec-46a5-b86d-220f086ed2cf" (UID: "e3c33642-ddec-46a5-b86d-220f086ed2cf"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.038635 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3c33642-ddec-46a5-b86d-220f086ed2cf-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e3c33642-ddec-46a5-b86d-220f086ed2cf" (UID: "e3c33642-ddec-46a5-b86d-220f086ed2cf"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.044556 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3c33642-ddec-46a5-b86d-220f086ed2cf-inventory" (OuterVolumeSpecName: "inventory") pod "e3c33642-ddec-46a5-b86d-220f086ed2cf" (UID: "e3c33642-ddec-46a5-b86d-220f086ed2cf"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.100585 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9qv4\" (UniqueName: \"kubernetes.io/projected/e3c33642-ddec-46a5-b86d-220f086ed2cf-kube-api-access-v9qv4\") on node \"crc\" DevicePath \"\"" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.100644 4790 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e3c33642-ddec-46a5-b86d-220f086ed2cf-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.100655 4790 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3c33642-ddec-46a5-b86d-220f086ed2cf-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.100666 4790 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3c33642-ddec-46a5-b86d-220f086ed2cf-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.267611 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-d8www" event={"ID":"e3c33642-ddec-46a5-b86d-220f086ed2cf","Type":"ContainerDied","Data":"414801491e66d0e1da71c5262c63c1d8632f0bb0d2b71975423c736522cdc3b8"} Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.268061 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="414801491e66d0e1da71c5262c63c1d8632f0bb0d2b71975423c736522cdc3b8" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.268165 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-d8www" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.355903 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="730d2a27-6ef4-4a97-8e6b-5272e26aa247" path="/var/lib/kubelet/pods/730d2a27-6ef4-4a97-8e6b-5272e26aa247/volumes" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.359920 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-cfxxv"] Nov 24 15:15:26 crc kubenswrapper[4790]: E1124 15:15:26.360409 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3c33642-ddec-46a5-b86d-220f086ed2cf" containerName="download-cache-openstack-openstack-cell1" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.360448 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3c33642-ddec-46a5-b86d-220f086ed2cf" containerName="download-cache-openstack-openstack-cell1" Nov 24 15:15:26 crc kubenswrapper[4790]: E1124 15:15:26.360482 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="730d2a27-6ef4-4a97-8e6b-5272e26aa247" containerName="extract-utilities" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.360488 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="730d2a27-6ef4-4a97-8e6b-5272e26aa247" containerName="extract-utilities" Nov 24 15:15:26 crc kubenswrapper[4790]: E1124 15:15:26.360502 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="730d2a27-6ef4-4a97-8e6b-5272e26aa247" containerName="registry-server" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.360528 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="730d2a27-6ef4-4a97-8e6b-5272e26aa247" containerName="registry-server" Nov 24 15:15:26 crc kubenswrapper[4790]: E1124 15:15:26.360553 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="833d9b1e-3d1a-4269-8dbb-26091e6479be" containerName="collect-profiles" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.360559 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="833d9b1e-3d1a-4269-8dbb-26091e6479be" containerName="collect-profiles" Nov 24 15:15:26 crc kubenswrapper[4790]: E1124 15:15:26.360573 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="730d2a27-6ef4-4a97-8e6b-5272e26aa247" containerName="extract-content" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.360578 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="730d2a27-6ef4-4a97-8e6b-5272e26aa247" containerName="extract-content" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.360852 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="730d2a27-6ef4-4a97-8e6b-5272e26aa247" containerName="registry-server" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.360905 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3c33642-ddec-46a5-b86d-220f086ed2cf" containerName="download-cache-openstack-openstack-cell1" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.360924 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="833d9b1e-3d1a-4269-8dbb-26091e6479be" containerName="collect-profiles" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.361794 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-cfxxv" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.365523 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fc497" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.365734 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.367521 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.367804 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.378137 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-cfxxv"] Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.508825 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/81c6adc7-3f2c-471d-a3ea-f2a947deb94a-ssh-key\") pod \"configure-network-openstack-openstack-cell1-cfxxv\" (UID: \"81c6adc7-3f2c-471d-a3ea-f2a947deb94a\") " pod="openstack/configure-network-openstack-openstack-cell1-cfxxv" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.508953 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/81c6adc7-3f2c-471d-a3ea-f2a947deb94a-ceph\") pod \"configure-network-openstack-openstack-cell1-cfxxv\" (UID: \"81c6adc7-3f2c-471d-a3ea-f2a947deb94a\") " pod="openstack/configure-network-openstack-openstack-cell1-cfxxv" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.508984 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/81c6adc7-3f2c-471d-a3ea-f2a947deb94a-inventory\") pod \"configure-network-openstack-openstack-cell1-cfxxv\" (UID: \"81c6adc7-3f2c-471d-a3ea-f2a947deb94a\") " pod="openstack/configure-network-openstack-openstack-cell1-cfxxv" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.509110 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7nqj\" (UniqueName: \"kubernetes.io/projected/81c6adc7-3f2c-471d-a3ea-f2a947deb94a-kube-api-access-s7nqj\") pod \"configure-network-openstack-openstack-cell1-cfxxv\" (UID: \"81c6adc7-3f2c-471d-a3ea-f2a947deb94a\") " pod="openstack/configure-network-openstack-openstack-cell1-cfxxv" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.611543 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/81c6adc7-3f2c-471d-a3ea-f2a947deb94a-ssh-key\") pod \"configure-network-openstack-openstack-cell1-cfxxv\" (UID: \"81c6adc7-3f2c-471d-a3ea-f2a947deb94a\") " pod="openstack/configure-network-openstack-openstack-cell1-cfxxv" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.611598 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/81c6adc7-3f2c-471d-a3ea-f2a947deb94a-ceph\") pod \"configure-network-openstack-openstack-cell1-cfxxv\" (UID: \"81c6adc7-3f2c-471d-a3ea-f2a947deb94a\") " pod="openstack/configure-network-openstack-openstack-cell1-cfxxv" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.611666 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/81c6adc7-3f2c-471d-a3ea-f2a947deb94a-inventory\") pod \"configure-network-openstack-openstack-cell1-cfxxv\" (UID: \"81c6adc7-3f2c-471d-a3ea-f2a947deb94a\") " pod="openstack/configure-network-openstack-openstack-cell1-cfxxv" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.612493 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7nqj\" (UniqueName: \"kubernetes.io/projected/81c6adc7-3f2c-471d-a3ea-f2a947deb94a-kube-api-access-s7nqj\") pod \"configure-network-openstack-openstack-cell1-cfxxv\" (UID: \"81c6adc7-3f2c-471d-a3ea-f2a947deb94a\") " pod="openstack/configure-network-openstack-openstack-cell1-cfxxv" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.616729 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/81c6adc7-3f2c-471d-a3ea-f2a947deb94a-ssh-key\") pod \"configure-network-openstack-openstack-cell1-cfxxv\" (UID: \"81c6adc7-3f2c-471d-a3ea-f2a947deb94a\") " pod="openstack/configure-network-openstack-openstack-cell1-cfxxv" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.617183 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/81c6adc7-3f2c-471d-a3ea-f2a947deb94a-inventory\") pod \"configure-network-openstack-openstack-cell1-cfxxv\" (UID: \"81c6adc7-3f2c-471d-a3ea-f2a947deb94a\") " pod="openstack/configure-network-openstack-openstack-cell1-cfxxv" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.617487 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/81c6adc7-3f2c-471d-a3ea-f2a947deb94a-ceph\") pod \"configure-network-openstack-openstack-cell1-cfxxv\" (UID: \"81c6adc7-3f2c-471d-a3ea-f2a947deb94a\") " pod="openstack/configure-network-openstack-openstack-cell1-cfxxv" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.629661 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7nqj\" (UniqueName: \"kubernetes.io/projected/81c6adc7-3f2c-471d-a3ea-f2a947deb94a-kube-api-access-s7nqj\") pod \"configure-network-openstack-openstack-cell1-cfxxv\" (UID: \"81c6adc7-3f2c-471d-a3ea-f2a947deb94a\") " pod="openstack/configure-network-openstack-openstack-cell1-cfxxv" Nov 24 15:15:26 crc kubenswrapper[4790]: I1124 15:15:26.706513 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-cfxxv" Nov 24 15:15:27 crc kubenswrapper[4790]: I1124 15:15:27.276248 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-cfxxv"] Nov 24 15:15:27 crc kubenswrapper[4790]: W1124 15:15:27.277466 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81c6adc7_3f2c_471d_a3ea_f2a947deb94a.slice/crio-4d1273db6979d723df2d7e8def82e35eb8e045a749bca8ac625288c78385509e WatchSource:0}: Error finding container 4d1273db6979d723df2d7e8def82e35eb8e045a749bca8ac625288c78385509e: Status 404 returned error can't find the container with id 4d1273db6979d723df2d7e8def82e35eb8e045a749bca8ac625288c78385509e Nov 24 15:15:28 crc kubenswrapper[4790]: I1124 15:15:28.294229 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-cfxxv" event={"ID":"81c6adc7-3f2c-471d-a3ea-f2a947deb94a","Type":"ContainerStarted","Data":"d620d6d1e16c9e8f54642b212260f97496e2ad5169f3a8b819798348370be481"} Nov 24 15:15:28 crc kubenswrapper[4790]: I1124 15:15:28.294561 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-cfxxv" event={"ID":"81c6adc7-3f2c-471d-a3ea-f2a947deb94a","Type":"ContainerStarted","Data":"4d1273db6979d723df2d7e8def82e35eb8e045a749bca8ac625288c78385509e"} Nov 24 15:15:28 crc kubenswrapper[4790]: I1124 15:15:28.323067 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-openstack-openstack-cell1-cfxxv" podStartSLOduration=2.113308525 podStartE2EDuration="2.323037196s" podCreationTimestamp="2025-11-24 15:15:26 +0000 UTC" firstStartedPulling="2025-11-24 15:15:27.280032749 +0000 UTC m=+7375.659926411" lastFinishedPulling="2025-11-24 15:15:27.48976142 +0000 UTC m=+7375.869655082" observedRunningTime="2025-11-24 15:15:28.311109309 +0000 UTC m=+7376.691002991" watchObservedRunningTime="2025-11-24 15:15:28.323037196 +0000 UTC m=+7376.702930858" Nov 24 15:15:43 crc kubenswrapper[4790]: I1124 15:15:43.938358 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:15:43 crc kubenswrapper[4790]: I1124 15:15:43.940648 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:16:03 crc kubenswrapper[4790]: I1124 15:16:03.283376 4790 scope.go:117] "RemoveContainer" containerID="6f24b1d494b226385df8a3047b512d4f54717c8d0cd839083dcb2d3da90e23eb" Nov 24 15:16:13 crc kubenswrapper[4790]: I1124 15:16:13.939348 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:16:13 crc kubenswrapper[4790]: I1124 15:16:13.939929 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:16:25 crc kubenswrapper[4790]: I1124 15:16:25.934059 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xzl7r"] Nov 24 15:16:25 crc kubenswrapper[4790]: I1124 15:16:25.936728 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xzl7r" Nov 24 15:16:25 crc kubenswrapper[4790]: I1124 15:16:25.950805 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xzl7r"] Nov 24 15:16:26 crc kubenswrapper[4790]: I1124 15:16:26.024730 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e88aa8a6-e145-452d-9362-3ee38c5f2839-utilities\") pod \"redhat-operators-xzl7r\" (UID: \"e88aa8a6-e145-452d-9362-3ee38c5f2839\") " pod="openshift-marketplace/redhat-operators-xzl7r" Nov 24 15:16:26 crc kubenswrapper[4790]: I1124 15:16:26.024968 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e88aa8a6-e145-452d-9362-3ee38c5f2839-catalog-content\") pod \"redhat-operators-xzl7r\" (UID: \"e88aa8a6-e145-452d-9362-3ee38c5f2839\") " pod="openshift-marketplace/redhat-operators-xzl7r" Nov 24 15:16:26 crc kubenswrapper[4790]: I1124 15:16:26.025967 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqdpt\" (UniqueName: \"kubernetes.io/projected/e88aa8a6-e145-452d-9362-3ee38c5f2839-kube-api-access-lqdpt\") pod \"redhat-operators-xzl7r\" (UID: \"e88aa8a6-e145-452d-9362-3ee38c5f2839\") " pod="openshift-marketplace/redhat-operators-xzl7r" Nov 24 15:16:26 crc kubenswrapper[4790]: I1124 15:16:26.128574 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e88aa8a6-e145-452d-9362-3ee38c5f2839-utilities\") pod \"redhat-operators-xzl7r\" (UID: \"e88aa8a6-e145-452d-9362-3ee38c5f2839\") " pod="openshift-marketplace/redhat-operators-xzl7r" Nov 24 15:16:26 crc kubenswrapper[4790]: I1124 15:16:26.128735 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e88aa8a6-e145-452d-9362-3ee38c5f2839-catalog-content\") pod \"redhat-operators-xzl7r\" (UID: \"e88aa8a6-e145-452d-9362-3ee38c5f2839\") " pod="openshift-marketplace/redhat-operators-xzl7r" Nov 24 15:16:26 crc kubenswrapper[4790]: I1124 15:16:26.128821 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqdpt\" (UniqueName: \"kubernetes.io/projected/e88aa8a6-e145-452d-9362-3ee38c5f2839-kube-api-access-lqdpt\") pod \"redhat-operators-xzl7r\" (UID: \"e88aa8a6-e145-452d-9362-3ee38c5f2839\") " pod="openshift-marketplace/redhat-operators-xzl7r" Nov 24 15:16:26 crc kubenswrapper[4790]: I1124 15:16:26.129122 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e88aa8a6-e145-452d-9362-3ee38c5f2839-utilities\") pod \"redhat-operators-xzl7r\" (UID: \"e88aa8a6-e145-452d-9362-3ee38c5f2839\") " pod="openshift-marketplace/redhat-operators-xzl7r" Nov 24 15:16:26 crc kubenswrapper[4790]: I1124 15:16:26.129547 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e88aa8a6-e145-452d-9362-3ee38c5f2839-catalog-content\") pod \"redhat-operators-xzl7r\" (UID: \"e88aa8a6-e145-452d-9362-3ee38c5f2839\") " pod="openshift-marketplace/redhat-operators-xzl7r" Nov 24 15:16:26 crc kubenswrapper[4790]: I1124 15:16:26.151598 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqdpt\" (UniqueName: \"kubernetes.io/projected/e88aa8a6-e145-452d-9362-3ee38c5f2839-kube-api-access-lqdpt\") pod \"redhat-operators-xzl7r\" (UID: \"e88aa8a6-e145-452d-9362-3ee38c5f2839\") " pod="openshift-marketplace/redhat-operators-xzl7r" Nov 24 15:16:26 crc kubenswrapper[4790]: I1124 15:16:26.265785 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xzl7r" Nov 24 15:16:26 crc kubenswrapper[4790]: I1124 15:16:26.728667 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xzl7r"] Nov 24 15:16:26 crc kubenswrapper[4790]: I1124 15:16:26.989400 4790 generic.go:334] "Generic (PLEG): container finished" podID="e88aa8a6-e145-452d-9362-3ee38c5f2839" containerID="0d73df8523115ace5f098a7068ebbb11f872070e09b27b1c8c249d6de7fe92ab" exitCode=0 Nov 24 15:16:26 crc kubenswrapper[4790]: I1124 15:16:26.989501 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xzl7r" event={"ID":"e88aa8a6-e145-452d-9362-3ee38c5f2839","Type":"ContainerDied","Data":"0d73df8523115ace5f098a7068ebbb11f872070e09b27b1c8c249d6de7fe92ab"} Nov 24 15:16:26 crc kubenswrapper[4790]: I1124 15:16:26.989639 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xzl7r" event={"ID":"e88aa8a6-e145-452d-9362-3ee38c5f2839","Type":"ContainerStarted","Data":"2a443e876e640c273dbccbe05b11c4cfd1ae2ee31284f356478d2202c2664027"} Nov 24 15:16:26 crc kubenswrapper[4790]: I1124 15:16:26.991472 4790 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 15:16:29 crc kubenswrapper[4790]: I1124 15:16:29.017138 4790 generic.go:334] "Generic (PLEG): container finished" podID="e88aa8a6-e145-452d-9362-3ee38c5f2839" containerID="855163a89f77bccc7ff43b3e7d99b798b1b0adb090308d7879729b32f5df8c2c" exitCode=0 Nov 24 15:16:29 crc kubenswrapper[4790]: I1124 15:16:29.017223 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xzl7r" event={"ID":"e88aa8a6-e145-452d-9362-3ee38c5f2839","Type":"ContainerDied","Data":"855163a89f77bccc7ff43b3e7d99b798b1b0adb090308d7879729b32f5df8c2c"} Nov 24 15:16:30 crc kubenswrapper[4790]: I1124 15:16:30.057920 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xzl7r" event={"ID":"e88aa8a6-e145-452d-9362-3ee38c5f2839","Type":"ContainerStarted","Data":"f384b1173b88b1e67030e8ab81e89361bad446ae543f51078c79f8b899172f09"} Nov 24 15:16:30 crc kubenswrapper[4790]: I1124 15:16:30.086297 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xzl7r" podStartSLOduration=2.624578368 podStartE2EDuration="5.086274458s" podCreationTimestamp="2025-11-24 15:16:25 +0000 UTC" firstStartedPulling="2025-11-24 15:16:26.991220565 +0000 UTC m=+7435.371114227" lastFinishedPulling="2025-11-24 15:16:29.452916615 +0000 UTC m=+7437.832810317" observedRunningTime="2025-11-24 15:16:30.078545479 +0000 UTC m=+7438.458439161" watchObservedRunningTime="2025-11-24 15:16:30.086274458 +0000 UTC m=+7438.466168120" Nov 24 15:16:36 crc kubenswrapper[4790]: I1124 15:16:36.266794 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xzl7r" Nov 24 15:16:36 crc kubenswrapper[4790]: I1124 15:16:36.267514 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xzl7r" Nov 24 15:16:36 crc kubenswrapper[4790]: I1124 15:16:36.336385 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xzl7r" Nov 24 15:16:37 crc kubenswrapper[4790]: I1124 15:16:37.227504 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xzl7r" Nov 24 15:16:37 crc kubenswrapper[4790]: I1124 15:16:37.284302 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xzl7r"] Nov 24 15:16:38 crc kubenswrapper[4790]: I1124 15:16:38.994653 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8rnj4"] Nov 24 15:16:39 crc kubenswrapper[4790]: I1124 15:16:39.005018 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8rnj4" Nov 24 15:16:39 crc kubenswrapper[4790]: I1124 15:16:39.020810 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8rnj4"] Nov 24 15:16:39 crc kubenswrapper[4790]: I1124 15:16:39.033136 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81fb48ee-289e-4442-81e5-ecba753eaecc-utilities\") pod \"redhat-marketplace-8rnj4\" (UID: \"81fb48ee-289e-4442-81e5-ecba753eaecc\") " pod="openshift-marketplace/redhat-marketplace-8rnj4" Nov 24 15:16:39 crc kubenswrapper[4790]: I1124 15:16:39.033209 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nh9c\" (UniqueName: \"kubernetes.io/projected/81fb48ee-289e-4442-81e5-ecba753eaecc-kube-api-access-2nh9c\") pod \"redhat-marketplace-8rnj4\" (UID: \"81fb48ee-289e-4442-81e5-ecba753eaecc\") " pod="openshift-marketplace/redhat-marketplace-8rnj4" Nov 24 15:16:39 crc kubenswrapper[4790]: I1124 15:16:39.033246 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81fb48ee-289e-4442-81e5-ecba753eaecc-catalog-content\") pod \"redhat-marketplace-8rnj4\" (UID: \"81fb48ee-289e-4442-81e5-ecba753eaecc\") " pod="openshift-marketplace/redhat-marketplace-8rnj4" Nov 24 15:16:39 crc kubenswrapper[4790]: I1124 15:16:39.135661 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81fb48ee-289e-4442-81e5-ecba753eaecc-utilities\") pod \"redhat-marketplace-8rnj4\" (UID: \"81fb48ee-289e-4442-81e5-ecba753eaecc\") " pod="openshift-marketplace/redhat-marketplace-8rnj4" Nov 24 15:16:39 crc kubenswrapper[4790]: I1124 15:16:39.136071 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nh9c\" (UniqueName: \"kubernetes.io/projected/81fb48ee-289e-4442-81e5-ecba753eaecc-kube-api-access-2nh9c\") pod \"redhat-marketplace-8rnj4\" (UID: \"81fb48ee-289e-4442-81e5-ecba753eaecc\") " pod="openshift-marketplace/redhat-marketplace-8rnj4" Nov 24 15:16:39 crc kubenswrapper[4790]: I1124 15:16:39.136099 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81fb48ee-289e-4442-81e5-ecba753eaecc-catalog-content\") pod \"redhat-marketplace-8rnj4\" (UID: \"81fb48ee-289e-4442-81e5-ecba753eaecc\") " pod="openshift-marketplace/redhat-marketplace-8rnj4" Nov 24 15:16:39 crc kubenswrapper[4790]: I1124 15:16:39.136332 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81fb48ee-289e-4442-81e5-ecba753eaecc-utilities\") pod \"redhat-marketplace-8rnj4\" (UID: \"81fb48ee-289e-4442-81e5-ecba753eaecc\") " pod="openshift-marketplace/redhat-marketplace-8rnj4" Nov 24 15:16:39 crc kubenswrapper[4790]: I1124 15:16:39.136594 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81fb48ee-289e-4442-81e5-ecba753eaecc-catalog-content\") pod \"redhat-marketplace-8rnj4\" (UID: \"81fb48ee-289e-4442-81e5-ecba753eaecc\") " pod="openshift-marketplace/redhat-marketplace-8rnj4" Nov 24 15:16:39 crc kubenswrapper[4790]: I1124 15:16:39.156740 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nh9c\" (UniqueName: \"kubernetes.io/projected/81fb48ee-289e-4442-81e5-ecba753eaecc-kube-api-access-2nh9c\") pod \"redhat-marketplace-8rnj4\" (UID: \"81fb48ee-289e-4442-81e5-ecba753eaecc\") " pod="openshift-marketplace/redhat-marketplace-8rnj4" Nov 24 15:16:39 crc kubenswrapper[4790]: I1124 15:16:39.164731 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xzl7r" podUID="e88aa8a6-e145-452d-9362-3ee38c5f2839" containerName="registry-server" containerID="cri-o://f384b1173b88b1e67030e8ab81e89361bad446ae543f51078c79f8b899172f09" gracePeriod=2 Nov 24 15:16:39 crc kubenswrapper[4790]: I1124 15:16:39.338973 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8rnj4" Nov 24 15:16:39 crc kubenswrapper[4790]: I1124 15:16:39.656003 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xzl7r" Nov 24 15:16:39 crc kubenswrapper[4790]: I1124 15:16:39.754447 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e88aa8a6-e145-452d-9362-3ee38c5f2839-utilities\") pod \"e88aa8a6-e145-452d-9362-3ee38c5f2839\" (UID: \"e88aa8a6-e145-452d-9362-3ee38c5f2839\") " Nov 24 15:16:39 crc kubenswrapper[4790]: I1124 15:16:39.754536 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqdpt\" (UniqueName: \"kubernetes.io/projected/e88aa8a6-e145-452d-9362-3ee38c5f2839-kube-api-access-lqdpt\") pod \"e88aa8a6-e145-452d-9362-3ee38c5f2839\" (UID: \"e88aa8a6-e145-452d-9362-3ee38c5f2839\") " Nov 24 15:16:39 crc kubenswrapper[4790]: I1124 15:16:39.754613 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e88aa8a6-e145-452d-9362-3ee38c5f2839-catalog-content\") pod \"e88aa8a6-e145-452d-9362-3ee38c5f2839\" (UID: \"e88aa8a6-e145-452d-9362-3ee38c5f2839\") " Nov 24 15:16:39 crc kubenswrapper[4790]: I1124 15:16:39.756123 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e88aa8a6-e145-452d-9362-3ee38c5f2839-utilities" (OuterVolumeSpecName: "utilities") pod "e88aa8a6-e145-452d-9362-3ee38c5f2839" (UID: "e88aa8a6-e145-452d-9362-3ee38c5f2839"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:16:39 crc kubenswrapper[4790]: I1124 15:16:39.761686 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e88aa8a6-e145-452d-9362-3ee38c5f2839-kube-api-access-lqdpt" (OuterVolumeSpecName: "kube-api-access-lqdpt") pod "e88aa8a6-e145-452d-9362-3ee38c5f2839" (UID: "e88aa8a6-e145-452d-9362-3ee38c5f2839"). InnerVolumeSpecName "kube-api-access-lqdpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:16:39 crc kubenswrapper[4790]: I1124 15:16:39.818376 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8rnj4"] Nov 24 15:16:39 crc kubenswrapper[4790]: I1124 15:16:39.839632 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e88aa8a6-e145-452d-9362-3ee38c5f2839-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e88aa8a6-e145-452d-9362-3ee38c5f2839" (UID: "e88aa8a6-e145-452d-9362-3ee38c5f2839"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:16:39 crc kubenswrapper[4790]: I1124 15:16:39.857440 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqdpt\" (UniqueName: \"kubernetes.io/projected/e88aa8a6-e145-452d-9362-3ee38c5f2839-kube-api-access-lqdpt\") on node \"crc\" DevicePath \"\"" Nov 24 15:16:39 crc kubenswrapper[4790]: I1124 15:16:39.857486 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e88aa8a6-e145-452d-9362-3ee38c5f2839-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 15:16:39 crc kubenswrapper[4790]: I1124 15:16:39.857498 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e88aa8a6-e145-452d-9362-3ee38c5f2839-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 15:16:40 crc kubenswrapper[4790]: I1124 15:16:40.183311 4790 generic.go:334] "Generic (PLEG): container finished" podID="e88aa8a6-e145-452d-9362-3ee38c5f2839" containerID="f384b1173b88b1e67030e8ab81e89361bad446ae543f51078c79f8b899172f09" exitCode=0 Nov 24 15:16:40 crc kubenswrapper[4790]: I1124 15:16:40.183374 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xzl7r" event={"ID":"e88aa8a6-e145-452d-9362-3ee38c5f2839","Type":"ContainerDied","Data":"f384b1173b88b1e67030e8ab81e89361bad446ae543f51078c79f8b899172f09"} Nov 24 15:16:40 crc kubenswrapper[4790]: I1124 15:16:40.184330 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xzl7r" event={"ID":"e88aa8a6-e145-452d-9362-3ee38c5f2839","Type":"ContainerDied","Data":"2a443e876e640c273dbccbe05b11c4cfd1ae2ee31284f356478d2202c2664027"} Nov 24 15:16:40 crc kubenswrapper[4790]: I1124 15:16:40.183422 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xzl7r" Nov 24 15:16:40 crc kubenswrapper[4790]: I1124 15:16:40.184362 4790 scope.go:117] "RemoveContainer" containerID="f384b1173b88b1e67030e8ab81e89361bad446ae543f51078c79f8b899172f09" Nov 24 15:16:40 crc kubenswrapper[4790]: I1124 15:16:40.188850 4790 generic.go:334] "Generic (PLEG): container finished" podID="81fb48ee-289e-4442-81e5-ecba753eaecc" containerID="5bb1f0afed91d0ed27b5c0be9d7e98bb556e9c6c27940c49f486bb1bf8951fc0" exitCode=0 Nov 24 15:16:40 crc kubenswrapper[4790]: I1124 15:16:40.188926 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8rnj4" event={"ID":"81fb48ee-289e-4442-81e5-ecba753eaecc","Type":"ContainerDied","Data":"5bb1f0afed91d0ed27b5c0be9d7e98bb556e9c6c27940c49f486bb1bf8951fc0"} Nov 24 15:16:40 crc kubenswrapper[4790]: I1124 15:16:40.188962 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8rnj4" event={"ID":"81fb48ee-289e-4442-81e5-ecba753eaecc","Type":"ContainerStarted","Data":"693cbb58e0f4e64e11588950873e88c59fa49401993b5553f17602a32451afb4"} Nov 24 15:16:40 crc kubenswrapper[4790]: I1124 15:16:40.253502 4790 scope.go:117] "RemoveContainer" containerID="855163a89f77bccc7ff43b3e7d99b798b1b0adb090308d7879729b32f5df8c2c" Nov 24 15:16:40 crc kubenswrapper[4790]: I1124 15:16:40.265830 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xzl7r"] Nov 24 15:16:40 crc kubenswrapper[4790]: I1124 15:16:40.276189 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xzl7r"] Nov 24 15:16:40 crc kubenswrapper[4790]: I1124 15:16:40.286852 4790 scope.go:117] "RemoveContainer" containerID="0d73df8523115ace5f098a7068ebbb11f872070e09b27b1c8c249d6de7fe92ab" Nov 24 15:16:40 crc kubenswrapper[4790]: I1124 15:16:40.333058 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e88aa8a6-e145-452d-9362-3ee38c5f2839" path="/var/lib/kubelet/pods/e88aa8a6-e145-452d-9362-3ee38c5f2839/volumes" Nov 24 15:16:40 crc kubenswrapper[4790]: I1124 15:16:40.346068 4790 scope.go:117] "RemoveContainer" containerID="f384b1173b88b1e67030e8ab81e89361bad446ae543f51078c79f8b899172f09" Nov 24 15:16:40 crc kubenswrapper[4790]: E1124 15:16:40.346969 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f384b1173b88b1e67030e8ab81e89361bad446ae543f51078c79f8b899172f09\": container with ID starting with f384b1173b88b1e67030e8ab81e89361bad446ae543f51078c79f8b899172f09 not found: ID does not exist" containerID="f384b1173b88b1e67030e8ab81e89361bad446ae543f51078c79f8b899172f09" Nov 24 15:16:40 crc kubenswrapper[4790]: I1124 15:16:40.347010 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f384b1173b88b1e67030e8ab81e89361bad446ae543f51078c79f8b899172f09"} err="failed to get container status \"f384b1173b88b1e67030e8ab81e89361bad446ae543f51078c79f8b899172f09\": rpc error: code = NotFound desc = could not find container \"f384b1173b88b1e67030e8ab81e89361bad446ae543f51078c79f8b899172f09\": container with ID starting with f384b1173b88b1e67030e8ab81e89361bad446ae543f51078c79f8b899172f09 not found: ID does not exist" Nov 24 15:16:40 crc kubenswrapper[4790]: I1124 15:16:40.347047 4790 scope.go:117] "RemoveContainer" containerID="855163a89f77bccc7ff43b3e7d99b798b1b0adb090308d7879729b32f5df8c2c" Nov 24 15:16:40 crc kubenswrapper[4790]: E1124 15:16:40.347516 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"855163a89f77bccc7ff43b3e7d99b798b1b0adb090308d7879729b32f5df8c2c\": container with ID starting with 855163a89f77bccc7ff43b3e7d99b798b1b0adb090308d7879729b32f5df8c2c not found: ID does not exist" containerID="855163a89f77bccc7ff43b3e7d99b798b1b0adb090308d7879729b32f5df8c2c" Nov 24 15:16:40 crc kubenswrapper[4790]: I1124 15:16:40.347641 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"855163a89f77bccc7ff43b3e7d99b798b1b0adb090308d7879729b32f5df8c2c"} err="failed to get container status \"855163a89f77bccc7ff43b3e7d99b798b1b0adb090308d7879729b32f5df8c2c\": rpc error: code = NotFound desc = could not find container \"855163a89f77bccc7ff43b3e7d99b798b1b0adb090308d7879729b32f5df8c2c\": container with ID starting with 855163a89f77bccc7ff43b3e7d99b798b1b0adb090308d7879729b32f5df8c2c not found: ID does not exist" Nov 24 15:16:40 crc kubenswrapper[4790]: I1124 15:16:40.347750 4790 scope.go:117] "RemoveContainer" containerID="0d73df8523115ace5f098a7068ebbb11f872070e09b27b1c8c249d6de7fe92ab" Nov 24 15:16:40 crc kubenswrapper[4790]: E1124 15:16:40.348486 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d73df8523115ace5f098a7068ebbb11f872070e09b27b1c8c249d6de7fe92ab\": container with ID starting with 0d73df8523115ace5f098a7068ebbb11f872070e09b27b1c8c249d6de7fe92ab not found: ID does not exist" containerID="0d73df8523115ace5f098a7068ebbb11f872070e09b27b1c8c249d6de7fe92ab" Nov 24 15:16:40 crc kubenswrapper[4790]: I1124 15:16:40.348614 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d73df8523115ace5f098a7068ebbb11f872070e09b27b1c8c249d6de7fe92ab"} err="failed to get container status \"0d73df8523115ace5f098a7068ebbb11f872070e09b27b1c8c249d6de7fe92ab\": rpc error: code = NotFound desc = could not find container \"0d73df8523115ace5f098a7068ebbb11f872070e09b27b1c8c249d6de7fe92ab\": container with ID starting with 0d73df8523115ace5f098a7068ebbb11f872070e09b27b1c8c249d6de7fe92ab not found: ID does not exist" Nov 24 15:16:41 crc kubenswrapper[4790]: I1124 15:16:41.201698 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8rnj4" event={"ID":"81fb48ee-289e-4442-81e5-ecba753eaecc","Type":"ContainerStarted","Data":"11c2e22609e83f6e97712bdef0783f8b1fc18976f3eef11ffab7660623ec9711"} Nov 24 15:16:42 crc kubenswrapper[4790]: I1124 15:16:42.216996 4790 generic.go:334] "Generic (PLEG): container finished" podID="81fb48ee-289e-4442-81e5-ecba753eaecc" containerID="11c2e22609e83f6e97712bdef0783f8b1fc18976f3eef11ffab7660623ec9711" exitCode=0 Nov 24 15:16:42 crc kubenswrapper[4790]: I1124 15:16:42.217075 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8rnj4" event={"ID":"81fb48ee-289e-4442-81e5-ecba753eaecc","Type":"ContainerDied","Data":"11c2e22609e83f6e97712bdef0783f8b1fc18976f3eef11ffab7660623ec9711"} Nov 24 15:16:43 crc kubenswrapper[4790]: I1124 15:16:43.239673 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8rnj4" event={"ID":"81fb48ee-289e-4442-81e5-ecba753eaecc","Type":"ContainerStarted","Data":"feb0fcb30a47340bb1cfc861b4198354fe08afa08ac76395e9929a5cc3deb9af"} Nov 24 15:16:43 crc kubenswrapper[4790]: I1124 15:16:43.277186 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8rnj4" podStartSLOduration=2.824793777 podStartE2EDuration="5.277157203s" podCreationTimestamp="2025-11-24 15:16:38 +0000 UTC" firstStartedPulling="2025-11-24 15:16:40.193922165 +0000 UTC m=+7448.573815867" lastFinishedPulling="2025-11-24 15:16:42.646285621 +0000 UTC m=+7451.026179293" observedRunningTime="2025-11-24 15:16:43.269251339 +0000 UTC m=+7451.649145011" watchObservedRunningTime="2025-11-24 15:16:43.277157203 +0000 UTC m=+7451.657050875" Nov 24 15:16:43 crc kubenswrapper[4790]: I1124 15:16:43.938293 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:16:43 crc kubenswrapper[4790]: I1124 15:16:43.938344 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:16:43 crc kubenswrapper[4790]: I1124 15:16:43.938400 4790 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 15:16:43 crc kubenswrapper[4790]: I1124 15:16:43.939177 4790 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ed032daa21005f5603fc87ffbe2ed53274be2bbefd5cbb4153e64e56dd530b81"} pod="openshift-machine-config-operator/machine-config-daemon-xz49t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 15:16:43 crc kubenswrapper[4790]: I1124 15:16:43.939246 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" containerID="cri-o://ed032daa21005f5603fc87ffbe2ed53274be2bbefd5cbb4153e64e56dd530b81" gracePeriod=600 Nov 24 15:16:44 crc kubenswrapper[4790]: E1124 15:16:44.075227 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:16:44 crc kubenswrapper[4790]: I1124 15:16:44.252833 4790 generic.go:334] "Generic (PLEG): container finished" podID="0d73b133-48f0-455f-8f6a-742e633f631a" containerID="ed032daa21005f5603fc87ffbe2ed53274be2bbefd5cbb4153e64e56dd530b81" exitCode=0 Nov 24 15:16:44 crc kubenswrapper[4790]: I1124 15:16:44.252914 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerDied","Data":"ed032daa21005f5603fc87ffbe2ed53274be2bbefd5cbb4153e64e56dd530b81"} Nov 24 15:16:44 crc kubenswrapper[4790]: I1124 15:16:44.253370 4790 scope.go:117] "RemoveContainer" containerID="f34c5fa54378b1b7f2049ef63829861f39b73c8e72d607198fb71c136fcce70c" Nov 24 15:16:44 crc kubenswrapper[4790]: I1124 15:16:44.253968 4790 scope.go:117] "RemoveContainer" containerID="ed032daa21005f5603fc87ffbe2ed53274be2bbefd5cbb4153e64e56dd530b81" Nov 24 15:16:44 crc kubenswrapper[4790]: E1124 15:16:44.254287 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:16:49 crc kubenswrapper[4790]: I1124 15:16:49.340140 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8rnj4" Nov 24 15:16:49 crc kubenswrapper[4790]: I1124 15:16:49.341449 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8rnj4" Nov 24 15:16:49 crc kubenswrapper[4790]: I1124 15:16:49.424735 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8rnj4" Nov 24 15:16:50 crc kubenswrapper[4790]: I1124 15:16:50.379029 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8rnj4" Nov 24 15:16:50 crc kubenswrapper[4790]: I1124 15:16:50.431936 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8rnj4"] Nov 24 15:16:52 crc kubenswrapper[4790]: I1124 15:16:52.357440 4790 generic.go:334] "Generic (PLEG): container finished" podID="81c6adc7-3f2c-471d-a3ea-f2a947deb94a" containerID="d620d6d1e16c9e8f54642b212260f97496e2ad5169f3a8b819798348370be481" exitCode=0 Nov 24 15:16:52 crc kubenswrapper[4790]: I1124 15:16:52.358278 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8rnj4" podUID="81fb48ee-289e-4442-81e5-ecba753eaecc" containerName="registry-server" containerID="cri-o://feb0fcb30a47340bb1cfc861b4198354fe08afa08ac76395e9929a5cc3deb9af" gracePeriod=2 Nov 24 15:16:52 crc kubenswrapper[4790]: I1124 15:16:52.357579 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-cfxxv" event={"ID":"81c6adc7-3f2c-471d-a3ea-f2a947deb94a","Type":"ContainerDied","Data":"d620d6d1e16c9e8f54642b212260f97496e2ad5169f3a8b819798348370be481"} Nov 24 15:16:52 crc kubenswrapper[4790]: I1124 15:16:52.856014 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8rnj4" Nov 24 15:16:53 crc kubenswrapper[4790]: I1124 15:16:53.050085 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81fb48ee-289e-4442-81e5-ecba753eaecc-catalog-content\") pod \"81fb48ee-289e-4442-81e5-ecba753eaecc\" (UID: \"81fb48ee-289e-4442-81e5-ecba753eaecc\") " Nov 24 15:16:53 crc kubenswrapper[4790]: I1124 15:16:53.050159 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nh9c\" (UniqueName: \"kubernetes.io/projected/81fb48ee-289e-4442-81e5-ecba753eaecc-kube-api-access-2nh9c\") pod \"81fb48ee-289e-4442-81e5-ecba753eaecc\" (UID: \"81fb48ee-289e-4442-81e5-ecba753eaecc\") " Nov 24 15:16:53 crc kubenswrapper[4790]: I1124 15:16:53.050281 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81fb48ee-289e-4442-81e5-ecba753eaecc-utilities\") pod \"81fb48ee-289e-4442-81e5-ecba753eaecc\" (UID: \"81fb48ee-289e-4442-81e5-ecba753eaecc\") " Nov 24 15:16:53 crc kubenswrapper[4790]: I1124 15:16:53.052178 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81fb48ee-289e-4442-81e5-ecba753eaecc-utilities" (OuterVolumeSpecName: "utilities") pod "81fb48ee-289e-4442-81e5-ecba753eaecc" (UID: "81fb48ee-289e-4442-81e5-ecba753eaecc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:16:53 crc kubenswrapper[4790]: I1124 15:16:53.057992 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81fb48ee-289e-4442-81e5-ecba753eaecc-kube-api-access-2nh9c" (OuterVolumeSpecName: "kube-api-access-2nh9c") pod "81fb48ee-289e-4442-81e5-ecba753eaecc" (UID: "81fb48ee-289e-4442-81e5-ecba753eaecc"). InnerVolumeSpecName "kube-api-access-2nh9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:16:53 crc kubenswrapper[4790]: I1124 15:16:53.083777 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81fb48ee-289e-4442-81e5-ecba753eaecc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "81fb48ee-289e-4442-81e5-ecba753eaecc" (UID: "81fb48ee-289e-4442-81e5-ecba753eaecc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:16:53 crc kubenswrapper[4790]: I1124 15:16:53.153939 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81fb48ee-289e-4442-81e5-ecba753eaecc-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 15:16:53 crc kubenswrapper[4790]: I1124 15:16:53.153978 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81fb48ee-289e-4442-81e5-ecba753eaecc-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 15:16:53 crc kubenswrapper[4790]: I1124 15:16:53.153994 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nh9c\" (UniqueName: \"kubernetes.io/projected/81fb48ee-289e-4442-81e5-ecba753eaecc-kube-api-access-2nh9c\") on node \"crc\" DevicePath \"\"" Nov 24 15:16:53 crc kubenswrapper[4790]: I1124 15:16:53.373114 4790 generic.go:334] "Generic (PLEG): container finished" podID="81fb48ee-289e-4442-81e5-ecba753eaecc" containerID="feb0fcb30a47340bb1cfc861b4198354fe08afa08ac76395e9929a5cc3deb9af" exitCode=0 Nov 24 15:16:53 crc kubenswrapper[4790]: I1124 15:16:53.373202 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8rnj4" event={"ID":"81fb48ee-289e-4442-81e5-ecba753eaecc","Type":"ContainerDied","Data":"feb0fcb30a47340bb1cfc861b4198354fe08afa08ac76395e9929a5cc3deb9af"} Nov 24 15:16:53 crc kubenswrapper[4790]: I1124 15:16:53.373234 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8rnj4" Nov 24 15:16:53 crc kubenswrapper[4790]: I1124 15:16:53.373263 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8rnj4" event={"ID":"81fb48ee-289e-4442-81e5-ecba753eaecc","Type":"ContainerDied","Data":"693cbb58e0f4e64e11588950873e88c59fa49401993b5553f17602a32451afb4"} Nov 24 15:16:53 crc kubenswrapper[4790]: I1124 15:16:53.373283 4790 scope.go:117] "RemoveContainer" containerID="feb0fcb30a47340bb1cfc861b4198354fe08afa08ac76395e9929a5cc3deb9af" Nov 24 15:16:53 crc kubenswrapper[4790]: I1124 15:16:53.415953 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8rnj4"] Nov 24 15:16:53 crc kubenswrapper[4790]: I1124 15:16:53.425969 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8rnj4"] Nov 24 15:16:53 crc kubenswrapper[4790]: I1124 15:16:53.426970 4790 scope.go:117] "RemoveContainer" containerID="11c2e22609e83f6e97712bdef0783f8b1fc18976f3eef11ffab7660623ec9711" Nov 24 15:16:53 crc kubenswrapper[4790]: I1124 15:16:53.457973 4790 scope.go:117] "RemoveContainer" containerID="5bb1f0afed91d0ed27b5c0be9d7e98bb556e9c6c27940c49f486bb1bf8951fc0" Nov 24 15:16:53 crc kubenswrapper[4790]: I1124 15:16:53.507739 4790 scope.go:117] "RemoveContainer" containerID="feb0fcb30a47340bb1cfc861b4198354fe08afa08ac76395e9929a5cc3deb9af" Nov 24 15:16:53 crc kubenswrapper[4790]: E1124 15:16:53.508247 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"feb0fcb30a47340bb1cfc861b4198354fe08afa08ac76395e9929a5cc3deb9af\": container with ID starting with feb0fcb30a47340bb1cfc861b4198354fe08afa08ac76395e9929a5cc3deb9af not found: ID does not exist" containerID="feb0fcb30a47340bb1cfc861b4198354fe08afa08ac76395e9929a5cc3deb9af" Nov 24 15:16:53 crc kubenswrapper[4790]: I1124 15:16:53.508303 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"feb0fcb30a47340bb1cfc861b4198354fe08afa08ac76395e9929a5cc3deb9af"} err="failed to get container status \"feb0fcb30a47340bb1cfc861b4198354fe08afa08ac76395e9929a5cc3deb9af\": rpc error: code = NotFound desc = could not find container \"feb0fcb30a47340bb1cfc861b4198354fe08afa08ac76395e9929a5cc3deb9af\": container with ID starting with feb0fcb30a47340bb1cfc861b4198354fe08afa08ac76395e9929a5cc3deb9af not found: ID does not exist" Nov 24 15:16:53 crc kubenswrapper[4790]: I1124 15:16:53.508355 4790 scope.go:117] "RemoveContainer" containerID="11c2e22609e83f6e97712bdef0783f8b1fc18976f3eef11ffab7660623ec9711" Nov 24 15:16:53 crc kubenswrapper[4790]: E1124 15:16:53.508954 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11c2e22609e83f6e97712bdef0783f8b1fc18976f3eef11ffab7660623ec9711\": container with ID starting with 11c2e22609e83f6e97712bdef0783f8b1fc18976f3eef11ffab7660623ec9711 not found: ID does not exist" containerID="11c2e22609e83f6e97712bdef0783f8b1fc18976f3eef11ffab7660623ec9711" Nov 24 15:16:53 crc kubenswrapper[4790]: I1124 15:16:53.509003 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11c2e22609e83f6e97712bdef0783f8b1fc18976f3eef11ffab7660623ec9711"} err="failed to get container status \"11c2e22609e83f6e97712bdef0783f8b1fc18976f3eef11ffab7660623ec9711\": rpc error: code = NotFound desc = could not find container \"11c2e22609e83f6e97712bdef0783f8b1fc18976f3eef11ffab7660623ec9711\": container with ID starting with 11c2e22609e83f6e97712bdef0783f8b1fc18976f3eef11ffab7660623ec9711 not found: ID does not exist" Nov 24 15:16:53 crc kubenswrapper[4790]: I1124 15:16:53.509035 4790 scope.go:117] "RemoveContainer" containerID="5bb1f0afed91d0ed27b5c0be9d7e98bb556e9c6c27940c49f486bb1bf8951fc0" Nov 24 15:16:53 crc kubenswrapper[4790]: E1124 15:16:53.509380 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bb1f0afed91d0ed27b5c0be9d7e98bb556e9c6c27940c49f486bb1bf8951fc0\": container with ID starting with 5bb1f0afed91d0ed27b5c0be9d7e98bb556e9c6c27940c49f486bb1bf8951fc0 not found: ID does not exist" containerID="5bb1f0afed91d0ed27b5c0be9d7e98bb556e9c6c27940c49f486bb1bf8951fc0" Nov 24 15:16:53 crc kubenswrapper[4790]: I1124 15:16:53.509434 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bb1f0afed91d0ed27b5c0be9d7e98bb556e9c6c27940c49f486bb1bf8951fc0"} err="failed to get container status \"5bb1f0afed91d0ed27b5c0be9d7e98bb556e9c6c27940c49f486bb1bf8951fc0\": rpc error: code = NotFound desc = could not find container \"5bb1f0afed91d0ed27b5c0be9d7e98bb556e9c6c27940c49f486bb1bf8951fc0\": container with ID starting with 5bb1f0afed91d0ed27b5c0be9d7e98bb556e9c6c27940c49f486bb1bf8951fc0 not found: ID does not exist" Nov 24 15:16:53 crc kubenswrapper[4790]: I1124 15:16:53.822316 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-cfxxv" Nov 24 15:16:53 crc kubenswrapper[4790]: I1124 15:16:53.978666 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7nqj\" (UniqueName: \"kubernetes.io/projected/81c6adc7-3f2c-471d-a3ea-f2a947deb94a-kube-api-access-s7nqj\") pod \"81c6adc7-3f2c-471d-a3ea-f2a947deb94a\" (UID: \"81c6adc7-3f2c-471d-a3ea-f2a947deb94a\") " Nov 24 15:16:53 crc kubenswrapper[4790]: I1124 15:16:53.979130 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/81c6adc7-3f2c-471d-a3ea-f2a947deb94a-inventory\") pod \"81c6adc7-3f2c-471d-a3ea-f2a947deb94a\" (UID: \"81c6adc7-3f2c-471d-a3ea-f2a947deb94a\") " Nov 24 15:16:53 crc kubenswrapper[4790]: I1124 15:16:53.979211 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/81c6adc7-3f2c-471d-a3ea-f2a947deb94a-ssh-key\") pod \"81c6adc7-3f2c-471d-a3ea-f2a947deb94a\" (UID: \"81c6adc7-3f2c-471d-a3ea-f2a947deb94a\") " Nov 24 15:16:53 crc kubenswrapper[4790]: I1124 15:16:53.979414 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/81c6adc7-3f2c-471d-a3ea-f2a947deb94a-ceph\") pod \"81c6adc7-3f2c-471d-a3ea-f2a947deb94a\" (UID: \"81c6adc7-3f2c-471d-a3ea-f2a947deb94a\") " Nov 24 15:16:53 crc kubenswrapper[4790]: I1124 15:16:53.984361 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81c6adc7-3f2c-471d-a3ea-f2a947deb94a-kube-api-access-s7nqj" (OuterVolumeSpecName: "kube-api-access-s7nqj") pod "81c6adc7-3f2c-471d-a3ea-f2a947deb94a" (UID: "81c6adc7-3f2c-471d-a3ea-f2a947deb94a"). InnerVolumeSpecName "kube-api-access-s7nqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:16:53 crc kubenswrapper[4790]: I1124 15:16:53.989992 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81c6adc7-3f2c-471d-a3ea-f2a947deb94a-ceph" (OuterVolumeSpecName: "ceph") pod "81c6adc7-3f2c-471d-a3ea-f2a947deb94a" (UID: "81c6adc7-3f2c-471d-a3ea-f2a947deb94a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.010341 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81c6adc7-3f2c-471d-a3ea-f2a947deb94a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "81c6adc7-3f2c-471d-a3ea-f2a947deb94a" (UID: "81c6adc7-3f2c-471d-a3ea-f2a947deb94a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.033094 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81c6adc7-3f2c-471d-a3ea-f2a947deb94a-inventory" (OuterVolumeSpecName: "inventory") pod "81c6adc7-3f2c-471d-a3ea-f2a947deb94a" (UID: "81c6adc7-3f2c-471d-a3ea-f2a947deb94a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.083101 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7nqj\" (UniqueName: \"kubernetes.io/projected/81c6adc7-3f2c-471d-a3ea-f2a947deb94a-kube-api-access-s7nqj\") on node \"crc\" DevicePath \"\"" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.083245 4790 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/81c6adc7-3f2c-471d-a3ea-f2a947deb94a-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.083353 4790 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/81c6adc7-3f2c-471d-a3ea-f2a947deb94a-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.083439 4790 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/81c6adc7-3f2c-471d-a3ea-f2a947deb94a-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.333770 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81fb48ee-289e-4442-81e5-ecba753eaecc" path="/var/lib/kubelet/pods/81fb48ee-289e-4442-81e5-ecba753eaecc/volumes" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.392547 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-cfxxv" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.392559 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-cfxxv" event={"ID":"81c6adc7-3f2c-471d-a3ea-f2a947deb94a","Type":"ContainerDied","Data":"4d1273db6979d723df2d7e8def82e35eb8e045a749bca8ac625288c78385509e"} Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.394001 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d1273db6979d723df2d7e8def82e35eb8e045a749bca8ac625288c78385509e" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.513851 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-jswk7"] Nov 24 15:16:54 crc kubenswrapper[4790]: E1124 15:16:54.514580 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e88aa8a6-e145-452d-9362-3ee38c5f2839" containerName="extract-utilities" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.514681 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="e88aa8a6-e145-452d-9362-3ee38c5f2839" containerName="extract-utilities" Nov 24 15:16:54 crc kubenswrapper[4790]: E1124 15:16:54.514775 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e88aa8a6-e145-452d-9362-3ee38c5f2839" containerName="registry-server" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.514850 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="e88aa8a6-e145-452d-9362-3ee38c5f2839" containerName="registry-server" Nov 24 15:16:54 crc kubenswrapper[4790]: E1124 15:16:54.514962 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81fb48ee-289e-4442-81e5-ecba753eaecc" containerName="extract-content" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.515039 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="81fb48ee-289e-4442-81e5-ecba753eaecc" containerName="extract-content" Nov 24 15:16:54 crc kubenswrapper[4790]: E1124 15:16:54.515125 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81fb48ee-289e-4442-81e5-ecba753eaecc" containerName="extract-utilities" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.515209 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="81fb48ee-289e-4442-81e5-ecba753eaecc" containerName="extract-utilities" Nov 24 15:16:54 crc kubenswrapper[4790]: E1124 15:16:54.515294 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81fb48ee-289e-4442-81e5-ecba753eaecc" containerName="registry-server" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.515373 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="81fb48ee-289e-4442-81e5-ecba753eaecc" containerName="registry-server" Nov 24 15:16:54 crc kubenswrapper[4790]: E1124 15:16:54.515512 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e88aa8a6-e145-452d-9362-3ee38c5f2839" containerName="extract-content" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.515599 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="e88aa8a6-e145-452d-9362-3ee38c5f2839" containerName="extract-content" Nov 24 15:16:54 crc kubenswrapper[4790]: E1124 15:16:54.515691 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81c6adc7-3f2c-471d-a3ea-f2a947deb94a" containerName="configure-network-openstack-openstack-cell1" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.515771 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="81c6adc7-3f2c-471d-a3ea-f2a947deb94a" containerName="configure-network-openstack-openstack-cell1" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.516123 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="81fb48ee-289e-4442-81e5-ecba753eaecc" containerName="registry-server" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.516230 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="e88aa8a6-e145-452d-9362-3ee38c5f2839" containerName="registry-server" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.516359 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="81c6adc7-3f2c-471d-a3ea-f2a947deb94a" containerName="configure-network-openstack-openstack-cell1" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.517157 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-jswk7" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.522620 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.522843 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.522972 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fc497" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.523025 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.528774 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-jswk7"] Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.594870 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dc65\" (UniqueName: \"kubernetes.io/projected/abfebf54-171c-46e3-bbb3-882266bca305-kube-api-access-2dc65\") pod \"validate-network-openstack-openstack-cell1-jswk7\" (UID: \"abfebf54-171c-46e3-bbb3-882266bca305\") " pod="openstack/validate-network-openstack-openstack-cell1-jswk7" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.594923 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/abfebf54-171c-46e3-bbb3-882266bca305-ceph\") pod \"validate-network-openstack-openstack-cell1-jswk7\" (UID: \"abfebf54-171c-46e3-bbb3-882266bca305\") " pod="openstack/validate-network-openstack-openstack-cell1-jswk7" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.594954 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abfebf54-171c-46e3-bbb3-882266bca305-inventory\") pod \"validate-network-openstack-openstack-cell1-jswk7\" (UID: \"abfebf54-171c-46e3-bbb3-882266bca305\") " pod="openstack/validate-network-openstack-openstack-cell1-jswk7" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.595028 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abfebf54-171c-46e3-bbb3-882266bca305-ssh-key\") pod \"validate-network-openstack-openstack-cell1-jswk7\" (UID: \"abfebf54-171c-46e3-bbb3-882266bca305\") " pod="openstack/validate-network-openstack-openstack-cell1-jswk7" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.697242 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dc65\" (UniqueName: \"kubernetes.io/projected/abfebf54-171c-46e3-bbb3-882266bca305-kube-api-access-2dc65\") pod \"validate-network-openstack-openstack-cell1-jswk7\" (UID: \"abfebf54-171c-46e3-bbb3-882266bca305\") " pod="openstack/validate-network-openstack-openstack-cell1-jswk7" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.697351 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/abfebf54-171c-46e3-bbb3-882266bca305-ceph\") pod \"validate-network-openstack-openstack-cell1-jswk7\" (UID: \"abfebf54-171c-46e3-bbb3-882266bca305\") " pod="openstack/validate-network-openstack-openstack-cell1-jswk7" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.697444 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abfebf54-171c-46e3-bbb3-882266bca305-inventory\") pod \"validate-network-openstack-openstack-cell1-jswk7\" (UID: \"abfebf54-171c-46e3-bbb3-882266bca305\") " pod="openstack/validate-network-openstack-openstack-cell1-jswk7" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.697699 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abfebf54-171c-46e3-bbb3-882266bca305-ssh-key\") pod \"validate-network-openstack-openstack-cell1-jswk7\" (UID: \"abfebf54-171c-46e3-bbb3-882266bca305\") " pod="openstack/validate-network-openstack-openstack-cell1-jswk7" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.705471 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abfebf54-171c-46e3-bbb3-882266bca305-inventory\") pod \"validate-network-openstack-openstack-cell1-jswk7\" (UID: \"abfebf54-171c-46e3-bbb3-882266bca305\") " pod="openstack/validate-network-openstack-openstack-cell1-jswk7" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.709440 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/abfebf54-171c-46e3-bbb3-882266bca305-ceph\") pod \"validate-network-openstack-openstack-cell1-jswk7\" (UID: \"abfebf54-171c-46e3-bbb3-882266bca305\") " pod="openstack/validate-network-openstack-openstack-cell1-jswk7" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.710567 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abfebf54-171c-46e3-bbb3-882266bca305-ssh-key\") pod \"validate-network-openstack-openstack-cell1-jswk7\" (UID: \"abfebf54-171c-46e3-bbb3-882266bca305\") " pod="openstack/validate-network-openstack-openstack-cell1-jswk7" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.730586 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dc65\" (UniqueName: \"kubernetes.io/projected/abfebf54-171c-46e3-bbb3-882266bca305-kube-api-access-2dc65\") pod \"validate-network-openstack-openstack-cell1-jswk7\" (UID: \"abfebf54-171c-46e3-bbb3-882266bca305\") " pod="openstack/validate-network-openstack-openstack-cell1-jswk7" Nov 24 15:16:54 crc kubenswrapper[4790]: I1124 15:16:54.837010 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-jswk7" Nov 24 15:16:55 crc kubenswrapper[4790]: I1124 15:16:55.427938 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-jswk7"] Nov 24 15:16:56 crc kubenswrapper[4790]: I1124 15:16:56.416821 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-jswk7" event={"ID":"abfebf54-171c-46e3-bbb3-882266bca305","Type":"ContainerStarted","Data":"ecbecb671606feda38934d5c43ec676f919d6cc15012040c2efd63cbf4a28995"} Nov 24 15:16:56 crc kubenswrapper[4790]: I1124 15:16:56.417630 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-jswk7" event={"ID":"abfebf54-171c-46e3-bbb3-882266bca305","Type":"ContainerStarted","Data":"3031d95f8b58efba1d9e99feb2ea79c55fbb0381cc58645ea151cfb1c4393778"} Nov 24 15:16:56 crc kubenswrapper[4790]: I1124 15:16:56.448218 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-openstack-openstack-cell1-jswk7" podStartSLOduration=2.241917485 podStartE2EDuration="2.448191759s" podCreationTimestamp="2025-11-24 15:16:54 +0000 UTC" firstStartedPulling="2025-11-24 15:16:55.434379047 +0000 UTC m=+7463.814272719" lastFinishedPulling="2025-11-24 15:16:55.640653301 +0000 UTC m=+7464.020546993" observedRunningTime="2025-11-24 15:16:56.43233303 +0000 UTC m=+7464.812226692" watchObservedRunningTime="2025-11-24 15:16:56.448191759 +0000 UTC m=+7464.828085451" Nov 24 15:16:59 crc kubenswrapper[4790]: I1124 15:16:59.315517 4790 scope.go:117] "RemoveContainer" containerID="ed032daa21005f5603fc87ffbe2ed53274be2bbefd5cbb4153e64e56dd530b81" Nov 24 15:16:59 crc kubenswrapper[4790]: E1124 15:16:59.317016 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:17:01 crc kubenswrapper[4790]: I1124 15:17:01.483522 4790 generic.go:334] "Generic (PLEG): container finished" podID="abfebf54-171c-46e3-bbb3-882266bca305" containerID="ecbecb671606feda38934d5c43ec676f919d6cc15012040c2efd63cbf4a28995" exitCode=0 Nov 24 15:17:01 crc kubenswrapper[4790]: I1124 15:17:01.483610 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-jswk7" event={"ID":"abfebf54-171c-46e3-bbb3-882266bca305","Type":"ContainerDied","Data":"ecbecb671606feda38934d5c43ec676f919d6cc15012040c2efd63cbf4a28995"} Nov 24 15:17:02 crc kubenswrapper[4790]: I1124 15:17:02.979662 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-jswk7" Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.095137 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/abfebf54-171c-46e3-bbb3-882266bca305-ceph\") pod \"abfebf54-171c-46e3-bbb3-882266bca305\" (UID: \"abfebf54-171c-46e3-bbb3-882266bca305\") " Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.095327 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abfebf54-171c-46e3-bbb3-882266bca305-ssh-key\") pod \"abfebf54-171c-46e3-bbb3-882266bca305\" (UID: \"abfebf54-171c-46e3-bbb3-882266bca305\") " Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.096168 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dc65\" (UniqueName: \"kubernetes.io/projected/abfebf54-171c-46e3-bbb3-882266bca305-kube-api-access-2dc65\") pod \"abfebf54-171c-46e3-bbb3-882266bca305\" (UID: \"abfebf54-171c-46e3-bbb3-882266bca305\") " Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.096252 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abfebf54-171c-46e3-bbb3-882266bca305-inventory\") pod \"abfebf54-171c-46e3-bbb3-882266bca305\" (UID: \"abfebf54-171c-46e3-bbb3-882266bca305\") " Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.112268 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abfebf54-171c-46e3-bbb3-882266bca305-kube-api-access-2dc65" (OuterVolumeSpecName: "kube-api-access-2dc65") pod "abfebf54-171c-46e3-bbb3-882266bca305" (UID: "abfebf54-171c-46e3-bbb3-882266bca305"). InnerVolumeSpecName "kube-api-access-2dc65". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.116283 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abfebf54-171c-46e3-bbb3-882266bca305-ceph" (OuterVolumeSpecName: "ceph") pod "abfebf54-171c-46e3-bbb3-882266bca305" (UID: "abfebf54-171c-46e3-bbb3-882266bca305"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.124663 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abfebf54-171c-46e3-bbb3-882266bca305-inventory" (OuterVolumeSpecName: "inventory") pod "abfebf54-171c-46e3-bbb3-882266bca305" (UID: "abfebf54-171c-46e3-bbb3-882266bca305"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.130953 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abfebf54-171c-46e3-bbb3-882266bca305-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "abfebf54-171c-46e3-bbb3-882266bca305" (UID: "abfebf54-171c-46e3-bbb3-882266bca305"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.199098 4790 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abfebf54-171c-46e3-bbb3-882266bca305-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.199134 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dc65\" (UniqueName: \"kubernetes.io/projected/abfebf54-171c-46e3-bbb3-882266bca305-kube-api-access-2dc65\") on node \"crc\" DevicePath \"\"" Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.199147 4790 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abfebf54-171c-46e3-bbb3-882266bca305-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.199157 4790 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/abfebf54-171c-46e3-bbb3-882266bca305-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.514704 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-jswk7" event={"ID":"abfebf54-171c-46e3-bbb3-882266bca305","Type":"ContainerDied","Data":"3031d95f8b58efba1d9e99feb2ea79c55fbb0381cc58645ea151cfb1c4393778"} Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.515056 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3031d95f8b58efba1d9e99feb2ea79c55fbb0381cc58645ea151cfb1c4393778" Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.514759 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-jswk7" Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.592791 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-openstack-openstack-cell1-cb4w5"] Nov 24 15:17:03 crc kubenswrapper[4790]: E1124 15:17:03.593569 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abfebf54-171c-46e3-bbb3-882266bca305" containerName="validate-network-openstack-openstack-cell1" Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.593593 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="abfebf54-171c-46e3-bbb3-882266bca305" containerName="validate-network-openstack-openstack-cell1" Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.593871 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="abfebf54-171c-46e3-bbb3-882266bca305" containerName="validate-network-openstack-openstack-cell1" Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.595128 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-cb4w5" Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.597550 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fc497" Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.599684 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.600072 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.601377 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.603809 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-cb4w5"] Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.709983 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f039968e-b352-4952-8420-71a78ab6cc49-inventory\") pod \"install-os-openstack-openstack-cell1-cb4w5\" (UID: \"f039968e-b352-4952-8420-71a78ab6cc49\") " pod="openstack/install-os-openstack-openstack-cell1-cb4w5" Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.710048 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxxsc\" (UniqueName: \"kubernetes.io/projected/f039968e-b352-4952-8420-71a78ab6cc49-kube-api-access-bxxsc\") pod \"install-os-openstack-openstack-cell1-cb4w5\" (UID: \"f039968e-b352-4952-8420-71a78ab6cc49\") " pod="openstack/install-os-openstack-openstack-cell1-cb4w5" Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.710096 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f039968e-b352-4952-8420-71a78ab6cc49-ssh-key\") pod \"install-os-openstack-openstack-cell1-cb4w5\" (UID: \"f039968e-b352-4952-8420-71a78ab6cc49\") " pod="openstack/install-os-openstack-openstack-cell1-cb4w5" Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.710384 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f039968e-b352-4952-8420-71a78ab6cc49-ceph\") pod \"install-os-openstack-openstack-cell1-cb4w5\" (UID: \"f039968e-b352-4952-8420-71a78ab6cc49\") " pod="openstack/install-os-openstack-openstack-cell1-cb4w5" Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.811930 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f039968e-b352-4952-8420-71a78ab6cc49-inventory\") pod \"install-os-openstack-openstack-cell1-cb4w5\" (UID: \"f039968e-b352-4952-8420-71a78ab6cc49\") " pod="openstack/install-os-openstack-openstack-cell1-cb4w5" Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.811971 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxxsc\" (UniqueName: \"kubernetes.io/projected/f039968e-b352-4952-8420-71a78ab6cc49-kube-api-access-bxxsc\") pod \"install-os-openstack-openstack-cell1-cb4w5\" (UID: \"f039968e-b352-4952-8420-71a78ab6cc49\") " pod="openstack/install-os-openstack-openstack-cell1-cb4w5" Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.812003 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f039968e-b352-4952-8420-71a78ab6cc49-ssh-key\") pod \"install-os-openstack-openstack-cell1-cb4w5\" (UID: \"f039968e-b352-4952-8420-71a78ab6cc49\") " pod="openstack/install-os-openstack-openstack-cell1-cb4w5" Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.812076 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f039968e-b352-4952-8420-71a78ab6cc49-ceph\") pod \"install-os-openstack-openstack-cell1-cb4w5\" (UID: \"f039968e-b352-4952-8420-71a78ab6cc49\") " pod="openstack/install-os-openstack-openstack-cell1-cb4w5" Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.818001 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f039968e-b352-4952-8420-71a78ab6cc49-inventory\") pod \"install-os-openstack-openstack-cell1-cb4w5\" (UID: \"f039968e-b352-4952-8420-71a78ab6cc49\") " pod="openstack/install-os-openstack-openstack-cell1-cb4w5" Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.819535 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f039968e-b352-4952-8420-71a78ab6cc49-ssh-key\") pod \"install-os-openstack-openstack-cell1-cb4w5\" (UID: \"f039968e-b352-4952-8420-71a78ab6cc49\") " pod="openstack/install-os-openstack-openstack-cell1-cb4w5" Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.820349 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f039968e-b352-4952-8420-71a78ab6cc49-ceph\") pod \"install-os-openstack-openstack-cell1-cb4w5\" (UID: \"f039968e-b352-4952-8420-71a78ab6cc49\") " pod="openstack/install-os-openstack-openstack-cell1-cb4w5" Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.840685 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxxsc\" (UniqueName: \"kubernetes.io/projected/f039968e-b352-4952-8420-71a78ab6cc49-kube-api-access-bxxsc\") pod \"install-os-openstack-openstack-cell1-cb4w5\" (UID: \"f039968e-b352-4952-8420-71a78ab6cc49\") " pod="openstack/install-os-openstack-openstack-cell1-cb4w5" Nov 24 15:17:03 crc kubenswrapper[4790]: I1124 15:17:03.924979 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-cb4w5" Nov 24 15:17:04 crc kubenswrapper[4790]: I1124 15:17:04.519071 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-cb4w5"] Nov 24 15:17:05 crc kubenswrapper[4790]: I1124 15:17:05.544653 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-cb4w5" event={"ID":"f039968e-b352-4952-8420-71a78ab6cc49","Type":"ContainerStarted","Data":"35e27ba3bd809bcf40605231f903991a8fd7e6150f054495b543407d089449cc"} Nov 24 15:17:05 crc kubenswrapper[4790]: I1124 15:17:05.545426 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-cb4w5" event={"ID":"f039968e-b352-4952-8420-71a78ab6cc49","Type":"ContainerStarted","Data":"8cef2db4324a8d3ea5f9d041a73d0a6326dfa712843f5c32fe75fe8189613ed7"} Nov 24 15:17:05 crc kubenswrapper[4790]: I1124 15:17:05.579666 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-openstack-openstack-cell1-cb4w5" podStartSLOduration=2.404469295 podStartE2EDuration="2.579642249s" podCreationTimestamp="2025-11-24 15:17:03 +0000 UTC" firstStartedPulling="2025-11-24 15:17:04.526324839 +0000 UTC m=+7472.906218511" lastFinishedPulling="2025-11-24 15:17:04.701497803 +0000 UTC m=+7473.081391465" observedRunningTime="2025-11-24 15:17:05.567286239 +0000 UTC m=+7473.947179951" watchObservedRunningTime="2025-11-24 15:17:05.579642249 +0000 UTC m=+7473.959535951" Nov 24 15:17:13 crc kubenswrapper[4790]: I1124 15:17:13.314310 4790 scope.go:117] "RemoveContainer" containerID="ed032daa21005f5603fc87ffbe2ed53274be2bbefd5cbb4153e64e56dd530b81" Nov 24 15:17:13 crc kubenswrapper[4790]: E1124 15:17:13.314995 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:17:25 crc kubenswrapper[4790]: I1124 15:17:25.314996 4790 scope.go:117] "RemoveContainer" containerID="ed032daa21005f5603fc87ffbe2ed53274be2bbefd5cbb4153e64e56dd530b81" Nov 24 15:17:25 crc kubenswrapper[4790]: E1124 15:17:25.316123 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:17:39 crc kubenswrapper[4790]: I1124 15:17:39.314624 4790 scope.go:117] "RemoveContainer" containerID="ed032daa21005f5603fc87ffbe2ed53274be2bbefd5cbb4153e64e56dd530b81" Nov 24 15:17:39 crc kubenswrapper[4790]: E1124 15:17:39.315342 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:17:50 crc kubenswrapper[4790]: I1124 15:17:50.314989 4790 scope.go:117] "RemoveContainer" containerID="ed032daa21005f5603fc87ffbe2ed53274be2bbefd5cbb4153e64e56dd530b81" Nov 24 15:17:50 crc kubenswrapper[4790]: E1124 15:17:50.317149 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:17:52 crc kubenswrapper[4790]: I1124 15:17:52.088745 4790 generic.go:334] "Generic (PLEG): container finished" podID="f039968e-b352-4952-8420-71a78ab6cc49" containerID="35e27ba3bd809bcf40605231f903991a8fd7e6150f054495b543407d089449cc" exitCode=0 Nov 24 15:17:52 crc kubenswrapper[4790]: I1124 15:17:52.088857 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-cb4w5" event={"ID":"f039968e-b352-4952-8420-71a78ab6cc49","Type":"ContainerDied","Data":"35e27ba3bd809bcf40605231f903991a8fd7e6150f054495b543407d089449cc"} Nov 24 15:17:53 crc kubenswrapper[4790]: I1124 15:17:53.661203 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-cb4w5" Nov 24 15:17:53 crc kubenswrapper[4790]: I1124 15:17:53.804169 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f039968e-b352-4952-8420-71a78ab6cc49-ceph\") pod \"f039968e-b352-4952-8420-71a78ab6cc49\" (UID: \"f039968e-b352-4952-8420-71a78ab6cc49\") " Nov 24 15:17:53 crc kubenswrapper[4790]: I1124 15:17:53.804295 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxxsc\" (UniqueName: \"kubernetes.io/projected/f039968e-b352-4952-8420-71a78ab6cc49-kube-api-access-bxxsc\") pod \"f039968e-b352-4952-8420-71a78ab6cc49\" (UID: \"f039968e-b352-4952-8420-71a78ab6cc49\") " Nov 24 15:17:53 crc kubenswrapper[4790]: I1124 15:17:53.804554 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f039968e-b352-4952-8420-71a78ab6cc49-ssh-key\") pod \"f039968e-b352-4952-8420-71a78ab6cc49\" (UID: \"f039968e-b352-4952-8420-71a78ab6cc49\") " Nov 24 15:17:53 crc kubenswrapper[4790]: I1124 15:17:53.804696 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f039968e-b352-4952-8420-71a78ab6cc49-inventory\") pod \"f039968e-b352-4952-8420-71a78ab6cc49\" (UID: \"f039968e-b352-4952-8420-71a78ab6cc49\") " Nov 24 15:17:53 crc kubenswrapper[4790]: I1124 15:17:53.810766 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f039968e-b352-4952-8420-71a78ab6cc49-kube-api-access-bxxsc" (OuterVolumeSpecName: "kube-api-access-bxxsc") pod "f039968e-b352-4952-8420-71a78ab6cc49" (UID: "f039968e-b352-4952-8420-71a78ab6cc49"). InnerVolumeSpecName "kube-api-access-bxxsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:17:53 crc kubenswrapper[4790]: I1124 15:17:53.812074 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f039968e-b352-4952-8420-71a78ab6cc49-ceph" (OuterVolumeSpecName: "ceph") pod "f039968e-b352-4952-8420-71a78ab6cc49" (UID: "f039968e-b352-4952-8420-71a78ab6cc49"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:17:53 crc kubenswrapper[4790]: I1124 15:17:53.837368 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f039968e-b352-4952-8420-71a78ab6cc49-inventory" (OuterVolumeSpecName: "inventory") pod "f039968e-b352-4952-8420-71a78ab6cc49" (UID: "f039968e-b352-4952-8420-71a78ab6cc49"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:17:53 crc kubenswrapper[4790]: I1124 15:17:53.850442 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f039968e-b352-4952-8420-71a78ab6cc49-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f039968e-b352-4952-8420-71a78ab6cc49" (UID: "f039968e-b352-4952-8420-71a78ab6cc49"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:17:53 crc kubenswrapper[4790]: I1124 15:17:53.924873 4790 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f039968e-b352-4952-8420-71a78ab6cc49-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 15:17:53 crc kubenswrapper[4790]: I1124 15:17:53.925348 4790 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f039968e-b352-4952-8420-71a78ab6cc49-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 15:17:53 crc kubenswrapper[4790]: I1124 15:17:53.925371 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxxsc\" (UniqueName: \"kubernetes.io/projected/f039968e-b352-4952-8420-71a78ab6cc49-kube-api-access-bxxsc\") on node \"crc\" DevicePath \"\"" Nov 24 15:17:53 crc kubenswrapper[4790]: I1124 15:17:53.925395 4790 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f039968e-b352-4952-8420-71a78ab6cc49-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 15:17:54 crc kubenswrapper[4790]: I1124 15:17:54.115188 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-cb4w5" event={"ID":"f039968e-b352-4952-8420-71a78ab6cc49","Type":"ContainerDied","Data":"8cef2db4324a8d3ea5f9d041a73d0a6326dfa712843f5c32fe75fe8189613ed7"} Nov 24 15:17:54 crc kubenswrapper[4790]: I1124 15:17:54.115232 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8cef2db4324a8d3ea5f9d041a73d0a6326dfa712843f5c32fe75fe8189613ed7" Nov 24 15:17:54 crc kubenswrapper[4790]: I1124 15:17:54.115270 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-cb4w5" Nov 24 15:17:54 crc kubenswrapper[4790]: I1124 15:17:54.205385 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-xfkr7"] Nov 24 15:17:54 crc kubenswrapper[4790]: E1124 15:17:54.206010 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f039968e-b352-4952-8420-71a78ab6cc49" containerName="install-os-openstack-openstack-cell1" Nov 24 15:17:54 crc kubenswrapper[4790]: I1124 15:17:54.206031 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="f039968e-b352-4952-8420-71a78ab6cc49" containerName="install-os-openstack-openstack-cell1" Nov 24 15:17:54 crc kubenswrapper[4790]: I1124 15:17:54.206305 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="f039968e-b352-4952-8420-71a78ab6cc49" containerName="install-os-openstack-openstack-cell1" Nov 24 15:17:54 crc kubenswrapper[4790]: I1124 15:17:54.207330 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-xfkr7" Nov 24 15:17:54 crc kubenswrapper[4790]: I1124 15:17:54.210625 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 15:17:54 crc kubenswrapper[4790]: I1124 15:17:54.210808 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fc497" Nov 24 15:17:54 crc kubenswrapper[4790]: I1124 15:17:54.210934 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 15:17:54 crc kubenswrapper[4790]: I1124 15:17:54.211855 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 15:17:54 crc kubenswrapper[4790]: I1124 15:17:54.217176 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-xfkr7"] Nov 24 15:17:54 crc kubenswrapper[4790]: I1124 15:17:54.233322 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8aebdda2-47dc-4465-b3b1-6de4cb11c07c-ssh-key\") pod \"configure-os-openstack-openstack-cell1-xfkr7\" (UID: \"8aebdda2-47dc-4465-b3b1-6de4cb11c07c\") " pod="openstack/configure-os-openstack-openstack-cell1-xfkr7" Nov 24 15:17:54 crc kubenswrapper[4790]: I1124 15:17:54.233380 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8aebdda2-47dc-4465-b3b1-6de4cb11c07c-inventory\") pod \"configure-os-openstack-openstack-cell1-xfkr7\" (UID: \"8aebdda2-47dc-4465-b3b1-6de4cb11c07c\") " pod="openstack/configure-os-openstack-openstack-cell1-xfkr7" Nov 24 15:17:54 crc kubenswrapper[4790]: I1124 15:17:54.233427 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8aebdda2-47dc-4465-b3b1-6de4cb11c07c-ceph\") pod \"configure-os-openstack-openstack-cell1-xfkr7\" (UID: \"8aebdda2-47dc-4465-b3b1-6de4cb11c07c\") " pod="openstack/configure-os-openstack-openstack-cell1-xfkr7" Nov 24 15:17:54 crc kubenswrapper[4790]: I1124 15:17:54.233485 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5kpc\" (UniqueName: \"kubernetes.io/projected/8aebdda2-47dc-4465-b3b1-6de4cb11c07c-kube-api-access-m5kpc\") pod \"configure-os-openstack-openstack-cell1-xfkr7\" (UID: \"8aebdda2-47dc-4465-b3b1-6de4cb11c07c\") " pod="openstack/configure-os-openstack-openstack-cell1-xfkr7" Nov 24 15:17:54 crc kubenswrapper[4790]: I1124 15:17:54.336997 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8aebdda2-47dc-4465-b3b1-6de4cb11c07c-ssh-key\") pod \"configure-os-openstack-openstack-cell1-xfkr7\" (UID: \"8aebdda2-47dc-4465-b3b1-6de4cb11c07c\") " pod="openstack/configure-os-openstack-openstack-cell1-xfkr7" Nov 24 15:17:54 crc kubenswrapper[4790]: I1124 15:17:54.337099 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8aebdda2-47dc-4465-b3b1-6de4cb11c07c-inventory\") pod \"configure-os-openstack-openstack-cell1-xfkr7\" (UID: \"8aebdda2-47dc-4465-b3b1-6de4cb11c07c\") " pod="openstack/configure-os-openstack-openstack-cell1-xfkr7" Nov 24 15:17:54 crc kubenswrapper[4790]: I1124 15:17:54.337187 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8aebdda2-47dc-4465-b3b1-6de4cb11c07c-ceph\") pod \"configure-os-openstack-openstack-cell1-xfkr7\" (UID: \"8aebdda2-47dc-4465-b3b1-6de4cb11c07c\") " pod="openstack/configure-os-openstack-openstack-cell1-xfkr7" Nov 24 15:17:54 crc kubenswrapper[4790]: I1124 15:17:54.337252 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5kpc\" (UniqueName: \"kubernetes.io/projected/8aebdda2-47dc-4465-b3b1-6de4cb11c07c-kube-api-access-m5kpc\") pod \"configure-os-openstack-openstack-cell1-xfkr7\" (UID: \"8aebdda2-47dc-4465-b3b1-6de4cb11c07c\") " pod="openstack/configure-os-openstack-openstack-cell1-xfkr7" Nov 24 15:17:54 crc kubenswrapper[4790]: I1124 15:17:54.343161 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8aebdda2-47dc-4465-b3b1-6de4cb11c07c-ssh-key\") pod \"configure-os-openstack-openstack-cell1-xfkr7\" (UID: \"8aebdda2-47dc-4465-b3b1-6de4cb11c07c\") " pod="openstack/configure-os-openstack-openstack-cell1-xfkr7" Nov 24 15:17:54 crc kubenswrapper[4790]: I1124 15:17:54.344603 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8aebdda2-47dc-4465-b3b1-6de4cb11c07c-inventory\") pod \"configure-os-openstack-openstack-cell1-xfkr7\" (UID: \"8aebdda2-47dc-4465-b3b1-6de4cb11c07c\") " pod="openstack/configure-os-openstack-openstack-cell1-xfkr7" Nov 24 15:17:54 crc kubenswrapper[4790]: I1124 15:17:54.345556 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8aebdda2-47dc-4465-b3b1-6de4cb11c07c-ceph\") pod \"configure-os-openstack-openstack-cell1-xfkr7\" (UID: \"8aebdda2-47dc-4465-b3b1-6de4cb11c07c\") " pod="openstack/configure-os-openstack-openstack-cell1-xfkr7" Nov 24 15:17:54 crc kubenswrapper[4790]: I1124 15:17:54.359463 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5kpc\" (UniqueName: \"kubernetes.io/projected/8aebdda2-47dc-4465-b3b1-6de4cb11c07c-kube-api-access-m5kpc\") pod \"configure-os-openstack-openstack-cell1-xfkr7\" (UID: \"8aebdda2-47dc-4465-b3b1-6de4cb11c07c\") " pod="openstack/configure-os-openstack-openstack-cell1-xfkr7" Nov 24 15:17:54 crc kubenswrapper[4790]: I1124 15:17:54.531618 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-xfkr7" Nov 24 15:17:55 crc kubenswrapper[4790]: I1124 15:17:55.204972 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-xfkr7"] Nov 24 15:17:56 crc kubenswrapper[4790]: I1124 15:17:56.140083 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-xfkr7" event={"ID":"8aebdda2-47dc-4465-b3b1-6de4cb11c07c","Type":"ContainerStarted","Data":"2d0a195340306dd3c0a646d3e886c190a9b966e561885d84f48ded4d8ebe6be7"} Nov 24 15:17:57 crc kubenswrapper[4790]: I1124 15:17:57.153917 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-xfkr7" event={"ID":"8aebdda2-47dc-4465-b3b1-6de4cb11c07c","Type":"ContainerStarted","Data":"648f0789be093e0f56429cb3a94264ca50b60b60efde3a01c74d34369aba41fa"} Nov 24 15:17:57 crc kubenswrapper[4790]: I1124 15:17:57.177763 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-cell1-xfkr7" podStartSLOduration=2.381767162 podStartE2EDuration="3.177743713s" podCreationTimestamp="2025-11-24 15:17:54 +0000 UTC" firstStartedPulling="2025-11-24 15:17:55.205343571 +0000 UTC m=+7523.585237273" lastFinishedPulling="2025-11-24 15:17:56.001320152 +0000 UTC m=+7524.381213824" observedRunningTime="2025-11-24 15:17:57.17377627 +0000 UTC m=+7525.553669942" watchObservedRunningTime="2025-11-24 15:17:57.177743713 +0000 UTC m=+7525.557637375" Nov 24 15:18:01 crc kubenswrapper[4790]: I1124 15:18:01.315463 4790 scope.go:117] "RemoveContainer" containerID="ed032daa21005f5603fc87ffbe2ed53274be2bbefd5cbb4153e64e56dd530b81" Nov 24 15:18:01 crc kubenswrapper[4790]: E1124 15:18:01.316159 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:18:13 crc kubenswrapper[4790]: I1124 15:18:13.315408 4790 scope.go:117] "RemoveContainer" containerID="ed032daa21005f5603fc87ffbe2ed53274be2bbefd5cbb4153e64e56dd530b81" Nov 24 15:18:13 crc kubenswrapper[4790]: E1124 15:18:13.316642 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:18:28 crc kubenswrapper[4790]: I1124 15:18:28.315850 4790 scope.go:117] "RemoveContainer" containerID="ed032daa21005f5603fc87ffbe2ed53274be2bbefd5cbb4153e64e56dd530b81" Nov 24 15:18:28 crc kubenswrapper[4790]: E1124 15:18:28.317274 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:18:42 crc kubenswrapper[4790]: I1124 15:18:42.331989 4790 scope.go:117] "RemoveContainer" containerID="ed032daa21005f5603fc87ffbe2ed53274be2bbefd5cbb4153e64e56dd530b81" Nov 24 15:18:42 crc kubenswrapper[4790]: E1124 15:18:42.333411 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:18:43 crc kubenswrapper[4790]: I1124 15:18:43.706803 4790 generic.go:334] "Generic (PLEG): container finished" podID="8aebdda2-47dc-4465-b3b1-6de4cb11c07c" containerID="648f0789be093e0f56429cb3a94264ca50b60b60efde3a01c74d34369aba41fa" exitCode=0 Nov 24 15:18:43 crc kubenswrapper[4790]: I1124 15:18:43.706897 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-xfkr7" event={"ID":"8aebdda2-47dc-4465-b3b1-6de4cb11c07c","Type":"ContainerDied","Data":"648f0789be093e0f56429cb3a94264ca50b60b60efde3a01c74d34369aba41fa"} Nov 24 15:18:45 crc kubenswrapper[4790]: I1124 15:18:45.214013 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-xfkr7" Nov 24 15:18:45 crc kubenswrapper[4790]: I1124 15:18:45.350453 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5kpc\" (UniqueName: \"kubernetes.io/projected/8aebdda2-47dc-4465-b3b1-6de4cb11c07c-kube-api-access-m5kpc\") pod \"8aebdda2-47dc-4465-b3b1-6de4cb11c07c\" (UID: \"8aebdda2-47dc-4465-b3b1-6de4cb11c07c\") " Nov 24 15:18:45 crc kubenswrapper[4790]: I1124 15:18:45.350780 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8aebdda2-47dc-4465-b3b1-6de4cb11c07c-ceph\") pod \"8aebdda2-47dc-4465-b3b1-6de4cb11c07c\" (UID: \"8aebdda2-47dc-4465-b3b1-6de4cb11c07c\") " Nov 24 15:18:45 crc kubenswrapper[4790]: I1124 15:18:45.350804 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8aebdda2-47dc-4465-b3b1-6de4cb11c07c-ssh-key\") pod \"8aebdda2-47dc-4465-b3b1-6de4cb11c07c\" (UID: \"8aebdda2-47dc-4465-b3b1-6de4cb11c07c\") " Nov 24 15:18:45 crc kubenswrapper[4790]: I1124 15:18:45.350828 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8aebdda2-47dc-4465-b3b1-6de4cb11c07c-inventory\") pod \"8aebdda2-47dc-4465-b3b1-6de4cb11c07c\" (UID: \"8aebdda2-47dc-4465-b3b1-6de4cb11c07c\") " Nov 24 15:18:45 crc kubenswrapper[4790]: I1124 15:18:45.359586 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8aebdda2-47dc-4465-b3b1-6de4cb11c07c-kube-api-access-m5kpc" (OuterVolumeSpecName: "kube-api-access-m5kpc") pod "8aebdda2-47dc-4465-b3b1-6de4cb11c07c" (UID: "8aebdda2-47dc-4465-b3b1-6de4cb11c07c"). InnerVolumeSpecName "kube-api-access-m5kpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:18:45 crc kubenswrapper[4790]: I1124 15:18:45.367250 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8aebdda2-47dc-4465-b3b1-6de4cb11c07c-ceph" (OuterVolumeSpecName: "ceph") pod "8aebdda2-47dc-4465-b3b1-6de4cb11c07c" (UID: "8aebdda2-47dc-4465-b3b1-6de4cb11c07c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:18:45 crc kubenswrapper[4790]: I1124 15:18:45.382717 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8aebdda2-47dc-4465-b3b1-6de4cb11c07c-inventory" (OuterVolumeSpecName: "inventory") pod "8aebdda2-47dc-4465-b3b1-6de4cb11c07c" (UID: "8aebdda2-47dc-4465-b3b1-6de4cb11c07c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:18:45 crc kubenswrapper[4790]: I1124 15:18:45.400714 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8aebdda2-47dc-4465-b3b1-6de4cb11c07c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8aebdda2-47dc-4465-b3b1-6de4cb11c07c" (UID: "8aebdda2-47dc-4465-b3b1-6de4cb11c07c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:18:45 crc kubenswrapper[4790]: I1124 15:18:45.454359 4790 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8aebdda2-47dc-4465-b3b1-6de4cb11c07c-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 15:18:45 crc kubenswrapper[4790]: I1124 15:18:45.454391 4790 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8aebdda2-47dc-4465-b3b1-6de4cb11c07c-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 15:18:45 crc kubenswrapper[4790]: I1124 15:18:45.454400 4790 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8aebdda2-47dc-4465-b3b1-6de4cb11c07c-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 15:18:45 crc kubenswrapper[4790]: I1124 15:18:45.454409 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5kpc\" (UniqueName: \"kubernetes.io/projected/8aebdda2-47dc-4465-b3b1-6de4cb11c07c-kube-api-access-m5kpc\") on node \"crc\" DevicePath \"\"" Nov 24 15:18:45 crc kubenswrapper[4790]: I1124 15:18:45.740938 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-xfkr7" event={"ID":"8aebdda2-47dc-4465-b3b1-6de4cb11c07c","Type":"ContainerDied","Data":"2d0a195340306dd3c0a646d3e886c190a9b966e561885d84f48ded4d8ebe6be7"} Nov 24 15:18:45 crc kubenswrapper[4790]: I1124 15:18:45.740988 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d0a195340306dd3c0a646d3e886c190a9b966e561885d84f48ded4d8ebe6be7" Nov 24 15:18:45 crc kubenswrapper[4790]: I1124 15:18:45.741111 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-xfkr7" Nov 24 15:18:45 crc kubenswrapper[4790]: I1124 15:18:45.853686 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-openstack-h9qxq"] Nov 24 15:18:45 crc kubenswrapper[4790]: E1124 15:18:45.854545 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8aebdda2-47dc-4465-b3b1-6de4cb11c07c" containerName="configure-os-openstack-openstack-cell1" Nov 24 15:18:45 crc kubenswrapper[4790]: I1124 15:18:45.854665 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="8aebdda2-47dc-4465-b3b1-6de4cb11c07c" containerName="configure-os-openstack-openstack-cell1" Nov 24 15:18:45 crc kubenswrapper[4790]: I1124 15:18:45.855072 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="8aebdda2-47dc-4465-b3b1-6de4cb11c07c" containerName="configure-os-openstack-openstack-cell1" Nov 24 15:18:45 crc kubenswrapper[4790]: I1124 15:18:45.856376 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-h9qxq" Nov 24 15:18:45 crc kubenswrapper[4790]: I1124 15:18:45.858805 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 15:18:45 crc kubenswrapper[4790]: I1124 15:18:45.859016 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 15:18:45 crc kubenswrapper[4790]: I1124 15:18:45.859156 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fc497" Nov 24 15:18:45 crc kubenswrapper[4790]: I1124 15:18:45.859715 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 15:18:45 crc kubenswrapper[4790]: I1124 15:18:45.892965 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-h9qxq"] Nov 24 15:18:45 crc kubenswrapper[4790]: I1124 15:18:45.966563 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9b974f0b-e04c-4394-85fa-14b57cff7504-ceph\") pod \"ssh-known-hosts-openstack-h9qxq\" (UID: \"9b974f0b-e04c-4394-85fa-14b57cff7504\") " pod="openstack/ssh-known-hosts-openstack-h9qxq" Nov 24 15:18:45 crc kubenswrapper[4790]: I1124 15:18:45.967415 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/9b974f0b-e04c-4394-85fa-14b57cff7504-inventory-0\") pod \"ssh-known-hosts-openstack-h9qxq\" (UID: \"9b974f0b-e04c-4394-85fa-14b57cff7504\") " pod="openstack/ssh-known-hosts-openstack-h9qxq" Nov 24 15:18:45 crc kubenswrapper[4790]: I1124 15:18:45.967510 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/9b974f0b-e04c-4394-85fa-14b57cff7504-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-h9qxq\" (UID: \"9b974f0b-e04c-4394-85fa-14b57cff7504\") " pod="openstack/ssh-known-hosts-openstack-h9qxq" Nov 24 15:18:45 crc kubenswrapper[4790]: I1124 15:18:45.967969 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2xm8\" (UniqueName: \"kubernetes.io/projected/9b974f0b-e04c-4394-85fa-14b57cff7504-kube-api-access-w2xm8\") pod \"ssh-known-hosts-openstack-h9qxq\" (UID: \"9b974f0b-e04c-4394-85fa-14b57cff7504\") " pod="openstack/ssh-known-hosts-openstack-h9qxq" Nov 24 15:18:46 crc kubenswrapper[4790]: I1124 15:18:46.069581 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/9b974f0b-e04c-4394-85fa-14b57cff7504-inventory-0\") pod \"ssh-known-hosts-openstack-h9qxq\" (UID: \"9b974f0b-e04c-4394-85fa-14b57cff7504\") " pod="openstack/ssh-known-hosts-openstack-h9qxq" Nov 24 15:18:46 crc kubenswrapper[4790]: I1124 15:18:46.069685 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/9b974f0b-e04c-4394-85fa-14b57cff7504-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-h9qxq\" (UID: \"9b974f0b-e04c-4394-85fa-14b57cff7504\") " pod="openstack/ssh-known-hosts-openstack-h9qxq" Nov 24 15:18:46 crc kubenswrapper[4790]: I1124 15:18:46.069830 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2xm8\" (UniqueName: \"kubernetes.io/projected/9b974f0b-e04c-4394-85fa-14b57cff7504-kube-api-access-w2xm8\") pod \"ssh-known-hosts-openstack-h9qxq\" (UID: \"9b974f0b-e04c-4394-85fa-14b57cff7504\") " pod="openstack/ssh-known-hosts-openstack-h9qxq" Nov 24 15:18:46 crc kubenswrapper[4790]: I1124 15:18:46.069988 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9b974f0b-e04c-4394-85fa-14b57cff7504-ceph\") pod \"ssh-known-hosts-openstack-h9qxq\" (UID: \"9b974f0b-e04c-4394-85fa-14b57cff7504\") " pod="openstack/ssh-known-hosts-openstack-h9qxq" Nov 24 15:18:46 crc kubenswrapper[4790]: I1124 15:18:46.076395 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/9b974f0b-e04c-4394-85fa-14b57cff7504-inventory-0\") pod \"ssh-known-hosts-openstack-h9qxq\" (UID: \"9b974f0b-e04c-4394-85fa-14b57cff7504\") " pod="openstack/ssh-known-hosts-openstack-h9qxq" Nov 24 15:18:46 crc kubenswrapper[4790]: I1124 15:18:46.079924 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9b974f0b-e04c-4394-85fa-14b57cff7504-ceph\") pod \"ssh-known-hosts-openstack-h9qxq\" (UID: \"9b974f0b-e04c-4394-85fa-14b57cff7504\") " pod="openstack/ssh-known-hosts-openstack-h9qxq" Nov 24 15:18:46 crc kubenswrapper[4790]: I1124 15:18:46.082308 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/9b974f0b-e04c-4394-85fa-14b57cff7504-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-h9qxq\" (UID: \"9b974f0b-e04c-4394-85fa-14b57cff7504\") " pod="openstack/ssh-known-hosts-openstack-h9qxq" Nov 24 15:18:46 crc kubenswrapper[4790]: I1124 15:18:46.089140 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2xm8\" (UniqueName: \"kubernetes.io/projected/9b974f0b-e04c-4394-85fa-14b57cff7504-kube-api-access-w2xm8\") pod \"ssh-known-hosts-openstack-h9qxq\" (UID: \"9b974f0b-e04c-4394-85fa-14b57cff7504\") " pod="openstack/ssh-known-hosts-openstack-h9qxq" Nov 24 15:18:46 crc kubenswrapper[4790]: I1124 15:18:46.191145 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-h9qxq" Nov 24 15:18:46 crc kubenswrapper[4790]: I1124 15:18:46.896620 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-h9qxq"] Nov 24 15:18:47 crc kubenswrapper[4790]: I1124 15:18:47.761891 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-h9qxq" event={"ID":"9b974f0b-e04c-4394-85fa-14b57cff7504","Type":"ContainerStarted","Data":"4933315fe5f716a845e1b4fc01f2ad76a6b0f6693d614936375eb69b7d6542ac"} Nov 24 15:18:47 crc kubenswrapper[4790]: I1124 15:18:47.762600 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-h9qxq" event={"ID":"9b974f0b-e04c-4394-85fa-14b57cff7504","Type":"ContainerStarted","Data":"5a5b8aad23f937a0c59487a72756a877335438c84e3626a725d96990f06fcaee"} Nov 24 15:18:47 crc kubenswrapper[4790]: I1124 15:18:47.790444 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-openstack-h9qxq" podStartSLOduration=2.596616177 podStartE2EDuration="2.790420679s" podCreationTimestamp="2025-11-24 15:18:45 +0000 UTC" firstStartedPulling="2025-11-24 15:18:46.913480607 +0000 UTC m=+7575.293374269" lastFinishedPulling="2025-11-24 15:18:47.107285099 +0000 UTC m=+7575.487178771" observedRunningTime="2025-11-24 15:18:47.783270817 +0000 UTC m=+7576.163164479" watchObservedRunningTime="2025-11-24 15:18:47.790420679 +0000 UTC m=+7576.170314351" Nov 24 15:18:54 crc kubenswrapper[4790]: I1124 15:18:54.315679 4790 scope.go:117] "RemoveContainer" containerID="ed032daa21005f5603fc87ffbe2ed53274be2bbefd5cbb4153e64e56dd530b81" Nov 24 15:18:54 crc kubenswrapper[4790]: E1124 15:18:54.316963 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:18:56 crc kubenswrapper[4790]: I1124 15:18:56.882087 4790 generic.go:334] "Generic (PLEG): container finished" podID="9b974f0b-e04c-4394-85fa-14b57cff7504" containerID="4933315fe5f716a845e1b4fc01f2ad76a6b0f6693d614936375eb69b7d6542ac" exitCode=0 Nov 24 15:18:56 crc kubenswrapper[4790]: I1124 15:18:56.882198 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-h9qxq" event={"ID":"9b974f0b-e04c-4394-85fa-14b57cff7504","Type":"ContainerDied","Data":"4933315fe5f716a845e1b4fc01f2ad76a6b0f6693d614936375eb69b7d6542ac"} Nov 24 15:18:58 crc kubenswrapper[4790]: I1124 15:18:58.468811 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-h9qxq" Nov 24 15:18:58 crc kubenswrapper[4790]: I1124 15:18:58.553489 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2xm8\" (UniqueName: \"kubernetes.io/projected/9b974f0b-e04c-4394-85fa-14b57cff7504-kube-api-access-w2xm8\") pod \"9b974f0b-e04c-4394-85fa-14b57cff7504\" (UID: \"9b974f0b-e04c-4394-85fa-14b57cff7504\") " Nov 24 15:18:58 crc kubenswrapper[4790]: I1124 15:18:58.553560 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9b974f0b-e04c-4394-85fa-14b57cff7504-ceph\") pod \"9b974f0b-e04c-4394-85fa-14b57cff7504\" (UID: \"9b974f0b-e04c-4394-85fa-14b57cff7504\") " Nov 24 15:18:58 crc kubenswrapper[4790]: I1124 15:18:58.553619 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/9b974f0b-e04c-4394-85fa-14b57cff7504-ssh-key-openstack-cell1\") pod \"9b974f0b-e04c-4394-85fa-14b57cff7504\" (UID: \"9b974f0b-e04c-4394-85fa-14b57cff7504\") " Nov 24 15:18:58 crc kubenswrapper[4790]: I1124 15:18:58.553765 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/9b974f0b-e04c-4394-85fa-14b57cff7504-inventory-0\") pod \"9b974f0b-e04c-4394-85fa-14b57cff7504\" (UID: \"9b974f0b-e04c-4394-85fa-14b57cff7504\") " Nov 24 15:18:58 crc kubenswrapper[4790]: I1124 15:18:58.560371 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b974f0b-e04c-4394-85fa-14b57cff7504-ceph" (OuterVolumeSpecName: "ceph") pod "9b974f0b-e04c-4394-85fa-14b57cff7504" (UID: "9b974f0b-e04c-4394-85fa-14b57cff7504"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:18:58 crc kubenswrapper[4790]: I1124 15:18:58.560648 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b974f0b-e04c-4394-85fa-14b57cff7504-kube-api-access-w2xm8" (OuterVolumeSpecName: "kube-api-access-w2xm8") pod "9b974f0b-e04c-4394-85fa-14b57cff7504" (UID: "9b974f0b-e04c-4394-85fa-14b57cff7504"). InnerVolumeSpecName "kube-api-access-w2xm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:18:58 crc kubenswrapper[4790]: I1124 15:18:58.586875 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b974f0b-e04c-4394-85fa-14b57cff7504-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "9b974f0b-e04c-4394-85fa-14b57cff7504" (UID: "9b974f0b-e04c-4394-85fa-14b57cff7504"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:18:58 crc kubenswrapper[4790]: I1124 15:18:58.605908 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b974f0b-e04c-4394-85fa-14b57cff7504-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "9b974f0b-e04c-4394-85fa-14b57cff7504" (UID: "9b974f0b-e04c-4394-85fa-14b57cff7504"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:18:58 crc kubenswrapper[4790]: I1124 15:18:58.656401 4790 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/9b974f0b-e04c-4394-85fa-14b57cff7504-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 24 15:18:58 crc kubenswrapper[4790]: I1124 15:18:58.656445 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2xm8\" (UniqueName: \"kubernetes.io/projected/9b974f0b-e04c-4394-85fa-14b57cff7504-kube-api-access-w2xm8\") on node \"crc\" DevicePath \"\"" Nov 24 15:18:58 crc kubenswrapper[4790]: I1124 15:18:58.656457 4790 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9b974f0b-e04c-4394-85fa-14b57cff7504-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 15:18:58 crc kubenswrapper[4790]: I1124 15:18:58.656469 4790 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/9b974f0b-e04c-4394-85fa-14b57cff7504-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Nov 24 15:18:59 crc kubenswrapper[4790]: I1124 15:18:59.084367 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-h9qxq" event={"ID":"9b974f0b-e04c-4394-85fa-14b57cff7504","Type":"ContainerDied","Data":"5a5b8aad23f937a0c59487a72756a877335438c84e3626a725d96990f06fcaee"} Nov 24 15:18:59 crc kubenswrapper[4790]: I1124 15:18:59.084417 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a5b8aad23f937a0c59487a72756a877335438c84e3626a725d96990f06fcaee" Nov 24 15:18:59 crc kubenswrapper[4790]: I1124 15:18:59.084475 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-h9qxq" Nov 24 15:18:59 crc kubenswrapper[4790]: I1124 15:18:59.131616 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-openstack-openstack-cell1-v7fx7"] Nov 24 15:18:59 crc kubenswrapper[4790]: E1124 15:18:59.132147 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b974f0b-e04c-4394-85fa-14b57cff7504" containerName="ssh-known-hosts-openstack" Nov 24 15:18:59 crc kubenswrapper[4790]: I1124 15:18:59.132164 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b974f0b-e04c-4394-85fa-14b57cff7504" containerName="ssh-known-hosts-openstack" Nov 24 15:18:59 crc kubenswrapper[4790]: I1124 15:18:59.132377 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b974f0b-e04c-4394-85fa-14b57cff7504" containerName="ssh-known-hosts-openstack" Nov 24 15:18:59 crc kubenswrapper[4790]: I1124 15:18:59.133179 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-v7fx7" Nov 24 15:18:59 crc kubenswrapper[4790]: I1124 15:18:59.137861 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 15:18:59 crc kubenswrapper[4790]: I1124 15:18:59.138154 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fc497" Nov 24 15:18:59 crc kubenswrapper[4790]: I1124 15:18:59.138191 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 15:18:59 crc kubenswrapper[4790]: I1124 15:18:59.138157 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 15:18:59 crc kubenswrapper[4790]: I1124 15:18:59.154571 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-v7fx7"] Nov 24 15:18:59 crc kubenswrapper[4790]: I1124 15:18:59.270108 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c4b8d2c-3996-4673-9a36-b21c45d58dec-ssh-key\") pod \"run-os-openstack-openstack-cell1-v7fx7\" (UID: \"2c4b8d2c-3996-4673-9a36-b21c45d58dec\") " pod="openstack/run-os-openstack-openstack-cell1-v7fx7" Nov 24 15:18:59 crc kubenswrapper[4790]: I1124 15:18:59.270626 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c4b8d2c-3996-4673-9a36-b21c45d58dec-inventory\") pod \"run-os-openstack-openstack-cell1-v7fx7\" (UID: \"2c4b8d2c-3996-4673-9a36-b21c45d58dec\") " pod="openstack/run-os-openstack-openstack-cell1-v7fx7" Nov 24 15:18:59 crc kubenswrapper[4790]: I1124 15:18:59.270821 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whltc\" (UniqueName: \"kubernetes.io/projected/2c4b8d2c-3996-4673-9a36-b21c45d58dec-kube-api-access-whltc\") pod \"run-os-openstack-openstack-cell1-v7fx7\" (UID: \"2c4b8d2c-3996-4673-9a36-b21c45d58dec\") " pod="openstack/run-os-openstack-openstack-cell1-v7fx7" Nov 24 15:18:59 crc kubenswrapper[4790]: I1124 15:18:59.271063 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2c4b8d2c-3996-4673-9a36-b21c45d58dec-ceph\") pod \"run-os-openstack-openstack-cell1-v7fx7\" (UID: \"2c4b8d2c-3996-4673-9a36-b21c45d58dec\") " pod="openstack/run-os-openstack-openstack-cell1-v7fx7" Nov 24 15:18:59 crc kubenswrapper[4790]: I1124 15:18:59.373068 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2c4b8d2c-3996-4673-9a36-b21c45d58dec-ceph\") pod \"run-os-openstack-openstack-cell1-v7fx7\" (UID: \"2c4b8d2c-3996-4673-9a36-b21c45d58dec\") " pod="openstack/run-os-openstack-openstack-cell1-v7fx7" Nov 24 15:18:59 crc kubenswrapper[4790]: I1124 15:18:59.373365 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c4b8d2c-3996-4673-9a36-b21c45d58dec-ssh-key\") pod \"run-os-openstack-openstack-cell1-v7fx7\" (UID: \"2c4b8d2c-3996-4673-9a36-b21c45d58dec\") " pod="openstack/run-os-openstack-openstack-cell1-v7fx7" Nov 24 15:18:59 crc kubenswrapper[4790]: I1124 15:18:59.373409 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c4b8d2c-3996-4673-9a36-b21c45d58dec-inventory\") pod \"run-os-openstack-openstack-cell1-v7fx7\" (UID: \"2c4b8d2c-3996-4673-9a36-b21c45d58dec\") " pod="openstack/run-os-openstack-openstack-cell1-v7fx7" Nov 24 15:18:59 crc kubenswrapper[4790]: I1124 15:18:59.373518 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whltc\" (UniqueName: \"kubernetes.io/projected/2c4b8d2c-3996-4673-9a36-b21c45d58dec-kube-api-access-whltc\") pod \"run-os-openstack-openstack-cell1-v7fx7\" (UID: \"2c4b8d2c-3996-4673-9a36-b21c45d58dec\") " pod="openstack/run-os-openstack-openstack-cell1-v7fx7" Nov 24 15:18:59 crc kubenswrapper[4790]: I1124 15:18:59.380925 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c4b8d2c-3996-4673-9a36-b21c45d58dec-inventory\") pod \"run-os-openstack-openstack-cell1-v7fx7\" (UID: \"2c4b8d2c-3996-4673-9a36-b21c45d58dec\") " pod="openstack/run-os-openstack-openstack-cell1-v7fx7" Nov 24 15:18:59 crc kubenswrapper[4790]: I1124 15:18:59.381220 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2c4b8d2c-3996-4673-9a36-b21c45d58dec-ceph\") pod \"run-os-openstack-openstack-cell1-v7fx7\" (UID: \"2c4b8d2c-3996-4673-9a36-b21c45d58dec\") " pod="openstack/run-os-openstack-openstack-cell1-v7fx7" Nov 24 15:18:59 crc kubenswrapper[4790]: I1124 15:18:59.382698 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c4b8d2c-3996-4673-9a36-b21c45d58dec-ssh-key\") pod \"run-os-openstack-openstack-cell1-v7fx7\" (UID: \"2c4b8d2c-3996-4673-9a36-b21c45d58dec\") " pod="openstack/run-os-openstack-openstack-cell1-v7fx7" Nov 24 15:18:59 crc kubenswrapper[4790]: I1124 15:18:59.391076 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whltc\" (UniqueName: \"kubernetes.io/projected/2c4b8d2c-3996-4673-9a36-b21c45d58dec-kube-api-access-whltc\") pod \"run-os-openstack-openstack-cell1-v7fx7\" (UID: \"2c4b8d2c-3996-4673-9a36-b21c45d58dec\") " pod="openstack/run-os-openstack-openstack-cell1-v7fx7" Nov 24 15:18:59 crc kubenswrapper[4790]: I1124 15:18:59.479164 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-v7fx7" Nov 24 15:19:00 crc kubenswrapper[4790]: I1124 15:19:00.081528 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-v7fx7"] Nov 24 15:19:01 crc kubenswrapper[4790]: I1124 15:19:01.112040 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-v7fx7" event={"ID":"2c4b8d2c-3996-4673-9a36-b21c45d58dec","Type":"ContainerStarted","Data":"8193f0352da309c069fb737a57686d9edb3e25f8acd5e1c087760dd55d08fe25"} Nov 24 15:19:01 crc kubenswrapper[4790]: I1124 15:19:01.112990 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-v7fx7" event={"ID":"2c4b8d2c-3996-4673-9a36-b21c45d58dec","Type":"ContainerStarted","Data":"d224a89cba958f6b404a95c7fc8a1619e1a5d3163712d62066d9a70bb5df1e1c"} Nov 24 15:19:04 crc kubenswrapper[4790]: I1124 15:19:04.598946 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-openstack-openstack-cell1-v7fx7" podStartSLOduration=5.412278506 podStartE2EDuration="5.598919695s" podCreationTimestamp="2025-11-24 15:18:59 +0000 UTC" firstStartedPulling="2025-11-24 15:19:00.093290651 +0000 UTC m=+7588.473184313" lastFinishedPulling="2025-11-24 15:19:00.27993183 +0000 UTC m=+7588.659825502" observedRunningTime="2025-11-24 15:19:01.134787895 +0000 UTC m=+7589.514681557" watchObservedRunningTime="2025-11-24 15:19:04.598919695 +0000 UTC m=+7592.978813357" Nov 24 15:19:04 crc kubenswrapper[4790]: I1124 15:19:04.607801 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-l4m79"] Nov 24 15:19:04 crc kubenswrapper[4790]: I1124 15:19:04.611714 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l4m79" Nov 24 15:19:04 crc kubenswrapper[4790]: I1124 15:19:04.622100 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l4m79"] Nov 24 15:19:04 crc kubenswrapper[4790]: I1124 15:19:04.712558 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a00262f8-d9ac-41db-9e38-9bcb935227b4-catalog-content\") pod \"community-operators-l4m79\" (UID: \"a00262f8-d9ac-41db-9e38-9bcb935227b4\") " pod="openshift-marketplace/community-operators-l4m79" Nov 24 15:19:04 crc kubenswrapper[4790]: I1124 15:19:04.712625 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fm8bw\" (UniqueName: \"kubernetes.io/projected/a00262f8-d9ac-41db-9e38-9bcb935227b4-kube-api-access-fm8bw\") pod \"community-operators-l4m79\" (UID: \"a00262f8-d9ac-41db-9e38-9bcb935227b4\") " pod="openshift-marketplace/community-operators-l4m79" Nov 24 15:19:04 crc kubenswrapper[4790]: I1124 15:19:04.712781 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a00262f8-d9ac-41db-9e38-9bcb935227b4-utilities\") pod \"community-operators-l4m79\" (UID: \"a00262f8-d9ac-41db-9e38-9bcb935227b4\") " pod="openshift-marketplace/community-operators-l4m79" Nov 24 15:19:04 crc kubenswrapper[4790]: I1124 15:19:04.815085 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a00262f8-d9ac-41db-9e38-9bcb935227b4-utilities\") pod \"community-operators-l4m79\" (UID: \"a00262f8-d9ac-41db-9e38-9bcb935227b4\") " pod="openshift-marketplace/community-operators-l4m79" Nov 24 15:19:04 crc kubenswrapper[4790]: I1124 15:19:04.815154 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a00262f8-d9ac-41db-9e38-9bcb935227b4-catalog-content\") pod \"community-operators-l4m79\" (UID: \"a00262f8-d9ac-41db-9e38-9bcb935227b4\") " pod="openshift-marketplace/community-operators-l4m79" Nov 24 15:19:04 crc kubenswrapper[4790]: I1124 15:19:04.815213 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fm8bw\" (UniqueName: \"kubernetes.io/projected/a00262f8-d9ac-41db-9e38-9bcb935227b4-kube-api-access-fm8bw\") pod \"community-operators-l4m79\" (UID: \"a00262f8-d9ac-41db-9e38-9bcb935227b4\") " pod="openshift-marketplace/community-operators-l4m79" Nov 24 15:19:04 crc kubenswrapper[4790]: I1124 15:19:04.816088 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a00262f8-d9ac-41db-9e38-9bcb935227b4-utilities\") pod \"community-operators-l4m79\" (UID: \"a00262f8-d9ac-41db-9e38-9bcb935227b4\") " pod="openshift-marketplace/community-operators-l4m79" Nov 24 15:19:04 crc kubenswrapper[4790]: I1124 15:19:04.816390 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a00262f8-d9ac-41db-9e38-9bcb935227b4-catalog-content\") pod \"community-operators-l4m79\" (UID: \"a00262f8-d9ac-41db-9e38-9bcb935227b4\") " pod="openshift-marketplace/community-operators-l4m79" Nov 24 15:19:04 crc kubenswrapper[4790]: I1124 15:19:04.835994 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fm8bw\" (UniqueName: \"kubernetes.io/projected/a00262f8-d9ac-41db-9e38-9bcb935227b4-kube-api-access-fm8bw\") pod \"community-operators-l4m79\" (UID: \"a00262f8-d9ac-41db-9e38-9bcb935227b4\") " pod="openshift-marketplace/community-operators-l4m79" Nov 24 15:19:04 crc kubenswrapper[4790]: I1124 15:19:04.956353 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l4m79" Nov 24 15:19:05 crc kubenswrapper[4790]: I1124 15:19:05.534865 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l4m79"] Nov 24 15:19:06 crc kubenswrapper[4790]: I1124 15:19:06.182693 4790 generic.go:334] "Generic (PLEG): container finished" podID="a00262f8-d9ac-41db-9e38-9bcb935227b4" containerID="1d4ce50ba805064f66621e89e0a7c5325d5a6661fd055aeb3097760dd7cbf0a0" exitCode=0 Nov 24 15:19:06 crc kubenswrapper[4790]: I1124 15:19:06.183773 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l4m79" event={"ID":"a00262f8-d9ac-41db-9e38-9bcb935227b4","Type":"ContainerDied","Data":"1d4ce50ba805064f66621e89e0a7c5325d5a6661fd055aeb3097760dd7cbf0a0"} Nov 24 15:19:06 crc kubenswrapper[4790]: I1124 15:19:06.183839 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l4m79" event={"ID":"a00262f8-d9ac-41db-9e38-9bcb935227b4","Type":"ContainerStarted","Data":"8e83188ae44192b917a30fe569098a76219e16673bc8ddb31f78af5f06ac825f"} Nov 24 15:19:07 crc kubenswrapper[4790]: I1124 15:19:07.195511 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l4m79" event={"ID":"a00262f8-d9ac-41db-9e38-9bcb935227b4","Type":"ContainerStarted","Data":"c39b8955f2815e90c10fa948dd42a96afa4de6d94d185e137dc4c95b026af1cd"} Nov 24 15:19:08 crc kubenswrapper[4790]: I1124 15:19:08.314821 4790 scope.go:117] "RemoveContainer" containerID="ed032daa21005f5603fc87ffbe2ed53274be2bbefd5cbb4153e64e56dd530b81" Nov 24 15:19:08 crc kubenswrapper[4790]: E1124 15:19:08.315438 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:19:09 crc kubenswrapper[4790]: I1124 15:19:09.220129 4790 generic.go:334] "Generic (PLEG): container finished" podID="a00262f8-d9ac-41db-9e38-9bcb935227b4" containerID="c39b8955f2815e90c10fa948dd42a96afa4de6d94d185e137dc4c95b026af1cd" exitCode=0 Nov 24 15:19:09 crc kubenswrapper[4790]: I1124 15:19:09.220187 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l4m79" event={"ID":"a00262f8-d9ac-41db-9e38-9bcb935227b4","Type":"ContainerDied","Data":"c39b8955f2815e90c10fa948dd42a96afa4de6d94d185e137dc4c95b026af1cd"} Nov 24 15:19:10 crc kubenswrapper[4790]: I1124 15:19:10.235412 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l4m79" event={"ID":"a00262f8-d9ac-41db-9e38-9bcb935227b4","Type":"ContainerStarted","Data":"f771096b2bbcae92c66773e47c76660c4530dc7c461a7ef0bfba50b3b8c716f2"} Nov 24 15:19:10 crc kubenswrapper[4790]: I1124 15:19:10.239771 4790 generic.go:334] "Generic (PLEG): container finished" podID="2c4b8d2c-3996-4673-9a36-b21c45d58dec" containerID="8193f0352da309c069fb737a57686d9edb3e25f8acd5e1c087760dd55d08fe25" exitCode=0 Nov 24 15:19:10 crc kubenswrapper[4790]: I1124 15:19:10.239837 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-v7fx7" event={"ID":"2c4b8d2c-3996-4673-9a36-b21c45d58dec","Type":"ContainerDied","Data":"8193f0352da309c069fb737a57686d9edb3e25f8acd5e1c087760dd55d08fe25"} Nov 24 15:19:10 crc kubenswrapper[4790]: I1124 15:19:10.264411 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-l4m79" podStartSLOduration=2.594751911 podStartE2EDuration="6.264385973s" podCreationTimestamp="2025-11-24 15:19:04 +0000 UTC" firstStartedPulling="2025-11-24 15:19:06.187207145 +0000 UTC m=+7594.567100807" lastFinishedPulling="2025-11-24 15:19:09.856841167 +0000 UTC m=+7598.236734869" observedRunningTime="2025-11-24 15:19:10.262100038 +0000 UTC m=+7598.641993760" watchObservedRunningTime="2025-11-24 15:19:10.264385973 +0000 UTC m=+7598.644279675" Nov 24 15:19:11 crc kubenswrapper[4790]: I1124 15:19:11.761692 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-v7fx7" Nov 24 15:19:11 crc kubenswrapper[4790]: I1124 15:19:11.878348 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c4b8d2c-3996-4673-9a36-b21c45d58dec-ssh-key\") pod \"2c4b8d2c-3996-4673-9a36-b21c45d58dec\" (UID: \"2c4b8d2c-3996-4673-9a36-b21c45d58dec\") " Nov 24 15:19:11 crc kubenswrapper[4790]: I1124 15:19:11.878599 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c4b8d2c-3996-4673-9a36-b21c45d58dec-inventory\") pod \"2c4b8d2c-3996-4673-9a36-b21c45d58dec\" (UID: \"2c4b8d2c-3996-4673-9a36-b21c45d58dec\") " Nov 24 15:19:11 crc kubenswrapper[4790]: I1124 15:19:11.878669 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whltc\" (UniqueName: \"kubernetes.io/projected/2c4b8d2c-3996-4673-9a36-b21c45d58dec-kube-api-access-whltc\") pod \"2c4b8d2c-3996-4673-9a36-b21c45d58dec\" (UID: \"2c4b8d2c-3996-4673-9a36-b21c45d58dec\") " Nov 24 15:19:11 crc kubenswrapper[4790]: I1124 15:19:11.878771 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2c4b8d2c-3996-4673-9a36-b21c45d58dec-ceph\") pod \"2c4b8d2c-3996-4673-9a36-b21c45d58dec\" (UID: \"2c4b8d2c-3996-4673-9a36-b21c45d58dec\") " Nov 24 15:19:11 crc kubenswrapper[4790]: I1124 15:19:11.883992 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c4b8d2c-3996-4673-9a36-b21c45d58dec-kube-api-access-whltc" (OuterVolumeSpecName: "kube-api-access-whltc") pod "2c4b8d2c-3996-4673-9a36-b21c45d58dec" (UID: "2c4b8d2c-3996-4673-9a36-b21c45d58dec"). InnerVolumeSpecName "kube-api-access-whltc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:19:11 crc kubenswrapper[4790]: I1124 15:19:11.884597 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c4b8d2c-3996-4673-9a36-b21c45d58dec-ceph" (OuterVolumeSpecName: "ceph") pod "2c4b8d2c-3996-4673-9a36-b21c45d58dec" (UID: "2c4b8d2c-3996-4673-9a36-b21c45d58dec"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:19:11 crc kubenswrapper[4790]: I1124 15:19:11.910288 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c4b8d2c-3996-4673-9a36-b21c45d58dec-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2c4b8d2c-3996-4673-9a36-b21c45d58dec" (UID: "2c4b8d2c-3996-4673-9a36-b21c45d58dec"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:19:11 crc kubenswrapper[4790]: I1124 15:19:11.922004 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c4b8d2c-3996-4673-9a36-b21c45d58dec-inventory" (OuterVolumeSpecName: "inventory") pod "2c4b8d2c-3996-4673-9a36-b21c45d58dec" (UID: "2c4b8d2c-3996-4673-9a36-b21c45d58dec"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:19:11 crc kubenswrapper[4790]: I1124 15:19:11.981312 4790 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c4b8d2c-3996-4673-9a36-b21c45d58dec-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 15:19:11 crc kubenswrapper[4790]: I1124 15:19:11.981353 4790 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c4b8d2c-3996-4673-9a36-b21c45d58dec-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 15:19:11 crc kubenswrapper[4790]: I1124 15:19:11.981362 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whltc\" (UniqueName: \"kubernetes.io/projected/2c4b8d2c-3996-4673-9a36-b21c45d58dec-kube-api-access-whltc\") on node \"crc\" DevicePath \"\"" Nov 24 15:19:11 crc kubenswrapper[4790]: I1124 15:19:11.981372 4790 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2c4b8d2c-3996-4673-9a36-b21c45d58dec-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 15:19:12 crc kubenswrapper[4790]: I1124 15:19:12.266129 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-v7fx7" event={"ID":"2c4b8d2c-3996-4673-9a36-b21c45d58dec","Type":"ContainerDied","Data":"d224a89cba958f6b404a95c7fc8a1619e1a5d3163712d62066d9a70bb5df1e1c"} Nov 24 15:19:12 crc kubenswrapper[4790]: I1124 15:19:12.266193 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d224a89cba958f6b404a95c7fc8a1619e1a5d3163712d62066d9a70bb5df1e1c" Nov 24 15:19:12 crc kubenswrapper[4790]: I1124 15:19:12.266304 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-v7fx7" Nov 24 15:19:12 crc kubenswrapper[4790]: I1124 15:19:12.389105 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-2wk4p"] Nov 24 15:19:12 crc kubenswrapper[4790]: E1124 15:19:12.390020 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c4b8d2c-3996-4673-9a36-b21c45d58dec" containerName="run-os-openstack-openstack-cell1" Nov 24 15:19:12 crc kubenswrapper[4790]: I1124 15:19:12.390044 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c4b8d2c-3996-4673-9a36-b21c45d58dec" containerName="run-os-openstack-openstack-cell1" Nov 24 15:19:12 crc kubenswrapper[4790]: I1124 15:19:12.394386 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c4b8d2c-3996-4673-9a36-b21c45d58dec" containerName="run-os-openstack-openstack-cell1" Nov 24 15:19:12 crc kubenswrapper[4790]: I1124 15:19:12.395706 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-2wk4p" Nov 24 15:19:12 crc kubenswrapper[4790]: I1124 15:19:12.399401 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 15:19:12 crc kubenswrapper[4790]: I1124 15:19:12.399415 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fc497" Nov 24 15:19:12 crc kubenswrapper[4790]: I1124 15:19:12.399842 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 15:19:12 crc kubenswrapper[4790]: I1124 15:19:12.400007 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 15:19:12 crc kubenswrapper[4790]: I1124 15:19:12.417334 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-2wk4p"] Nov 24 15:19:12 crc kubenswrapper[4790]: I1124 15:19:12.491613 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6689103-8cb5-494e-995f-36f3a9e1f83c-inventory\") pod \"reboot-os-openstack-openstack-cell1-2wk4p\" (UID: \"f6689103-8cb5-494e-995f-36f3a9e1f83c\") " pod="openstack/reboot-os-openstack-openstack-cell1-2wk4p" Nov 24 15:19:12 crc kubenswrapper[4790]: I1124 15:19:12.491810 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6689103-8cb5-494e-995f-36f3a9e1f83c-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-2wk4p\" (UID: \"f6689103-8cb5-494e-995f-36f3a9e1f83c\") " pod="openstack/reboot-os-openstack-openstack-cell1-2wk4p" Nov 24 15:19:12 crc kubenswrapper[4790]: I1124 15:19:12.494529 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5rqw\" (UniqueName: \"kubernetes.io/projected/f6689103-8cb5-494e-995f-36f3a9e1f83c-kube-api-access-h5rqw\") pod \"reboot-os-openstack-openstack-cell1-2wk4p\" (UID: \"f6689103-8cb5-494e-995f-36f3a9e1f83c\") " pod="openstack/reboot-os-openstack-openstack-cell1-2wk4p" Nov 24 15:19:12 crc kubenswrapper[4790]: I1124 15:19:12.495578 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f6689103-8cb5-494e-995f-36f3a9e1f83c-ceph\") pod \"reboot-os-openstack-openstack-cell1-2wk4p\" (UID: \"f6689103-8cb5-494e-995f-36f3a9e1f83c\") " pod="openstack/reboot-os-openstack-openstack-cell1-2wk4p" Nov 24 15:19:12 crc kubenswrapper[4790]: I1124 15:19:12.597989 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f6689103-8cb5-494e-995f-36f3a9e1f83c-ceph\") pod \"reboot-os-openstack-openstack-cell1-2wk4p\" (UID: \"f6689103-8cb5-494e-995f-36f3a9e1f83c\") " pod="openstack/reboot-os-openstack-openstack-cell1-2wk4p" Nov 24 15:19:12 crc kubenswrapper[4790]: I1124 15:19:12.598072 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6689103-8cb5-494e-995f-36f3a9e1f83c-inventory\") pod \"reboot-os-openstack-openstack-cell1-2wk4p\" (UID: \"f6689103-8cb5-494e-995f-36f3a9e1f83c\") " pod="openstack/reboot-os-openstack-openstack-cell1-2wk4p" Nov 24 15:19:12 crc kubenswrapper[4790]: I1124 15:19:12.598112 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6689103-8cb5-494e-995f-36f3a9e1f83c-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-2wk4p\" (UID: \"f6689103-8cb5-494e-995f-36f3a9e1f83c\") " pod="openstack/reboot-os-openstack-openstack-cell1-2wk4p" Nov 24 15:19:12 crc kubenswrapper[4790]: I1124 15:19:12.598195 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5rqw\" (UniqueName: \"kubernetes.io/projected/f6689103-8cb5-494e-995f-36f3a9e1f83c-kube-api-access-h5rqw\") pod \"reboot-os-openstack-openstack-cell1-2wk4p\" (UID: \"f6689103-8cb5-494e-995f-36f3a9e1f83c\") " pod="openstack/reboot-os-openstack-openstack-cell1-2wk4p" Nov 24 15:19:12 crc kubenswrapper[4790]: I1124 15:19:12.611296 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6689103-8cb5-494e-995f-36f3a9e1f83c-inventory\") pod \"reboot-os-openstack-openstack-cell1-2wk4p\" (UID: \"f6689103-8cb5-494e-995f-36f3a9e1f83c\") " pod="openstack/reboot-os-openstack-openstack-cell1-2wk4p" Nov 24 15:19:12 crc kubenswrapper[4790]: I1124 15:19:12.611301 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6689103-8cb5-494e-995f-36f3a9e1f83c-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-2wk4p\" (UID: \"f6689103-8cb5-494e-995f-36f3a9e1f83c\") " pod="openstack/reboot-os-openstack-openstack-cell1-2wk4p" Nov 24 15:19:12 crc kubenswrapper[4790]: I1124 15:19:12.611659 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f6689103-8cb5-494e-995f-36f3a9e1f83c-ceph\") pod \"reboot-os-openstack-openstack-cell1-2wk4p\" (UID: \"f6689103-8cb5-494e-995f-36f3a9e1f83c\") " pod="openstack/reboot-os-openstack-openstack-cell1-2wk4p" Nov 24 15:19:12 crc kubenswrapper[4790]: I1124 15:19:12.614468 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5rqw\" (UniqueName: \"kubernetes.io/projected/f6689103-8cb5-494e-995f-36f3a9e1f83c-kube-api-access-h5rqw\") pod \"reboot-os-openstack-openstack-cell1-2wk4p\" (UID: \"f6689103-8cb5-494e-995f-36f3a9e1f83c\") " pod="openstack/reboot-os-openstack-openstack-cell1-2wk4p" Nov 24 15:19:12 crc kubenswrapper[4790]: I1124 15:19:12.722354 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-2wk4p" Nov 24 15:19:13 crc kubenswrapper[4790]: I1124 15:19:13.441520 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-2wk4p"] Nov 24 15:19:13 crc kubenswrapper[4790]: W1124 15:19:13.444078 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6689103_8cb5_494e_995f_36f3a9e1f83c.slice/crio-c53005ae536f0f364538d45b5427b12ab57da6af2cbc0aba93dc049fb15adfe7 WatchSource:0}: Error finding container c53005ae536f0f364538d45b5427b12ab57da6af2cbc0aba93dc049fb15adfe7: Status 404 returned error can't find the container with id c53005ae536f0f364538d45b5427b12ab57da6af2cbc0aba93dc049fb15adfe7 Nov 24 15:19:14 crc kubenswrapper[4790]: I1124 15:19:14.337507 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-2wk4p" event={"ID":"f6689103-8cb5-494e-995f-36f3a9e1f83c","Type":"ContainerStarted","Data":"efe3c7ad8c6caed2a2abfcf2354ca76afa976806ed4d203a7d5e67e1e33d800a"} Nov 24 15:19:14 crc kubenswrapper[4790]: I1124 15:19:14.338674 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-2wk4p" event={"ID":"f6689103-8cb5-494e-995f-36f3a9e1f83c","Type":"ContainerStarted","Data":"c53005ae536f0f364538d45b5427b12ab57da6af2cbc0aba93dc049fb15adfe7"} Nov 24 15:19:14 crc kubenswrapper[4790]: I1124 15:19:14.362299 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-openstack-openstack-cell1-2wk4p" podStartSLOduration=2.178882161 podStartE2EDuration="2.362271767s" podCreationTimestamp="2025-11-24 15:19:12 +0000 UTC" firstStartedPulling="2025-11-24 15:19:13.447196937 +0000 UTC m=+7601.827090609" lastFinishedPulling="2025-11-24 15:19:13.630586513 +0000 UTC m=+7602.010480215" observedRunningTime="2025-11-24 15:19:14.352946863 +0000 UTC m=+7602.732840565" watchObservedRunningTime="2025-11-24 15:19:14.362271767 +0000 UTC m=+7602.742165469" Nov 24 15:19:14 crc kubenswrapper[4790]: I1124 15:19:14.957394 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-l4m79" Nov 24 15:19:14 crc kubenswrapper[4790]: I1124 15:19:14.957508 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-l4m79" Nov 24 15:19:15 crc kubenswrapper[4790]: I1124 15:19:15.067443 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-l4m79" Nov 24 15:19:15 crc kubenswrapper[4790]: I1124 15:19:15.435874 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-l4m79" Nov 24 15:19:15 crc kubenswrapper[4790]: I1124 15:19:15.514142 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l4m79"] Nov 24 15:19:17 crc kubenswrapper[4790]: I1124 15:19:17.373658 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-l4m79" podUID="a00262f8-d9ac-41db-9e38-9bcb935227b4" containerName="registry-server" containerID="cri-o://f771096b2bbcae92c66773e47c76660c4530dc7c461a7ef0bfba50b3b8c716f2" gracePeriod=2 Nov 24 15:19:17 crc kubenswrapper[4790]: I1124 15:19:17.942184 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l4m79" Nov 24 15:19:18 crc kubenswrapper[4790]: I1124 15:19:18.035464 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a00262f8-d9ac-41db-9e38-9bcb935227b4-utilities\") pod \"a00262f8-d9ac-41db-9e38-9bcb935227b4\" (UID: \"a00262f8-d9ac-41db-9e38-9bcb935227b4\") " Nov 24 15:19:18 crc kubenswrapper[4790]: I1124 15:19:18.035654 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a00262f8-d9ac-41db-9e38-9bcb935227b4-catalog-content\") pod \"a00262f8-d9ac-41db-9e38-9bcb935227b4\" (UID: \"a00262f8-d9ac-41db-9e38-9bcb935227b4\") " Nov 24 15:19:18 crc kubenswrapper[4790]: I1124 15:19:18.035723 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fm8bw\" (UniqueName: \"kubernetes.io/projected/a00262f8-d9ac-41db-9e38-9bcb935227b4-kube-api-access-fm8bw\") pod \"a00262f8-d9ac-41db-9e38-9bcb935227b4\" (UID: \"a00262f8-d9ac-41db-9e38-9bcb935227b4\") " Nov 24 15:19:18 crc kubenswrapper[4790]: I1124 15:19:18.040304 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a00262f8-d9ac-41db-9e38-9bcb935227b4-utilities" (OuterVolumeSpecName: "utilities") pod "a00262f8-d9ac-41db-9e38-9bcb935227b4" (UID: "a00262f8-d9ac-41db-9e38-9bcb935227b4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:19:18 crc kubenswrapper[4790]: I1124 15:19:18.042902 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a00262f8-d9ac-41db-9e38-9bcb935227b4-kube-api-access-fm8bw" (OuterVolumeSpecName: "kube-api-access-fm8bw") pod "a00262f8-d9ac-41db-9e38-9bcb935227b4" (UID: "a00262f8-d9ac-41db-9e38-9bcb935227b4"). InnerVolumeSpecName "kube-api-access-fm8bw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:19:18 crc kubenswrapper[4790]: I1124 15:19:18.089005 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a00262f8-d9ac-41db-9e38-9bcb935227b4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a00262f8-d9ac-41db-9e38-9bcb935227b4" (UID: "a00262f8-d9ac-41db-9e38-9bcb935227b4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:19:18 crc kubenswrapper[4790]: I1124 15:19:18.138599 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a00262f8-d9ac-41db-9e38-9bcb935227b4-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 15:19:18 crc kubenswrapper[4790]: I1124 15:19:18.138628 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a00262f8-d9ac-41db-9e38-9bcb935227b4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 15:19:18 crc kubenswrapper[4790]: I1124 15:19:18.138639 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fm8bw\" (UniqueName: \"kubernetes.io/projected/a00262f8-d9ac-41db-9e38-9bcb935227b4-kube-api-access-fm8bw\") on node \"crc\" DevicePath \"\"" Nov 24 15:19:18 crc kubenswrapper[4790]: I1124 15:19:18.390043 4790 generic.go:334] "Generic (PLEG): container finished" podID="a00262f8-d9ac-41db-9e38-9bcb935227b4" containerID="f771096b2bbcae92c66773e47c76660c4530dc7c461a7ef0bfba50b3b8c716f2" exitCode=0 Nov 24 15:19:18 crc kubenswrapper[4790]: I1124 15:19:18.390105 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l4m79" event={"ID":"a00262f8-d9ac-41db-9e38-9bcb935227b4","Type":"ContainerDied","Data":"f771096b2bbcae92c66773e47c76660c4530dc7c461a7ef0bfba50b3b8c716f2"} Nov 24 15:19:18 crc kubenswrapper[4790]: I1124 15:19:18.390143 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l4m79" event={"ID":"a00262f8-d9ac-41db-9e38-9bcb935227b4","Type":"ContainerDied","Data":"8e83188ae44192b917a30fe569098a76219e16673bc8ddb31f78af5f06ac825f"} Nov 24 15:19:18 crc kubenswrapper[4790]: I1124 15:19:18.390191 4790 scope.go:117] "RemoveContainer" containerID="f771096b2bbcae92c66773e47c76660c4530dc7c461a7ef0bfba50b3b8c716f2" Nov 24 15:19:18 crc kubenswrapper[4790]: I1124 15:19:18.390385 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l4m79" Nov 24 15:19:18 crc kubenswrapper[4790]: I1124 15:19:18.428703 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l4m79"] Nov 24 15:19:18 crc kubenswrapper[4790]: I1124 15:19:18.438310 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-l4m79"] Nov 24 15:19:18 crc kubenswrapper[4790]: I1124 15:19:18.442520 4790 scope.go:117] "RemoveContainer" containerID="c39b8955f2815e90c10fa948dd42a96afa4de6d94d185e137dc4c95b026af1cd" Nov 24 15:19:18 crc kubenswrapper[4790]: I1124 15:19:18.495390 4790 scope.go:117] "RemoveContainer" containerID="1d4ce50ba805064f66621e89e0a7c5325d5a6661fd055aeb3097760dd7cbf0a0" Nov 24 15:19:18 crc kubenswrapper[4790]: I1124 15:19:18.533947 4790 scope.go:117] "RemoveContainer" containerID="f771096b2bbcae92c66773e47c76660c4530dc7c461a7ef0bfba50b3b8c716f2" Nov 24 15:19:18 crc kubenswrapper[4790]: E1124 15:19:18.534350 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f771096b2bbcae92c66773e47c76660c4530dc7c461a7ef0bfba50b3b8c716f2\": container with ID starting with f771096b2bbcae92c66773e47c76660c4530dc7c461a7ef0bfba50b3b8c716f2 not found: ID does not exist" containerID="f771096b2bbcae92c66773e47c76660c4530dc7c461a7ef0bfba50b3b8c716f2" Nov 24 15:19:18 crc kubenswrapper[4790]: I1124 15:19:18.534388 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f771096b2bbcae92c66773e47c76660c4530dc7c461a7ef0bfba50b3b8c716f2"} err="failed to get container status \"f771096b2bbcae92c66773e47c76660c4530dc7c461a7ef0bfba50b3b8c716f2\": rpc error: code = NotFound desc = could not find container \"f771096b2bbcae92c66773e47c76660c4530dc7c461a7ef0bfba50b3b8c716f2\": container with ID starting with f771096b2bbcae92c66773e47c76660c4530dc7c461a7ef0bfba50b3b8c716f2 not found: ID does not exist" Nov 24 15:19:18 crc kubenswrapper[4790]: I1124 15:19:18.534415 4790 scope.go:117] "RemoveContainer" containerID="c39b8955f2815e90c10fa948dd42a96afa4de6d94d185e137dc4c95b026af1cd" Nov 24 15:19:18 crc kubenswrapper[4790]: E1124 15:19:18.534737 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c39b8955f2815e90c10fa948dd42a96afa4de6d94d185e137dc4c95b026af1cd\": container with ID starting with c39b8955f2815e90c10fa948dd42a96afa4de6d94d185e137dc4c95b026af1cd not found: ID does not exist" containerID="c39b8955f2815e90c10fa948dd42a96afa4de6d94d185e137dc4c95b026af1cd" Nov 24 15:19:18 crc kubenswrapper[4790]: I1124 15:19:18.534791 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c39b8955f2815e90c10fa948dd42a96afa4de6d94d185e137dc4c95b026af1cd"} err="failed to get container status \"c39b8955f2815e90c10fa948dd42a96afa4de6d94d185e137dc4c95b026af1cd\": rpc error: code = NotFound desc = could not find container \"c39b8955f2815e90c10fa948dd42a96afa4de6d94d185e137dc4c95b026af1cd\": container with ID starting with c39b8955f2815e90c10fa948dd42a96afa4de6d94d185e137dc4c95b026af1cd not found: ID does not exist" Nov 24 15:19:18 crc kubenswrapper[4790]: I1124 15:19:18.534829 4790 scope.go:117] "RemoveContainer" containerID="1d4ce50ba805064f66621e89e0a7c5325d5a6661fd055aeb3097760dd7cbf0a0" Nov 24 15:19:18 crc kubenswrapper[4790]: E1124 15:19:18.536258 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d4ce50ba805064f66621e89e0a7c5325d5a6661fd055aeb3097760dd7cbf0a0\": container with ID starting with 1d4ce50ba805064f66621e89e0a7c5325d5a6661fd055aeb3097760dd7cbf0a0 not found: ID does not exist" containerID="1d4ce50ba805064f66621e89e0a7c5325d5a6661fd055aeb3097760dd7cbf0a0" Nov 24 15:19:18 crc kubenswrapper[4790]: I1124 15:19:18.536287 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d4ce50ba805064f66621e89e0a7c5325d5a6661fd055aeb3097760dd7cbf0a0"} err="failed to get container status \"1d4ce50ba805064f66621e89e0a7c5325d5a6661fd055aeb3097760dd7cbf0a0\": rpc error: code = NotFound desc = could not find container \"1d4ce50ba805064f66621e89e0a7c5325d5a6661fd055aeb3097760dd7cbf0a0\": container with ID starting with 1d4ce50ba805064f66621e89e0a7c5325d5a6661fd055aeb3097760dd7cbf0a0 not found: ID does not exist" Nov 24 15:19:19 crc kubenswrapper[4790]: I1124 15:19:19.315672 4790 scope.go:117] "RemoveContainer" containerID="ed032daa21005f5603fc87ffbe2ed53274be2bbefd5cbb4153e64e56dd530b81" Nov 24 15:19:19 crc kubenswrapper[4790]: E1124 15:19:19.316359 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:19:20 crc kubenswrapper[4790]: I1124 15:19:20.334155 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a00262f8-d9ac-41db-9e38-9bcb935227b4" path="/var/lib/kubelet/pods/a00262f8-d9ac-41db-9e38-9bcb935227b4/volumes" Nov 24 15:19:30 crc kubenswrapper[4790]: I1124 15:19:30.555143 4790 generic.go:334] "Generic (PLEG): container finished" podID="f6689103-8cb5-494e-995f-36f3a9e1f83c" containerID="efe3c7ad8c6caed2a2abfcf2354ca76afa976806ed4d203a7d5e67e1e33d800a" exitCode=0 Nov 24 15:19:30 crc kubenswrapper[4790]: I1124 15:19:30.555229 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-2wk4p" event={"ID":"f6689103-8cb5-494e-995f-36f3a9e1f83c","Type":"ContainerDied","Data":"efe3c7ad8c6caed2a2abfcf2354ca76afa976806ed4d203a7d5e67e1e33d800a"} Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.104311 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-2wk4p" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.186589 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6689103-8cb5-494e-995f-36f3a9e1f83c-inventory\") pod \"f6689103-8cb5-494e-995f-36f3a9e1f83c\" (UID: \"f6689103-8cb5-494e-995f-36f3a9e1f83c\") " Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.187091 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f6689103-8cb5-494e-995f-36f3a9e1f83c-ceph\") pod \"f6689103-8cb5-494e-995f-36f3a9e1f83c\" (UID: \"f6689103-8cb5-494e-995f-36f3a9e1f83c\") " Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.187152 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5rqw\" (UniqueName: \"kubernetes.io/projected/f6689103-8cb5-494e-995f-36f3a9e1f83c-kube-api-access-h5rqw\") pod \"f6689103-8cb5-494e-995f-36f3a9e1f83c\" (UID: \"f6689103-8cb5-494e-995f-36f3a9e1f83c\") " Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.187249 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6689103-8cb5-494e-995f-36f3a9e1f83c-ssh-key\") pod \"f6689103-8cb5-494e-995f-36f3a9e1f83c\" (UID: \"f6689103-8cb5-494e-995f-36f3a9e1f83c\") " Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.193399 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6689103-8cb5-494e-995f-36f3a9e1f83c-ceph" (OuterVolumeSpecName: "ceph") pod "f6689103-8cb5-494e-995f-36f3a9e1f83c" (UID: "f6689103-8cb5-494e-995f-36f3a9e1f83c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.193719 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6689103-8cb5-494e-995f-36f3a9e1f83c-kube-api-access-h5rqw" (OuterVolumeSpecName: "kube-api-access-h5rqw") pod "f6689103-8cb5-494e-995f-36f3a9e1f83c" (UID: "f6689103-8cb5-494e-995f-36f3a9e1f83c"). InnerVolumeSpecName "kube-api-access-h5rqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.217607 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6689103-8cb5-494e-995f-36f3a9e1f83c-inventory" (OuterVolumeSpecName: "inventory") pod "f6689103-8cb5-494e-995f-36f3a9e1f83c" (UID: "f6689103-8cb5-494e-995f-36f3a9e1f83c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.220964 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6689103-8cb5-494e-995f-36f3a9e1f83c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f6689103-8cb5-494e-995f-36f3a9e1f83c" (UID: "f6689103-8cb5-494e-995f-36f3a9e1f83c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.296966 4790 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6689103-8cb5-494e-995f-36f3a9e1f83c-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.297316 4790 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f6689103-8cb5-494e-995f-36f3a9e1f83c-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.297485 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5rqw\" (UniqueName: \"kubernetes.io/projected/f6689103-8cb5-494e-995f-36f3a9e1f83c-kube-api-access-h5rqw\") on node \"crc\" DevicePath \"\"" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.297617 4790 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6689103-8cb5-494e-995f-36f3a9e1f83c-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.580222 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-2wk4p" event={"ID":"f6689103-8cb5-494e-995f-36f3a9e1f83c","Type":"ContainerDied","Data":"c53005ae536f0f364538d45b5427b12ab57da6af2cbc0aba93dc049fb15adfe7"} Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.580275 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c53005ae536f0f364538d45b5427b12ab57da6af2cbc0aba93dc049fb15adfe7" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.580340 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-2wk4p" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.764946 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-xrpsr"] Nov 24 15:19:32 crc kubenswrapper[4790]: E1124 15:19:32.765501 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a00262f8-d9ac-41db-9e38-9bcb935227b4" containerName="extract-content" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.765522 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="a00262f8-d9ac-41db-9e38-9bcb935227b4" containerName="extract-content" Nov 24 15:19:32 crc kubenswrapper[4790]: E1124 15:19:32.765557 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a00262f8-d9ac-41db-9e38-9bcb935227b4" containerName="extract-utilities" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.765567 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="a00262f8-d9ac-41db-9e38-9bcb935227b4" containerName="extract-utilities" Nov 24 15:19:32 crc kubenswrapper[4790]: E1124 15:19:32.765589 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a00262f8-d9ac-41db-9e38-9bcb935227b4" containerName="registry-server" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.765597 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="a00262f8-d9ac-41db-9e38-9bcb935227b4" containerName="registry-server" Nov 24 15:19:32 crc kubenswrapper[4790]: E1124 15:19:32.765612 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6689103-8cb5-494e-995f-36f3a9e1f83c" containerName="reboot-os-openstack-openstack-cell1" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.765621 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6689103-8cb5-494e-995f-36f3a9e1f83c" containerName="reboot-os-openstack-openstack-cell1" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.765930 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6689103-8cb5-494e-995f-36f3a9e1f83c" containerName="reboot-os-openstack-openstack-cell1" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.765969 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="a00262f8-d9ac-41db-9e38-9bcb935227b4" containerName="registry-server" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.766968 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.769280 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fc497" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.770971 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.771367 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.772368 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.780165 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-xrpsr"] Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.912001 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7zm6\" (UniqueName: \"kubernetes.io/projected/08f58547-542f-4a8f-a1a5-c433c2747d70-kube-api-access-x7zm6\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.912068 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.912094 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-inventory\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.912249 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.912489 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.912603 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.912640 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.912810 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-ceph\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.912961 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.913491 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-ssh-key\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.913997 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:32 crc kubenswrapper[4790]: I1124 15:19:32.914041 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:33 crc kubenswrapper[4790]: I1124 15:19:33.015668 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-ssh-key\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:33 crc kubenswrapper[4790]: I1124 15:19:33.016138 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:33 crc kubenswrapper[4790]: I1124 15:19:33.016164 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:33 crc kubenswrapper[4790]: I1124 15:19:33.016228 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7zm6\" (UniqueName: \"kubernetes.io/projected/08f58547-542f-4a8f-a1a5-c433c2747d70-kube-api-access-x7zm6\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:33 crc kubenswrapper[4790]: I1124 15:19:33.016267 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:33 crc kubenswrapper[4790]: I1124 15:19:33.016290 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-inventory\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:33 crc kubenswrapper[4790]: I1124 15:19:33.016314 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:33 crc kubenswrapper[4790]: I1124 15:19:33.016360 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:33 crc kubenswrapper[4790]: I1124 15:19:33.016396 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:33 crc kubenswrapper[4790]: I1124 15:19:33.016422 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:33 crc kubenswrapper[4790]: I1124 15:19:33.016489 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-ceph\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:33 crc kubenswrapper[4790]: I1124 15:19:33.016525 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:33 crc kubenswrapper[4790]: I1124 15:19:33.023387 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-inventory\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:33 crc kubenswrapper[4790]: I1124 15:19:33.024537 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:33 crc kubenswrapper[4790]: I1124 15:19:33.026053 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:33 crc kubenswrapper[4790]: I1124 15:19:33.026775 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-ssh-key\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:33 crc kubenswrapper[4790]: I1124 15:19:33.026993 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:33 crc kubenswrapper[4790]: I1124 15:19:33.028052 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:33 crc kubenswrapper[4790]: I1124 15:19:33.029462 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-ceph\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:33 crc kubenswrapper[4790]: I1124 15:19:33.030801 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:33 crc kubenswrapper[4790]: I1124 15:19:33.033337 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:33 crc kubenswrapper[4790]: I1124 15:19:33.033844 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:33 crc kubenswrapper[4790]: I1124 15:19:33.047364 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7zm6\" (UniqueName: \"kubernetes.io/projected/08f58547-542f-4a8f-a1a5-c433c2747d70-kube-api-access-x7zm6\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:33 crc kubenswrapper[4790]: I1124 15:19:33.065193 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-xrpsr\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:33 crc kubenswrapper[4790]: I1124 15:19:33.088516 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:33 crc kubenswrapper[4790]: I1124 15:19:33.315587 4790 scope.go:117] "RemoveContainer" containerID="ed032daa21005f5603fc87ffbe2ed53274be2bbefd5cbb4153e64e56dd530b81" Nov 24 15:19:33 crc kubenswrapper[4790]: E1124 15:19:33.316585 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:19:33 crc kubenswrapper[4790]: I1124 15:19:33.763594 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-xrpsr"] Nov 24 15:19:33 crc kubenswrapper[4790]: W1124 15:19:33.769307 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08f58547_542f_4a8f_a1a5_c433c2747d70.slice/crio-acc1f8ccee20a2a3dcac75a831e203efd4ca875fc290c02537fa7adf0b465561 WatchSource:0}: Error finding container acc1f8ccee20a2a3dcac75a831e203efd4ca875fc290c02537fa7adf0b465561: Status 404 returned error can't find the container with id acc1f8ccee20a2a3dcac75a831e203efd4ca875fc290c02537fa7adf0b465561 Nov 24 15:19:34 crc kubenswrapper[4790]: I1124 15:19:34.599375 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" event={"ID":"08f58547-542f-4a8f-a1a5-c433c2747d70","Type":"ContainerStarted","Data":"e311579a46e267d62d0614b117f0af99190965002a357b7af5e1beb3f52b0b4c"} Nov 24 15:19:34 crc kubenswrapper[4790]: I1124 15:19:34.599970 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" event={"ID":"08f58547-542f-4a8f-a1a5-c433c2747d70","Type":"ContainerStarted","Data":"acc1f8ccee20a2a3dcac75a831e203efd4ca875fc290c02537fa7adf0b465561"} Nov 24 15:19:39 crc kubenswrapper[4790]: E1124 15:19:39.679463 4790 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6689103_8cb5_494e_995f_36f3a9e1f83c.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6689103_8cb5_494e_995f_36f3a9e1f83c.slice/crio-c53005ae536f0f364538d45b5427b12ab57da6af2cbc0aba93dc049fb15adfe7\": RecentStats: unable to find data in memory cache]" Nov 24 15:19:48 crc kubenswrapper[4790]: I1124 15:19:48.315600 4790 scope.go:117] "RemoveContainer" containerID="ed032daa21005f5603fc87ffbe2ed53274be2bbefd5cbb4153e64e56dd530b81" Nov 24 15:19:48 crc kubenswrapper[4790]: E1124 15:19:48.317033 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:19:50 crc kubenswrapper[4790]: E1124 15:19:50.001102 4790 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6689103_8cb5_494e_995f_36f3a9e1f83c.slice/crio-c53005ae536f0f364538d45b5427b12ab57da6af2cbc0aba93dc049fb15adfe7\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6689103_8cb5_494e_995f_36f3a9e1f83c.slice\": RecentStats: unable to find data in memory cache]" Nov 24 15:19:54 crc kubenswrapper[4790]: I1124 15:19:54.824997 4790 generic.go:334] "Generic (PLEG): container finished" podID="08f58547-542f-4a8f-a1a5-c433c2747d70" containerID="e311579a46e267d62d0614b117f0af99190965002a357b7af5e1beb3f52b0b4c" exitCode=0 Nov 24 15:19:54 crc kubenswrapper[4790]: I1124 15:19:54.825120 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" event={"ID":"08f58547-542f-4a8f-a1a5-c433c2747d70","Type":"ContainerDied","Data":"e311579a46e267d62d0614b117f0af99190965002a357b7af5e1beb3f52b0b4c"} Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.433698 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.542088 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zm6\" (UniqueName: \"kubernetes.io/projected/08f58547-542f-4a8f-a1a5-c433c2747d70-kube-api-access-x7zm6\") pod \"08f58547-542f-4a8f-a1a5-c433c2747d70\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.542209 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-neutron-metadata-combined-ca-bundle\") pod \"08f58547-542f-4a8f-a1a5-c433c2747d70\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.542279 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-neutron-dhcp-combined-ca-bundle\") pod \"08f58547-542f-4a8f-a1a5-c433c2747d70\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.542380 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-inventory\") pod \"08f58547-542f-4a8f-a1a5-c433c2747d70\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.543071 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-nova-combined-ca-bundle\") pod \"08f58547-542f-4a8f-a1a5-c433c2747d70\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.543198 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-bootstrap-combined-ca-bundle\") pod \"08f58547-542f-4a8f-a1a5-c433c2747d70\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.543376 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-ssh-key\") pod \"08f58547-542f-4a8f-a1a5-c433c2747d70\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.543485 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-libvirt-combined-ca-bundle\") pod \"08f58547-542f-4a8f-a1a5-c433c2747d70\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.543602 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-neutron-sriov-combined-ca-bundle\") pod \"08f58547-542f-4a8f-a1a5-c433c2747d70\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.544184 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-ovn-combined-ca-bundle\") pod \"08f58547-542f-4a8f-a1a5-c433c2747d70\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.544390 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-telemetry-combined-ca-bundle\") pod \"08f58547-542f-4a8f-a1a5-c433c2747d70\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.544484 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-ceph\") pod \"08f58547-542f-4a8f-a1a5-c433c2747d70\" (UID: \"08f58547-542f-4a8f-a1a5-c433c2747d70\") " Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.549146 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "08f58547-542f-4a8f-a1a5-c433c2747d70" (UID: "08f58547-542f-4a8f-a1a5-c433c2747d70"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.549794 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "08f58547-542f-4a8f-a1a5-c433c2747d70" (UID: "08f58547-542f-4a8f-a1a5-c433c2747d70"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.550921 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "08f58547-542f-4a8f-a1a5-c433c2747d70" (UID: "08f58547-542f-4a8f-a1a5-c433c2747d70"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.551007 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "08f58547-542f-4a8f-a1a5-c433c2747d70" (UID: "08f58547-542f-4a8f-a1a5-c433c2747d70"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.551021 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08f58547-542f-4a8f-a1a5-c433c2747d70-kube-api-access-x7zm6" (OuterVolumeSpecName: "kube-api-access-x7zm6") pod "08f58547-542f-4a8f-a1a5-c433c2747d70" (UID: "08f58547-542f-4a8f-a1a5-c433c2747d70"). InnerVolumeSpecName "kube-api-access-x7zm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.551514 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "08f58547-542f-4a8f-a1a5-c433c2747d70" (UID: "08f58547-542f-4a8f-a1a5-c433c2747d70"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.553583 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-ceph" (OuterVolumeSpecName: "ceph") pod "08f58547-542f-4a8f-a1a5-c433c2747d70" (UID: "08f58547-542f-4a8f-a1a5-c433c2747d70"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.553705 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "08f58547-542f-4a8f-a1a5-c433c2747d70" (UID: "08f58547-542f-4a8f-a1a5-c433c2747d70"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.554826 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "08f58547-542f-4a8f-a1a5-c433c2747d70" (UID: "08f58547-542f-4a8f-a1a5-c433c2747d70"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.563960 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "08f58547-542f-4a8f-a1a5-c433c2747d70" (UID: "08f58547-542f-4a8f-a1a5-c433c2747d70"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.595139 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "08f58547-542f-4a8f-a1a5-c433c2747d70" (UID: "08f58547-542f-4a8f-a1a5-c433c2747d70"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.595640 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-inventory" (OuterVolumeSpecName: "inventory") pod "08f58547-542f-4a8f-a1a5-c433c2747d70" (UID: "08f58547-542f-4a8f-a1a5-c433c2747d70"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.649049 4790 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.649333 4790 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.649342 4790 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.649351 4790 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.649360 4790 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.649369 4790 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.649381 4790 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.649393 4790 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.649402 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zm6\" (UniqueName: \"kubernetes.io/projected/08f58547-542f-4a8f-a1a5-c433c2747d70-kube-api-access-x7zm6\") on node \"crc\" DevicePath \"\"" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.649413 4790 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.649421 4790 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.649430 4790 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08f58547-542f-4a8f-a1a5-c433c2747d70-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.848396 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" event={"ID":"08f58547-542f-4a8f-a1a5-c433c2747d70","Type":"ContainerDied","Data":"acc1f8ccee20a2a3dcac75a831e203efd4ca875fc290c02537fa7adf0b465561"} Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.848436 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="acc1f8ccee20a2a3dcac75a831e203efd4ca875fc290c02537fa7adf0b465561" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.848514 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-xrpsr" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.959582 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-4mkwj"] Nov 24 15:19:56 crc kubenswrapper[4790]: E1124 15:19:56.960028 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08f58547-542f-4a8f-a1a5-c433c2747d70" containerName="install-certs-openstack-openstack-cell1" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.960044 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="08f58547-542f-4a8f-a1a5-c433c2747d70" containerName="install-certs-openstack-openstack-cell1" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.960271 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="08f58547-542f-4a8f-a1a5-c433c2747d70" containerName="install-certs-openstack-openstack-cell1" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.961070 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-4mkwj" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.970355 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.970589 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.970728 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.970869 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fc497" Nov 24 15:19:56 crc kubenswrapper[4790]: I1124 15:19:56.980872 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-4mkwj"] Nov 24 15:19:57 crc kubenswrapper[4790]: I1124 15:19:57.059166 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpqqd\" (UniqueName: \"kubernetes.io/projected/31455824-5727-44a5-b2c9-9eb68e20c305-kube-api-access-xpqqd\") pod \"ceph-client-openstack-openstack-cell1-4mkwj\" (UID: \"31455824-5727-44a5-b2c9-9eb68e20c305\") " pod="openstack/ceph-client-openstack-openstack-cell1-4mkwj" Nov 24 15:19:57 crc kubenswrapper[4790]: I1124 15:19:57.059217 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31455824-5727-44a5-b2c9-9eb68e20c305-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-4mkwj\" (UID: \"31455824-5727-44a5-b2c9-9eb68e20c305\") " pod="openstack/ceph-client-openstack-openstack-cell1-4mkwj" Nov 24 15:19:57 crc kubenswrapper[4790]: I1124 15:19:57.059283 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31455824-5727-44a5-b2c9-9eb68e20c305-inventory\") pod \"ceph-client-openstack-openstack-cell1-4mkwj\" (UID: \"31455824-5727-44a5-b2c9-9eb68e20c305\") " pod="openstack/ceph-client-openstack-openstack-cell1-4mkwj" Nov 24 15:19:57 crc kubenswrapper[4790]: I1124 15:19:57.059379 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/31455824-5727-44a5-b2c9-9eb68e20c305-ceph\") pod \"ceph-client-openstack-openstack-cell1-4mkwj\" (UID: \"31455824-5727-44a5-b2c9-9eb68e20c305\") " pod="openstack/ceph-client-openstack-openstack-cell1-4mkwj" Nov 24 15:19:57 crc kubenswrapper[4790]: I1124 15:19:57.160977 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpqqd\" (UniqueName: \"kubernetes.io/projected/31455824-5727-44a5-b2c9-9eb68e20c305-kube-api-access-xpqqd\") pod \"ceph-client-openstack-openstack-cell1-4mkwj\" (UID: \"31455824-5727-44a5-b2c9-9eb68e20c305\") " pod="openstack/ceph-client-openstack-openstack-cell1-4mkwj" Nov 24 15:19:57 crc kubenswrapper[4790]: I1124 15:19:57.161053 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31455824-5727-44a5-b2c9-9eb68e20c305-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-4mkwj\" (UID: \"31455824-5727-44a5-b2c9-9eb68e20c305\") " pod="openstack/ceph-client-openstack-openstack-cell1-4mkwj" Nov 24 15:19:57 crc kubenswrapper[4790]: I1124 15:19:57.161175 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31455824-5727-44a5-b2c9-9eb68e20c305-inventory\") pod \"ceph-client-openstack-openstack-cell1-4mkwj\" (UID: \"31455824-5727-44a5-b2c9-9eb68e20c305\") " pod="openstack/ceph-client-openstack-openstack-cell1-4mkwj" Nov 24 15:19:57 crc kubenswrapper[4790]: I1124 15:19:57.161363 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/31455824-5727-44a5-b2c9-9eb68e20c305-ceph\") pod \"ceph-client-openstack-openstack-cell1-4mkwj\" (UID: \"31455824-5727-44a5-b2c9-9eb68e20c305\") " pod="openstack/ceph-client-openstack-openstack-cell1-4mkwj" Nov 24 15:19:57 crc kubenswrapper[4790]: I1124 15:19:57.165935 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31455824-5727-44a5-b2c9-9eb68e20c305-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-4mkwj\" (UID: \"31455824-5727-44a5-b2c9-9eb68e20c305\") " pod="openstack/ceph-client-openstack-openstack-cell1-4mkwj" Nov 24 15:19:57 crc kubenswrapper[4790]: I1124 15:19:57.173367 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31455824-5727-44a5-b2c9-9eb68e20c305-inventory\") pod \"ceph-client-openstack-openstack-cell1-4mkwj\" (UID: \"31455824-5727-44a5-b2c9-9eb68e20c305\") " pod="openstack/ceph-client-openstack-openstack-cell1-4mkwj" Nov 24 15:19:57 crc kubenswrapper[4790]: I1124 15:19:57.179475 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/31455824-5727-44a5-b2c9-9eb68e20c305-ceph\") pod \"ceph-client-openstack-openstack-cell1-4mkwj\" (UID: \"31455824-5727-44a5-b2c9-9eb68e20c305\") " pod="openstack/ceph-client-openstack-openstack-cell1-4mkwj" Nov 24 15:19:57 crc kubenswrapper[4790]: I1124 15:19:57.191050 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpqqd\" (UniqueName: \"kubernetes.io/projected/31455824-5727-44a5-b2c9-9eb68e20c305-kube-api-access-xpqqd\") pod \"ceph-client-openstack-openstack-cell1-4mkwj\" (UID: \"31455824-5727-44a5-b2c9-9eb68e20c305\") " pod="openstack/ceph-client-openstack-openstack-cell1-4mkwj" Nov 24 15:19:57 crc kubenswrapper[4790]: I1124 15:19:57.282755 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-4mkwj" Nov 24 15:19:57 crc kubenswrapper[4790]: I1124 15:19:57.934438 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-4mkwj"] Nov 24 15:19:57 crc kubenswrapper[4790]: W1124 15:19:57.939698 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31455824_5727_44a5_b2c9_9eb68e20c305.slice/crio-ebc4dec3e5835a766fcffd7eb5ccc5b3b20897dce9ba2e23d92e13c82084353b WatchSource:0}: Error finding container ebc4dec3e5835a766fcffd7eb5ccc5b3b20897dce9ba2e23d92e13c82084353b: Status 404 returned error can't find the container with id ebc4dec3e5835a766fcffd7eb5ccc5b3b20897dce9ba2e23d92e13c82084353b Nov 24 15:19:58 crc kubenswrapper[4790]: I1124 15:19:58.876033 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-4mkwj" event={"ID":"31455824-5727-44a5-b2c9-9eb68e20c305","Type":"ContainerStarted","Data":"c60b13a8384aadbfdefcfe0beec103ca910989798355101c030e190545875780"} Nov 24 15:19:58 crc kubenswrapper[4790]: I1124 15:19:58.877075 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-4mkwj" event={"ID":"31455824-5727-44a5-b2c9-9eb68e20c305","Type":"ContainerStarted","Data":"ebc4dec3e5835a766fcffd7eb5ccc5b3b20897dce9ba2e23d92e13c82084353b"} Nov 24 15:19:58 crc kubenswrapper[4790]: I1124 15:19:58.905631 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-openstack-openstack-cell1-4mkwj" podStartSLOduration=2.712845113 podStartE2EDuration="2.905608245s" podCreationTimestamp="2025-11-24 15:19:56 +0000 UTC" firstStartedPulling="2025-11-24 15:19:57.943505216 +0000 UTC m=+7646.323398888" lastFinishedPulling="2025-11-24 15:19:58.136268348 +0000 UTC m=+7646.516162020" observedRunningTime="2025-11-24 15:19:58.894337036 +0000 UTC m=+7647.274230708" watchObservedRunningTime="2025-11-24 15:19:58.905608245 +0000 UTC m=+7647.285501917" Nov 24 15:19:59 crc kubenswrapper[4790]: I1124 15:19:59.318785 4790 scope.go:117] "RemoveContainer" containerID="ed032daa21005f5603fc87ffbe2ed53274be2bbefd5cbb4153e64e56dd530b81" Nov 24 15:19:59 crc kubenswrapper[4790]: E1124 15:19:59.319255 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:20:00 crc kubenswrapper[4790]: E1124 15:20:00.318522 4790 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6689103_8cb5_494e_995f_36f3a9e1f83c.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6689103_8cb5_494e_995f_36f3a9e1f83c.slice/crio-c53005ae536f0f364538d45b5427b12ab57da6af2cbc0aba93dc049fb15adfe7\": RecentStats: unable to find data in memory cache]" Nov 24 15:20:03 crc kubenswrapper[4790]: I1124 15:20:03.955462 4790 generic.go:334] "Generic (PLEG): container finished" podID="31455824-5727-44a5-b2c9-9eb68e20c305" containerID="c60b13a8384aadbfdefcfe0beec103ca910989798355101c030e190545875780" exitCode=0 Nov 24 15:20:03 crc kubenswrapper[4790]: I1124 15:20:03.955595 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-4mkwj" event={"ID":"31455824-5727-44a5-b2c9-9eb68e20c305","Type":"ContainerDied","Data":"c60b13a8384aadbfdefcfe0beec103ca910989798355101c030e190545875780"} Nov 24 15:20:05 crc kubenswrapper[4790]: I1124 15:20:05.504118 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-4mkwj" Nov 24 15:20:05 crc kubenswrapper[4790]: I1124 15:20:05.552723 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31455824-5727-44a5-b2c9-9eb68e20c305-ssh-key\") pod \"31455824-5727-44a5-b2c9-9eb68e20c305\" (UID: \"31455824-5727-44a5-b2c9-9eb68e20c305\") " Nov 24 15:20:05 crc kubenswrapper[4790]: I1124 15:20:05.552957 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpqqd\" (UniqueName: \"kubernetes.io/projected/31455824-5727-44a5-b2c9-9eb68e20c305-kube-api-access-xpqqd\") pod \"31455824-5727-44a5-b2c9-9eb68e20c305\" (UID: \"31455824-5727-44a5-b2c9-9eb68e20c305\") " Nov 24 15:20:05 crc kubenswrapper[4790]: I1124 15:20:05.553055 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31455824-5727-44a5-b2c9-9eb68e20c305-inventory\") pod \"31455824-5727-44a5-b2c9-9eb68e20c305\" (UID: \"31455824-5727-44a5-b2c9-9eb68e20c305\") " Nov 24 15:20:05 crc kubenswrapper[4790]: I1124 15:20:05.553097 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/31455824-5727-44a5-b2c9-9eb68e20c305-ceph\") pod \"31455824-5727-44a5-b2c9-9eb68e20c305\" (UID: \"31455824-5727-44a5-b2c9-9eb68e20c305\") " Nov 24 15:20:05 crc kubenswrapper[4790]: I1124 15:20:05.562978 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31455824-5727-44a5-b2c9-9eb68e20c305-kube-api-access-xpqqd" (OuterVolumeSpecName: "kube-api-access-xpqqd") pod "31455824-5727-44a5-b2c9-9eb68e20c305" (UID: "31455824-5727-44a5-b2c9-9eb68e20c305"). InnerVolumeSpecName "kube-api-access-xpqqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:20:05 crc kubenswrapper[4790]: I1124 15:20:05.564587 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31455824-5727-44a5-b2c9-9eb68e20c305-ceph" (OuterVolumeSpecName: "ceph") pod "31455824-5727-44a5-b2c9-9eb68e20c305" (UID: "31455824-5727-44a5-b2c9-9eb68e20c305"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:20:05 crc kubenswrapper[4790]: I1124 15:20:05.588018 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31455824-5727-44a5-b2c9-9eb68e20c305-inventory" (OuterVolumeSpecName: "inventory") pod "31455824-5727-44a5-b2c9-9eb68e20c305" (UID: "31455824-5727-44a5-b2c9-9eb68e20c305"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:20:05 crc kubenswrapper[4790]: I1124 15:20:05.592717 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31455824-5727-44a5-b2c9-9eb68e20c305-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "31455824-5727-44a5-b2c9-9eb68e20c305" (UID: "31455824-5727-44a5-b2c9-9eb68e20c305"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:20:05 crc kubenswrapper[4790]: I1124 15:20:05.656762 4790 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31455824-5727-44a5-b2c9-9eb68e20c305-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 15:20:05 crc kubenswrapper[4790]: I1124 15:20:05.657123 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpqqd\" (UniqueName: \"kubernetes.io/projected/31455824-5727-44a5-b2c9-9eb68e20c305-kube-api-access-xpqqd\") on node \"crc\" DevicePath \"\"" Nov 24 15:20:05 crc kubenswrapper[4790]: I1124 15:20:05.657138 4790 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31455824-5727-44a5-b2c9-9eb68e20c305-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 15:20:05 crc kubenswrapper[4790]: I1124 15:20:05.657152 4790 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/31455824-5727-44a5-b2c9-9eb68e20c305-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 15:20:05 crc kubenswrapper[4790]: I1124 15:20:05.987187 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-4mkwj" event={"ID":"31455824-5727-44a5-b2c9-9eb68e20c305","Type":"ContainerDied","Data":"ebc4dec3e5835a766fcffd7eb5ccc5b3b20897dce9ba2e23d92e13c82084353b"} Nov 24 15:20:05 crc kubenswrapper[4790]: I1124 15:20:05.987244 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-4mkwj" Nov 24 15:20:05 crc kubenswrapper[4790]: I1124 15:20:05.987267 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ebc4dec3e5835a766fcffd7eb5ccc5b3b20897dce9ba2e23d92e13c82084353b" Nov 24 15:20:06 crc kubenswrapper[4790]: I1124 15:20:06.086972 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-openstack-openstack-cell1-znrmt"] Nov 24 15:20:06 crc kubenswrapper[4790]: E1124 15:20:06.087637 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31455824-5727-44a5-b2c9-9eb68e20c305" containerName="ceph-client-openstack-openstack-cell1" Nov 24 15:20:06 crc kubenswrapper[4790]: I1124 15:20:06.087658 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="31455824-5727-44a5-b2c9-9eb68e20c305" containerName="ceph-client-openstack-openstack-cell1" Nov 24 15:20:06 crc kubenswrapper[4790]: I1124 15:20:06.087981 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="31455824-5727-44a5-b2c9-9eb68e20c305" containerName="ceph-client-openstack-openstack-cell1" Nov 24 15:20:06 crc kubenswrapper[4790]: I1124 15:20:06.089042 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-znrmt" Nov 24 15:20:06 crc kubenswrapper[4790]: I1124 15:20:06.093412 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 15:20:06 crc kubenswrapper[4790]: I1124 15:20:06.093502 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 15:20:06 crc kubenswrapper[4790]: I1124 15:20:06.093417 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 15:20:06 crc kubenswrapper[4790]: I1124 15:20:06.093734 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Nov 24 15:20:06 crc kubenswrapper[4790]: I1124 15:20:06.095643 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fc497" Nov 24 15:20:06 crc kubenswrapper[4790]: I1124 15:20:06.108822 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-znrmt"] Nov 24 15:20:06 crc kubenswrapper[4790]: I1124 15:20:06.168177 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6036a536-d292-481d-b3f3-22b8ca768c8e-inventory\") pod \"ovn-openstack-openstack-cell1-znrmt\" (UID: \"6036a536-d292-481d-b3f3-22b8ca768c8e\") " pod="openstack/ovn-openstack-openstack-cell1-znrmt" Nov 24 15:20:06 crc kubenswrapper[4790]: I1124 15:20:06.168253 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6036a536-d292-481d-b3f3-22b8ca768c8e-ssh-key\") pod \"ovn-openstack-openstack-cell1-znrmt\" (UID: \"6036a536-d292-481d-b3f3-22b8ca768c8e\") " pod="openstack/ovn-openstack-openstack-cell1-znrmt" Nov 24 15:20:06 crc kubenswrapper[4790]: I1124 15:20:06.168349 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6036a536-d292-481d-b3f3-22b8ca768c8e-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-znrmt\" (UID: \"6036a536-d292-481d-b3f3-22b8ca768c8e\") " pod="openstack/ovn-openstack-openstack-cell1-znrmt" Nov 24 15:20:06 crc kubenswrapper[4790]: I1124 15:20:06.168639 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgw2w\" (UniqueName: \"kubernetes.io/projected/6036a536-d292-481d-b3f3-22b8ca768c8e-kube-api-access-fgw2w\") pod \"ovn-openstack-openstack-cell1-znrmt\" (UID: \"6036a536-d292-481d-b3f3-22b8ca768c8e\") " pod="openstack/ovn-openstack-openstack-cell1-znrmt" Nov 24 15:20:06 crc kubenswrapper[4790]: I1124 15:20:06.168720 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6036a536-d292-481d-b3f3-22b8ca768c8e-ceph\") pod \"ovn-openstack-openstack-cell1-znrmt\" (UID: \"6036a536-d292-481d-b3f3-22b8ca768c8e\") " pod="openstack/ovn-openstack-openstack-cell1-znrmt" Nov 24 15:20:06 crc kubenswrapper[4790]: I1124 15:20:06.168795 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6036a536-d292-481d-b3f3-22b8ca768c8e-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-znrmt\" (UID: \"6036a536-d292-481d-b3f3-22b8ca768c8e\") " pod="openstack/ovn-openstack-openstack-cell1-znrmt" Nov 24 15:20:06 crc kubenswrapper[4790]: I1124 15:20:06.270022 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6036a536-d292-481d-b3f3-22b8ca768c8e-inventory\") pod \"ovn-openstack-openstack-cell1-znrmt\" (UID: \"6036a536-d292-481d-b3f3-22b8ca768c8e\") " pod="openstack/ovn-openstack-openstack-cell1-znrmt" Nov 24 15:20:06 crc kubenswrapper[4790]: I1124 15:20:06.270082 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6036a536-d292-481d-b3f3-22b8ca768c8e-ssh-key\") pod \"ovn-openstack-openstack-cell1-znrmt\" (UID: \"6036a536-d292-481d-b3f3-22b8ca768c8e\") " pod="openstack/ovn-openstack-openstack-cell1-znrmt" Nov 24 15:20:06 crc kubenswrapper[4790]: I1124 15:20:06.270114 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6036a536-d292-481d-b3f3-22b8ca768c8e-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-znrmt\" (UID: \"6036a536-d292-481d-b3f3-22b8ca768c8e\") " pod="openstack/ovn-openstack-openstack-cell1-znrmt" Nov 24 15:20:06 crc kubenswrapper[4790]: I1124 15:20:06.270195 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgw2w\" (UniqueName: \"kubernetes.io/projected/6036a536-d292-481d-b3f3-22b8ca768c8e-kube-api-access-fgw2w\") pod \"ovn-openstack-openstack-cell1-znrmt\" (UID: \"6036a536-d292-481d-b3f3-22b8ca768c8e\") " pod="openstack/ovn-openstack-openstack-cell1-znrmt" Nov 24 15:20:06 crc kubenswrapper[4790]: I1124 15:20:06.270229 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6036a536-d292-481d-b3f3-22b8ca768c8e-ceph\") pod \"ovn-openstack-openstack-cell1-znrmt\" (UID: \"6036a536-d292-481d-b3f3-22b8ca768c8e\") " pod="openstack/ovn-openstack-openstack-cell1-znrmt" Nov 24 15:20:06 crc kubenswrapper[4790]: I1124 15:20:06.270268 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6036a536-d292-481d-b3f3-22b8ca768c8e-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-znrmt\" (UID: \"6036a536-d292-481d-b3f3-22b8ca768c8e\") " pod="openstack/ovn-openstack-openstack-cell1-znrmt" Nov 24 15:20:06 crc kubenswrapper[4790]: I1124 15:20:06.272409 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6036a536-d292-481d-b3f3-22b8ca768c8e-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-znrmt\" (UID: \"6036a536-d292-481d-b3f3-22b8ca768c8e\") " pod="openstack/ovn-openstack-openstack-cell1-znrmt" Nov 24 15:20:06 crc kubenswrapper[4790]: I1124 15:20:06.277286 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6036a536-d292-481d-b3f3-22b8ca768c8e-ssh-key\") pod \"ovn-openstack-openstack-cell1-znrmt\" (UID: \"6036a536-d292-481d-b3f3-22b8ca768c8e\") " pod="openstack/ovn-openstack-openstack-cell1-znrmt" Nov 24 15:20:06 crc kubenswrapper[4790]: I1124 15:20:06.277782 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6036a536-d292-481d-b3f3-22b8ca768c8e-inventory\") pod \"ovn-openstack-openstack-cell1-znrmt\" (UID: \"6036a536-d292-481d-b3f3-22b8ca768c8e\") " pod="openstack/ovn-openstack-openstack-cell1-znrmt" Nov 24 15:20:06 crc kubenswrapper[4790]: I1124 15:20:06.285579 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6036a536-d292-481d-b3f3-22b8ca768c8e-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-znrmt\" (UID: \"6036a536-d292-481d-b3f3-22b8ca768c8e\") " pod="openstack/ovn-openstack-openstack-cell1-znrmt" Nov 24 15:20:06 crc kubenswrapper[4790]: I1124 15:20:06.287558 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6036a536-d292-481d-b3f3-22b8ca768c8e-ceph\") pod \"ovn-openstack-openstack-cell1-znrmt\" (UID: \"6036a536-d292-481d-b3f3-22b8ca768c8e\") " pod="openstack/ovn-openstack-openstack-cell1-znrmt" Nov 24 15:20:06 crc kubenswrapper[4790]: I1124 15:20:06.289517 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgw2w\" (UniqueName: \"kubernetes.io/projected/6036a536-d292-481d-b3f3-22b8ca768c8e-kube-api-access-fgw2w\") pod \"ovn-openstack-openstack-cell1-znrmt\" (UID: \"6036a536-d292-481d-b3f3-22b8ca768c8e\") " pod="openstack/ovn-openstack-openstack-cell1-znrmt" Nov 24 15:20:06 crc kubenswrapper[4790]: I1124 15:20:06.424132 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-znrmt" Nov 24 15:20:06 crc kubenswrapper[4790]: I1124 15:20:06.995142 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-znrmt"] Nov 24 15:20:08 crc kubenswrapper[4790]: I1124 15:20:08.021022 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-znrmt" event={"ID":"6036a536-d292-481d-b3f3-22b8ca768c8e","Type":"ContainerStarted","Data":"37eafcd552232ce2abb820bf337617b02084d67ab06dbda2cceeafe631a8ccad"} Nov 24 15:20:08 crc kubenswrapper[4790]: I1124 15:20:08.022810 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-znrmt" event={"ID":"6036a536-d292-481d-b3f3-22b8ca768c8e","Type":"ContainerStarted","Data":"b1573daaebe3d6d9829c833188dc0c4463a693edb8a84fd3dfccb4f79f599775"} Nov 24 15:20:08 crc kubenswrapper[4790]: I1124 15:20:08.061095 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-openstack-openstack-cell1-znrmt" podStartSLOduration=1.800648459 podStartE2EDuration="2.061074904s" podCreationTimestamp="2025-11-24 15:20:06 +0000 UTC" firstStartedPulling="2025-11-24 15:20:07.004305857 +0000 UTC m=+7655.384199529" lastFinishedPulling="2025-11-24 15:20:07.264732302 +0000 UTC m=+7655.644625974" observedRunningTime="2025-11-24 15:20:08.048818187 +0000 UTC m=+7656.428711859" watchObservedRunningTime="2025-11-24 15:20:08.061074904 +0000 UTC m=+7656.440968576" Nov 24 15:20:10 crc kubenswrapper[4790]: E1124 15:20:10.610831 4790 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6689103_8cb5_494e_995f_36f3a9e1f83c.slice/crio-c53005ae536f0f364538d45b5427b12ab57da6af2cbc0aba93dc049fb15adfe7\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6689103_8cb5_494e_995f_36f3a9e1f83c.slice\": RecentStats: unable to find data in memory cache]" Nov 24 15:20:12 crc kubenswrapper[4790]: I1124 15:20:12.325879 4790 scope.go:117] "RemoveContainer" containerID="ed032daa21005f5603fc87ffbe2ed53274be2bbefd5cbb4153e64e56dd530b81" Nov 24 15:20:12 crc kubenswrapper[4790]: E1124 15:20:12.326923 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:20:20 crc kubenswrapper[4790]: E1124 15:20:20.918005 4790 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6689103_8cb5_494e_995f_36f3a9e1f83c.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6689103_8cb5_494e_995f_36f3a9e1f83c.slice/crio-c53005ae536f0f364538d45b5427b12ab57da6af2cbc0aba93dc049fb15adfe7\": RecentStats: unable to find data in memory cache]" Nov 24 15:20:27 crc kubenswrapper[4790]: I1124 15:20:27.315412 4790 scope.go:117] "RemoveContainer" containerID="ed032daa21005f5603fc87ffbe2ed53274be2bbefd5cbb4153e64e56dd530b81" Nov 24 15:20:27 crc kubenswrapper[4790]: E1124 15:20:27.316719 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:20:31 crc kubenswrapper[4790]: E1124 15:20:31.244185 4790 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6689103_8cb5_494e_995f_36f3a9e1f83c.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6689103_8cb5_494e_995f_36f3a9e1f83c.slice/crio-c53005ae536f0f364538d45b5427b12ab57da6af2cbc0aba93dc049fb15adfe7\": RecentStats: unable to find data in memory cache]" Nov 24 15:20:41 crc kubenswrapper[4790]: I1124 15:20:41.315564 4790 scope.go:117] "RemoveContainer" containerID="ed032daa21005f5603fc87ffbe2ed53274be2bbefd5cbb4153e64e56dd530b81" Nov 24 15:20:41 crc kubenswrapper[4790]: E1124 15:20:41.316470 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:20:54 crc kubenswrapper[4790]: I1124 15:20:54.315798 4790 scope.go:117] "RemoveContainer" containerID="ed032daa21005f5603fc87ffbe2ed53274be2bbefd5cbb4153e64e56dd530b81" Nov 24 15:20:54 crc kubenswrapper[4790]: E1124 15:20:54.316933 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:21:09 crc kubenswrapper[4790]: I1124 15:21:09.314822 4790 scope.go:117] "RemoveContainer" containerID="ed032daa21005f5603fc87ffbe2ed53274be2bbefd5cbb4153e64e56dd530b81" Nov 24 15:21:09 crc kubenswrapper[4790]: E1124 15:21:09.316034 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:21:17 crc kubenswrapper[4790]: I1124 15:21:17.858450 4790 generic.go:334] "Generic (PLEG): container finished" podID="6036a536-d292-481d-b3f3-22b8ca768c8e" containerID="37eafcd552232ce2abb820bf337617b02084d67ab06dbda2cceeafe631a8ccad" exitCode=0 Nov 24 15:21:17 crc kubenswrapper[4790]: I1124 15:21:17.859042 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-znrmt" event={"ID":"6036a536-d292-481d-b3f3-22b8ca768c8e","Type":"ContainerDied","Data":"37eafcd552232ce2abb820bf337617b02084d67ab06dbda2cceeafe631a8ccad"} Nov 24 15:21:19 crc kubenswrapper[4790]: I1124 15:21:19.406211 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-znrmt" Nov 24 15:21:19 crc kubenswrapper[4790]: I1124 15:21:19.533724 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6036a536-d292-481d-b3f3-22b8ca768c8e-ceph\") pod \"6036a536-d292-481d-b3f3-22b8ca768c8e\" (UID: \"6036a536-d292-481d-b3f3-22b8ca768c8e\") " Nov 24 15:21:19 crc kubenswrapper[4790]: I1124 15:21:19.533845 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6036a536-d292-481d-b3f3-22b8ca768c8e-inventory\") pod \"6036a536-d292-481d-b3f3-22b8ca768c8e\" (UID: \"6036a536-d292-481d-b3f3-22b8ca768c8e\") " Nov 24 15:21:19 crc kubenswrapper[4790]: I1124 15:21:19.533992 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6036a536-d292-481d-b3f3-22b8ca768c8e-ovncontroller-config-0\") pod \"6036a536-d292-481d-b3f3-22b8ca768c8e\" (UID: \"6036a536-d292-481d-b3f3-22b8ca768c8e\") " Nov 24 15:21:19 crc kubenswrapper[4790]: I1124 15:21:19.534136 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6036a536-d292-481d-b3f3-22b8ca768c8e-ssh-key\") pod \"6036a536-d292-481d-b3f3-22b8ca768c8e\" (UID: \"6036a536-d292-481d-b3f3-22b8ca768c8e\") " Nov 24 15:21:19 crc kubenswrapper[4790]: I1124 15:21:19.534785 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgw2w\" (UniqueName: \"kubernetes.io/projected/6036a536-d292-481d-b3f3-22b8ca768c8e-kube-api-access-fgw2w\") pod \"6036a536-d292-481d-b3f3-22b8ca768c8e\" (UID: \"6036a536-d292-481d-b3f3-22b8ca768c8e\") " Nov 24 15:21:19 crc kubenswrapper[4790]: I1124 15:21:19.534851 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6036a536-d292-481d-b3f3-22b8ca768c8e-ovn-combined-ca-bundle\") pod \"6036a536-d292-481d-b3f3-22b8ca768c8e\" (UID: \"6036a536-d292-481d-b3f3-22b8ca768c8e\") " Nov 24 15:21:19 crc kubenswrapper[4790]: I1124 15:21:19.539797 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6036a536-d292-481d-b3f3-22b8ca768c8e-ceph" (OuterVolumeSpecName: "ceph") pod "6036a536-d292-481d-b3f3-22b8ca768c8e" (UID: "6036a536-d292-481d-b3f3-22b8ca768c8e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:21:19 crc kubenswrapper[4790]: I1124 15:21:19.539920 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6036a536-d292-481d-b3f3-22b8ca768c8e-kube-api-access-fgw2w" (OuterVolumeSpecName: "kube-api-access-fgw2w") pod "6036a536-d292-481d-b3f3-22b8ca768c8e" (UID: "6036a536-d292-481d-b3f3-22b8ca768c8e"). InnerVolumeSpecName "kube-api-access-fgw2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:21:19 crc kubenswrapper[4790]: I1124 15:21:19.540560 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6036a536-d292-481d-b3f3-22b8ca768c8e-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "6036a536-d292-481d-b3f3-22b8ca768c8e" (UID: "6036a536-d292-481d-b3f3-22b8ca768c8e"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:21:19 crc kubenswrapper[4790]: I1124 15:21:19.568156 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6036a536-d292-481d-b3f3-22b8ca768c8e-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "6036a536-d292-481d-b3f3-22b8ca768c8e" (UID: "6036a536-d292-481d-b3f3-22b8ca768c8e"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 15:21:19 crc kubenswrapper[4790]: I1124 15:21:19.568990 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6036a536-d292-481d-b3f3-22b8ca768c8e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6036a536-d292-481d-b3f3-22b8ca768c8e" (UID: "6036a536-d292-481d-b3f3-22b8ca768c8e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:21:19 crc kubenswrapper[4790]: I1124 15:21:19.572376 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6036a536-d292-481d-b3f3-22b8ca768c8e-inventory" (OuterVolumeSpecName: "inventory") pod "6036a536-d292-481d-b3f3-22b8ca768c8e" (UID: "6036a536-d292-481d-b3f3-22b8ca768c8e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:21:19 crc kubenswrapper[4790]: I1124 15:21:19.637644 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgw2w\" (UniqueName: \"kubernetes.io/projected/6036a536-d292-481d-b3f3-22b8ca768c8e-kube-api-access-fgw2w\") on node \"crc\" DevicePath \"\"" Nov 24 15:21:19 crc kubenswrapper[4790]: I1124 15:21:19.637680 4790 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6036a536-d292-481d-b3f3-22b8ca768c8e-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:21:19 crc kubenswrapper[4790]: I1124 15:21:19.637690 4790 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6036a536-d292-481d-b3f3-22b8ca768c8e-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 15:21:19 crc kubenswrapper[4790]: I1124 15:21:19.637699 4790 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6036a536-d292-481d-b3f3-22b8ca768c8e-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 15:21:19 crc kubenswrapper[4790]: I1124 15:21:19.637711 4790 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6036a536-d292-481d-b3f3-22b8ca768c8e-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 15:21:19 crc kubenswrapper[4790]: I1124 15:21:19.637718 4790 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6036a536-d292-481d-b3f3-22b8ca768c8e-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 15:21:19 crc kubenswrapper[4790]: I1124 15:21:19.888009 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-znrmt" event={"ID":"6036a536-d292-481d-b3f3-22b8ca768c8e","Type":"ContainerDied","Data":"b1573daaebe3d6d9829c833188dc0c4463a693edb8a84fd3dfccb4f79f599775"} Nov 24 15:21:19 crc kubenswrapper[4790]: I1124 15:21:19.888356 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1573daaebe3d6d9829c833188dc0c4463a693edb8a84fd3dfccb4f79f599775" Nov 24 15:21:19 crc kubenswrapper[4790]: I1124 15:21:19.888143 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-znrmt" Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.001943 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-68gxl"] Nov 24 15:21:20 crc kubenswrapper[4790]: E1124 15:21:20.002596 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6036a536-d292-481d-b3f3-22b8ca768c8e" containerName="ovn-openstack-openstack-cell1" Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.002670 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="6036a536-d292-481d-b3f3-22b8ca768c8e" containerName="ovn-openstack-openstack-cell1" Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.002936 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="6036a536-d292-481d-b3f3-22b8ca768c8e" containerName="ovn-openstack-openstack-cell1" Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.005743 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-68gxl" Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.009351 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fc497" Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.009408 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.009670 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.009784 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.009897 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.011624 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.040963 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-68gxl"] Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.147844 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kf6vs\" (UniqueName: \"kubernetes.io/projected/052884d0-d104-44a6-bcf8-86ac5a67b6ff-kube-api-access-kf6vs\") pod \"neutron-metadata-openstack-openstack-cell1-68gxl\" (UID: \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-68gxl" Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.147945 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-68gxl\" (UID: \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-68gxl" Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.147990 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-68gxl\" (UID: \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-68gxl" Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.148108 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-68gxl\" (UID: \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-68gxl" Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.148137 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-68gxl\" (UID: \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-68gxl" Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.148178 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-68gxl\" (UID: \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-68gxl" Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.148278 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-68gxl\" (UID: \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-68gxl" Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.250659 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-68gxl\" (UID: \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-68gxl" Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.250781 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kf6vs\" (UniqueName: \"kubernetes.io/projected/052884d0-d104-44a6-bcf8-86ac5a67b6ff-kube-api-access-kf6vs\") pod \"neutron-metadata-openstack-openstack-cell1-68gxl\" (UID: \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-68gxl" Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.250822 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-68gxl\" (UID: \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-68gxl" Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.250848 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-68gxl\" (UID: \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-68gxl" Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.250965 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-68gxl\" (UID: \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-68gxl" Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.250998 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-68gxl\" (UID: \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-68gxl" Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.251042 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-68gxl\" (UID: \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-68gxl" Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.258912 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-68gxl\" (UID: \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-68gxl" Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.259060 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-68gxl\" (UID: \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-68gxl" Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.259103 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-68gxl\" (UID: \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-68gxl" Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.259242 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-68gxl\" (UID: \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-68gxl" Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.259546 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-68gxl\" (UID: \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-68gxl" Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.262449 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-68gxl\" (UID: \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-68gxl" Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.278142 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kf6vs\" (UniqueName: \"kubernetes.io/projected/052884d0-d104-44a6-bcf8-86ac5a67b6ff-kube-api-access-kf6vs\") pod \"neutron-metadata-openstack-openstack-cell1-68gxl\" (UID: \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-68gxl" Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.343526 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-68gxl" Nov 24 15:21:20 crc kubenswrapper[4790]: I1124 15:21:20.984824 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-68gxl"] Nov 24 15:21:21 crc kubenswrapper[4790]: I1124 15:21:21.915378 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-68gxl" event={"ID":"052884d0-d104-44a6-bcf8-86ac5a67b6ff","Type":"ContainerStarted","Data":"1ab5da592c9b2581cb7e78773c307f35548b1500752f88532c440ae7e86bae03"} Nov 24 15:21:21 crc kubenswrapper[4790]: I1124 15:21:21.915830 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-68gxl" event={"ID":"052884d0-d104-44a6-bcf8-86ac5a67b6ff","Type":"ContainerStarted","Data":"637ca584ae88d92ed41d73b6d9aae22f43b55de70694b41f960fb8647392c260"} Nov 24 15:21:21 crc kubenswrapper[4790]: I1124 15:21:21.934116 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-openstack-openstack-cell1-68gxl" podStartSLOduration=2.612965842 podStartE2EDuration="2.934097204s" podCreationTimestamp="2025-11-24 15:21:19 +0000 UTC" firstStartedPulling="2025-11-24 15:21:20.996909569 +0000 UTC m=+7729.376803231" lastFinishedPulling="2025-11-24 15:21:21.318040921 +0000 UTC m=+7729.697934593" observedRunningTime="2025-11-24 15:21:21.93149695 +0000 UTC m=+7730.311390612" watchObservedRunningTime="2025-11-24 15:21:21.934097204 +0000 UTC m=+7730.313990866" Nov 24 15:21:23 crc kubenswrapper[4790]: I1124 15:21:23.314415 4790 scope.go:117] "RemoveContainer" containerID="ed032daa21005f5603fc87ffbe2ed53274be2bbefd5cbb4153e64e56dd530b81" Nov 24 15:21:23 crc kubenswrapper[4790]: E1124 15:21:23.315750 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:21:38 crc kubenswrapper[4790]: I1124 15:21:38.315600 4790 scope.go:117] "RemoveContainer" containerID="ed032daa21005f5603fc87ffbe2ed53274be2bbefd5cbb4153e64e56dd530b81" Nov 24 15:21:38 crc kubenswrapper[4790]: E1124 15:21:38.316695 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:21:51 crc kubenswrapper[4790]: I1124 15:21:51.316053 4790 scope.go:117] "RemoveContainer" containerID="ed032daa21005f5603fc87ffbe2ed53274be2bbefd5cbb4153e64e56dd530b81" Nov 24 15:21:52 crc kubenswrapper[4790]: I1124 15:21:52.396776 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"7191c6329f823582f9ae00d9da816e469d4f902ed9ef90979520daad145e04e4"} Nov 24 15:22:17 crc kubenswrapper[4790]: I1124 15:22:17.719287 4790 generic.go:334] "Generic (PLEG): container finished" podID="052884d0-d104-44a6-bcf8-86ac5a67b6ff" containerID="1ab5da592c9b2581cb7e78773c307f35548b1500752f88532c440ae7e86bae03" exitCode=0 Nov 24 15:22:17 crc kubenswrapper[4790]: I1124 15:22:17.719380 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-68gxl" event={"ID":"052884d0-d104-44a6-bcf8-86ac5a67b6ff","Type":"ContainerDied","Data":"1ab5da592c9b2581cb7e78773c307f35548b1500752f88532c440ae7e86bae03"} Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.296828 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-68gxl" Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.427731 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-inventory\") pod \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\" (UID: \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\") " Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.427936 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-ssh-key\") pod \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\" (UID: \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\") " Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.427966 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-nova-metadata-neutron-config-0\") pod \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\" (UID: \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\") " Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.428058 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-neutron-ovn-metadata-agent-neutron-config-0\") pod \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\" (UID: \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\") " Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.428132 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-ceph\") pod \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\" (UID: \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\") " Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.428185 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-neutron-metadata-combined-ca-bundle\") pod \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\" (UID: \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\") " Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.428284 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kf6vs\" (UniqueName: \"kubernetes.io/projected/052884d0-d104-44a6-bcf8-86ac5a67b6ff-kube-api-access-kf6vs\") pod \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\" (UID: \"052884d0-d104-44a6-bcf8-86ac5a67b6ff\") " Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.434351 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-ceph" (OuterVolumeSpecName: "ceph") pod "052884d0-d104-44a6-bcf8-86ac5a67b6ff" (UID: "052884d0-d104-44a6-bcf8-86ac5a67b6ff"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.435199 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "052884d0-d104-44a6-bcf8-86ac5a67b6ff" (UID: "052884d0-d104-44a6-bcf8-86ac5a67b6ff"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.436308 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/052884d0-d104-44a6-bcf8-86ac5a67b6ff-kube-api-access-kf6vs" (OuterVolumeSpecName: "kube-api-access-kf6vs") pod "052884d0-d104-44a6-bcf8-86ac5a67b6ff" (UID: "052884d0-d104-44a6-bcf8-86ac5a67b6ff"). InnerVolumeSpecName "kube-api-access-kf6vs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.459163 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "052884d0-d104-44a6-bcf8-86ac5a67b6ff" (UID: "052884d0-d104-44a6-bcf8-86ac5a67b6ff"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.460626 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-inventory" (OuterVolumeSpecName: "inventory") pod "052884d0-d104-44a6-bcf8-86ac5a67b6ff" (UID: "052884d0-d104-44a6-bcf8-86ac5a67b6ff"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.475005 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "052884d0-d104-44a6-bcf8-86ac5a67b6ff" (UID: "052884d0-d104-44a6-bcf8-86ac5a67b6ff"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.487548 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "052884d0-d104-44a6-bcf8-86ac5a67b6ff" (UID: "052884d0-d104-44a6-bcf8-86ac5a67b6ff"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.532491 4790 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.532513 4790 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.532525 4790 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.532536 4790 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.532544 4790 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.532554 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kf6vs\" (UniqueName: \"kubernetes.io/projected/052884d0-d104-44a6-bcf8-86ac5a67b6ff-kube-api-access-kf6vs\") on node \"crc\" DevicePath \"\"" Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.532563 4790 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/052884d0-d104-44a6-bcf8-86ac5a67b6ff-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.748310 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-68gxl" event={"ID":"052884d0-d104-44a6-bcf8-86ac5a67b6ff","Type":"ContainerDied","Data":"637ca584ae88d92ed41d73b6d9aae22f43b55de70694b41f960fb8647392c260"} Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.748365 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="637ca584ae88d92ed41d73b6d9aae22f43b55de70694b41f960fb8647392c260" Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.748403 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-68gxl" Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.852525 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-zgjq6"] Nov 24 15:22:19 crc kubenswrapper[4790]: E1124 15:22:19.853364 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="052884d0-d104-44a6-bcf8-86ac5a67b6ff" containerName="neutron-metadata-openstack-openstack-cell1" Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.853406 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="052884d0-d104-44a6-bcf8-86ac5a67b6ff" containerName="neutron-metadata-openstack-openstack-cell1" Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.853952 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="052884d0-d104-44a6-bcf8-86ac5a67b6ff" containerName="neutron-metadata-openstack-openstack-cell1" Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.855978 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-zgjq6" Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.859239 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.859275 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fc497" Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.859477 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.860210 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.863430 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-zgjq6"] Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.864738 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.940245 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abbde07e-75cb-4db6-9f73-726b7fe5f03c-inventory\") pod \"libvirt-openstack-openstack-cell1-zgjq6\" (UID: \"abbde07e-75cb-4db6-9f73-726b7fe5f03c\") " pod="openstack/libvirt-openstack-openstack-cell1-zgjq6" Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.940341 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abbde07e-75cb-4db6-9f73-726b7fe5f03c-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-zgjq6\" (UID: \"abbde07e-75cb-4db6-9f73-726b7fe5f03c\") " pod="openstack/libvirt-openstack-openstack-cell1-zgjq6" Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.940366 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/abbde07e-75cb-4db6-9f73-726b7fe5f03c-ceph\") pod \"libvirt-openstack-openstack-cell1-zgjq6\" (UID: \"abbde07e-75cb-4db6-9f73-726b7fe5f03c\") " pod="openstack/libvirt-openstack-openstack-cell1-zgjq6" Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.940407 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59zzp\" (UniqueName: \"kubernetes.io/projected/abbde07e-75cb-4db6-9f73-726b7fe5f03c-kube-api-access-59zzp\") pod \"libvirt-openstack-openstack-cell1-zgjq6\" (UID: \"abbde07e-75cb-4db6-9f73-726b7fe5f03c\") " pod="openstack/libvirt-openstack-openstack-cell1-zgjq6" Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.940459 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abbde07e-75cb-4db6-9f73-726b7fe5f03c-ssh-key\") pod \"libvirt-openstack-openstack-cell1-zgjq6\" (UID: \"abbde07e-75cb-4db6-9f73-726b7fe5f03c\") " pod="openstack/libvirt-openstack-openstack-cell1-zgjq6" Nov 24 15:22:19 crc kubenswrapper[4790]: I1124 15:22:19.940823 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/abbde07e-75cb-4db6-9f73-726b7fe5f03c-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-zgjq6\" (UID: \"abbde07e-75cb-4db6-9f73-726b7fe5f03c\") " pod="openstack/libvirt-openstack-openstack-cell1-zgjq6" Nov 24 15:22:20 crc kubenswrapper[4790]: I1124 15:22:20.042934 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59zzp\" (UniqueName: \"kubernetes.io/projected/abbde07e-75cb-4db6-9f73-726b7fe5f03c-kube-api-access-59zzp\") pod \"libvirt-openstack-openstack-cell1-zgjq6\" (UID: \"abbde07e-75cb-4db6-9f73-726b7fe5f03c\") " pod="openstack/libvirt-openstack-openstack-cell1-zgjq6" Nov 24 15:22:20 crc kubenswrapper[4790]: I1124 15:22:20.043107 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abbde07e-75cb-4db6-9f73-726b7fe5f03c-ssh-key\") pod \"libvirt-openstack-openstack-cell1-zgjq6\" (UID: \"abbde07e-75cb-4db6-9f73-726b7fe5f03c\") " pod="openstack/libvirt-openstack-openstack-cell1-zgjq6" Nov 24 15:22:20 crc kubenswrapper[4790]: I1124 15:22:20.043249 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/abbde07e-75cb-4db6-9f73-726b7fe5f03c-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-zgjq6\" (UID: \"abbde07e-75cb-4db6-9f73-726b7fe5f03c\") " pod="openstack/libvirt-openstack-openstack-cell1-zgjq6" Nov 24 15:22:20 crc kubenswrapper[4790]: I1124 15:22:20.043323 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abbde07e-75cb-4db6-9f73-726b7fe5f03c-inventory\") pod \"libvirt-openstack-openstack-cell1-zgjq6\" (UID: \"abbde07e-75cb-4db6-9f73-726b7fe5f03c\") " pod="openstack/libvirt-openstack-openstack-cell1-zgjq6" Nov 24 15:22:20 crc kubenswrapper[4790]: I1124 15:22:20.043471 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abbde07e-75cb-4db6-9f73-726b7fe5f03c-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-zgjq6\" (UID: \"abbde07e-75cb-4db6-9f73-726b7fe5f03c\") " pod="openstack/libvirt-openstack-openstack-cell1-zgjq6" Nov 24 15:22:20 crc kubenswrapper[4790]: I1124 15:22:20.043525 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/abbde07e-75cb-4db6-9f73-726b7fe5f03c-ceph\") pod \"libvirt-openstack-openstack-cell1-zgjq6\" (UID: \"abbde07e-75cb-4db6-9f73-726b7fe5f03c\") " pod="openstack/libvirt-openstack-openstack-cell1-zgjq6" Nov 24 15:22:20 crc kubenswrapper[4790]: I1124 15:22:20.047871 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/abbde07e-75cb-4db6-9f73-726b7fe5f03c-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-zgjq6\" (UID: \"abbde07e-75cb-4db6-9f73-726b7fe5f03c\") " pod="openstack/libvirt-openstack-openstack-cell1-zgjq6" Nov 24 15:22:20 crc kubenswrapper[4790]: I1124 15:22:20.048179 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abbde07e-75cb-4db6-9f73-726b7fe5f03c-inventory\") pod \"libvirt-openstack-openstack-cell1-zgjq6\" (UID: \"abbde07e-75cb-4db6-9f73-726b7fe5f03c\") " pod="openstack/libvirt-openstack-openstack-cell1-zgjq6" Nov 24 15:22:20 crc kubenswrapper[4790]: I1124 15:22:20.048775 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/abbde07e-75cb-4db6-9f73-726b7fe5f03c-ceph\") pod \"libvirt-openstack-openstack-cell1-zgjq6\" (UID: \"abbde07e-75cb-4db6-9f73-726b7fe5f03c\") " pod="openstack/libvirt-openstack-openstack-cell1-zgjq6" Nov 24 15:22:20 crc kubenswrapper[4790]: I1124 15:22:20.050417 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abbde07e-75cb-4db6-9f73-726b7fe5f03c-ssh-key\") pod \"libvirt-openstack-openstack-cell1-zgjq6\" (UID: \"abbde07e-75cb-4db6-9f73-726b7fe5f03c\") " pod="openstack/libvirt-openstack-openstack-cell1-zgjq6" Nov 24 15:22:20 crc kubenswrapper[4790]: I1124 15:22:20.057564 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abbde07e-75cb-4db6-9f73-726b7fe5f03c-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-zgjq6\" (UID: \"abbde07e-75cb-4db6-9f73-726b7fe5f03c\") " pod="openstack/libvirt-openstack-openstack-cell1-zgjq6" Nov 24 15:22:20 crc kubenswrapper[4790]: I1124 15:22:20.058687 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59zzp\" (UniqueName: \"kubernetes.io/projected/abbde07e-75cb-4db6-9f73-726b7fe5f03c-kube-api-access-59zzp\") pod \"libvirt-openstack-openstack-cell1-zgjq6\" (UID: \"abbde07e-75cb-4db6-9f73-726b7fe5f03c\") " pod="openstack/libvirt-openstack-openstack-cell1-zgjq6" Nov 24 15:22:20 crc kubenswrapper[4790]: I1124 15:22:20.191038 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-zgjq6" Nov 24 15:22:20 crc kubenswrapper[4790]: I1124 15:22:20.834992 4790 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 15:22:20 crc kubenswrapper[4790]: I1124 15:22:20.843171 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-zgjq6"] Nov 24 15:22:21 crc kubenswrapper[4790]: I1124 15:22:21.771760 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-zgjq6" event={"ID":"abbde07e-75cb-4db6-9f73-726b7fe5f03c","Type":"ContainerStarted","Data":"39bced2b024c1d0cf88624982a13db7876622a2008d5d0da7ccf8a8360ea0eb9"} Nov 24 15:22:21 crc kubenswrapper[4790]: I1124 15:22:21.772338 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-zgjq6" event={"ID":"abbde07e-75cb-4db6-9f73-726b7fe5f03c","Type":"ContainerStarted","Data":"50c8f3dbead07db34e598721d09e1c60a331e8acb89379b01e965906702db764"} Nov 24 15:24:13 crc kubenswrapper[4790]: I1124 15:24:13.938444 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:24:13 crc kubenswrapper[4790]: I1124 15:24:13.939383 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:24:43 crc kubenswrapper[4790]: I1124 15:24:43.939207 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:24:43 crc kubenswrapper[4790]: I1124 15:24:43.940212 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:25:13 crc kubenswrapper[4790]: I1124 15:25:13.938441 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:25:13 crc kubenswrapper[4790]: I1124 15:25:13.938816 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:25:13 crc kubenswrapper[4790]: I1124 15:25:13.938868 4790 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 15:25:13 crc kubenswrapper[4790]: I1124 15:25:13.939811 4790 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7191c6329f823582f9ae00d9da816e469d4f902ed9ef90979520daad145e04e4"} pod="openshift-machine-config-operator/machine-config-daemon-xz49t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 15:25:13 crc kubenswrapper[4790]: I1124 15:25:13.939872 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" containerID="cri-o://7191c6329f823582f9ae00d9da816e469d4f902ed9ef90979520daad145e04e4" gracePeriod=600 Nov 24 15:25:14 crc kubenswrapper[4790]: I1124 15:25:14.077443 4790 generic.go:334] "Generic (PLEG): container finished" podID="0d73b133-48f0-455f-8f6a-742e633f631a" containerID="7191c6329f823582f9ae00d9da816e469d4f902ed9ef90979520daad145e04e4" exitCode=0 Nov 24 15:25:14 crc kubenswrapper[4790]: I1124 15:25:14.077559 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerDied","Data":"7191c6329f823582f9ae00d9da816e469d4f902ed9ef90979520daad145e04e4"} Nov 24 15:25:14 crc kubenswrapper[4790]: I1124 15:25:14.077865 4790 scope.go:117] "RemoveContainer" containerID="ed032daa21005f5603fc87ffbe2ed53274be2bbefd5cbb4153e64e56dd530b81" Nov 24 15:25:15 crc kubenswrapper[4790]: I1124 15:25:15.094149 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"87c151c65c6ec7e6e0150da1b299aebe7cb33ee5e3738cff003c27747ab5c270"} Nov 24 15:25:15 crc kubenswrapper[4790]: I1124 15:25:15.113158 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-openstack-openstack-cell1-zgjq6" podStartSLOduration=175.88830939 podStartE2EDuration="2m56.113143719s" podCreationTimestamp="2025-11-24 15:22:19 +0000 UTC" firstStartedPulling="2025-11-24 15:22:20.834686724 +0000 UTC m=+7789.214580386" lastFinishedPulling="2025-11-24 15:22:21.059521053 +0000 UTC m=+7789.439414715" observedRunningTime="2025-11-24 15:22:21.793055998 +0000 UTC m=+7790.172949660" watchObservedRunningTime="2025-11-24 15:25:15.113143719 +0000 UTC m=+7963.493037381" Nov 24 15:27:20 crc kubenswrapper[4790]: I1124 15:27:20.505674 4790 generic.go:334] "Generic (PLEG): container finished" podID="abbde07e-75cb-4db6-9f73-726b7fe5f03c" containerID="39bced2b024c1d0cf88624982a13db7876622a2008d5d0da7ccf8a8360ea0eb9" exitCode=0 Nov 24 15:27:20 crc kubenswrapper[4790]: I1124 15:27:20.505842 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-zgjq6" event={"ID":"abbde07e-75cb-4db6-9f73-726b7fe5f03c","Type":"ContainerDied","Data":"39bced2b024c1d0cf88624982a13db7876622a2008d5d0da7ccf8a8360ea0eb9"} Nov 24 15:27:21 crc kubenswrapper[4790]: I1124 15:27:21.709014 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pn5gv"] Nov 24 15:27:21 crc kubenswrapper[4790]: I1124 15:27:21.712637 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pn5gv" Nov 24 15:27:21 crc kubenswrapper[4790]: I1124 15:27:21.732447 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pn5gv"] Nov 24 15:27:21 crc kubenswrapper[4790]: I1124 15:27:21.856344 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4-catalog-content\") pod \"redhat-operators-pn5gv\" (UID: \"6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4\") " pod="openshift-marketplace/redhat-operators-pn5gv" Nov 24 15:27:21 crc kubenswrapper[4790]: I1124 15:27:21.858006 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4-utilities\") pod \"redhat-operators-pn5gv\" (UID: \"6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4\") " pod="openshift-marketplace/redhat-operators-pn5gv" Nov 24 15:27:21 crc kubenswrapper[4790]: I1124 15:27:21.858113 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9pl5\" (UniqueName: \"kubernetes.io/projected/6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4-kube-api-access-c9pl5\") pod \"redhat-operators-pn5gv\" (UID: \"6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4\") " pod="openshift-marketplace/redhat-operators-pn5gv" Nov 24 15:27:21 crc kubenswrapper[4790]: I1124 15:27:21.959598 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4-catalog-content\") pod \"redhat-operators-pn5gv\" (UID: \"6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4\") " pod="openshift-marketplace/redhat-operators-pn5gv" Nov 24 15:27:21 crc kubenswrapper[4790]: I1124 15:27:21.959670 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4-utilities\") pod \"redhat-operators-pn5gv\" (UID: \"6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4\") " pod="openshift-marketplace/redhat-operators-pn5gv" Nov 24 15:27:21 crc kubenswrapper[4790]: I1124 15:27:21.959737 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9pl5\" (UniqueName: \"kubernetes.io/projected/6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4-kube-api-access-c9pl5\") pod \"redhat-operators-pn5gv\" (UID: \"6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4\") " pod="openshift-marketplace/redhat-operators-pn5gv" Nov 24 15:27:21 crc kubenswrapper[4790]: I1124 15:27:21.960266 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4-catalog-content\") pod \"redhat-operators-pn5gv\" (UID: \"6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4\") " pod="openshift-marketplace/redhat-operators-pn5gv" Nov 24 15:27:21 crc kubenswrapper[4790]: I1124 15:27:21.960326 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4-utilities\") pod \"redhat-operators-pn5gv\" (UID: \"6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4\") " pod="openshift-marketplace/redhat-operators-pn5gv" Nov 24 15:27:21 crc kubenswrapper[4790]: I1124 15:27:21.983256 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9pl5\" (UniqueName: \"kubernetes.io/projected/6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4-kube-api-access-c9pl5\") pod \"redhat-operators-pn5gv\" (UID: \"6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4\") " pod="openshift-marketplace/redhat-operators-pn5gv" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.038779 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pn5gv" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.125346 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-zgjq6" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.265419 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abbde07e-75cb-4db6-9f73-726b7fe5f03c-libvirt-combined-ca-bundle\") pod \"abbde07e-75cb-4db6-9f73-726b7fe5f03c\" (UID: \"abbde07e-75cb-4db6-9f73-726b7fe5f03c\") " Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.265562 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/abbde07e-75cb-4db6-9f73-726b7fe5f03c-libvirt-secret-0\") pod \"abbde07e-75cb-4db6-9f73-726b7fe5f03c\" (UID: \"abbde07e-75cb-4db6-9f73-726b7fe5f03c\") " Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.265604 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abbde07e-75cb-4db6-9f73-726b7fe5f03c-inventory\") pod \"abbde07e-75cb-4db6-9f73-726b7fe5f03c\" (UID: \"abbde07e-75cb-4db6-9f73-726b7fe5f03c\") " Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.265758 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abbde07e-75cb-4db6-9f73-726b7fe5f03c-ssh-key\") pod \"abbde07e-75cb-4db6-9f73-726b7fe5f03c\" (UID: \"abbde07e-75cb-4db6-9f73-726b7fe5f03c\") " Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.265876 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/abbde07e-75cb-4db6-9f73-726b7fe5f03c-ceph\") pod \"abbde07e-75cb-4db6-9f73-726b7fe5f03c\" (UID: \"abbde07e-75cb-4db6-9f73-726b7fe5f03c\") " Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.266039 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59zzp\" (UniqueName: \"kubernetes.io/projected/abbde07e-75cb-4db6-9f73-726b7fe5f03c-kube-api-access-59zzp\") pod \"abbde07e-75cb-4db6-9f73-726b7fe5f03c\" (UID: \"abbde07e-75cb-4db6-9f73-726b7fe5f03c\") " Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.272228 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abbde07e-75cb-4db6-9f73-726b7fe5f03c-ceph" (OuterVolumeSpecName: "ceph") pod "abbde07e-75cb-4db6-9f73-726b7fe5f03c" (UID: "abbde07e-75cb-4db6-9f73-726b7fe5f03c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.275213 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abbde07e-75cb-4db6-9f73-726b7fe5f03c-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "abbde07e-75cb-4db6-9f73-726b7fe5f03c" (UID: "abbde07e-75cb-4db6-9f73-726b7fe5f03c"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.275260 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abbde07e-75cb-4db6-9f73-726b7fe5f03c-kube-api-access-59zzp" (OuterVolumeSpecName: "kube-api-access-59zzp") pod "abbde07e-75cb-4db6-9f73-726b7fe5f03c" (UID: "abbde07e-75cb-4db6-9f73-726b7fe5f03c"). InnerVolumeSpecName "kube-api-access-59zzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.310946 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abbde07e-75cb-4db6-9f73-726b7fe5f03c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "abbde07e-75cb-4db6-9f73-726b7fe5f03c" (UID: "abbde07e-75cb-4db6-9f73-726b7fe5f03c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.312305 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abbde07e-75cb-4db6-9f73-726b7fe5f03c-inventory" (OuterVolumeSpecName: "inventory") pod "abbde07e-75cb-4db6-9f73-726b7fe5f03c" (UID: "abbde07e-75cb-4db6-9f73-726b7fe5f03c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.341171 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abbde07e-75cb-4db6-9f73-726b7fe5f03c-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "abbde07e-75cb-4db6-9f73-726b7fe5f03c" (UID: "abbde07e-75cb-4db6-9f73-726b7fe5f03c"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.368697 4790 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/abbde07e-75cb-4db6-9f73-726b7fe5f03c-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.368727 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59zzp\" (UniqueName: \"kubernetes.io/projected/abbde07e-75cb-4db6-9f73-726b7fe5f03c-kube-api-access-59zzp\") on node \"crc\" DevicePath \"\"" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.368737 4790 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abbde07e-75cb-4db6-9f73-726b7fe5f03c-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.368746 4790 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/abbde07e-75cb-4db6-9f73-726b7fe5f03c-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.368755 4790 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abbde07e-75cb-4db6-9f73-726b7fe5f03c-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.368763 4790 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abbde07e-75cb-4db6-9f73-726b7fe5f03c-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.504794 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pn5gv"] Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.530538 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pn5gv" event={"ID":"6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4","Type":"ContainerStarted","Data":"1dbba0e849f840c0a5bbaf128e04c6e29c0ccfc3669409a30b8ba9f6b4b0949d"} Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.535653 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-zgjq6" event={"ID":"abbde07e-75cb-4db6-9f73-726b7fe5f03c","Type":"ContainerDied","Data":"50c8f3dbead07db34e598721d09e1c60a331e8acb89379b01e965906702db764"} Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.535680 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50c8f3dbead07db34e598721d09e1c60a331e8acb89379b01e965906702db764" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.535747 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-zgjq6" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.655594 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-k6gv2"] Nov 24 15:27:22 crc kubenswrapper[4790]: E1124 15:27:22.658829 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abbde07e-75cb-4db6-9f73-726b7fe5f03c" containerName="libvirt-openstack-openstack-cell1" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.658984 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="abbde07e-75cb-4db6-9f73-726b7fe5f03c" containerName="libvirt-openstack-openstack-cell1" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.659361 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="abbde07e-75cb-4db6-9f73-726b7fe5f03c" containerName="libvirt-openstack-openstack-cell1" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.660376 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.664579 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.664764 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.664891 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.664992 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.665130 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fc497" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.665284 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.665442 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.681733 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-k6gv2"] Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.682227 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-k6gv2\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.682264 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9r9zg\" (UniqueName: \"kubernetes.io/projected/1feac502-08da-479d-8efe-2c84747f9d0a-kube-api-access-9r9zg\") pod \"nova-cell1-openstack-openstack-cell1-k6gv2\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.682293 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-k6gv2\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.682310 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-k6gv2\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.682339 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-ceph\") pod \"nova-cell1-openstack-openstack-cell1-k6gv2\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.682367 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-k6gv2\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.682385 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-k6gv2\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.682414 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-inventory\") pod \"nova-cell1-openstack-openstack-cell1-k6gv2\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.682442 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/1feac502-08da-479d-8efe-2c84747f9d0a-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-k6gv2\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.682487 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-k6gv2\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.682528 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/1feac502-08da-479d-8efe-2c84747f9d0a-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-k6gv2\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.783959 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-k6gv2\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.784029 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9r9zg\" (UniqueName: \"kubernetes.io/projected/1feac502-08da-479d-8efe-2c84747f9d0a-kube-api-access-9r9zg\") pod \"nova-cell1-openstack-openstack-cell1-k6gv2\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.784061 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-k6gv2\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.784086 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-k6gv2\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.784122 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-ceph\") pod \"nova-cell1-openstack-openstack-cell1-k6gv2\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.784153 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-k6gv2\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.784171 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-k6gv2\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.784205 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-inventory\") pod \"nova-cell1-openstack-openstack-cell1-k6gv2\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.784238 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/1feac502-08da-479d-8efe-2c84747f9d0a-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-k6gv2\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.784293 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-k6gv2\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.784332 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/1feac502-08da-479d-8efe-2c84747f9d0a-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-k6gv2\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.785190 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/1feac502-08da-479d-8efe-2c84747f9d0a-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-k6gv2\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.786440 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/1feac502-08da-479d-8efe-2c84747f9d0a-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-k6gv2\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.792017 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-inventory\") pod \"nova-cell1-openstack-openstack-cell1-k6gv2\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.792098 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-k6gv2\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.792333 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-k6gv2\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.792568 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-k6gv2\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.794294 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-k6gv2\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.794766 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-k6gv2\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.795064 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-k6gv2\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.795181 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-ceph\") pod \"nova-cell1-openstack-openstack-cell1-k6gv2\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.805691 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9r9zg\" (UniqueName: \"kubernetes.io/projected/1feac502-08da-479d-8efe-2c84747f9d0a-kube-api-access-9r9zg\") pod \"nova-cell1-openstack-openstack-cell1-k6gv2\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:22 crc kubenswrapper[4790]: I1124 15:27:22.987691 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:27:23 crc kubenswrapper[4790]: I1124 15:27:23.550511 4790 generic.go:334] "Generic (PLEG): container finished" podID="6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4" containerID="8379ac50b8e46fd5cec715d74dc65a59b08b5ec63c57752f2e01d6c00c54f124" exitCode=0 Nov 24 15:27:23 crc kubenswrapper[4790]: I1124 15:27:23.550610 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pn5gv" event={"ID":"6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4","Type":"ContainerDied","Data":"8379ac50b8e46fd5cec715d74dc65a59b08b5ec63c57752f2e01d6c00c54f124"} Nov 24 15:27:23 crc kubenswrapper[4790]: I1124 15:27:23.554684 4790 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 15:27:23 crc kubenswrapper[4790]: I1124 15:27:23.617750 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-k6gv2"] Nov 24 15:27:23 crc kubenswrapper[4790]: W1124 15:27:23.619594 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1feac502_08da_479d_8efe_2c84747f9d0a.slice/crio-13d8efd54fc6bb20c0b69310aae3f6a21ce05e07d4c84b1bc00edb04657d0ea9 WatchSource:0}: Error finding container 13d8efd54fc6bb20c0b69310aae3f6a21ce05e07d4c84b1bc00edb04657d0ea9: Status 404 returned error can't find the container with id 13d8efd54fc6bb20c0b69310aae3f6a21ce05e07d4c84b1bc00edb04657d0ea9 Nov 24 15:27:24 crc kubenswrapper[4790]: I1124 15:27:24.566553 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" event={"ID":"1feac502-08da-479d-8efe-2c84747f9d0a","Type":"ContainerStarted","Data":"c72b3a5553f2022dd6113bbfeff2f7885c48c54603c20ed696b35eca99e72728"} Nov 24 15:27:24 crc kubenswrapper[4790]: I1124 15:27:24.567306 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" event={"ID":"1feac502-08da-479d-8efe-2c84747f9d0a","Type":"ContainerStarted","Data":"13d8efd54fc6bb20c0b69310aae3f6a21ce05e07d4c84b1bc00edb04657d0ea9"} Nov 24 15:27:24 crc kubenswrapper[4790]: I1124 15:27:24.597151 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" podStartSLOduration=2.404257848 podStartE2EDuration="2.597129188s" podCreationTimestamp="2025-11-24 15:27:22 +0000 UTC" firstStartedPulling="2025-11-24 15:27:23.622799244 +0000 UTC m=+8092.002692906" lastFinishedPulling="2025-11-24 15:27:23.815670574 +0000 UTC m=+8092.195564246" observedRunningTime="2025-11-24 15:27:24.587551965 +0000 UTC m=+8092.967445627" watchObservedRunningTime="2025-11-24 15:27:24.597129188 +0000 UTC m=+8092.977022850" Nov 24 15:27:25 crc kubenswrapper[4790]: I1124 15:27:25.583475 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pn5gv" event={"ID":"6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4","Type":"ContainerStarted","Data":"897062d5a95f52ffbf65fb7cadffe6ea45e0fd91cf9d5aeaeec6acfcbcecb271"} Nov 24 15:27:26 crc kubenswrapper[4790]: I1124 15:27:26.596716 4790 generic.go:334] "Generic (PLEG): container finished" podID="6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4" containerID="897062d5a95f52ffbf65fb7cadffe6ea45e0fd91cf9d5aeaeec6acfcbcecb271" exitCode=0 Nov 24 15:27:26 crc kubenswrapper[4790]: I1124 15:27:26.596786 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pn5gv" event={"ID":"6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4","Type":"ContainerDied","Data":"897062d5a95f52ffbf65fb7cadffe6ea45e0fd91cf9d5aeaeec6acfcbcecb271"} Nov 24 15:27:27 crc kubenswrapper[4790]: I1124 15:27:27.610993 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pn5gv" event={"ID":"6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4","Type":"ContainerStarted","Data":"1973d9522fb8a7d3ab66a612a96207ca83eccb228705adedb53fd3188a604a07"} Nov 24 15:27:28 crc kubenswrapper[4790]: I1124 15:27:28.661755 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pn5gv" podStartSLOduration=4.079102407 podStartE2EDuration="7.66161737s" podCreationTimestamp="2025-11-24 15:27:21 +0000 UTC" firstStartedPulling="2025-11-24 15:27:23.55439554 +0000 UTC m=+8091.934289202" lastFinishedPulling="2025-11-24 15:27:27.136910503 +0000 UTC m=+8095.516804165" observedRunningTime="2025-11-24 15:27:28.65111574 +0000 UTC m=+8097.031009412" watchObservedRunningTime="2025-11-24 15:27:28.66161737 +0000 UTC m=+8097.041511042" Nov 24 15:27:32 crc kubenswrapper[4790]: I1124 15:27:32.039584 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pn5gv" Nov 24 15:27:32 crc kubenswrapper[4790]: I1124 15:27:32.039939 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pn5gv" Nov 24 15:27:33 crc kubenswrapper[4790]: I1124 15:27:33.123462 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pn5gv" podUID="6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4" containerName="registry-server" probeResult="failure" output=< Nov 24 15:27:33 crc kubenswrapper[4790]: timeout: failed to connect service ":50051" within 1s Nov 24 15:27:33 crc kubenswrapper[4790]: > Nov 24 15:27:42 crc kubenswrapper[4790]: I1124 15:27:42.128730 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pn5gv" Nov 24 15:27:42 crc kubenswrapper[4790]: I1124 15:27:42.208825 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pn5gv" Nov 24 15:27:42 crc kubenswrapper[4790]: I1124 15:27:42.377319 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pn5gv"] Nov 24 15:27:43 crc kubenswrapper[4790]: I1124 15:27:43.818003 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pn5gv" podUID="6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4" containerName="registry-server" containerID="cri-o://1973d9522fb8a7d3ab66a612a96207ca83eccb228705adedb53fd3188a604a07" gracePeriod=2 Nov 24 15:27:43 crc kubenswrapper[4790]: I1124 15:27:43.938312 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:27:43 crc kubenswrapper[4790]: I1124 15:27:43.938425 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:27:44 crc kubenswrapper[4790]: I1124 15:27:44.409703 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pn5gv" Nov 24 15:27:44 crc kubenswrapper[4790]: I1124 15:27:44.602075 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9pl5\" (UniqueName: \"kubernetes.io/projected/6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4-kube-api-access-c9pl5\") pod \"6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4\" (UID: \"6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4\") " Nov 24 15:27:44 crc kubenswrapper[4790]: I1124 15:27:44.602220 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4-catalog-content\") pod \"6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4\" (UID: \"6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4\") " Nov 24 15:27:44 crc kubenswrapper[4790]: I1124 15:27:44.602364 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4-utilities\") pod \"6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4\" (UID: \"6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4\") " Nov 24 15:27:44 crc kubenswrapper[4790]: I1124 15:27:44.618111 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4-utilities" (OuterVolumeSpecName: "utilities") pod "6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4" (UID: "6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:27:44 crc kubenswrapper[4790]: I1124 15:27:44.639960 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4-kube-api-access-c9pl5" (OuterVolumeSpecName: "kube-api-access-c9pl5") pod "6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4" (UID: "6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4"). InnerVolumeSpecName "kube-api-access-c9pl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:27:44 crc kubenswrapper[4790]: I1124 15:27:44.711529 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9pl5\" (UniqueName: \"kubernetes.io/projected/6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4-kube-api-access-c9pl5\") on node \"crc\" DevicePath \"\"" Nov 24 15:27:44 crc kubenswrapper[4790]: I1124 15:27:44.712158 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 15:27:44 crc kubenswrapper[4790]: I1124 15:27:44.720349 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4" (UID: "6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:27:44 crc kubenswrapper[4790]: I1124 15:27:44.814323 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 15:27:44 crc kubenswrapper[4790]: I1124 15:27:44.829007 4790 generic.go:334] "Generic (PLEG): container finished" podID="6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4" containerID="1973d9522fb8a7d3ab66a612a96207ca83eccb228705adedb53fd3188a604a07" exitCode=0 Nov 24 15:27:44 crc kubenswrapper[4790]: I1124 15:27:44.829050 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pn5gv" event={"ID":"6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4","Type":"ContainerDied","Data":"1973d9522fb8a7d3ab66a612a96207ca83eccb228705adedb53fd3188a604a07"} Nov 24 15:27:44 crc kubenswrapper[4790]: I1124 15:27:44.829076 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pn5gv" event={"ID":"6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4","Type":"ContainerDied","Data":"1dbba0e849f840c0a5bbaf128e04c6e29c0ccfc3669409a30b8ba9f6b4b0949d"} Nov 24 15:27:44 crc kubenswrapper[4790]: I1124 15:27:44.829092 4790 scope.go:117] "RemoveContainer" containerID="1973d9522fb8a7d3ab66a612a96207ca83eccb228705adedb53fd3188a604a07" Nov 24 15:27:44 crc kubenswrapper[4790]: I1124 15:27:44.829268 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pn5gv" Nov 24 15:27:44 crc kubenswrapper[4790]: I1124 15:27:44.873850 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pn5gv"] Nov 24 15:27:44 crc kubenswrapper[4790]: I1124 15:27:44.883106 4790 scope.go:117] "RemoveContainer" containerID="897062d5a95f52ffbf65fb7cadffe6ea45e0fd91cf9d5aeaeec6acfcbcecb271" Nov 24 15:27:44 crc kubenswrapper[4790]: I1124 15:27:44.888086 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pn5gv"] Nov 24 15:27:44 crc kubenswrapper[4790]: I1124 15:27:44.936693 4790 scope.go:117] "RemoveContainer" containerID="8379ac50b8e46fd5cec715d74dc65a59b08b5ec63c57752f2e01d6c00c54f124" Nov 24 15:27:44 crc kubenswrapper[4790]: I1124 15:27:44.988463 4790 scope.go:117] "RemoveContainer" containerID="1973d9522fb8a7d3ab66a612a96207ca83eccb228705adedb53fd3188a604a07" Nov 24 15:27:44 crc kubenswrapper[4790]: E1124 15:27:44.989001 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1973d9522fb8a7d3ab66a612a96207ca83eccb228705adedb53fd3188a604a07\": container with ID starting with 1973d9522fb8a7d3ab66a612a96207ca83eccb228705adedb53fd3188a604a07 not found: ID does not exist" containerID="1973d9522fb8a7d3ab66a612a96207ca83eccb228705adedb53fd3188a604a07" Nov 24 15:27:44 crc kubenswrapper[4790]: I1124 15:27:44.989050 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1973d9522fb8a7d3ab66a612a96207ca83eccb228705adedb53fd3188a604a07"} err="failed to get container status \"1973d9522fb8a7d3ab66a612a96207ca83eccb228705adedb53fd3188a604a07\": rpc error: code = NotFound desc = could not find container \"1973d9522fb8a7d3ab66a612a96207ca83eccb228705adedb53fd3188a604a07\": container with ID starting with 1973d9522fb8a7d3ab66a612a96207ca83eccb228705adedb53fd3188a604a07 not found: ID does not exist" Nov 24 15:27:44 crc kubenswrapper[4790]: I1124 15:27:44.989080 4790 scope.go:117] "RemoveContainer" containerID="897062d5a95f52ffbf65fb7cadffe6ea45e0fd91cf9d5aeaeec6acfcbcecb271" Nov 24 15:27:44 crc kubenswrapper[4790]: E1124 15:27:44.989640 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"897062d5a95f52ffbf65fb7cadffe6ea45e0fd91cf9d5aeaeec6acfcbcecb271\": container with ID starting with 897062d5a95f52ffbf65fb7cadffe6ea45e0fd91cf9d5aeaeec6acfcbcecb271 not found: ID does not exist" containerID="897062d5a95f52ffbf65fb7cadffe6ea45e0fd91cf9d5aeaeec6acfcbcecb271" Nov 24 15:27:44 crc kubenswrapper[4790]: I1124 15:27:44.989685 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"897062d5a95f52ffbf65fb7cadffe6ea45e0fd91cf9d5aeaeec6acfcbcecb271"} err="failed to get container status \"897062d5a95f52ffbf65fb7cadffe6ea45e0fd91cf9d5aeaeec6acfcbcecb271\": rpc error: code = NotFound desc = could not find container \"897062d5a95f52ffbf65fb7cadffe6ea45e0fd91cf9d5aeaeec6acfcbcecb271\": container with ID starting with 897062d5a95f52ffbf65fb7cadffe6ea45e0fd91cf9d5aeaeec6acfcbcecb271 not found: ID does not exist" Nov 24 15:27:44 crc kubenswrapper[4790]: I1124 15:27:44.989712 4790 scope.go:117] "RemoveContainer" containerID="8379ac50b8e46fd5cec715d74dc65a59b08b5ec63c57752f2e01d6c00c54f124" Nov 24 15:27:44 crc kubenswrapper[4790]: E1124 15:27:44.990124 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8379ac50b8e46fd5cec715d74dc65a59b08b5ec63c57752f2e01d6c00c54f124\": container with ID starting with 8379ac50b8e46fd5cec715d74dc65a59b08b5ec63c57752f2e01d6c00c54f124 not found: ID does not exist" containerID="8379ac50b8e46fd5cec715d74dc65a59b08b5ec63c57752f2e01d6c00c54f124" Nov 24 15:27:44 crc kubenswrapper[4790]: I1124 15:27:44.990176 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8379ac50b8e46fd5cec715d74dc65a59b08b5ec63c57752f2e01d6c00c54f124"} err="failed to get container status \"8379ac50b8e46fd5cec715d74dc65a59b08b5ec63c57752f2e01d6c00c54f124\": rpc error: code = NotFound desc = could not find container \"8379ac50b8e46fd5cec715d74dc65a59b08b5ec63c57752f2e01d6c00c54f124\": container with ID starting with 8379ac50b8e46fd5cec715d74dc65a59b08b5ec63c57752f2e01d6c00c54f124 not found: ID does not exist" Nov 24 15:27:46 crc kubenswrapper[4790]: I1124 15:27:46.339225 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4" path="/var/lib/kubelet/pods/6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4/volumes" Nov 24 15:28:13 crc kubenswrapper[4790]: I1124 15:28:13.939169 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:28:13 crc kubenswrapper[4790]: I1124 15:28:13.940014 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:28:43 crc kubenswrapper[4790]: I1124 15:28:43.938449 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:28:43 crc kubenswrapper[4790]: I1124 15:28:43.939198 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:28:43 crc kubenswrapper[4790]: I1124 15:28:43.939265 4790 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 15:28:43 crc kubenswrapper[4790]: I1124 15:28:43.940441 4790 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"87c151c65c6ec7e6e0150da1b299aebe7cb33ee5e3738cff003c27747ab5c270"} pod="openshift-machine-config-operator/machine-config-daemon-xz49t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 15:28:43 crc kubenswrapper[4790]: I1124 15:28:43.940575 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" containerID="cri-o://87c151c65c6ec7e6e0150da1b299aebe7cb33ee5e3738cff003c27747ab5c270" gracePeriod=600 Nov 24 15:28:44 crc kubenswrapper[4790]: E1124 15:28:44.073703 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:28:44 crc kubenswrapper[4790]: I1124 15:28:44.680908 4790 generic.go:334] "Generic (PLEG): container finished" podID="0d73b133-48f0-455f-8f6a-742e633f631a" containerID="87c151c65c6ec7e6e0150da1b299aebe7cb33ee5e3738cff003c27747ab5c270" exitCode=0 Nov 24 15:28:44 crc kubenswrapper[4790]: I1124 15:28:44.680981 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerDied","Data":"87c151c65c6ec7e6e0150da1b299aebe7cb33ee5e3738cff003c27747ab5c270"} Nov 24 15:28:44 crc kubenswrapper[4790]: I1124 15:28:44.681332 4790 scope.go:117] "RemoveContainer" containerID="7191c6329f823582f9ae00d9da816e469d4f902ed9ef90979520daad145e04e4" Nov 24 15:28:44 crc kubenswrapper[4790]: I1124 15:28:44.682530 4790 scope.go:117] "RemoveContainer" containerID="87c151c65c6ec7e6e0150da1b299aebe7cb33ee5e3738cff003c27747ab5c270" Nov 24 15:28:44 crc kubenswrapper[4790]: E1124 15:28:44.684057 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:28:59 crc kubenswrapper[4790]: I1124 15:28:59.315225 4790 scope.go:117] "RemoveContainer" containerID="87c151c65c6ec7e6e0150da1b299aebe7cb33ee5e3738cff003c27747ab5c270" Nov 24 15:28:59 crc kubenswrapper[4790]: E1124 15:28:59.316308 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:29:14 crc kubenswrapper[4790]: I1124 15:29:14.315176 4790 scope.go:117] "RemoveContainer" containerID="87c151c65c6ec7e6e0150da1b299aebe7cb33ee5e3738cff003c27747ab5c270" Nov 24 15:29:14 crc kubenswrapper[4790]: E1124 15:29:14.316239 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:29:29 crc kubenswrapper[4790]: I1124 15:29:29.315467 4790 scope.go:117] "RemoveContainer" containerID="87c151c65c6ec7e6e0150da1b299aebe7cb33ee5e3738cff003c27747ab5c270" Nov 24 15:29:29 crc kubenswrapper[4790]: E1124 15:29:29.316373 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:29:41 crc kubenswrapper[4790]: I1124 15:29:41.314488 4790 scope.go:117] "RemoveContainer" containerID="87c151c65c6ec7e6e0150da1b299aebe7cb33ee5e3738cff003c27747ab5c270" Nov 24 15:29:41 crc kubenswrapper[4790]: E1124 15:29:41.315392 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:29:54 crc kubenswrapper[4790]: I1124 15:29:54.315118 4790 scope.go:117] "RemoveContainer" containerID="87c151c65c6ec7e6e0150da1b299aebe7cb33ee5e3738cff003c27747ab5c270" Nov 24 15:29:54 crc kubenswrapper[4790]: E1124 15:29:54.316478 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:30:00 crc kubenswrapper[4790]: I1124 15:30:00.145783 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399970-8hnq4"] Nov 24 15:30:00 crc kubenswrapper[4790]: E1124 15:30:00.147045 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4" containerName="registry-server" Nov 24 15:30:00 crc kubenswrapper[4790]: I1124 15:30:00.147061 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4" containerName="registry-server" Nov 24 15:30:00 crc kubenswrapper[4790]: E1124 15:30:00.147082 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4" containerName="extract-utilities" Nov 24 15:30:00 crc kubenswrapper[4790]: I1124 15:30:00.147090 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4" containerName="extract-utilities" Nov 24 15:30:00 crc kubenswrapper[4790]: E1124 15:30:00.147130 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4" containerName="extract-content" Nov 24 15:30:00 crc kubenswrapper[4790]: I1124 15:30:00.147138 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4" containerName="extract-content" Nov 24 15:30:00 crc kubenswrapper[4790]: I1124 15:30:00.147419 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bd7d467-3891-4bcd-9c84-5e2d9fe6b5f4" containerName="registry-server" Nov 24 15:30:00 crc kubenswrapper[4790]: I1124 15:30:00.148374 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399970-8hnq4" Nov 24 15:30:00 crc kubenswrapper[4790]: I1124 15:30:00.150947 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 15:30:00 crc kubenswrapper[4790]: I1124 15:30:00.155247 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 15:30:00 crc kubenswrapper[4790]: I1124 15:30:00.171018 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399970-8hnq4"] Nov 24 15:30:00 crc kubenswrapper[4790]: I1124 15:30:00.213815 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8cp2\" (UniqueName: \"kubernetes.io/projected/b4154d10-d7a6-4cb3-9b9c-682555fa0fa9-kube-api-access-l8cp2\") pod \"collect-profiles-29399970-8hnq4\" (UID: \"b4154d10-d7a6-4cb3-9b9c-682555fa0fa9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399970-8hnq4" Nov 24 15:30:00 crc kubenswrapper[4790]: I1124 15:30:00.213931 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b4154d10-d7a6-4cb3-9b9c-682555fa0fa9-secret-volume\") pod \"collect-profiles-29399970-8hnq4\" (UID: \"b4154d10-d7a6-4cb3-9b9c-682555fa0fa9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399970-8hnq4" Nov 24 15:30:00 crc kubenswrapper[4790]: I1124 15:30:00.214119 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b4154d10-d7a6-4cb3-9b9c-682555fa0fa9-config-volume\") pod \"collect-profiles-29399970-8hnq4\" (UID: \"b4154d10-d7a6-4cb3-9b9c-682555fa0fa9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399970-8hnq4" Nov 24 15:30:00 crc kubenswrapper[4790]: I1124 15:30:00.315657 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8cp2\" (UniqueName: \"kubernetes.io/projected/b4154d10-d7a6-4cb3-9b9c-682555fa0fa9-kube-api-access-l8cp2\") pod \"collect-profiles-29399970-8hnq4\" (UID: \"b4154d10-d7a6-4cb3-9b9c-682555fa0fa9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399970-8hnq4" Nov 24 15:30:00 crc kubenswrapper[4790]: I1124 15:30:00.315790 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b4154d10-d7a6-4cb3-9b9c-682555fa0fa9-secret-volume\") pod \"collect-profiles-29399970-8hnq4\" (UID: \"b4154d10-d7a6-4cb3-9b9c-682555fa0fa9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399970-8hnq4" Nov 24 15:30:00 crc kubenswrapper[4790]: I1124 15:30:00.315910 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b4154d10-d7a6-4cb3-9b9c-682555fa0fa9-config-volume\") pod \"collect-profiles-29399970-8hnq4\" (UID: \"b4154d10-d7a6-4cb3-9b9c-682555fa0fa9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399970-8hnq4" Nov 24 15:30:00 crc kubenswrapper[4790]: I1124 15:30:00.316847 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b4154d10-d7a6-4cb3-9b9c-682555fa0fa9-config-volume\") pod \"collect-profiles-29399970-8hnq4\" (UID: \"b4154d10-d7a6-4cb3-9b9c-682555fa0fa9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399970-8hnq4" Nov 24 15:30:00 crc kubenswrapper[4790]: I1124 15:30:00.321255 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b4154d10-d7a6-4cb3-9b9c-682555fa0fa9-secret-volume\") pod \"collect-profiles-29399970-8hnq4\" (UID: \"b4154d10-d7a6-4cb3-9b9c-682555fa0fa9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399970-8hnq4" Nov 24 15:30:00 crc kubenswrapper[4790]: I1124 15:30:00.337556 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8cp2\" (UniqueName: \"kubernetes.io/projected/b4154d10-d7a6-4cb3-9b9c-682555fa0fa9-kube-api-access-l8cp2\") pod \"collect-profiles-29399970-8hnq4\" (UID: \"b4154d10-d7a6-4cb3-9b9c-682555fa0fa9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399970-8hnq4" Nov 24 15:30:00 crc kubenswrapper[4790]: I1124 15:30:00.479308 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399970-8hnq4" Nov 24 15:30:00 crc kubenswrapper[4790]: I1124 15:30:00.966193 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399970-8hnq4"] Nov 24 15:30:01 crc kubenswrapper[4790]: I1124 15:30:01.733446 4790 generic.go:334] "Generic (PLEG): container finished" podID="b4154d10-d7a6-4cb3-9b9c-682555fa0fa9" containerID="02e67cb9aa809e2cf28a38ce49ce3d49f1123388099ce84c3ffedddd47fa576c" exitCode=0 Nov 24 15:30:01 crc kubenswrapper[4790]: I1124 15:30:01.733562 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399970-8hnq4" event={"ID":"b4154d10-d7a6-4cb3-9b9c-682555fa0fa9","Type":"ContainerDied","Data":"02e67cb9aa809e2cf28a38ce49ce3d49f1123388099ce84c3ffedddd47fa576c"} Nov 24 15:30:01 crc kubenswrapper[4790]: I1124 15:30:01.733972 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399970-8hnq4" event={"ID":"b4154d10-d7a6-4cb3-9b9c-682555fa0fa9","Type":"ContainerStarted","Data":"555a65d658f201cdfcb48a9897b5d1a35087b7eeaeb0e5f9f429d7159d925493"} Nov 24 15:30:03 crc kubenswrapper[4790]: I1124 15:30:03.180875 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399970-8hnq4" Nov 24 15:30:03 crc kubenswrapper[4790]: I1124 15:30:03.296395 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8cp2\" (UniqueName: \"kubernetes.io/projected/b4154d10-d7a6-4cb3-9b9c-682555fa0fa9-kube-api-access-l8cp2\") pod \"b4154d10-d7a6-4cb3-9b9c-682555fa0fa9\" (UID: \"b4154d10-d7a6-4cb3-9b9c-682555fa0fa9\") " Nov 24 15:30:03 crc kubenswrapper[4790]: I1124 15:30:03.296496 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b4154d10-d7a6-4cb3-9b9c-682555fa0fa9-secret-volume\") pod \"b4154d10-d7a6-4cb3-9b9c-682555fa0fa9\" (UID: \"b4154d10-d7a6-4cb3-9b9c-682555fa0fa9\") " Nov 24 15:30:03 crc kubenswrapper[4790]: I1124 15:30:03.296576 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b4154d10-d7a6-4cb3-9b9c-682555fa0fa9-config-volume\") pod \"b4154d10-d7a6-4cb3-9b9c-682555fa0fa9\" (UID: \"b4154d10-d7a6-4cb3-9b9c-682555fa0fa9\") " Nov 24 15:30:03 crc kubenswrapper[4790]: I1124 15:30:03.298009 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4154d10-d7a6-4cb3-9b9c-682555fa0fa9-config-volume" (OuterVolumeSpecName: "config-volume") pod "b4154d10-d7a6-4cb3-9b9c-682555fa0fa9" (UID: "b4154d10-d7a6-4cb3-9b9c-682555fa0fa9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 15:30:03 crc kubenswrapper[4790]: I1124 15:30:03.303869 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4154d10-d7a6-4cb3-9b9c-682555fa0fa9-kube-api-access-l8cp2" (OuterVolumeSpecName: "kube-api-access-l8cp2") pod "b4154d10-d7a6-4cb3-9b9c-682555fa0fa9" (UID: "b4154d10-d7a6-4cb3-9b9c-682555fa0fa9"). InnerVolumeSpecName "kube-api-access-l8cp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:30:03 crc kubenswrapper[4790]: I1124 15:30:03.310614 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4154d10-d7a6-4cb3-9b9c-682555fa0fa9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b4154d10-d7a6-4cb3-9b9c-682555fa0fa9" (UID: "b4154d10-d7a6-4cb3-9b9c-682555fa0fa9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:30:03 crc kubenswrapper[4790]: I1124 15:30:03.400139 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8cp2\" (UniqueName: \"kubernetes.io/projected/b4154d10-d7a6-4cb3-9b9c-682555fa0fa9-kube-api-access-l8cp2\") on node \"crc\" DevicePath \"\"" Nov 24 15:30:03 crc kubenswrapper[4790]: I1124 15:30:03.400457 4790 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b4154d10-d7a6-4cb3-9b9c-682555fa0fa9-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 15:30:03 crc kubenswrapper[4790]: I1124 15:30:03.400475 4790 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b4154d10-d7a6-4cb3-9b9c-682555fa0fa9-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 15:30:03 crc kubenswrapper[4790]: I1124 15:30:03.762208 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399970-8hnq4" event={"ID":"b4154d10-d7a6-4cb3-9b9c-682555fa0fa9","Type":"ContainerDied","Data":"555a65d658f201cdfcb48a9897b5d1a35087b7eeaeb0e5f9f429d7159d925493"} Nov 24 15:30:03 crc kubenswrapper[4790]: I1124 15:30:03.762257 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="555a65d658f201cdfcb48a9897b5d1a35087b7eeaeb0e5f9f429d7159d925493" Nov 24 15:30:03 crc kubenswrapper[4790]: I1124 15:30:03.762275 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399970-8hnq4" Nov 24 15:30:04 crc kubenswrapper[4790]: I1124 15:30:04.282931 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399925-xfp46"] Nov 24 15:30:04 crc kubenswrapper[4790]: I1124 15:30:04.296631 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399925-xfp46"] Nov 24 15:30:04 crc kubenswrapper[4790]: I1124 15:30:04.336865 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6cd3c64-493b-4d69-8cec-7465da8136a2" path="/var/lib/kubelet/pods/a6cd3c64-493b-4d69-8cec-7465da8136a2/volumes" Nov 24 15:30:07 crc kubenswrapper[4790]: I1124 15:30:07.315105 4790 scope.go:117] "RemoveContainer" containerID="87c151c65c6ec7e6e0150da1b299aebe7cb33ee5e3738cff003c27747ab5c270" Nov 24 15:30:07 crc kubenswrapper[4790]: E1124 15:30:07.316260 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:30:22 crc kubenswrapper[4790]: I1124 15:30:22.315666 4790 scope.go:117] "RemoveContainer" containerID="87c151c65c6ec7e6e0150da1b299aebe7cb33ee5e3738cff003c27747ab5c270" Nov 24 15:30:22 crc kubenswrapper[4790]: E1124 15:30:22.317289 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:30:36 crc kubenswrapper[4790]: I1124 15:30:36.315574 4790 scope.go:117] "RemoveContainer" containerID="87c151c65c6ec7e6e0150da1b299aebe7cb33ee5e3738cff003c27747ab5c270" Nov 24 15:30:36 crc kubenswrapper[4790]: E1124 15:30:36.316705 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:30:45 crc kubenswrapper[4790]: I1124 15:30:45.342389 4790 generic.go:334] "Generic (PLEG): container finished" podID="1feac502-08da-479d-8efe-2c84747f9d0a" containerID="c72b3a5553f2022dd6113bbfeff2f7885c48c54603c20ed696b35eca99e72728" exitCode=0 Nov 24 15:30:45 crc kubenswrapper[4790]: I1124 15:30:45.342626 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" event={"ID":"1feac502-08da-479d-8efe-2c84747f9d0a","Type":"ContainerDied","Data":"c72b3a5553f2022dd6113bbfeff2f7885c48c54603c20ed696b35eca99e72728"} Nov 24 15:30:46 crc kubenswrapper[4790]: I1124 15:30:46.838413 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.006408 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-ceph\") pod \"1feac502-08da-479d-8efe-2c84747f9d0a\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.006463 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-nova-migration-ssh-key-0\") pod \"1feac502-08da-479d-8efe-2c84747f9d0a\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.006582 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-ssh-key\") pod \"1feac502-08da-479d-8efe-2c84747f9d0a\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.006614 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/1feac502-08da-479d-8efe-2c84747f9d0a-nova-cells-global-config-0\") pod \"1feac502-08da-479d-8efe-2c84747f9d0a\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.006662 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-nova-cell1-compute-config-0\") pod \"1feac502-08da-479d-8efe-2c84747f9d0a\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.006748 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9r9zg\" (UniqueName: \"kubernetes.io/projected/1feac502-08da-479d-8efe-2c84747f9d0a-kube-api-access-9r9zg\") pod \"1feac502-08da-479d-8efe-2c84747f9d0a\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.006836 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-nova-migration-ssh-key-1\") pod \"1feac502-08da-479d-8efe-2c84747f9d0a\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.006875 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/1feac502-08da-479d-8efe-2c84747f9d0a-nova-cells-global-config-1\") pod \"1feac502-08da-479d-8efe-2c84747f9d0a\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.006934 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-nova-cell1-compute-config-1\") pod \"1feac502-08da-479d-8efe-2c84747f9d0a\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.006961 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-nova-cell1-combined-ca-bundle\") pod \"1feac502-08da-479d-8efe-2c84747f9d0a\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.006993 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-inventory\") pod \"1feac502-08da-479d-8efe-2c84747f9d0a\" (UID: \"1feac502-08da-479d-8efe-2c84747f9d0a\") " Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.013007 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "1feac502-08da-479d-8efe-2c84747f9d0a" (UID: "1feac502-08da-479d-8efe-2c84747f9d0a"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.015185 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-ceph" (OuterVolumeSpecName: "ceph") pod "1feac502-08da-479d-8efe-2c84747f9d0a" (UID: "1feac502-08da-479d-8efe-2c84747f9d0a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.015256 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1feac502-08da-479d-8efe-2c84747f9d0a-kube-api-access-9r9zg" (OuterVolumeSpecName: "kube-api-access-9r9zg") pod "1feac502-08da-479d-8efe-2c84747f9d0a" (UID: "1feac502-08da-479d-8efe-2c84747f9d0a"). InnerVolumeSpecName "kube-api-access-9r9zg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.043364 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "1feac502-08da-479d-8efe-2c84747f9d0a" (UID: "1feac502-08da-479d-8efe-2c84747f9d0a"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.044614 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "1feac502-08da-479d-8efe-2c84747f9d0a" (UID: "1feac502-08da-479d-8efe-2c84747f9d0a"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.045720 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "1feac502-08da-479d-8efe-2c84747f9d0a" (UID: "1feac502-08da-479d-8efe-2c84747f9d0a"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.047400 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1feac502-08da-479d-8efe-2c84747f9d0a-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "1feac502-08da-479d-8efe-2c84747f9d0a" (UID: "1feac502-08da-479d-8efe-2c84747f9d0a"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.048732 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1feac502-08da-479d-8efe-2c84747f9d0a-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "1feac502-08da-479d-8efe-2c84747f9d0a" (UID: "1feac502-08da-479d-8efe-2c84747f9d0a"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.052407 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-inventory" (OuterVolumeSpecName: "inventory") pod "1feac502-08da-479d-8efe-2c84747f9d0a" (UID: "1feac502-08da-479d-8efe-2c84747f9d0a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.056878 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1feac502-08da-479d-8efe-2c84747f9d0a" (UID: "1feac502-08da-479d-8efe-2c84747f9d0a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.065076 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "1feac502-08da-479d-8efe-2c84747f9d0a" (UID: "1feac502-08da-479d-8efe-2c84747f9d0a"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.109452 4790 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/1feac502-08da-479d-8efe-2c84747f9d0a-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.109500 4790 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.109521 4790 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.109540 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9r9zg\" (UniqueName: \"kubernetes.io/projected/1feac502-08da-479d-8efe-2c84747f9d0a-kube-api-access-9r9zg\") on node \"crc\" DevicePath \"\"" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.109560 4790 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.109578 4790 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/1feac502-08da-479d-8efe-2c84747f9d0a-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.109597 4790 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.109614 4790 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.109632 4790 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.109650 4790 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.109666 4790 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/1feac502-08da-479d-8efe-2c84747f9d0a-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.374169 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" event={"ID":"1feac502-08da-479d-8efe-2c84747f9d0a","Type":"ContainerDied","Data":"13d8efd54fc6bb20c0b69310aae3f6a21ce05e07d4c84b1bc00edb04657d0ea9"} Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.374643 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13d8efd54fc6bb20c0b69310aae3f6a21ce05e07d4c84b1bc00edb04657d0ea9" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.374755 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-k6gv2" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.514402 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-ncldn"] Nov 24 15:30:47 crc kubenswrapper[4790]: E1124 15:30:47.514818 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4154d10-d7a6-4cb3-9b9c-682555fa0fa9" containerName="collect-profiles" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.514831 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4154d10-d7a6-4cb3-9b9c-682555fa0fa9" containerName="collect-profiles" Nov 24 15:30:47 crc kubenswrapper[4790]: E1124 15:30:47.514856 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1feac502-08da-479d-8efe-2c84747f9d0a" containerName="nova-cell1-openstack-openstack-cell1" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.514862 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="1feac502-08da-479d-8efe-2c84747f9d0a" containerName="nova-cell1-openstack-openstack-cell1" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.515103 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="1feac502-08da-479d-8efe-2c84747f9d0a" containerName="nova-cell1-openstack-openstack-cell1" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.515115 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4154d10-d7a6-4cb3-9b9c-682555fa0fa9" containerName="collect-profiles" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.516608 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-ncldn" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.522625 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-ssh-key\") pod \"telemetry-openstack-openstack-cell1-ncldn\" (UID: \"d286ee05-e6fc-4945-b7a9-87d56a06486c\") " pod="openstack/telemetry-openstack-openstack-cell1-ncldn" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.522865 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-ncldn\" (UID: \"d286ee05-e6fc-4945-b7a9-87d56a06486c\") " pod="openstack/telemetry-openstack-openstack-cell1-ncldn" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.522981 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-ncldn\" (UID: \"d286ee05-e6fc-4945-b7a9-87d56a06486c\") " pod="openstack/telemetry-openstack-openstack-cell1-ncldn" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.523214 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-inventory\") pod \"telemetry-openstack-openstack-cell1-ncldn\" (UID: \"d286ee05-e6fc-4945-b7a9-87d56a06486c\") " pod="openstack/telemetry-openstack-openstack-cell1-ncldn" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.523281 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-ncldn\" (UID: \"d286ee05-e6fc-4945-b7a9-87d56a06486c\") " pod="openstack/telemetry-openstack-openstack-cell1-ncldn" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.523376 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-ceph\") pod \"telemetry-openstack-openstack-cell1-ncldn\" (UID: \"d286ee05-e6fc-4945-b7a9-87d56a06486c\") " pod="openstack/telemetry-openstack-openstack-cell1-ncldn" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.523518 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mbwh\" (UniqueName: \"kubernetes.io/projected/d286ee05-e6fc-4945-b7a9-87d56a06486c-kube-api-access-2mbwh\") pod \"telemetry-openstack-openstack-cell1-ncldn\" (UID: \"d286ee05-e6fc-4945-b7a9-87d56a06486c\") " pod="openstack/telemetry-openstack-openstack-cell1-ncldn" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.523602 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-ncldn\" (UID: \"d286ee05-e6fc-4945-b7a9-87d56a06486c\") " pod="openstack/telemetry-openstack-openstack-cell1-ncldn" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.525410 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.525567 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.525858 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.526047 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.526255 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fc497" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.537620 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-ncldn"] Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.625534 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mbwh\" (UniqueName: \"kubernetes.io/projected/d286ee05-e6fc-4945-b7a9-87d56a06486c-kube-api-access-2mbwh\") pod \"telemetry-openstack-openstack-cell1-ncldn\" (UID: \"d286ee05-e6fc-4945-b7a9-87d56a06486c\") " pod="openstack/telemetry-openstack-openstack-cell1-ncldn" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.625583 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-ncldn\" (UID: \"d286ee05-e6fc-4945-b7a9-87d56a06486c\") " pod="openstack/telemetry-openstack-openstack-cell1-ncldn" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.625618 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-ssh-key\") pod \"telemetry-openstack-openstack-cell1-ncldn\" (UID: \"d286ee05-e6fc-4945-b7a9-87d56a06486c\") " pod="openstack/telemetry-openstack-openstack-cell1-ncldn" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.625660 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-ncldn\" (UID: \"d286ee05-e6fc-4945-b7a9-87d56a06486c\") " pod="openstack/telemetry-openstack-openstack-cell1-ncldn" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.625687 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-ncldn\" (UID: \"d286ee05-e6fc-4945-b7a9-87d56a06486c\") " pod="openstack/telemetry-openstack-openstack-cell1-ncldn" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.625766 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-inventory\") pod \"telemetry-openstack-openstack-cell1-ncldn\" (UID: \"d286ee05-e6fc-4945-b7a9-87d56a06486c\") " pod="openstack/telemetry-openstack-openstack-cell1-ncldn" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.625787 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-ncldn\" (UID: \"d286ee05-e6fc-4945-b7a9-87d56a06486c\") " pod="openstack/telemetry-openstack-openstack-cell1-ncldn" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.625808 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-ceph\") pod \"telemetry-openstack-openstack-cell1-ncldn\" (UID: \"d286ee05-e6fc-4945-b7a9-87d56a06486c\") " pod="openstack/telemetry-openstack-openstack-cell1-ncldn" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.630939 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-ncldn\" (UID: \"d286ee05-e6fc-4945-b7a9-87d56a06486c\") " pod="openstack/telemetry-openstack-openstack-cell1-ncldn" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.633430 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-ceph\") pod \"telemetry-openstack-openstack-cell1-ncldn\" (UID: \"d286ee05-e6fc-4945-b7a9-87d56a06486c\") " pod="openstack/telemetry-openstack-openstack-cell1-ncldn" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.634197 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-ncldn\" (UID: \"d286ee05-e6fc-4945-b7a9-87d56a06486c\") " pod="openstack/telemetry-openstack-openstack-cell1-ncldn" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.634367 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-ssh-key\") pod \"telemetry-openstack-openstack-cell1-ncldn\" (UID: \"d286ee05-e6fc-4945-b7a9-87d56a06486c\") " pod="openstack/telemetry-openstack-openstack-cell1-ncldn" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.635310 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-inventory\") pod \"telemetry-openstack-openstack-cell1-ncldn\" (UID: \"d286ee05-e6fc-4945-b7a9-87d56a06486c\") " pod="openstack/telemetry-openstack-openstack-cell1-ncldn" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.635919 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-ncldn\" (UID: \"d286ee05-e6fc-4945-b7a9-87d56a06486c\") " pod="openstack/telemetry-openstack-openstack-cell1-ncldn" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.637722 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-ncldn\" (UID: \"d286ee05-e6fc-4945-b7a9-87d56a06486c\") " pod="openstack/telemetry-openstack-openstack-cell1-ncldn" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.653483 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mbwh\" (UniqueName: \"kubernetes.io/projected/d286ee05-e6fc-4945-b7a9-87d56a06486c-kube-api-access-2mbwh\") pod \"telemetry-openstack-openstack-cell1-ncldn\" (UID: \"d286ee05-e6fc-4945-b7a9-87d56a06486c\") " pod="openstack/telemetry-openstack-openstack-cell1-ncldn" Nov 24 15:30:47 crc kubenswrapper[4790]: I1124 15:30:47.851278 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-ncldn" Nov 24 15:30:48 crc kubenswrapper[4790]: I1124 15:30:48.429305 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-ncldn"] Nov 24 15:30:49 crc kubenswrapper[4790]: I1124 15:30:49.315818 4790 scope.go:117] "RemoveContainer" containerID="87c151c65c6ec7e6e0150da1b299aebe7cb33ee5e3738cff003c27747ab5c270" Nov 24 15:30:49 crc kubenswrapper[4790]: E1124 15:30:49.316682 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:30:49 crc kubenswrapper[4790]: I1124 15:30:49.402071 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-ncldn" event={"ID":"d286ee05-e6fc-4945-b7a9-87d56a06486c","Type":"ContainerStarted","Data":"c3ed390eaf9203ec0e138d68ce403b2d5f5341ad24f873fc2626d667c7d21d61"} Nov 24 15:30:49 crc kubenswrapper[4790]: I1124 15:30:49.402139 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-ncldn" event={"ID":"d286ee05-e6fc-4945-b7a9-87d56a06486c","Type":"ContainerStarted","Data":"11d584cbc6b006a34b84bed78480e5070bc9272ca1c38974fbe2d1812f071095"} Nov 24 15:30:49 crc kubenswrapper[4790]: I1124 15:30:49.443495 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-openstack-openstack-cell1-ncldn" podStartSLOduration=2.250190717 podStartE2EDuration="2.443466099s" podCreationTimestamp="2025-11-24 15:30:47 +0000 UTC" firstStartedPulling="2025-11-24 15:30:48.443678524 +0000 UTC m=+8296.823572196" lastFinishedPulling="2025-11-24 15:30:48.636953896 +0000 UTC m=+8297.016847578" observedRunningTime="2025-11-24 15:30:49.430254106 +0000 UTC m=+8297.810147768" watchObservedRunningTime="2025-11-24 15:30:49.443466099 +0000 UTC m=+8297.823359791" Nov 24 15:31:03 crc kubenswrapper[4790]: I1124 15:31:03.315165 4790 scope.go:117] "RemoveContainer" containerID="87c151c65c6ec7e6e0150da1b299aebe7cb33ee5e3738cff003c27747ab5c270" Nov 24 15:31:03 crc kubenswrapper[4790]: E1124 15:31:03.316241 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:31:03 crc kubenswrapper[4790]: I1124 15:31:03.802441 4790 scope.go:117] "RemoveContainer" containerID="aa0b084db76668da61051be1b9de078a5ea793b509937ffc68ca303214a318f9" Nov 24 15:31:18 crc kubenswrapper[4790]: I1124 15:31:18.315152 4790 scope.go:117] "RemoveContainer" containerID="87c151c65c6ec7e6e0150da1b299aebe7cb33ee5e3738cff003c27747ab5c270" Nov 24 15:31:18 crc kubenswrapper[4790]: E1124 15:31:18.316128 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:31:29 crc kubenswrapper[4790]: I1124 15:31:29.315103 4790 scope.go:117] "RemoveContainer" containerID="87c151c65c6ec7e6e0150da1b299aebe7cb33ee5e3738cff003c27747ab5c270" Nov 24 15:31:29 crc kubenswrapper[4790]: E1124 15:31:29.316268 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:31:44 crc kubenswrapper[4790]: I1124 15:31:44.315718 4790 scope.go:117] "RemoveContainer" containerID="87c151c65c6ec7e6e0150da1b299aebe7cb33ee5e3738cff003c27747ab5c270" Nov 24 15:31:44 crc kubenswrapper[4790]: E1124 15:31:44.318323 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:31:55 crc kubenswrapper[4790]: I1124 15:31:55.314977 4790 scope.go:117] "RemoveContainer" containerID="87c151c65c6ec7e6e0150da1b299aebe7cb33ee5e3738cff003c27747ab5c270" Nov 24 15:31:55 crc kubenswrapper[4790]: E1124 15:31:55.315705 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:32:08 crc kubenswrapper[4790]: I1124 15:32:08.315846 4790 scope.go:117] "RemoveContainer" containerID="87c151c65c6ec7e6e0150da1b299aebe7cb33ee5e3738cff003c27747ab5c270" Nov 24 15:32:08 crc kubenswrapper[4790]: E1124 15:32:08.317174 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:32:22 crc kubenswrapper[4790]: I1124 15:32:22.328297 4790 scope.go:117] "RemoveContainer" containerID="87c151c65c6ec7e6e0150da1b299aebe7cb33ee5e3738cff003c27747ab5c270" Nov 24 15:32:22 crc kubenswrapper[4790]: E1124 15:32:22.330257 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:32:37 crc kubenswrapper[4790]: I1124 15:32:37.314950 4790 scope.go:117] "RemoveContainer" containerID="87c151c65c6ec7e6e0150da1b299aebe7cb33ee5e3738cff003c27747ab5c270" Nov 24 15:32:37 crc kubenswrapper[4790]: E1124 15:32:37.315909 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:32:51 crc kubenswrapper[4790]: I1124 15:32:51.315262 4790 scope.go:117] "RemoveContainer" containerID="87c151c65c6ec7e6e0150da1b299aebe7cb33ee5e3738cff003c27747ab5c270" Nov 24 15:32:51 crc kubenswrapper[4790]: E1124 15:32:51.316410 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:33:05 crc kubenswrapper[4790]: I1124 15:33:05.315169 4790 scope.go:117] "RemoveContainer" containerID="87c151c65c6ec7e6e0150da1b299aebe7cb33ee5e3738cff003c27747ab5c270" Nov 24 15:33:05 crc kubenswrapper[4790]: E1124 15:33:05.316257 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:33:20 crc kubenswrapper[4790]: I1124 15:33:20.314853 4790 scope.go:117] "RemoveContainer" containerID="87c151c65c6ec7e6e0150da1b299aebe7cb33ee5e3738cff003c27747ab5c270" Nov 24 15:33:20 crc kubenswrapper[4790]: E1124 15:33:20.317681 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:33:32 crc kubenswrapper[4790]: I1124 15:33:32.329326 4790 scope.go:117] "RemoveContainer" containerID="87c151c65c6ec7e6e0150da1b299aebe7cb33ee5e3738cff003c27747ab5c270" Nov 24 15:33:32 crc kubenswrapper[4790]: E1124 15:33:32.330397 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:33:43 crc kubenswrapper[4790]: I1124 15:33:43.315021 4790 scope.go:117] "RemoveContainer" containerID="87c151c65c6ec7e6e0150da1b299aebe7cb33ee5e3738cff003c27747ab5c270" Nov 24 15:33:43 crc kubenswrapper[4790]: E1124 15:33:43.315973 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:33:56 crc kubenswrapper[4790]: I1124 15:33:56.315553 4790 scope.go:117] "RemoveContainer" containerID="87c151c65c6ec7e6e0150da1b299aebe7cb33ee5e3738cff003c27747ab5c270" Nov 24 15:33:57 crc kubenswrapper[4790]: I1124 15:33:57.037096 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"4237e8fdaaefb1b14cc56e0c19383d77cde5a1f384bd8ee6c5281df48485a3d9"} Nov 24 15:34:10 crc kubenswrapper[4790]: I1124 15:34:10.778172 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xfcfp"] Nov 24 15:34:10 crc kubenswrapper[4790]: I1124 15:34:10.782824 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xfcfp" Nov 24 15:34:10 crc kubenswrapper[4790]: I1124 15:34:10.815517 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xfcfp"] Nov 24 15:34:10 crc kubenswrapper[4790]: I1124 15:34:10.869202 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/155ff7f8-3c39-45f8-820a-a6a6e9edb190-utilities\") pod \"certified-operators-xfcfp\" (UID: \"155ff7f8-3c39-45f8-820a-a6a6e9edb190\") " pod="openshift-marketplace/certified-operators-xfcfp" Nov 24 15:34:10 crc kubenswrapper[4790]: I1124 15:34:10.869547 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/155ff7f8-3c39-45f8-820a-a6a6e9edb190-catalog-content\") pod \"certified-operators-xfcfp\" (UID: \"155ff7f8-3c39-45f8-820a-a6a6e9edb190\") " pod="openshift-marketplace/certified-operators-xfcfp" Nov 24 15:34:10 crc kubenswrapper[4790]: I1124 15:34:10.869662 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwkg2\" (UniqueName: \"kubernetes.io/projected/155ff7f8-3c39-45f8-820a-a6a6e9edb190-kube-api-access-kwkg2\") pod \"certified-operators-xfcfp\" (UID: \"155ff7f8-3c39-45f8-820a-a6a6e9edb190\") " pod="openshift-marketplace/certified-operators-xfcfp" Nov 24 15:34:10 crc kubenswrapper[4790]: I1124 15:34:10.971603 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/155ff7f8-3c39-45f8-820a-a6a6e9edb190-utilities\") pod \"certified-operators-xfcfp\" (UID: \"155ff7f8-3c39-45f8-820a-a6a6e9edb190\") " pod="openshift-marketplace/certified-operators-xfcfp" Nov 24 15:34:10 crc kubenswrapper[4790]: I1124 15:34:10.971701 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/155ff7f8-3c39-45f8-820a-a6a6e9edb190-catalog-content\") pod \"certified-operators-xfcfp\" (UID: \"155ff7f8-3c39-45f8-820a-a6a6e9edb190\") " pod="openshift-marketplace/certified-operators-xfcfp" Nov 24 15:34:10 crc kubenswrapper[4790]: I1124 15:34:10.971728 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwkg2\" (UniqueName: \"kubernetes.io/projected/155ff7f8-3c39-45f8-820a-a6a6e9edb190-kube-api-access-kwkg2\") pod \"certified-operators-xfcfp\" (UID: \"155ff7f8-3c39-45f8-820a-a6a6e9edb190\") " pod="openshift-marketplace/certified-operators-xfcfp" Nov 24 15:34:10 crc kubenswrapper[4790]: I1124 15:34:10.972358 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/155ff7f8-3c39-45f8-820a-a6a6e9edb190-utilities\") pod \"certified-operators-xfcfp\" (UID: \"155ff7f8-3c39-45f8-820a-a6a6e9edb190\") " pod="openshift-marketplace/certified-operators-xfcfp" Nov 24 15:34:10 crc kubenswrapper[4790]: I1124 15:34:10.972504 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/155ff7f8-3c39-45f8-820a-a6a6e9edb190-catalog-content\") pod \"certified-operators-xfcfp\" (UID: \"155ff7f8-3c39-45f8-820a-a6a6e9edb190\") " pod="openshift-marketplace/certified-operators-xfcfp" Nov 24 15:34:10 crc kubenswrapper[4790]: I1124 15:34:10.995422 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwkg2\" (UniqueName: \"kubernetes.io/projected/155ff7f8-3c39-45f8-820a-a6a6e9edb190-kube-api-access-kwkg2\") pod \"certified-operators-xfcfp\" (UID: \"155ff7f8-3c39-45f8-820a-a6a6e9edb190\") " pod="openshift-marketplace/certified-operators-xfcfp" Nov 24 15:34:11 crc kubenswrapper[4790]: I1124 15:34:11.150095 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xfcfp" Nov 24 15:34:11 crc kubenswrapper[4790]: I1124 15:34:11.671324 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xfcfp"] Nov 24 15:34:12 crc kubenswrapper[4790]: I1124 15:34:12.386751 4790 generic.go:334] "Generic (PLEG): container finished" podID="155ff7f8-3c39-45f8-820a-a6a6e9edb190" containerID="9051cd9a038c4557ec91b129e0092a8f416468e7fd6d61e15ba9624f390dbf4f" exitCode=0 Nov 24 15:34:12 crc kubenswrapper[4790]: I1124 15:34:12.386825 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xfcfp" event={"ID":"155ff7f8-3c39-45f8-820a-a6a6e9edb190","Type":"ContainerDied","Data":"9051cd9a038c4557ec91b129e0092a8f416468e7fd6d61e15ba9624f390dbf4f"} Nov 24 15:34:12 crc kubenswrapper[4790]: I1124 15:34:12.386925 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xfcfp" event={"ID":"155ff7f8-3c39-45f8-820a-a6a6e9edb190","Type":"ContainerStarted","Data":"3aae9ee1c1f7881f1dbba2dde0185081f6818c7ee3f0fefc01cdf372fa9f009e"} Nov 24 15:34:12 crc kubenswrapper[4790]: I1124 15:34:12.397293 4790 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 15:34:14 crc kubenswrapper[4790]: I1124 15:34:14.412819 4790 generic.go:334] "Generic (PLEG): container finished" podID="155ff7f8-3c39-45f8-820a-a6a6e9edb190" containerID="f9ce16a1d743796d3eca19bca2836cf439c6487a404b9d5f6e5adf23a1d180b8" exitCode=0 Nov 24 15:34:14 crc kubenswrapper[4790]: I1124 15:34:14.412910 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xfcfp" event={"ID":"155ff7f8-3c39-45f8-820a-a6a6e9edb190","Type":"ContainerDied","Data":"f9ce16a1d743796d3eca19bca2836cf439c6487a404b9d5f6e5adf23a1d180b8"} Nov 24 15:34:15 crc kubenswrapper[4790]: I1124 15:34:15.430038 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xfcfp" event={"ID":"155ff7f8-3c39-45f8-820a-a6a6e9edb190","Type":"ContainerStarted","Data":"5bb2c063355fc12f9f17f8bed6833b9e496d5e87406d4e517681176510832635"} Nov 24 15:34:15 crc kubenswrapper[4790]: I1124 15:34:15.461131 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xfcfp" podStartSLOduration=2.97716775 podStartE2EDuration="5.461113739s" podCreationTimestamp="2025-11-24 15:34:10 +0000 UTC" firstStartedPulling="2025-11-24 15:34:12.391216101 +0000 UTC m=+8500.771109803" lastFinishedPulling="2025-11-24 15:34:14.87516209 +0000 UTC m=+8503.255055792" observedRunningTime="2025-11-24 15:34:15.455546822 +0000 UTC m=+8503.835440514" watchObservedRunningTime="2025-11-24 15:34:15.461113739 +0000 UTC m=+8503.841007401" Nov 24 15:34:21 crc kubenswrapper[4790]: I1124 15:34:21.151212 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xfcfp" Nov 24 15:34:21 crc kubenswrapper[4790]: I1124 15:34:21.151693 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xfcfp" Nov 24 15:34:21 crc kubenswrapper[4790]: I1124 15:34:21.232492 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xfcfp" Nov 24 15:34:21 crc kubenswrapper[4790]: I1124 15:34:21.593096 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xfcfp" Nov 24 15:34:21 crc kubenswrapper[4790]: I1124 15:34:21.666601 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xfcfp"] Nov 24 15:34:23 crc kubenswrapper[4790]: I1124 15:34:23.535900 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xfcfp" podUID="155ff7f8-3c39-45f8-820a-a6a6e9edb190" containerName="registry-server" containerID="cri-o://5bb2c063355fc12f9f17f8bed6833b9e496d5e87406d4e517681176510832635" gracePeriod=2 Nov 24 15:34:24 crc kubenswrapper[4790]: I1124 15:34:24.044624 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xfcfp" Nov 24 15:34:24 crc kubenswrapper[4790]: I1124 15:34:24.191613 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwkg2\" (UniqueName: \"kubernetes.io/projected/155ff7f8-3c39-45f8-820a-a6a6e9edb190-kube-api-access-kwkg2\") pod \"155ff7f8-3c39-45f8-820a-a6a6e9edb190\" (UID: \"155ff7f8-3c39-45f8-820a-a6a6e9edb190\") " Nov 24 15:34:24 crc kubenswrapper[4790]: I1124 15:34:24.191760 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/155ff7f8-3c39-45f8-820a-a6a6e9edb190-utilities\") pod \"155ff7f8-3c39-45f8-820a-a6a6e9edb190\" (UID: \"155ff7f8-3c39-45f8-820a-a6a6e9edb190\") " Nov 24 15:34:24 crc kubenswrapper[4790]: I1124 15:34:24.191858 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/155ff7f8-3c39-45f8-820a-a6a6e9edb190-catalog-content\") pod \"155ff7f8-3c39-45f8-820a-a6a6e9edb190\" (UID: \"155ff7f8-3c39-45f8-820a-a6a6e9edb190\") " Nov 24 15:34:24 crc kubenswrapper[4790]: I1124 15:34:24.192449 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/155ff7f8-3c39-45f8-820a-a6a6e9edb190-utilities" (OuterVolumeSpecName: "utilities") pod "155ff7f8-3c39-45f8-820a-a6a6e9edb190" (UID: "155ff7f8-3c39-45f8-820a-a6a6e9edb190"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:34:24 crc kubenswrapper[4790]: I1124 15:34:24.197776 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/155ff7f8-3c39-45f8-820a-a6a6e9edb190-kube-api-access-kwkg2" (OuterVolumeSpecName: "kube-api-access-kwkg2") pod "155ff7f8-3c39-45f8-820a-a6a6e9edb190" (UID: "155ff7f8-3c39-45f8-820a-a6a6e9edb190"). InnerVolumeSpecName "kube-api-access-kwkg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:34:24 crc kubenswrapper[4790]: I1124 15:34:24.234410 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/155ff7f8-3c39-45f8-820a-a6a6e9edb190-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "155ff7f8-3c39-45f8-820a-a6a6e9edb190" (UID: "155ff7f8-3c39-45f8-820a-a6a6e9edb190"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:34:24 crc kubenswrapper[4790]: I1124 15:34:24.295481 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/155ff7f8-3c39-45f8-820a-a6a6e9edb190-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 15:34:24 crc kubenswrapper[4790]: I1124 15:34:24.295524 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/155ff7f8-3c39-45f8-820a-a6a6e9edb190-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 15:34:24 crc kubenswrapper[4790]: I1124 15:34:24.295540 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwkg2\" (UniqueName: \"kubernetes.io/projected/155ff7f8-3c39-45f8-820a-a6a6e9edb190-kube-api-access-kwkg2\") on node \"crc\" DevicePath \"\"" Nov 24 15:34:24 crc kubenswrapper[4790]: I1124 15:34:24.552526 4790 generic.go:334] "Generic (PLEG): container finished" podID="155ff7f8-3c39-45f8-820a-a6a6e9edb190" containerID="5bb2c063355fc12f9f17f8bed6833b9e496d5e87406d4e517681176510832635" exitCode=0 Nov 24 15:34:24 crc kubenswrapper[4790]: I1124 15:34:24.552717 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xfcfp" event={"ID":"155ff7f8-3c39-45f8-820a-a6a6e9edb190","Type":"ContainerDied","Data":"5bb2c063355fc12f9f17f8bed6833b9e496d5e87406d4e517681176510832635"} Nov 24 15:34:24 crc kubenswrapper[4790]: I1124 15:34:24.552952 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xfcfp" event={"ID":"155ff7f8-3c39-45f8-820a-a6a6e9edb190","Type":"ContainerDied","Data":"3aae9ee1c1f7881f1dbba2dde0185081f6818c7ee3f0fefc01cdf372fa9f009e"} Nov 24 15:34:24 crc kubenswrapper[4790]: I1124 15:34:24.552982 4790 scope.go:117] "RemoveContainer" containerID="5bb2c063355fc12f9f17f8bed6833b9e496d5e87406d4e517681176510832635" Nov 24 15:34:24 crc kubenswrapper[4790]: I1124 15:34:24.552776 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xfcfp" Nov 24 15:34:24 crc kubenswrapper[4790]: I1124 15:34:24.586190 4790 scope.go:117] "RemoveContainer" containerID="f9ce16a1d743796d3eca19bca2836cf439c6487a404b9d5f6e5adf23a1d180b8" Nov 24 15:34:24 crc kubenswrapper[4790]: I1124 15:34:24.587414 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xfcfp"] Nov 24 15:34:24 crc kubenswrapper[4790]: I1124 15:34:24.602690 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xfcfp"] Nov 24 15:34:24 crc kubenswrapper[4790]: I1124 15:34:24.623213 4790 scope.go:117] "RemoveContainer" containerID="9051cd9a038c4557ec91b129e0092a8f416468e7fd6d61e15ba9624f390dbf4f" Nov 24 15:34:24 crc kubenswrapper[4790]: I1124 15:34:24.657403 4790 scope.go:117] "RemoveContainer" containerID="5bb2c063355fc12f9f17f8bed6833b9e496d5e87406d4e517681176510832635" Nov 24 15:34:24 crc kubenswrapper[4790]: E1124 15:34:24.657851 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bb2c063355fc12f9f17f8bed6833b9e496d5e87406d4e517681176510832635\": container with ID starting with 5bb2c063355fc12f9f17f8bed6833b9e496d5e87406d4e517681176510832635 not found: ID does not exist" containerID="5bb2c063355fc12f9f17f8bed6833b9e496d5e87406d4e517681176510832635" Nov 24 15:34:24 crc kubenswrapper[4790]: I1124 15:34:24.657921 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bb2c063355fc12f9f17f8bed6833b9e496d5e87406d4e517681176510832635"} err="failed to get container status \"5bb2c063355fc12f9f17f8bed6833b9e496d5e87406d4e517681176510832635\": rpc error: code = NotFound desc = could not find container \"5bb2c063355fc12f9f17f8bed6833b9e496d5e87406d4e517681176510832635\": container with ID starting with 5bb2c063355fc12f9f17f8bed6833b9e496d5e87406d4e517681176510832635 not found: ID does not exist" Nov 24 15:34:24 crc kubenswrapper[4790]: I1124 15:34:24.657950 4790 scope.go:117] "RemoveContainer" containerID="f9ce16a1d743796d3eca19bca2836cf439c6487a404b9d5f6e5adf23a1d180b8" Nov 24 15:34:24 crc kubenswrapper[4790]: E1124 15:34:24.658208 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9ce16a1d743796d3eca19bca2836cf439c6487a404b9d5f6e5adf23a1d180b8\": container with ID starting with f9ce16a1d743796d3eca19bca2836cf439c6487a404b9d5f6e5adf23a1d180b8 not found: ID does not exist" containerID="f9ce16a1d743796d3eca19bca2836cf439c6487a404b9d5f6e5adf23a1d180b8" Nov 24 15:34:24 crc kubenswrapper[4790]: I1124 15:34:24.658234 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9ce16a1d743796d3eca19bca2836cf439c6487a404b9d5f6e5adf23a1d180b8"} err="failed to get container status \"f9ce16a1d743796d3eca19bca2836cf439c6487a404b9d5f6e5adf23a1d180b8\": rpc error: code = NotFound desc = could not find container \"f9ce16a1d743796d3eca19bca2836cf439c6487a404b9d5f6e5adf23a1d180b8\": container with ID starting with f9ce16a1d743796d3eca19bca2836cf439c6487a404b9d5f6e5adf23a1d180b8 not found: ID does not exist" Nov 24 15:34:24 crc kubenswrapper[4790]: I1124 15:34:24.658253 4790 scope.go:117] "RemoveContainer" containerID="9051cd9a038c4557ec91b129e0092a8f416468e7fd6d61e15ba9624f390dbf4f" Nov 24 15:34:24 crc kubenswrapper[4790]: E1124 15:34:24.659071 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9051cd9a038c4557ec91b129e0092a8f416468e7fd6d61e15ba9624f390dbf4f\": container with ID starting with 9051cd9a038c4557ec91b129e0092a8f416468e7fd6d61e15ba9624f390dbf4f not found: ID does not exist" containerID="9051cd9a038c4557ec91b129e0092a8f416468e7fd6d61e15ba9624f390dbf4f" Nov 24 15:34:24 crc kubenswrapper[4790]: I1124 15:34:24.659117 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9051cd9a038c4557ec91b129e0092a8f416468e7fd6d61e15ba9624f390dbf4f"} err="failed to get container status \"9051cd9a038c4557ec91b129e0092a8f416468e7fd6d61e15ba9624f390dbf4f\": rpc error: code = NotFound desc = could not find container \"9051cd9a038c4557ec91b129e0092a8f416468e7fd6d61e15ba9624f390dbf4f\": container with ID starting with 9051cd9a038c4557ec91b129e0092a8f416468e7fd6d61e15ba9624f390dbf4f not found: ID does not exist" Nov 24 15:34:26 crc kubenswrapper[4790]: I1124 15:34:26.337685 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="155ff7f8-3c39-45f8-820a-a6a6e9edb190" path="/var/lib/kubelet/pods/155ff7f8-3c39-45f8-820a-a6a6e9edb190/volumes" Nov 24 15:35:14 crc kubenswrapper[4790]: I1124 15:35:14.485970 4790 generic.go:334] "Generic (PLEG): container finished" podID="d286ee05-e6fc-4945-b7a9-87d56a06486c" containerID="c3ed390eaf9203ec0e138d68ce403b2d5f5341ad24f873fc2626d667c7d21d61" exitCode=0 Nov 24 15:35:14 crc kubenswrapper[4790]: I1124 15:35:14.486775 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-ncldn" event={"ID":"d286ee05-e6fc-4945-b7a9-87d56a06486c","Type":"ContainerDied","Data":"c3ed390eaf9203ec0e138d68ce403b2d5f5341ad24f873fc2626d667c7d21d61"} Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.171602 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-ncldn" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.253709 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-ssh-key\") pod \"d286ee05-e6fc-4945-b7a9-87d56a06486c\" (UID: \"d286ee05-e6fc-4945-b7a9-87d56a06486c\") " Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.253864 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-inventory\") pod \"d286ee05-e6fc-4945-b7a9-87d56a06486c\" (UID: \"d286ee05-e6fc-4945-b7a9-87d56a06486c\") " Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.253995 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-ceilometer-compute-config-data-1\") pod \"d286ee05-e6fc-4945-b7a9-87d56a06486c\" (UID: \"d286ee05-e6fc-4945-b7a9-87d56a06486c\") " Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.254044 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-ceilometer-compute-config-data-2\") pod \"d286ee05-e6fc-4945-b7a9-87d56a06486c\" (UID: \"d286ee05-e6fc-4945-b7a9-87d56a06486c\") " Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.254197 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-telemetry-combined-ca-bundle\") pod \"d286ee05-e6fc-4945-b7a9-87d56a06486c\" (UID: \"d286ee05-e6fc-4945-b7a9-87d56a06486c\") " Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.254408 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mbwh\" (UniqueName: \"kubernetes.io/projected/d286ee05-e6fc-4945-b7a9-87d56a06486c-kube-api-access-2mbwh\") pod \"d286ee05-e6fc-4945-b7a9-87d56a06486c\" (UID: \"d286ee05-e6fc-4945-b7a9-87d56a06486c\") " Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.254557 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-ceilometer-compute-config-data-0\") pod \"d286ee05-e6fc-4945-b7a9-87d56a06486c\" (UID: \"d286ee05-e6fc-4945-b7a9-87d56a06486c\") " Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.254622 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-ceph\") pod \"d286ee05-e6fc-4945-b7a9-87d56a06486c\" (UID: \"d286ee05-e6fc-4945-b7a9-87d56a06486c\") " Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.261835 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d286ee05-e6fc-4945-b7a9-87d56a06486c-kube-api-access-2mbwh" (OuterVolumeSpecName: "kube-api-access-2mbwh") pod "d286ee05-e6fc-4945-b7a9-87d56a06486c" (UID: "d286ee05-e6fc-4945-b7a9-87d56a06486c"). InnerVolumeSpecName "kube-api-access-2mbwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.263602 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-ceph" (OuterVolumeSpecName: "ceph") pod "d286ee05-e6fc-4945-b7a9-87d56a06486c" (UID: "d286ee05-e6fc-4945-b7a9-87d56a06486c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.281837 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "d286ee05-e6fc-4945-b7a9-87d56a06486c" (UID: "d286ee05-e6fc-4945-b7a9-87d56a06486c"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.300841 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d286ee05-e6fc-4945-b7a9-87d56a06486c" (UID: "d286ee05-e6fc-4945-b7a9-87d56a06486c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.301748 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "d286ee05-e6fc-4945-b7a9-87d56a06486c" (UID: "d286ee05-e6fc-4945-b7a9-87d56a06486c"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.303971 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "d286ee05-e6fc-4945-b7a9-87d56a06486c" (UID: "d286ee05-e6fc-4945-b7a9-87d56a06486c"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.307733 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-inventory" (OuterVolumeSpecName: "inventory") pod "d286ee05-e6fc-4945-b7a9-87d56a06486c" (UID: "d286ee05-e6fc-4945-b7a9-87d56a06486c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.310361 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "d286ee05-e6fc-4945-b7a9-87d56a06486c" (UID: "d286ee05-e6fc-4945-b7a9-87d56a06486c"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.358697 4790 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.358740 4790 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.358759 4790 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.358777 4790 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.358790 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mbwh\" (UniqueName: \"kubernetes.io/projected/d286ee05-e6fc-4945-b7a9-87d56a06486c-kube-api-access-2mbwh\") on node \"crc\" DevicePath \"\"" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.358803 4790 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.358815 4790 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.358827 4790 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d286ee05-e6fc-4945-b7a9-87d56a06486c-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.516590 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-ncldn" event={"ID":"d286ee05-e6fc-4945-b7a9-87d56a06486c","Type":"ContainerDied","Data":"11d584cbc6b006a34b84bed78480e5070bc9272ca1c38974fbe2d1812f071095"} Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.517031 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11d584cbc6b006a34b84bed78480e5070bc9272ca1c38974fbe2d1812f071095" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.516642 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-ncldn" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.651084 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-d9tc9"] Nov 24 15:35:16 crc kubenswrapper[4790]: E1124 15:35:16.651540 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="155ff7f8-3c39-45f8-820a-a6a6e9edb190" containerName="extract-utilities" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.651556 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="155ff7f8-3c39-45f8-820a-a6a6e9edb190" containerName="extract-utilities" Nov 24 15:35:16 crc kubenswrapper[4790]: E1124 15:35:16.651572 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="155ff7f8-3c39-45f8-820a-a6a6e9edb190" containerName="extract-content" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.651578 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="155ff7f8-3c39-45f8-820a-a6a6e9edb190" containerName="extract-content" Nov 24 15:35:16 crc kubenswrapper[4790]: E1124 15:35:16.651597 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="155ff7f8-3c39-45f8-820a-a6a6e9edb190" containerName="registry-server" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.651603 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="155ff7f8-3c39-45f8-820a-a6a6e9edb190" containerName="registry-server" Nov 24 15:35:16 crc kubenswrapper[4790]: E1124 15:35:16.651619 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d286ee05-e6fc-4945-b7a9-87d56a06486c" containerName="telemetry-openstack-openstack-cell1" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.651624 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="d286ee05-e6fc-4945-b7a9-87d56a06486c" containerName="telemetry-openstack-openstack-cell1" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.651812 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="d286ee05-e6fc-4945-b7a9-87d56a06486c" containerName="telemetry-openstack-openstack-cell1" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.651831 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="155ff7f8-3c39-45f8-820a-a6a6e9edb190" containerName="registry-server" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.652543 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-d9tc9" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.654719 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.655102 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.655759 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fc497" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.656369 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.656510 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-sriov-agent-neutron-config" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.679308 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-d9tc9"] Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.768757 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/367b480c-25ab-497e-aac9-26a29735867c-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-d9tc9\" (UID: \"367b480c-25ab-497e-aac9-26a29735867c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-d9tc9" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.768809 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/367b480c-25ab-497e-aac9-26a29735867c-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-d9tc9\" (UID: \"367b480c-25ab-497e-aac9-26a29735867c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-d9tc9" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.769131 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/367b480c-25ab-497e-aac9-26a29735867c-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-d9tc9\" (UID: \"367b480c-25ab-497e-aac9-26a29735867c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-d9tc9" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.769205 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8858n\" (UniqueName: \"kubernetes.io/projected/367b480c-25ab-497e-aac9-26a29735867c-kube-api-access-8858n\") pod \"neutron-sriov-openstack-openstack-cell1-d9tc9\" (UID: \"367b480c-25ab-497e-aac9-26a29735867c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-d9tc9" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.769255 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/367b480c-25ab-497e-aac9-26a29735867c-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-d9tc9\" (UID: \"367b480c-25ab-497e-aac9-26a29735867c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-d9tc9" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.769296 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/367b480c-25ab-497e-aac9-26a29735867c-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-d9tc9\" (UID: \"367b480c-25ab-497e-aac9-26a29735867c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-d9tc9" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.871107 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/367b480c-25ab-497e-aac9-26a29735867c-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-d9tc9\" (UID: \"367b480c-25ab-497e-aac9-26a29735867c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-d9tc9" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.871172 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8858n\" (UniqueName: \"kubernetes.io/projected/367b480c-25ab-497e-aac9-26a29735867c-kube-api-access-8858n\") pod \"neutron-sriov-openstack-openstack-cell1-d9tc9\" (UID: \"367b480c-25ab-497e-aac9-26a29735867c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-d9tc9" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.871212 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/367b480c-25ab-497e-aac9-26a29735867c-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-d9tc9\" (UID: \"367b480c-25ab-497e-aac9-26a29735867c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-d9tc9" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.871252 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/367b480c-25ab-497e-aac9-26a29735867c-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-d9tc9\" (UID: \"367b480c-25ab-497e-aac9-26a29735867c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-d9tc9" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.871367 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/367b480c-25ab-497e-aac9-26a29735867c-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-d9tc9\" (UID: \"367b480c-25ab-497e-aac9-26a29735867c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-d9tc9" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.871393 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/367b480c-25ab-497e-aac9-26a29735867c-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-d9tc9\" (UID: \"367b480c-25ab-497e-aac9-26a29735867c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-d9tc9" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.878614 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/367b480c-25ab-497e-aac9-26a29735867c-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-d9tc9\" (UID: \"367b480c-25ab-497e-aac9-26a29735867c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-d9tc9" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.878795 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/367b480c-25ab-497e-aac9-26a29735867c-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-d9tc9\" (UID: \"367b480c-25ab-497e-aac9-26a29735867c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-d9tc9" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.879291 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/367b480c-25ab-497e-aac9-26a29735867c-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-d9tc9\" (UID: \"367b480c-25ab-497e-aac9-26a29735867c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-d9tc9" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.879839 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/367b480c-25ab-497e-aac9-26a29735867c-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-d9tc9\" (UID: \"367b480c-25ab-497e-aac9-26a29735867c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-d9tc9" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.881725 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/367b480c-25ab-497e-aac9-26a29735867c-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-d9tc9\" (UID: \"367b480c-25ab-497e-aac9-26a29735867c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-d9tc9" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.890857 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8858n\" (UniqueName: \"kubernetes.io/projected/367b480c-25ab-497e-aac9-26a29735867c-kube-api-access-8858n\") pod \"neutron-sriov-openstack-openstack-cell1-d9tc9\" (UID: \"367b480c-25ab-497e-aac9-26a29735867c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-d9tc9" Nov 24 15:35:16 crc kubenswrapper[4790]: I1124 15:35:16.975216 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-d9tc9" Nov 24 15:35:17 crc kubenswrapper[4790]: I1124 15:35:17.593763 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-d9tc9"] Nov 24 15:35:18 crc kubenswrapper[4790]: I1124 15:35:18.566150 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-d9tc9" event={"ID":"367b480c-25ab-497e-aac9-26a29735867c","Type":"ContainerStarted","Data":"27994adea0ee37066e49077b73900bc48a6ce48b2cc8be2a0f093c305797bad9"} Nov 24 15:35:18 crc kubenswrapper[4790]: I1124 15:35:18.566482 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-d9tc9" event={"ID":"367b480c-25ab-497e-aac9-26a29735867c","Type":"ContainerStarted","Data":"620f85fa49154624103e70c0dda1796987e5b6db168f3ca42ec874563b18f5ae"} Nov 24 15:35:18 crc kubenswrapper[4790]: I1124 15:35:18.590021 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-sriov-openstack-openstack-cell1-d9tc9" podStartSLOduration=2.388621184 podStartE2EDuration="2.590000836s" podCreationTimestamp="2025-11-24 15:35:16 +0000 UTC" firstStartedPulling="2025-11-24 15:35:17.598974859 +0000 UTC m=+8565.978868531" lastFinishedPulling="2025-11-24 15:35:17.800354521 +0000 UTC m=+8566.180248183" observedRunningTime="2025-11-24 15:35:18.582818223 +0000 UTC m=+8566.962711885" watchObservedRunningTime="2025-11-24 15:35:18.590000836 +0000 UTC m=+8566.969894508" Nov 24 15:35:46 crc kubenswrapper[4790]: I1124 15:35:46.292954 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-n7v8f"] Nov 24 15:35:46 crc kubenswrapper[4790]: I1124 15:35:46.296169 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n7v8f" Nov 24 15:35:46 crc kubenswrapper[4790]: I1124 15:35:46.329518 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n7v8f"] Nov 24 15:35:46 crc kubenswrapper[4790]: I1124 15:35:46.383532 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60e7a2da-ad17-4a92-9608-c86b119962ee-catalog-content\") pod \"community-operators-n7v8f\" (UID: \"60e7a2da-ad17-4a92-9608-c86b119962ee\") " pod="openshift-marketplace/community-operators-n7v8f" Nov 24 15:35:46 crc kubenswrapper[4790]: I1124 15:35:46.383588 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s84jk\" (UniqueName: \"kubernetes.io/projected/60e7a2da-ad17-4a92-9608-c86b119962ee-kube-api-access-s84jk\") pod \"community-operators-n7v8f\" (UID: \"60e7a2da-ad17-4a92-9608-c86b119962ee\") " pod="openshift-marketplace/community-operators-n7v8f" Nov 24 15:35:46 crc kubenswrapper[4790]: I1124 15:35:46.383631 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60e7a2da-ad17-4a92-9608-c86b119962ee-utilities\") pod \"community-operators-n7v8f\" (UID: \"60e7a2da-ad17-4a92-9608-c86b119962ee\") " pod="openshift-marketplace/community-operators-n7v8f" Nov 24 15:35:46 crc kubenswrapper[4790]: I1124 15:35:46.485085 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60e7a2da-ad17-4a92-9608-c86b119962ee-catalog-content\") pod \"community-operators-n7v8f\" (UID: \"60e7a2da-ad17-4a92-9608-c86b119962ee\") " pod="openshift-marketplace/community-operators-n7v8f" Nov 24 15:35:46 crc kubenswrapper[4790]: I1124 15:35:46.485390 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s84jk\" (UniqueName: \"kubernetes.io/projected/60e7a2da-ad17-4a92-9608-c86b119962ee-kube-api-access-s84jk\") pod \"community-operators-n7v8f\" (UID: \"60e7a2da-ad17-4a92-9608-c86b119962ee\") " pod="openshift-marketplace/community-operators-n7v8f" Nov 24 15:35:46 crc kubenswrapper[4790]: I1124 15:35:46.485426 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60e7a2da-ad17-4a92-9608-c86b119962ee-utilities\") pod \"community-operators-n7v8f\" (UID: \"60e7a2da-ad17-4a92-9608-c86b119962ee\") " pod="openshift-marketplace/community-operators-n7v8f" Nov 24 15:35:46 crc kubenswrapper[4790]: I1124 15:35:46.485713 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60e7a2da-ad17-4a92-9608-c86b119962ee-catalog-content\") pod \"community-operators-n7v8f\" (UID: \"60e7a2da-ad17-4a92-9608-c86b119962ee\") " pod="openshift-marketplace/community-operators-n7v8f" Nov 24 15:35:46 crc kubenswrapper[4790]: I1124 15:35:46.485898 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60e7a2da-ad17-4a92-9608-c86b119962ee-utilities\") pod \"community-operators-n7v8f\" (UID: \"60e7a2da-ad17-4a92-9608-c86b119962ee\") " pod="openshift-marketplace/community-operators-n7v8f" Nov 24 15:35:46 crc kubenswrapper[4790]: I1124 15:35:46.506069 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s84jk\" (UniqueName: \"kubernetes.io/projected/60e7a2da-ad17-4a92-9608-c86b119962ee-kube-api-access-s84jk\") pod \"community-operators-n7v8f\" (UID: \"60e7a2da-ad17-4a92-9608-c86b119962ee\") " pod="openshift-marketplace/community-operators-n7v8f" Nov 24 15:35:46 crc kubenswrapper[4790]: I1124 15:35:46.617873 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n7v8f" Nov 24 15:35:47 crc kubenswrapper[4790]: I1124 15:35:47.206095 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n7v8f"] Nov 24 15:35:47 crc kubenswrapper[4790]: I1124 15:35:47.992098 4790 generic.go:334] "Generic (PLEG): container finished" podID="60e7a2da-ad17-4a92-9608-c86b119962ee" containerID="4cfe0f277b568b34ca12035154c8d4873c94c79c13c9cd4ee73d75a53de89362" exitCode=0 Nov 24 15:35:47 crc kubenswrapper[4790]: I1124 15:35:47.992164 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n7v8f" event={"ID":"60e7a2da-ad17-4a92-9608-c86b119962ee","Type":"ContainerDied","Data":"4cfe0f277b568b34ca12035154c8d4873c94c79c13c9cd4ee73d75a53de89362"} Nov 24 15:35:47 crc kubenswrapper[4790]: I1124 15:35:47.992653 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n7v8f" event={"ID":"60e7a2da-ad17-4a92-9608-c86b119962ee","Type":"ContainerStarted","Data":"e3e9b4596143cfb7fb6e8fafeba2978a781322c18af52f0ba75b39e04c673499"} Nov 24 15:35:48 crc kubenswrapper[4790]: I1124 15:35:48.500155 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fg59n"] Nov 24 15:35:48 crc kubenswrapper[4790]: I1124 15:35:48.511257 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fg59n" Nov 24 15:35:48 crc kubenswrapper[4790]: I1124 15:35:48.535370 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fg59n"] Nov 24 15:35:48 crc kubenswrapper[4790]: I1124 15:35:48.554734 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1db7b347-aedc-410f-8f52-a92c615a0328-utilities\") pod \"redhat-marketplace-fg59n\" (UID: \"1db7b347-aedc-410f-8f52-a92c615a0328\") " pod="openshift-marketplace/redhat-marketplace-fg59n" Nov 24 15:35:48 crc kubenswrapper[4790]: I1124 15:35:48.554856 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1db7b347-aedc-410f-8f52-a92c615a0328-catalog-content\") pod \"redhat-marketplace-fg59n\" (UID: \"1db7b347-aedc-410f-8f52-a92c615a0328\") " pod="openshift-marketplace/redhat-marketplace-fg59n" Nov 24 15:35:48 crc kubenswrapper[4790]: I1124 15:35:48.554951 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggb5w\" (UniqueName: \"kubernetes.io/projected/1db7b347-aedc-410f-8f52-a92c615a0328-kube-api-access-ggb5w\") pod \"redhat-marketplace-fg59n\" (UID: \"1db7b347-aedc-410f-8f52-a92c615a0328\") " pod="openshift-marketplace/redhat-marketplace-fg59n" Nov 24 15:35:48 crc kubenswrapper[4790]: I1124 15:35:48.662209 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1db7b347-aedc-410f-8f52-a92c615a0328-utilities\") pod \"redhat-marketplace-fg59n\" (UID: \"1db7b347-aedc-410f-8f52-a92c615a0328\") " pod="openshift-marketplace/redhat-marketplace-fg59n" Nov 24 15:35:48 crc kubenswrapper[4790]: I1124 15:35:48.662253 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1db7b347-aedc-410f-8f52-a92c615a0328-catalog-content\") pod \"redhat-marketplace-fg59n\" (UID: \"1db7b347-aedc-410f-8f52-a92c615a0328\") " pod="openshift-marketplace/redhat-marketplace-fg59n" Nov 24 15:35:48 crc kubenswrapper[4790]: I1124 15:35:48.662271 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggb5w\" (UniqueName: \"kubernetes.io/projected/1db7b347-aedc-410f-8f52-a92c615a0328-kube-api-access-ggb5w\") pod \"redhat-marketplace-fg59n\" (UID: \"1db7b347-aedc-410f-8f52-a92c615a0328\") " pod="openshift-marketplace/redhat-marketplace-fg59n" Nov 24 15:35:48 crc kubenswrapper[4790]: I1124 15:35:48.662781 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1db7b347-aedc-410f-8f52-a92c615a0328-utilities\") pod \"redhat-marketplace-fg59n\" (UID: \"1db7b347-aedc-410f-8f52-a92c615a0328\") " pod="openshift-marketplace/redhat-marketplace-fg59n" Nov 24 15:35:48 crc kubenswrapper[4790]: I1124 15:35:48.662790 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1db7b347-aedc-410f-8f52-a92c615a0328-catalog-content\") pod \"redhat-marketplace-fg59n\" (UID: \"1db7b347-aedc-410f-8f52-a92c615a0328\") " pod="openshift-marketplace/redhat-marketplace-fg59n" Nov 24 15:35:48 crc kubenswrapper[4790]: I1124 15:35:48.691259 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggb5w\" (UniqueName: \"kubernetes.io/projected/1db7b347-aedc-410f-8f52-a92c615a0328-kube-api-access-ggb5w\") pod \"redhat-marketplace-fg59n\" (UID: \"1db7b347-aedc-410f-8f52-a92c615a0328\") " pod="openshift-marketplace/redhat-marketplace-fg59n" Nov 24 15:35:48 crc kubenswrapper[4790]: I1124 15:35:48.860847 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fg59n" Nov 24 15:35:49 crc kubenswrapper[4790]: I1124 15:35:49.014031 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n7v8f" event={"ID":"60e7a2da-ad17-4a92-9608-c86b119962ee","Type":"ContainerStarted","Data":"9eb398f1d67f46e0cd849f27da2cf4fddcc478aafd9598df8ff2053da6682fcf"} Nov 24 15:35:49 crc kubenswrapper[4790]: I1124 15:35:49.385987 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fg59n"] Nov 24 15:35:49 crc kubenswrapper[4790]: W1124 15:35:49.389291 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1db7b347_aedc_410f_8f52_a92c615a0328.slice/crio-397d8a52547ca161d69aff8ec18d07acec91a638376236cc45288c4871673dc1 WatchSource:0}: Error finding container 397d8a52547ca161d69aff8ec18d07acec91a638376236cc45288c4871673dc1: Status 404 returned error can't find the container with id 397d8a52547ca161d69aff8ec18d07acec91a638376236cc45288c4871673dc1 Nov 24 15:35:50 crc kubenswrapper[4790]: I1124 15:35:50.033513 4790 generic.go:334] "Generic (PLEG): container finished" podID="1db7b347-aedc-410f-8f52-a92c615a0328" containerID="bdffd2a1ca8c2d6b6fa410146b8b84e2a90286cd62bc1a99584420c60d455ec8" exitCode=0 Nov 24 15:35:50 crc kubenswrapper[4790]: I1124 15:35:50.033604 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fg59n" event={"ID":"1db7b347-aedc-410f-8f52-a92c615a0328","Type":"ContainerDied","Data":"bdffd2a1ca8c2d6b6fa410146b8b84e2a90286cd62bc1a99584420c60d455ec8"} Nov 24 15:35:50 crc kubenswrapper[4790]: I1124 15:35:50.034092 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fg59n" event={"ID":"1db7b347-aedc-410f-8f52-a92c615a0328","Type":"ContainerStarted","Data":"397d8a52547ca161d69aff8ec18d07acec91a638376236cc45288c4871673dc1"} Nov 24 15:35:50 crc kubenswrapper[4790]: I1124 15:35:50.037081 4790 generic.go:334] "Generic (PLEG): container finished" podID="60e7a2da-ad17-4a92-9608-c86b119962ee" containerID="9eb398f1d67f46e0cd849f27da2cf4fddcc478aafd9598df8ff2053da6682fcf" exitCode=0 Nov 24 15:35:50 crc kubenswrapper[4790]: I1124 15:35:50.037126 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n7v8f" event={"ID":"60e7a2da-ad17-4a92-9608-c86b119962ee","Type":"ContainerDied","Data":"9eb398f1d67f46e0cd849f27da2cf4fddcc478aafd9598df8ff2053da6682fcf"} Nov 24 15:35:51 crc kubenswrapper[4790]: I1124 15:35:51.059079 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n7v8f" event={"ID":"60e7a2da-ad17-4a92-9608-c86b119962ee","Type":"ContainerStarted","Data":"b3f4ec858c84b084748bfe596f4cf244783041add925abc8ebdcc6b2e5cdbdf3"} Nov 24 15:35:51 crc kubenswrapper[4790]: I1124 15:35:51.063272 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fg59n" event={"ID":"1db7b347-aedc-410f-8f52-a92c615a0328","Type":"ContainerStarted","Data":"d25827b9e15f274ca37efef49d2750cafbfe9a18c91e121b5ed1a07d296707df"} Nov 24 15:35:51 crc kubenswrapper[4790]: I1124 15:35:51.087134 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-n7v8f" podStartSLOduration=2.556229988 podStartE2EDuration="5.087109323s" podCreationTimestamp="2025-11-24 15:35:46 +0000 UTC" firstStartedPulling="2025-11-24 15:35:47.994673917 +0000 UTC m=+8596.374567579" lastFinishedPulling="2025-11-24 15:35:50.525553242 +0000 UTC m=+8598.905446914" observedRunningTime="2025-11-24 15:35:51.07818161 +0000 UTC m=+8599.458075282" watchObservedRunningTime="2025-11-24 15:35:51.087109323 +0000 UTC m=+8599.467002985" Nov 24 15:35:52 crc kubenswrapper[4790]: I1124 15:35:52.079945 4790 generic.go:334] "Generic (PLEG): container finished" podID="1db7b347-aedc-410f-8f52-a92c615a0328" containerID="d25827b9e15f274ca37efef49d2750cafbfe9a18c91e121b5ed1a07d296707df" exitCode=0 Nov 24 15:35:52 crc kubenswrapper[4790]: I1124 15:35:52.080083 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fg59n" event={"ID":"1db7b347-aedc-410f-8f52-a92c615a0328","Type":"ContainerDied","Data":"d25827b9e15f274ca37efef49d2750cafbfe9a18c91e121b5ed1a07d296707df"} Nov 24 15:35:53 crc kubenswrapper[4790]: I1124 15:35:53.094049 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fg59n" event={"ID":"1db7b347-aedc-410f-8f52-a92c615a0328","Type":"ContainerStarted","Data":"13587e6ed8818360200496fc5e87a6129bdd8cb48833306ed12cf2b26d9744a7"} Nov 24 15:35:53 crc kubenswrapper[4790]: I1124 15:35:53.118608 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fg59n" podStartSLOduration=2.649565708 podStartE2EDuration="5.118587855s" podCreationTimestamp="2025-11-24 15:35:48 +0000 UTC" firstStartedPulling="2025-11-24 15:35:50.035578876 +0000 UTC m=+8598.415472548" lastFinishedPulling="2025-11-24 15:35:52.504601033 +0000 UTC m=+8600.884494695" observedRunningTime="2025-11-24 15:35:53.114219232 +0000 UTC m=+8601.494112934" watchObservedRunningTime="2025-11-24 15:35:53.118587855 +0000 UTC m=+8601.498481527" Nov 24 15:35:56 crc kubenswrapper[4790]: I1124 15:35:56.618332 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-n7v8f" Nov 24 15:35:56 crc kubenswrapper[4790]: I1124 15:35:56.618783 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-n7v8f" Nov 24 15:35:57 crc kubenswrapper[4790]: I1124 15:35:57.689335 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-n7v8f" podUID="60e7a2da-ad17-4a92-9608-c86b119962ee" containerName="registry-server" probeResult="failure" output=< Nov 24 15:35:57 crc kubenswrapper[4790]: timeout: failed to connect service ":50051" within 1s Nov 24 15:35:57 crc kubenswrapper[4790]: > Nov 24 15:35:58 crc kubenswrapper[4790]: I1124 15:35:58.862200 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fg59n" Nov 24 15:35:58 crc kubenswrapper[4790]: I1124 15:35:58.862537 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fg59n" Nov 24 15:35:58 crc kubenswrapper[4790]: I1124 15:35:58.956791 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fg59n" Nov 24 15:35:59 crc kubenswrapper[4790]: I1124 15:35:59.232858 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fg59n" Nov 24 15:35:59 crc kubenswrapper[4790]: I1124 15:35:59.294146 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fg59n"] Nov 24 15:36:01 crc kubenswrapper[4790]: I1124 15:36:01.180564 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fg59n" podUID="1db7b347-aedc-410f-8f52-a92c615a0328" containerName="registry-server" containerID="cri-o://13587e6ed8818360200496fc5e87a6129bdd8cb48833306ed12cf2b26d9744a7" gracePeriod=2 Nov 24 15:36:01 crc kubenswrapper[4790]: I1124 15:36:01.748753 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fg59n" Nov 24 15:36:01 crc kubenswrapper[4790]: I1124 15:36:01.797390 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggb5w\" (UniqueName: \"kubernetes.io/projected/1db7b347-aedc-410f-8f52-a92c615a0328-kube-api-access-ggb5w\") pod \"1db7b347-aedc-410f-8f52-a92c615a0328\" (UID: \"1db7b347-aedc-410f-8f52-a92c615a0328\") " Nov 24 15:36:01 crc kubenswrapper[4790]: I1124 15:36:01.797652 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1db7b347-aedc-410f-8f52-a92c615a0328-catalog-content\") pod \"1db7b347-aedc-410f-8f52-a92c615a0328\" (UID: \"1db7b347-aedc-410f-8f52-a92c615a0328\") " Nov 24 15:36:01 crc kubenswrapper[4790]: I1124 15:36:01.797692 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1db7b347-aedc-410f-8f52-a92c615a0328-utilities\") pod \"1db7b347-aedc-410f-8f52-a92c615a0328\" (UID: \"1db7b347-aedc-410f-8f52-a92c615a0328\") " Nov 24 15:36:01 crc kubenswrapper[4790]: I1124 15:36:01.798665 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1db7b347-aedc-410f-8f52-a92c615a0328-utilities" (OuterVolumeSpecName: "utilities") pod "1db7b347-aedc-410f-8f52-a92c615a0328" (UID: "1db7b347-aedc-410f-8f52-a92c615a0328"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:36:01 crc kubenswrapper[4790]: I1124 15:36:01.807186 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1db7b347-aedc-410f-8f52-a92c615a0328-kube-api-access-ggb5w" (OuterVolumeSpecName: "kube-api-access-ggb5w") pod "1db7b347-aedc-410f-8f52-a92c615a0328" (UID: "1db7b347-aedc-410f-8f52-a92c615a0328"). InnerVolumeSpecName "kube-api-access-ggb5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:36:01 crc kubenswrapper[4790]: I1124 15:36:01.821377 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1db7b347-aedc-410f-8f52-a92c615a0328-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1db7b347-aedc-410f-8f52-a92c615a0328" (UID: "1db7b347-aedc-410f-8f52-a92c615a0328"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:36:01 crc kubenswrapper[4790]: I1124 15:36:01.900273 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggb5w\" (UniqueName: \"kubernetes.io/projected/1db7b347-aedc-410f-8f52-a92c615a0328-kube-api-access-ggb5w\") on node \"crc\" DevicePath \"\"" Nov 24 15:36:01 crc kubenswrapper[4790]: I1124 15:36:01.900329 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1db7b347-aedc-410f-8f52-a92c615a0328-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 15:36:01 crc kubenswrapper[4790]: I1124 15:36:01.900350 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1db7b347-aedc-410f-8f52-a92c615a0328-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 15:36:02 crc kubenswrapper[4790]: I1124 15:36:02.195580 4790 generic.go:334] "Generic (PLEG): container finished" podID="1db7b347-aedc-410f-8f52-a92c615a0328" containerID="13587e6ed8818360200496fc5e87a6129bdd8cb48833306ed12cf2b26d9744a7" exitCode=0 Nov 24 15:36:02 crc kubenswrapper[4790]: I1124 15:36:02.195652 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fg59n" Nov 24 15:36:02 crc kubenswrapper[4790]: I1124 15:36:02.195650 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fg59n" event={"ID":"1db7b347-aedc-410f-8f52-a92c615a0328","Type":"ContainerDied","Data":"13587e6ed8818360200496fc5e87a6129bdd8cb48833306ed12cf2b26d9744a7"} Nov 24 15:36:02 crc kubenswrapper[4790]: I1124 15:36:02.195711 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fg59n" event={"ID":"1db7b347-aedc-410f-8f52-a92c615a0328","Type":"ContainerDied","Data":"397d8a52547ca161d69aff8ec18d07acec91a638376236cc45288c4871673dc1"} Nov 24 15:36:02 crc kubenswrapper[4790]: I1124 15:36:02.195748 4790 scope.go:117] "RemoveContainer" containerID="13587e6ed8818360200496fc5e87a6129bdd8cb48833306ed12cf2b26d9744a7" Nov 24 15:36:02 crc kubenswrapper[4790]: I1124 15:36:02.236048 4790 scope.go:117] "RemoveContainer" containerID="d25827b9e15f274ca37efef49d2750cafbfe9a18c91e121b5ed1a07d296707df" Nov 24 15:36:02 crc kubenswrapper[4790]: I1124 15:36:02.244485 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fg59n"] Nov 24 15:36:02 crc kubenswrapper[4790]: I1124 15:36:02.254079 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fg59n"] Nov 24 15:36:02 crc kubenswrapper[4790]: I1124 15:36:02.272015 4790 scope.go:117] "RemoveContainer" containerID="bdffd2a1ca8c2d6b6fa410146b8b84e2a90286cd62bc1a99584420c60d455ec8" Nov 24 15:36:02 crc kubenswrapper[4790]: I1124 15:36:02.320450 4790 scope.go:117] "RemoveContainer" containerID="13587e6ed8818360200496fc5e87a6129bdd8cb48833306ed12cf2b26d9744a7" Nov 24 15:36:02 crc kubenswrapper[4790]: E1124 15:36:02.321094 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13587e6ed8818360200496fc5e87a6129bdd8cb48833306ed12cf2b26d9744a7\": container with ID starting with 13587e6ed8818360200496fc5e87a6129bdd8cb48833306ed12cf2b26d9744a7 not found: ID does not exist" containerID="13587e6ed8818360200496fc5e87a6129bdd8cb48833306ed12cf2b26d9744a7" Nov 24 15:36:02 crc kubenswrapper[4790]: I1124 15:36:02.321139 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13587e6ed8818360200496fc5e87a6129bdd8cb48833306ed12cf2b26d9744a7"} err="failed to get container status \"13587e6ed8818360200496fc5e87a6129bdd8cb48833306ed12cf2b26d9744a7\": rpc error: code = NotFound desc = could not find container \"13587e6ed8818360200496fc5e87a6129bdd8cb48833306ed12cf2b26d9744a7\": container with ID starting with 13587e6ed8818360200496fc5e87a6129bdd8cb48833306ed12cf2b26d9744a7 not found: ID does not exist" Nov 24 15:36:02 crc kubenswrapper[4790]: I1124 15:36:02.321167 4790 scope.go:117] "RemoveContainer" containerID="d25827b9e15f274ca37efef49d2750cafbfe9a18c91e121b5ed1a07d296707df" Nov 24 15:36:02 crc kubenswrapper[4790]: E1124 15:36:02.321604 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d25827b9e15f274ca37efef49d2750cafbfe9a18c91e121b5ed1a07d296707df\": container with ID starting with d25827b9e15f274ca37efef49d2750cafbfe9a18c91e121b5ed1a07d296707df not found: ID does not exist" containerID="d25827b9e15f274ca37efef49d2750cafbfe9a18c91e121b5ed1a07d296707df" Nov 24 15:36:02 crc kubenswrapper[4790]: I1124 15:36:02.321653 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d25827b9e15f274ca37efef49d2750cafbfe9a18c91e121b5ed1a07d296707df"} err="failed to get container status \"d25827b9e15f274ca37efef49d2750cafbfe9a18c91e121b5ed1a07d296707df\": rpc error: code = NotFound desc = could not find container \"d25827b9e15f274ca37efef49d2750cafbfe9a18c91e121b5ed1a07d296707df\": container with ID starting with d25827b9e15f274ca37efef49d2750cafbfe9a18c91e121b5ed1a07d296707df not found: ID does not exist" Nov 24 15:36:02 crc kubenswrapper[4790]: I1124 15:36:02.321687 4790 scope.go:117] "RemoveContainer" containerID="bdffd2a1ca8c2d6b6fa410146b8b84e2a90286cd62bc1a99584420c60d455ec8" Nov 24 15:36:02 crc kubenswrapper[4790]: E1124 15:36:02.322005 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdffd2a1ca8c2d6b6fa410146b8b84e2a90286cd62bc1a99584420c60d455ec8\": container with ID starting with bdffd2a1ca8c2d6b6fa410146b8b84e2a90286cd62bc1a99584420c60d455ec8 not found: ID does not exist" containerID="bdffd2a1ca8c2d6b6fa410146b8b84e2a90286cd62bc1a99584420c60d455ec8" Nov 24 15:36:02 crc kubenswrapper[4790]: I1124 15:36:02.322032 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdffd2a1ca8c2d6b6fa410146b8b84e2a90286cd62bc1a99584420c60d455ec8"} err="failed to get container status \"bdffd2a1ca8c2d6b6fa410146b8b84e2a90286cd62bc1a99584420c60d455ec8\": rpc error: code = NotFound desc = could not find container \"bdffd2a1ca8c2d6b6fa410146b8b84e2a90286cd62bc1a99584420c60d455ec8\": container with ID starting with bdffd2a1ca8c2d6b6fa410146b8b84e2a90286cd62bc1a99584420c60d455ec8 not found: ID does not exist" Nov 24 15:36:02 crc kubenswrapper[4790]: I1124 15:36:02.333201 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1db7b347-aedc-410f-8f52-a92c615a0328" path="/var/lib/kubelet/pods/1db7b347-aedc-410f-8f52-a92c615a0328/volumes" Nov 24 15:36:06 crc kubenswrapper[4790]: I1124 15:36:06.715634 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-n7v8f" Nov 24 15:36:06 crc kubenswrapper[4790]: I1124 15:36:06.803814 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-n7v8f" Nov 24 15:36:06 crc kubenswrapper[4790]: I1124 15:36:06.971100 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n7v8f"] Nov 24 15:36:08 crc kubenswrapper[4790]: I1124 15:36:08.266728 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-n7v8f" podUID="60e7a2da-ad17-4a92-9608-c86b119962ee" containerName="registry-server" containerID="cri-o://b3f4ec858c84b084748bfe596f4cf244783041add925abc8ebdcc6b2e5cdbdf3" gracePeriod=2 Nov 24 15:36:08 crc kubenswrapper[4790]: I1124 15:36:08.735623 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n7v8f" Nov 24 15:36:08 crc kubenswrapper[4790]: I1124 15:36:08.774956 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s84jk\" (UniqueName: \"kubernetes.io/projected/60e7a2da-ad17-4a92-9608-c86b119962ee-kube-api-access-s84jk\") pod \"60e7a2da-ad17-4a92-9608-c86b119962ee\" (UID: \"60e7a2da-ad17-4a92-9608-c86b119962ee\") " Nov 24 15:36:08 crc kubenswrapper[4790]: I1124 15:36:08.775177 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60e7a2da-ad17-4a92-9608-c86b119962ee-catalog-content\") pod \"60e7a2da-ad17-4a92-9608-c86b119962ee\" (UID: \"60e7a2da-ad17-4a92-9608-c86b119962ee\") " Nov 24 15:36:08 crc kubenswrapper[4790]: I1124 15:36:08.775428 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60e7a2da-ad17-4a92-9608-c86b119962ee-utilities\") pod \"60e7a2da-ad17-4a92-9608-c86b119962ee\" (UID: \"60e7a2da-ad17-4a92-9608-c86b119962ee\") " Nov 24 15:36:08 crc kubenswrapper[4790]: I1124 15:36:08.776799 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60e7a2da-ad17-4a92-9608-c86b119962ee-utilities" (OuterVolumeSpecName: "utilities") pod "60e7a2da-ad17-4a92-9608-c86b119962ee" (UID: "60e7a2da-ad17-4a92-9608-c86b119962ee"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:36:08 crc kubenswrapper[4790]: I1124 15:36:08.785200 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60e7a2da-ad17-4a92-9608-c86b119962ee-kube-api-access-s84jk" (OuterVolumeSpecName: "kube-api-access-s84jk") pod "60e7a2da-ad17-4a92-9608-c86b119962ee" (UID: "60e7a2da-ad17-4a92-9608-c86b119962ee"). InnerVolumeSpecName "kube-api-access-s84jk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:36:08 crc kubenswrapper[4790]: I1124 15:36:08.830415 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60e7a2da-ad17-4a92-9608-c86b119962ee-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "60e7a2da-ad17-4a92-9608-c86b119962ee" (UID: "60e7a2da-ad17-4a92-9608-c86b119962ee"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:36:08 crc kubenswrapper[4790]: I1124 15:36:08.877943 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60e7a2da-ad17-4a92-9608-c86b119962ee-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 15:36:08 crc kubenswrapper[4790]: I1124 15:36:08.877981 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60e7a2da-ad17-4a92-9608-c86b119962ee-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 15:36:08 crc kubenswrapper[4790]: I1124 15:36:08.877994 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s84jk\" (UniqueName: \"kubernetes.io/projected/60e7a2da-ad17-4a92-9608-c86b119962ee-kube-api-access-s84jk\") on node \"crc\" DevicePath \"\"" Nov 24 15:36:09 crc kubenswrapper[4790]: I1124 15:36:09.286318 4790 generic.go:334] "Generic (PLEG): container finished" podID="60e7a2da-ad17-4a92-9608-c86b119962ee" containerID="b3f4ec858c84b084748bfe596f4cf244783041add925abc8ebdcc6b2e5cdbdf3" exitCode=0 Nov 24 15:36:09 crc kubenswrapper[4790]: I1124 15:36:09.286388 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n7v8f" event={"ID":"60e7a2da-ad17-4a92-9608-c86b119962ee","Type":"ContainerDied","Data":"b3f4ec858c84b084748bfe596f4cf244783041add925abc8ebdcc6b2e5cdbdf3"} Nov 24 15:36:09 crc kubenswrapper[4790]: I1124 15:36:09.286435 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n7v8f" Nov 24 15:36:09 crc kubenswrapper[4790]: I1124 15:36:09.287845 4790 scope.go:117] "RemoveContainer" containerID="b3f4ec858c84b084748bfe596f4cf244783041add925abc8ebdcc6b2e5cdbdf3" Nov 24 15:36:09 crc kubenswrapper[4790]: I1124 15:36:09.287751 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n7v8f" event={"ID":"60e7a2da-ad17-4a92-9608-c86b119962ee","Type":"ContainerDied","Data":"e3e9b4596143cfb7fb6e8fafeba2978a781322c18af52f0ba75b39e04c673499"} Nov 24 15:36:09 crc kubenswrapper[4790]: I1124 15:36:09.334920 4790 scope.go:117] "RemoveContainer" containerID="9eb398f1d67f46e0cd849f27da2cf4fddcc478aafd9598df8ff2053da6682fcf" Nov 24 15:36:09 crc kubenswrapper[4790]: I1124 15:36:09.355948 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n7v8f"] Nov 24 15:36:09 crc kubenswrapper[4790]: I1124 15:36:09.371187 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-n7v8f"] Nov 24 15:36:09 crc kubenswrapper[4790]: I1124 15:36:09.425009 4790 scope.go:117] "RemoveContainer" containerID="4cfe0f277b568b34ca12035154c8d4873c94c79c13c9cd4ee73d75a53de89362" Nov 24 15:36:09 crc kubenswrapper[4790]: I1124 15:36:09.467612 4790 scope.go:117] "RemoveContainer" containerID="b3f4ec858c84b084748bfe596f4cf244783041add925abc8ebdcc6b2e5cdbdf3" Nov 24 15:36:09 crc kubenswrapper[4790]: E1124 15:36:09.468276 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3f4ec858c84b084748bfe596f4cf244783041add925abc8ebdcc6b2e5cdbdf3\": container with ID starting with b3f4ec858c84b084748bfe596f4cf244783041add925abc8ebdcc6b2e5cdbdf3 not found: ID does not exist" containerID="b3f4ec858c84b084748bfe596f4cf244783041add925abc8ebdcc6b2e5cdbdf3" Nov 24 15:36:09 crc kubenswrapper[4790]: I1124 15:36:09.468346 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3f4ec858c84b084748bfe596f4cf244783041add925abc8ebdcc6b2e5cdbdf3"} err="failed to get container status \"b3f4ec858c84b084748bfe596f4cf244783041add925abc8ebdcc6b2e5cdbdf3\": rpc error: code = NotFound desc = could not find container \"b3f4ec858c84b084748bfe596f4cf244783041add925abc8ebdcc6b2e5cdbdf3\": container with ID starting with b3f4ec858c84b084748bfe596f4cf244783041add925abc8ebdcc6b2e5cdbdf3 not found: ID does not exist" Nov 24 15:36:09 crc kubenswrapper[4790]: I1124 15:36:09.468445 4790 scope.go:117] "RemoveContainer" containerID="9eb398f1d67f46e0cd849f27da2cf4fddcc478aafd9598df8ff2053da6682fcf" Nov 24 15:36:09 crc kubenswrapper[4790]: E1124 15:36:09.468927 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9eb398f1d67f46e0cd849f27da2cf4fddcc478aafd9598df8ff2053da6682fcf\": container with ID starting with 9eb398f1d67f46e0cd849f27da2cf4fddcc478aafd9598df8ff2053da6682fcf not found: ID does not exist" containerID="9eb398f1d67f46e0cd849f27da2cf4fddcc478aafd9598df8ff2053da6682fcf" Nov 24 15:36:09 crc kubenswrapper[4790]: I1124 15:36:09.469017 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9eb398f1d67f46e0cd849f27da2cf4fddcc478aafd9598df8ff2053da6682fcf"} err="failed to get container status \"9eb398f1d67f46e0cd849f27da2cf4fddcc478aafd9598df8ff2053da6682fcf\": rpc error: code = NotFound desc = could not find container \"9eb398f1d67f46e0cd849f27da2cf4fddcc478aafd9598df8ff2053da6682fcf\": container with ID starting with 9eb398f1d67f46e0cd849f27da2cf4fddcc478aafd9598df8ff2053da6682fcf not found: ID does not exist" Nov 24 15:36:09 crc kubenswrapper[4790]: I1124 15:36:09.469061 4790 scope.go:117] "RemoveContainer" containerID="4cfe0f277b568b34ca12035154c8d4873c94c79c13c9cd4ee73d75a53de89362" Nov 24 15:36:09 crc kubenswrapper[4790]: E1124 15:36:09.469396 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cfe0f277b568b34ca12035154c8d4873c94c79c13c9cd4ee73d75a53de89362\": container with ID starting with 4cfe0f277b568b34ca12035154c8d4873c94c79c13c9cd4ee73d75a53de89362 not found: ID does not exist" containerID="4cfe0f277b568b34ca12035154c8d4873c94c79c13c9cd4ee73d75a53de89362" Nov 24 15:36:09 crc kubenswrapper[4790]: I1124 15:36:09.469432 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cfe0f277b568b34ca12035154c8d4873c94c79c13c9cd4ee73d75a53de89362"} err="failed to get container status \"4cfe0f277b568b34ca12035154c8d4873c94c79c13c9cd4ee73d75a53de89362\": rpc error: code = NotFound desc = could not find container \"4cfe0f277b568b34ca12035154c8d4873c94c79c13c9cd4ee73d75a53de89362\": container with ID starting with 4cfe0f277b568b34ca12035154c8d4873c94c79c13c9cd4ee73d75a53de89362 not found: ID does not exist" Nov 24 15:36:10 crc kubenswrapper[4790]: I1124 15:36:10.336876 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60e7a2da-ad17-4a92-9608-c86b119962ee" path="/var/lib/kubelet/pods/60e7a2da-ad17-4a92-9608-c86b119962ee/volumes" Nov 24 15:36:13 crc kubenswrapper[4790]: I1124 15:36:13.938979 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:36:13 crc kubenswrapper[4790]: I1124 15:36:13.939696 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:36:43 crc kubenswrapper[4790]: I1124 15:36:43.938580 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:36:43 crc kubenswrapper[4790]: I1124 15:36:43.939353 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:37:13 crc kubenswrapper[4790]: I1124 15:37:13.939023 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:37:13 crc kubenswrapper[4790]: I1124 15:37:13.939828 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:37:13 crc kubenswrapper[4790]: I1124 15:37:13.939953 4790 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 15:37:13 crc kubenswrapper[4790]: I1124 15:37:13.941324 4790 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4237e8fdaaefb1b14cc56e0c19383d77cde5a1f384bd8ee6c5281df48485a3d9"} pod="openshift-machine-config-operator/machine-config-daemon-xz49t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 15:37:13 crc kubenswrapper[4790]: I1124 15:37:13.941446 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" containerID="cri-o://4237e8fdaaefb1b14cc56e0c19383d77cde5a1f384bd8ee6c5281df48485a3d9" gracePeriod=600 Nov 24 15:37:14 crc kubenswrapper[4790]: I1124 15:37:14.196198 4790 generic.go:334] "Generic (PLEG): container finished" podID="0d73b133-48f0-455f-8f6a-742e633f631a" containerID="4237e8fdaaefb1b14cc56e0c19383d77cde5a1f384bd8ee6c5281df48485a3d9" exitCode=0 Nov 24 15:37:14 crc kubenswrapper[4790]: I1124 15:37:14.196300 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerDied","Data":"4237e8fdaaefb1b14cc56e0c19383d77cde5a1f384bd8ee6c5281df48485a3d9"} Nov 24 15:37:14 crc kubenswrapper[4790]: I1124 15:37:14.196687 4790 scope.go:117] "RemoveContainer" containerID="87c151c65c6ec7e6e0150da1b299aebe7cb33ee5e3738cff003c27747ab5c270" Nov 24 15:37:15 crc kubenswrapper[4790]: I1124 15:37:15.218277 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"c579c182b84b6d5bab1987b612c6237aee4927c7e2c1e98dc1b59eca4477811d"} Nov 24 15:38:20 crc kubenswrapper[4790]: I1124 15:38:20.544471 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-drgck"] Nov 24 15:38:20 crc kubenswrapper[4790]: E1124 15:38:20.545709 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1db7b347-aedc-410f-8f52-a92c615a0328" containerName="registry-server" Nov 24 15:38:20 crc kubenswrapper[4790]: I1124 15:38:20.545732 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="1db7b347-aedc-410f-8f52-a92c615a0328" containerName="registry-server" Nov 24 15:38:20 crc kubenswrapper[4790]: E1124 15:38:20.545777 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1db7b347-aedc-410f-8f52-a92c615a0328" containerName="extract-utilities" Nov 24 15:38:20 crc kubenswrapper[4790]: I1124 15:38:20.545788 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="1db7b347-aedc-410f-8f52-a92c615a0328" containerName="extract-utilities" Nov 24 15:38:20 crc kubenswrapper[4790]: E1124 15:38:20.545826 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1db7b347-aedc-410f-8f52-a92c615a0328" containerName="extract-content" Nov 24 15:38:20 crc kubenswrapper[4790]: I1124 15:38:20.545838 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="1db7b347-aedc-410f-8f52-a92c615a0328" containerName="extract-content" Nov 24 15:38:20 crc kubenswrapper[4790]: E1124 15:38:20.545863 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60e7a2da-ad17-4a92-9608-c86b119962ee" containerName="extract-utilities" Nov 24 15:38:20 crc kubenswrapper[4790]: I1124 15:38:20.545874 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="60e7a2da-ad17-4a92-9608-c86b119962ee" containerName="extract-utilities" Nov 24 15:38:20 crc kubenswrapper[4790]: E1124 15:38:20.545938 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60e7a2da-ad17-4a92-9608-c86b119962ee" containerName="extract-content" Nov 24 15:38:20 crc kubenswrapper[4790]: I1124 15:38:20.545953 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="60e7a2da-ad17-4a92-9608-c86b119962ee" containerName="extract-content" Nov 24 15:38:20 crc kubenswrapper[4790]: E1124 15:38:20.545979 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60e7a2da-ad17-4a92-9608-c86b119962ee" containerName="registry-server" Nov 24 15:38:20 crc kubenswrapper[4790]: I1124 15:38:20.545990 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="60e7a2da-ad17-4a92-9608-c86b119962ee" containerName="registry-server" Nov 24 15:38:20 crc kubenswrapper[4790]: I1124 15:38:20.546324 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="1db7b347-aedc-410f-8f52-a92c615a0328" containerName="registry-server" Nov 24 15:38:20 crc kubenswrapper[4790]: I1124 15:38:20.546382 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="60e7a2da-ad17-4a92-9608-c86b119962ee" containerName="registry-server" Nov 24 15:38:20 crc kubenswrapper[4790]: I1124 15:38:20.549430 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-drgck" Nov 24 15:38:20 crc kubenswrapper[4790]: I1124 15:38:20.553999 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-drgck"] Nov 24 15:38:20 crc kubenswrapper[4790]: I1124 15:38:20.574351 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fg7w\" (UniqueName: \"kubernetes.io/projected/456cbf15-f908-4e81-96f9-2397e3e654ca-kube-api-access-4fg7w\") pod \"redhat-operators-drgck\" (UID: \"456cbf15-f908-4e81-96f9-2397e3e654ca\") " pod="openshift-marketplace/redhat-operators-drgck" Nov 24 15:38:20 crc kubenswrapper[4790]: I1124 15:38:20.574507 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/456cbf15-f908-4e81-96f9-2397e3e654ca-utilities\") pod \"redhat-operators-drgck\" (UID: \"456cbf15-f908-4e81-96f9-2397e3e654ca\") " pod="openshift-marketplace/redhat-operators-drgck" Nov 24 15:38:20 crc kubenswrapper[4790]: I1124 15:38:20.574980 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/456cbf15-f908-4e81-96f9-2397e3e654ca-catalog-content\") pod \"redhat-operators-drgck\" (UID: \"456cbf15-f908-4e81-96f9-2397e3e654ca\") " pod="openshift-marketplace/redhat-operators-drgck" Nov 24 15:38:20 crc kubenswrapper[4790]: I1124 15:38:20.676985 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fg7w\" (UniqueName: \"kubernetes.io/projected/456cbf15-f908-4e81-96f9-2397e3e654ca-kube-api-access-4fg7w\") pod \"redhat-operators-drgck\" (UID: \"456cbf15-f908-4e81-96f9-2397e3e654ca\") " pod="openshift-marketplace/redhat-operators-drgck" Nov 24 15:38:20 crc kubenswrapper[4790]: I1124 15:38:20.677053 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/456cbf15-f908-4e81-96f9-2397e3e654ca-utilities\") pod \"redhat-operators-drgck\" (UID: \"456cbf15-f908-4e81-96f9-2397e3e654ca\") " pod="openshift-marketplace/redhat-operators-drgck" Nov 24 15:38:20 crc kubenswrapper[4790]: I1124 15:38:20.677182 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/456cbf15-f908-4e81-96f9-2397e3e654ca-catalog-content\") pod \"redhat-operators-drgck\" (UID: \"456cbf15-f908-4e81-96f9-2397e3e654ca\") " pod="openshift-marketplace/redhat-operators-drgck" Nov 24 15:38:20 crc kubenswrapper[4790]: I1124 15:38:20.677809 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/456cbf15-f908-4e81-96f9-2397e3e654ca-catalog-content\") pod \"redhat-operators-drgck\" (UID: \"456cbf15-f908-4e81-96f9-2397e3e654ca\") " pod="openshift-marketplace/redhat-operators-drgck" Nov 24 15:38:20 crc kubenswrapper[4790]: I1124 15:38:20.677818 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/456cbf15-f908-4e81-96f9-2397e3e654ca-utilities\") pod \"redhat-operators-drgck\" (UID: \"456cbf15-f908-4e81-96f9-2397e3e654ca\") " pod="openshift-marketplace/redhat-operators-drgck" Nov 24 15:38:20 crc kubenswrapper[4790]: I1124 15:38:20.707143 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fg7w\" (UniqueName: \"kubernetes.io/projected/456cbf15-f908-4e81-96f9-2397e3e654ca-kube-api-access-4fg7w\") pod \"redhat-operators-drgck\" (UID: \"456cbf15-f908-4e81-96f9-2397e3e654ca\") " pod="openshift-marketplace/redhat-operators-drgck" Nov 24 15:38:20 crc kubenswrapper[4790]: I1124 15:38:20.871836 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-drgck" Nov 24 15:38:21 crc kubenswrapper[4790]: I1124 15:38:21.394816 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-drgck"] Nov 24 15:38:21 crc kubenswrapper[4790]: W1124 15:38:21.414863 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod456cbf15_f908_4e81_96f9_2397e3e654ca.slice/crio-7998a175ebd85847ec9b2fce4a4b0390ad948fd7a078a72f54631773bb5eb016 WatchSource:0}: Error finding container 7998a175ebd85847ec9b2fce4a4b0390ad948fd7a078a72f54631773bb5eb016: Status 404 returned error can't find the container with id 7998a175ebd85847ec9b2fce4a4b0390ad948fd7a078a72f54631773bb5eb016 Nov 24 15:38:22 crc kubenswrapper[4790]: I1124 15:38:22.115209 4790 generic.go:334] "Generic (PLEG): container finished" podID="456cbf15-f908-4e81-96f9-2397e3e654ca" containerID="d0c961b759f1d1134ece0e6fa352cf9834fd485d8867c58dfcda05c8e9e00018" exitCode=0 Nov 24 15:38:22 crc kubenswrapper[4790]: I1124 15:38:22.115273 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-drgck" event={"ID":"456cbf15-f908-4e81-96f9-2397e3e654ca","Type":"ContainerDied","Data":"d0c961b759f1d1134ece0e6fa352cf9834fd485d8867c58dfcda05c8e9e00018"} Nov 24 15:38:22 crc kubenswrapper[4790]: I1124 15:38:22.115525 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-drgck" event={"ID":"456cbf15-f908-4e81-96f9-2397e3e654ca","Type":"ContainerStarted","Data":"7998a175ebd85847ec9b2fce4a4b0390ad948fd7a078a72f54631773bb5eb016"} Nov 24 15:38:24 crc kubenswrapper[4790]: I1124 15:38:24.154914 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-drgck" event={"ID":"456cbf15-f908-4e81-96f9-2397e3e654ca","Type":"ContainerStarted","Data":"ac5c9543d69ff66ad085119fc373c8c7bed2f6c99ec0aa2fb3a17fcb0ab42e86"} Nov 24 15:38:27 crc kubenswrapper[4790]: I1124 15:38:27.188138 4790 generic.go:334] "Generic (PLEG): container finished" podID="456cbf15-f908-4e81-96f9-2397e3e654ca" containerID="ac5c9543d69ff66ad085119fc373c8c7bed2f6c99ec0aa2fb3a17fcb0ab42e86" exitCode=0 Nov 24 15:38:27 crc kubenswrapper[4790]: I1124 15:38:27.188240 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-drgck" event={"ID":"456cbf15-f908-4e81-96f9-2397e3e654ca","Type":"ContainerDied","Data":"ac5c9543d69ff66ad085119fc373c8c7bed2f6c99ec0aa2fb3a17fcb0ab42e86"} Nov 24 15:38:28 crc kubenswrapper[4790]: I1124 15:38:28.201306 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-drgck" event={"ID":"456cbf15-f908-4e81-96f9-2397e3e654ca","Type":"ContainerStarted","Data":"7bd7b4ca6a808d8b4356df3caf417ddf132c22cc5703a1b2bac514ba5b29b13a"} Nov 24 15:38:28 crc kubenswrapper[4790]: I1124 15:38:28.228031 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-drgck" podStartSLOduration=2.591748882 podStartE2EDuration="8.22800776s" podCreationTimestamp="2025-11-24 15:38:20 +0000 UTC" firstStartedPulling="2025-11-24 15:38:22.117537701 +0000 UTC m=+8750.497431393" lastFinishedPulling="2025-11-24 15:38:27.753796609 +0000 UTC m=+8756.133690271" observedRunningTime="2025-11-24 15:38:28.221842096 +0000 UTC m=+8756.601735768" watchObservedRunningTime="2025-11-24 15:38:28.22800776 +0000 UTC m=+8756.607901432" Nov 24 15:38:30 crc kubenswrapper[4790]: I1124 15:38:30.872676 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-drgck" Nov 24 15:38:30 crc kubenswrapper[4790]: I1124 15:38:30.873196 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-drgck" Nov 24 15:38:31 crc kubenswrapper[4790]: I1124 15:38:31.937263 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-drgck" podUID="456cbf15-f908-4e81-96f9-2397e3e654ca" containerName="registry-server" probeResult="failure" output=< Nov 24 15:38:31 crc kubenswrapper[4790]: timeout: failed to connect service ":50051" within 1s Nov 24 15:38:31 crc kubenswrapper[4790]: > Nov 24 15:38:40 crc kubenswrapper[4790]: I1124 15:38:40.926139 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-drgck" Nov 24 15:38:40 crc kubenswrapper[4790]: I1124 15:38:40.978588 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-drgck" Nov 24 15:38:41 crc kubenswrapper[4790]: I1124 15:38:41.183279 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-drgck"] Nov 24 15:38:42 crc kubenswrapper[4790]: I1124 15:38:42.372397 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-drgck" podUID="456cbf15-f908-4e81-96f9-2397e3e654ca" containerName="registry-server" containerID="cri-o://7bd7b4ca6a808d8b4356df3caf417ddf132c22cc5703a1b2bac514ba5b29b13a" gracePeriod=2 Nov 24 15:38:42 crc kubenswrapper[4790]: I1124 15:38:42.939481 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-drgck" Nov 24 15:38:43 crc kubenswrapper[4790]: I1124 15:38:43.086764 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fg7w\" (UniqueName: \"kubernetes.io/projected/456cbf15-f908-4e81-96f9-2397e3e654ca-kube-api-access-4fg7w\") pod \"456cbf15-f908-4e81-96f9-2397e3e654ca\" (UID: \"456cbf15-f908-4e81-96f9-2397e3e654ca\") " Nov 24 15:38:43 crc kubenswrapper[4790]: I1124 15:38:43.087003 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/456cbf15-f908-4e81-96f9-2397e3e654ca-catalog-content\") pod \"456cbf15-f908-4e81-96f9-2397e3e654ca\" (UID: \"456cbf15-f908-4e81-96f9-2397e3e654ca\") " Nov 24 15:38:43 crc kubenswrapper[4790]: I1124 15:38:43.087309 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/456cbf15-f908-4e81-96f9-2397e3e654ca-utilities\") pod \"456cbf15-f908-4e81-96f9-2397e3e654ca\" (UID: \"456cbf15-f908-4e81-96f9-2397e3e654ca\") " Nov 24 15:38:43 crc kubenswrapper[4790]: I1124 15:38:43.088317 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/456cbf15-f908-4e81-96f9-2397e3e654ca-utilities" (OuterVolumeSpecName: "utilities") pod "456cbf15-f908-4e81-96f9-2397e3e654ca" (UID: "456cbf15-f908-4e81-96f9-2397e3e654ca"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:38:43 crc kubenswrapper[4790]: I1124 15:38:43.090228 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/456cbf15-f908-4e81-96f9-2397e3e654ca-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 15:38:43 crc kubenswrapper[4790]: I1124 15:38:43.094573 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/456cbf15-f908-4e81-96f9-2397e3e654ca-kube-api-access-4fg7w" (OuterVolumeSpecName: "kube-api-access-4fg7w") pod "456cbf15-f908-4e81-96f9-2397e3e654ca" (UID: "456cbf15-f908-4e81-96f9-2397e3e654ca"). InnerVolumeSpecName "kube-api-access-4fg7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:38:43 crc kubenswrapper[4790]: I1124 15:38:43.193608 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fg7w\" (UniqueName: \"kubernetes.io/projected/456cbf15-f908-4e81-96f9-2397e3e654ca-kube-api-access-4fg7w\") on node \"crc\" DevicePath \"\"" Nov 24 15:38:43 crc kubenswrapper[4790]: I1124 15:38:43.214137 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/456cbf15-f908-4e81-96f9-2397e3e654ca-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "456cbf15-f908-4e81-96f9-2397e3e654ca" (UID: "456cbf15-f908-4e81-96f9-2397e3e654ca"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:38:43 crc kubenswrapper[4790]: I1124 15:38:43.295416 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/456cbf15-f908-4e81-96f9-2397e3e654ca-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 15:38:43 crc kubenswrapper[4790]: I1124 15:38:43.389178 4790 generic.go:334] "Generic (PLEG): container finished" podID="456cbf15-f908-4e81-96f9-2397e3e654ca" containerID="7bd7b4ca6a808d8b4356df3caf417ddf132c22cc5703a1b2bac514ba5b29b13a" exitCode=0 Nov 24 15:38:43 crc kubenswrapper[4790]: I1124 15:38:43.389236 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-drgck" event={"ID":"456cbf15-f908-4e81-96f9-2397e3e654ca","Type":"ContainerDied","Data":"7bd7b4ca6a808d8b4356df3caf417ddf132c22cc5703a1b2bac514ba5b29b13a"} Nov 24 15:38:43 crc kubenswrapper[4790]: I1124 15:38:43.389281 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-drgck" event={"ID":"456cbf15-f908-4e81-96f9-2397e3e654ca","Type":"ContainerDied","Data":"7998a175ebd85847ec9b2fce4a4b0390ad948fd7a078a72f54631773bb5eb016"} Nov 24 15:38:43 crc kubenswrapper[4790]: I1124 15:38:43.389302 4790 scope.go:117] "RemoveContainer" containerID="7bd7b4ca6a808d8b4356df3caf417ddf132c22cc5703a1b2bac514ba5b29b13a" Nov 24 15:38:43 crc kubenswrapper[4790]: I1124 15:38:43.390821 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-drgck" Nov 24 15:38:43 crc kubenswrapper[4790]: I1124 15:38:43.427256 4790 scope.go:117] "RemoveContainer" containerID="ac5c9543d69ff66ad085119fc373c8c7bed2f6c99ec0aa2fb3a17fcb0ab42e86" Nov 24 15:38:43 crc kubenswrapper[4790]: I1124 15:38:43.441253 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-drgck"] Nov 24 15:38:43 crc kubenswrapper[4790]: I1124 15:38:43.451708 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-drgck"] Nov 24 15:38:43 crc kubenswrapper[4790]: I1124 15:38:43.452134 4790 scope.go:117] "RemoveContainer" containerID="d0c961b759f1d1134ece0e6fa352cf9834fd485d8867c58dfcda05c8e9e00018" Nov 24 15:38:43 crc kubenswrapper[4790]: I1124 15:38:43.521470 4790 scope.go:117] "RemoveContainer" containerID="7bd7b4ca6a808d8b4356df3caf417ddf132c22cc5703a1b2bac514ba5b29b13a" Nov 24 15:38:43 crc kubenswrapper[4790]: E1124 15:38:43.522025 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bd7b4ca6a808d8b4356df3caf417ddf132c22cc5703a1b2bac514ba5b29b13a\": container with ID starting with 7bd7b4ca6a808d8b4356df3caf417ddf132c22cc5703a1b2bac514ba5b29b13a not found: ID does not exist" containerID="7bd7b4ca6a808d8b4356df3caf417ddf132c22cc5703a1b2bac514ba5b29b13a" Nov 24 15:38:43 crc kubenswrapper[4790]: I1124 15:38:43.522072 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bd7b4ca6a808d8b4356df3caf417ddf132c22cc5703a1b2bac514ba5b29b13a"} err="failed to get container status \"7bd7b4ca6a808d8b4356df3caf417ddf132c22cc5703a1b2bac514ba5b29b13a\": rpc error: code = NotFound desc = could not find container \"7bd7b4ca6a808d8b4356df3caf417ddf132c22cc5703a1b2bac514ba5b29b13a\": container with ID starting with 7bd7b4ca6a808d8b4356df3caf417ddf132c22cc5703a1b2bac514ba5b29b13a not found: ID does not exist" Nov 24 15:38:43 crc kubenswrapper[4790]: I1124 15:38:43.522104 4790 scope.go:117] "RemoveContainer" containerID="ac5c9543d69ff66ad085119fc373c8c7bed2f6c99ec0aa2fb3a17fcb0ab42e86" Nov 24 15:38:43 crc kubenswrapper[4790]: E1124 15:38:43.530816 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac5c9543d69ff66ad085119fc373c8c7bed2f6c99ec0aa2fb3a17fcb0ab42e86\": container with ID starting with ac5c9543d69ff66ad085119fc373c8c7bed2f6c99ec0aa2fb3a17fcb0ab42e86 not found: ID does not exist" containerID="ac5c9543d69ff66ad085119fc373c8c7bed2f6c99ec0aa2fb3a17fcb0ab42e86" Nov 24 15:38:43 crc kubenswrapper[4790]: I1124 15:38:43.530844 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac5c9543d69ff66ad085119fc373c8c7bed2f6c99ec0aa2fb3a17fcb0ab42e86"} err="failed to get container status \"ac5c9543d69ff66ad085119fc373c8c7bed2f6c99ec0aa2fb3a17fcb0ab42e86\": rpc error: code = NotFound desc = could not find container \"ac5c9543d69ff66ad085119fc373c8c7bed2f6c99ec0aa2fb3a17fcb0ab42e86\": container with ID starting with ac5c9543d69ff66ad085119fc373c8c7bed2f6c99ec0aa2fb3a17fcb0ab42e86 not found: ID does not exist" Nov 24 15:38:43 crc kubenswrapper[4790]: I1124 15:38:43.530860 4790 scope.go:117] "RemoveContainer" containerID="d0c961b759f1d1134ece0e6fa352cf9834fd485d8867c58dfcda05c8e9e00018" Nov 24 15:38:43 crc kubenswrapper[4790]: E1124 15:38:43.531274 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0c961b759f1d1134ece0e6fa352cf9834fd485d8867c58dfcda05c8e9e00018\": container with ID starting with d0c961b759f1d1134ece0e6fa352cf9834fd485d8867c58dfcda05c8e9e00018 not found: ID does not exist" containerID="d0c961b759f1d1134ece0e6fa352cf9834fd485d8867c58dfcda05c8e9e00018" Nov 24 15:38:43 crc kubenswrapper[4790]: I1124 15:38:43.531397 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0c961b759f1d1134ece0e6fa352cf9834fd485d8867c58dfcda05c8e9e00018"} err="failed to get container status \"d0c961b759f1d1134ece0e6fa352cf9834fd485d8867c58dfcda05c8e9e00018\": rpc error: code = NotFound desc = could not find container \"d0c961b759f1d1134ece0e6fa352cf9834fd485d8867c58dfcda05c8e9e00018\": container with ID starting with d0c961b759f1d1134ece0e6fa352cf9834fd485d8867c58dfcda05c8e9e00018 not found: ID does not exist" Nov 24 15:38:44 crc kubenswrapper[4790]: I1124 15:38:44.337089 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="456cbf15-f908-4e81-96f9-2397e3e654ca" path="/var/lib/kubelet/pods/456cbf15-f908-4e81-96f9-2397e3e654ca/volumes" Nov 24 15:38:47 crc kubenswrapper[4790]: I1124 15:38:47.299041 4790 trace.go:236] Trace[1054999183]: "Calculate volume metrics of persistence for pod openstack/rabbitmq-server-0" (24-Nov-2025 15:38:46.176) (total time: 1122ms): Nov 24 15:38:47 crc kubenswrapper[4790]: Trace[1054999183]: [1.122153043s] [1.122153043s] END Nov 24 15:38:47 crc kubenswrapper[4790]: I1124 15:38:47.414667 4790 trace.go:236] Trace[1052955693]: "Calculate volume metrics of ovndbcluster-sb-etc-ovn for pod openstack/ovsdbserver-sb-0" (24-Nov-2025 15:38:46.354) (total time: 1059ms): Nov 24 15:38:47 crc kubenswrapper[4790]: Trace[1052955693]: [1.059680838s] [1.059680838s] END Nov 24 15:39:28 crc kubenswrapper[4790]: I1124 15:39:28.973739 4790 generic.go:334] "Generic (PLEG): container finished" podID="367b480c-25ab-497e-aac9-26a29735867c" containerID="27994adea0ee37066e49077b73900bc48a6ce48b2cc8be2a0f093c305797bad9" exitCode=0 Nov 24 15:39:28 crc kubenswrapper[4790]: I1124 15:39:28.974012 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-d9tc9" event={"ID":"367b480c-25ab-497e-aac9-26a29735867c","Type":"ContainerDied","Data":"27994adea0ee37066e49077b73900bc48a6ce48b2cc8be2a0f093c305797bad9"} Nov 24 15:39:30 crc kubenswrapper[4790]: I1124 15:39:30.568545 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-d9tc9" Nov 24 15:39:30 crc kubenswrapper[4790]: I1124 15:39:30.676572 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/367b480c-25ab-497e-aac9-26a29735867c-inventory\") pod \"367b480c-25ab-497e-aac9-26a29735867c\" (UID: \"367b480c-25ab-497e-aac9-26a29735867c\") " Nov 24 15:39:30 crc kubenswrapper[4790]: I1124 15:39:30.676744 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/367b480c-25ab-497e-aac9-26a29735867c-neutron-sriov-agent-neutron-config-0\") pod \"367b480c-25ab-497e-aac9-26a29735867c\" (UID: \"367b480c-25ab-497e-aac9-26a29735867c\") " Nov 24 15:39:30 crc kubenswrapper[4790]: I1124 15:39:30.676810 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/367b480c-25ab-497e-aac9-26a29735867c-ssh-key\") pod \"367b480c-25ab-497e-aac9-26a29735867c\" (UID: \"367b480c-25ab-497e-aac9-26a29735867c\") " Nov 24 15:39:30 crc kubenswrapper[4790]: I1124 15:39:30.676842 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/367b480c-25ab-497e-aac9-26a29735867c-ceph\") pod \"367b480c-25ab-497e-aac9-26a29735867c\" (UID: \"367b480c-25ab-497e-aac9-26a29735867c\") " Nov 24 15:39:30 crc kubenswrapper[4790]: I1124 15:39:30.676872 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8858n\" (UniqueName: \"kubernetes.io/projected/367b480c-25ab-497e-aac9-26a29735867c-kube-api-access-8858n\") pod \"367b480c-25ab-497e-aac9-26a29735867c\" (UID: \"367b480c-25ab-497e-aac9-26a29735867c\") " Nov 24 15:39:30 crc kubenswrapper[4790]: I1124 15:39:30.676933 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/367b480c-25ab-497e-aac9-26a29735867c-neutron-sriov-combined-ca-bundle\") pod \"367b480c-25ab-497e-aac9-26a29735867c\" (UID: \"367b480c-25ab-497e-aac9-26a29735867c\") " Nov 24 15:39:30 crc kubenswrapper[4790]: I1124 15:39:30.682629 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/367b480c-25ab-497e-aac9-26a29735867c-ceph" (OuterVolumeSpecName: "ceph") pod "367b480c-25ab-497e-aac9-26a29735867c" (UID: "367b480c-25ab-497e-aac9-26a29735867c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:39:30 crc kubenswrapper[4790]: I1124 15:39:30.683287 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/367b480c-25ab-497e-aac9-26a29735867c-kube-api-access-8858n" (OuterVolumeSpecName: "kube-api-access-8858n") pod "367b480c-25ab-497e-aac9-26a29735867c" (UID: "367b480c-25ab-497e-aac9-26a29735867c"). InnerVolumeSpecName "kube-api-access-8858n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:39:30 crc kubenswrapper[4790]: I1124 15:39:30.683470 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/367b480c-25ab-497e-aac9-26a29735867c-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "367b480c-25ab-497e-aac9-26a29735867c" (UID: "367b480c-25ab-497e-aac9-26a29735867c"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:39:30 crc kubenswrapper[4790]: I1124 15:39:30.709897 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/367b480c-25ab-497e-aac9-26a29735867c-neutron-sriov-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-sriov-agent-neutron-config-0") pod "367b480c-25ab-497e-aac9-26a29735867c" (UID: "367b480c-25ab-497e-aac9-26a29735867c"). InnerVolumeSpecName "neutron-sriov-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:39:30 crc kubenswrapper[4790]: I1124 15:39:30.711354 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/367b480c-25ab-497e-aac9-26a29735867c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "367b480c-25ab-497e-aac9-26a29735867c" (UID: "367b480c-25ab-497e-aac9-26a29735867c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:39:30 crc kubenswrapper[4790]: I1124 15:39:30.718377 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/367b480c-25ab-497e-aac9-26a29735867c-inventory" (OuterVolumeSpecName: "inventory") pod "367b480c-25ab-497e-aac9-26a29735867c" (UID: "367b480c-25ab-497e-aac9-26a29735867c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:39:30 crc kubenswrapper[4790]: I1124 15:39:30.779682 4790 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/367b480c-25ab-497e-aac9-26a29735867c-neutron-sriov-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 15:39:30 crc kubenswrapper[4790]: I1124 15:39:30.779718 4790 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/367b480c-25ab-497e-aac9-26a29735867c-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 15:39:30 crc kubenswrapper[4790]: I1124 15:39:30.779730 4790 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/367b480c-25ab-497e-aac9-26a29735867c-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 15:39:30 crc kubenswrapper[4790]: I1124 15:39:30.779739 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8858n\" (UniqueName: \"kubernetes.io/projected/367b480c-25ab-497e-aac9-26a29735867c-kube-api-access-8858n\") on node \"crc\" DevicePath \"\"" Nov 24 15:39:30 crc kubenswrapper[4790]: I1124 15:39:30.779750 4790 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/367b480c-25ab-497e-aac9-26a29735867c-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:39:30 crc kubenswrapper[4790]: I1124 15:39:30.779761 4790 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/367b480c-25ab-497e-aac9-26a29735867c-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.006664 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-d9tc9" event={"ID":"367b480c-25ab-497e-aac9-26a29735867c","Type":"ContainerDied","Data":"620f85fa49154624103e70c0dda1796987e5b6db168f3ca42ec874563b18f5ae"} Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.006730 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="620f85fa49154624103e70c0dda1796987e5b6db168f3ca42ec874563b18f5ae" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.006735 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-d9tc9" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.116025 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-n9962"] Nov 24 15:39:31 crc kubenswrapper[4790]: E1124 15:39:31.117009 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="456cbf15-f908-4e81-96f9-2397e3e654ca" containerName="extract-utilities" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.117255 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="456cbf15-f908-4e81-96f9-2397e3e654ca" containerName="extract-utilities" Nov 24 15:39:31 crc kubenswrapper[4790]: E1124 15:39:31.117292 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="456cbf15-f908-4e81-96f9-2397e3e654ca" containerName="registry-server" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.117305 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="456cbf15-f908-4e81-96f9-2397e3e654ca" containerName="registry-server" Nov 24 15:39:31 crc kubenswrapper[4790]: E1124 15:39:31.117342 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="367b480c-25ab-497e-aac9-26a29735867c" containerName="neutron-sriov-openstack-openstack-cell1" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.117357 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="367b480c-25ab-497e-aac9-26a29735867c" containerName="neutron-sriov-openstack-openstack-cell1" Nov 24 15:39:31 crc kubenswrapper[4790]: E1124 15:39:31.117380 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="456cbf15-f908-4e81-96f9-2397e3e654ca" containerName="extract-content" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.117392 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="456cbf15-f908-4e81-96f9-2397e3e654ca" containerName="extract-content" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.117814 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="456cbf15-f908-4e81-96f9-2397e3e654ca" containerName="registry-server" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.117853 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="367b480c-25ab-497e-aac9-26a29735867c" containerName="neutron-sriov-openstack-openstack-cell1" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.119167 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-n9962" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.123133 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-dhcp-agent-neutron-config" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.128421 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.134201 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fc497" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.134593 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.134622 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.146815 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-n9962"] Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.190154 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/515ed57f-53d0-470b-aafe-7deb07d527c9-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-n9962\" (UID: \"515ed57f-53d0-470b-aafe-7deb07d527c9\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-n9962" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.190356 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/515ed57f-53d0-470b-aafe-7deb07d527c9-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-n9962\" (UID: \"515ed57f-53d0-470b-aafe-7deb07d527c9\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-n9962" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.190429 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbl5q\" (UniqueName: \"kubernetes.io/projected/515ed57f-53d0-470b-aafe-7deb07d527c9-kube-api-access-rbl5q\") pod \"neutron-dhcp-openstack-openstack-cell1-n9962\" (UID: \"515ed57f-53d0-470b-aafe-7deb07d527c9\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-n9962" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.190565 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/515ed57f-53d0-470b-aafe-7deb07d527c9-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-n9962\" (UID: \"515ed57f-53d0-470b-aafe-7deb07d527c9\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-n9962" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.190654 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/515ed57f-53d0-470b-aafe-7deb07d527c9-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-n9962\" (UID: \"515ed57f-53d0-470b-aafe-7deb07d527c9\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-n9962" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.190735 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/515ed57f-53d0-470b-aafe-7deb07d527c9-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-n9962\" (UID: \"515ed57f-53d0-470b-aafe-7deb07d527c9\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-n9962" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.293800 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/515ed57f-53d0-470b-aafe-7deb07d527c9-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-n9962\" (UID: \"515ed57f-53d0-470b-aafe-7deb07d527c9\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-n9962" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.293993 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/515ed57f-53d0-470b-aafe-7deb07d527c9-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-n9962\" (UID: \"515ed57f-53d0-470b-aafe-7deb07d527c9\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-n9962" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.294046 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbl5q\" (UniqueName: \"kubernetes.io/projected/515ed57f-53d0-470b-aafe-7deb07d527c9-kube-api-access-rbl5q\") pod \"neutron-dhcp-openstack-openstack-cell1-n9962\" (UID: \"515ed57f-53d0-470b-aafe-7deb07d527c9\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-n9962" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.294096 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/515ed57f-53d0-470b-aafe-7deb07d527c9-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-n9962\" (UID: \"515ed57f-53d0-470b-aafe-7deb07d527c9\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-n9962" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.294141 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/515ed57f-53d0-470b-aafe-7deb07d527c9-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-n9962\" (UID: \"515ed57f-53d0-470b-aafe-7deb07d527c9\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-n9962" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.294192 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/515ed57f-53d0-470b-aafe-7deb07d527c9-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-n9962\" (UID: \"515ed57f-53d0-470b-aafe-7deb07d527c9\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-n9962" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.298042 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/515ed57f-53d0-470b-aafe-7deb07d527c9-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-n9962\" (UID: \"515ed57f-53d0-470b-aafe-7deb07d527c9\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-n9962" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.298289 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/515ed57f-53d0-470b-aafe-7deb07d527c9-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-n9962\" (UID: \"515ed57f-53d0-470b-aafe-7deb07d527c9\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-n9962" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.299464 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/515ed57f-53d0-470b-aafe-7deb07d527c9-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-n9962\" (UID: \"515ed57f-53d0-470b-aafe-7deb07d527c9\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-n9962" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.299737 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/515ed57f-53d0-470b-aafe-7deb07d527c9-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-n9962\" (UID: \"515ed57f-53d0-470b-aafe-7deb07d527c9\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-n9962" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.301268 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/515ed57f-53d0-470b-aafe-7deb07d527c9-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-n9962\" (UID: \"515ed57f-53d0-470b-aafe-7deb07d527c9\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-n9962" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.312837 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbl5q\" (UniqueName: \"kubernetes.io/projected/515ed57f-53d0-470b-aafe-7deb07d527c9-kube-api-access-rbl5q\") pod \"neutron-dhcp-openstack-openstack-cell1-n9962\" (UID: \"515ed57f-53d0-470b-aafe-7deb07d527c9\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-n9962" Nov 24 15:39:31 crc kubenswrapper[4790]: I1124 15:39:31.458933 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-n9962" Nov 24 15:39:32 crc kubenswrapper[4790]: I1124 15:39:32.071678 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-n9962"] Nov 24 15:39:32 crc kubenswrapper[4790]: I1124 15:39:32.080556 4790 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 15:39:33 crc kubenswrapper[4790]: I1124 15:39:33.033089 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-n9962" event={"ID":"515ed57f-53d0-470b-aafe-7deb07d527c9","Type":"ContainerStarted","Data":"522114cd527dfd31f5450ce7f9586288f25d8827aeeb648281ebc806e5bcdc6f"} Nov 24 15:39:33 crc kubenswrapper[4790]: I1124 15:39:33.033594 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-n9962" event={"ID":"515ed57f-53d0-470b-aafe-7deb07d527c9","Type":"ContainerStarted","Data":"8ebe7ec7e7e358eac1d12a770ccbf043a78c17d9149b2ac9db7d189c45780a6b"} Nov 24 15:39:33 crc kubenswrapper[4790]: I1124 15:39:33.070598 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dhcp-openstack-openstack-cell1-n9962" podStartSLOduration=1.860500212 podStartE2EDuration="2.070580308s" podCreationTimestamp="2025-11-24 15:39:31 +0000 UTC" firstStartedPulling="2025-11-24 15:39:32.080201169 +0000 UTC m=+8820.460094851" lastFinishedPulling="2025-11-24 15:39:32.290281245 +0000 UTC m=+8820.670174947" observedRunningTime="2025-11-24 15:39:33.059496354 +0000 UTC m=+8821.439390016" watchObservedRunningTime="2025-11-24 15:39:33.070580308 +0000 UTC m=+8821.450473980" Nov 24 15:39:43 crc kubenswrapper[4790]: I1124 15:39:43.939350 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:39:43 crc kubenswrapper[4790]: I1124 15:39:43.940132 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:40:13 crc kubenswrapper[4790]: I1124 15:40:13.939271 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:40:13 crc kubenswrapper[4790]: I1124 15:40:13.940123 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:40:43 crc kubenswrapper[4790]: I1124 15:40:43.939295 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:40:43 crc kubenswrapper[4790]: I1124 15:40:43.940199 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:40:43 crc kubenswrapper[4790]: I1124 15:40:43.940286 4790 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 15:40:43 crc kubenswrapper[4790]: I1124 15:40:43.941677 4790 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c579c182b84b6d5bab1987b612c6237aee4927c7e2c1e98dc1b59eca4477811d"} pod="openshift-machine-config-operator/machine-config-daemon-xz49t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 15:40:43 crc kubenswrapper[4790]: I1124 15:40:43.941781 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" containerID="cri-o://c579c182b84b6d5bab1987b612c6237aee4927c7e2c1e98dc1b59eca4477811d" gracePeriod=600 Nov 24 15:40:44 crc kubenswrapper[4790]: E1124 15:40:44.339799 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:40:45 crc kubenswrapper[4790]: I1124 15:40:45.013573 4790 generic.go:334] "Generic (PLEG): container finished" podID="0d73b133-48f0-455f-8f6a-742e633f631a" containerID="c579c182b84b6d5bab1987b612c6237aee4927c7e2c1e98dc1b59eca4477811d" exitCode=0 Nov 24 15:40:45 crc kubenswrapper[4790]: I1124 15:40:45.013951 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerDied","Data":"c579c182b84b6d5bab1987b612c6237aee4927c7e2c1e98dc1b59eca4477811d"} Nov 24 15:40:45 crc kubenswrapper[4790]: I1124 15:40:45.013998 4790 scope.go:117] "RemoveContainer" containerID="4237e8fdaaefb1b14cc56e0c19383d77cde5a1f384bd8ee6c5281df48485a3d9" Nov 24 15:40:45 crc kubenswrapper[4790]: I1124 15:40:45.015128 4790 scope.go:117] "RemoveContainer" containerID="c579c182b84b6d5bab1987b612c6237aee4927c7e2c1e98dc1b59eca4477811d" Nov 24 15:40:45 crc kubenswrapper[4790]: E1124 15:40:45.015648 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:40:58 crc kubenswrapper[4790]: I1124 15:40:58.315612 4790 scope.go:117] "RemoveContainer" containerID="c579c182b84b6d5bab1987b612c6237aee4927c7e2c1e98dc1b59eca4477811d" Nov 24 15:40:58 crc kubenswrapper[4790]: E1124 15:40:58.316693 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:41:13 crc kubenswrapper[4790]: I1124 15:41:13.314870 4790 scope.go:117] "RemoveContainer" containerID="c579c182b84b6d5bab1987b612c6237aee4927c7e2c1e98dc1b59eca4477811d" Nov 24 15:41:13 crc kubenswrapper[4790]: E1124 15:41:13.315678 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:41:26 crc kubenswrapper[4790]: I1124 15:41:26.315805 4790 scope.go:117] "RemoveContainer" containerID="c579c182b84b6d5bab1987b612c6237aee4927c7e2c1e98dc1b59eca4477811d" Nov 24 15:41:26 crc kubenswrapper[4790]: E1124 15:41:26.317363 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:41:40 crc kubenswrapper[4790]: I1124 15:41:40.314219 4790 scope.go:117] "RemoveContainer" containerID="c579c182b84b6d5bab1987b612c6237aee4927c7e2c1e98dc1b59eca4477811d" Nov 24 15:41:40 crc kubenswrapper[4790]: E1124 15:41:40.315214 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:41:51 crc kubenswrapper[4790]: I1124 15:41:51.315092 4790 scope.go:117] "RemoveContainer" containerID="c579c182b84b6d5bab1987b612c6237aee4927c7e2c1e98dc1b59eca4477811d" Nov 24 15:41:51 crc kubenswrapper[4790]: E1124 15:41:51.316032 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:42:02 crc kubenswrapper[4790]: I1124 15:42:02.328661 4790 scope.go:117] "RemoveContainer" containerID="c579c182b84b6d5bab1987b612c6237aee4927c7e2c1e98dc1b59eca4477811d" Nov 24 15:42:02 crc kubenswrapper[4790]: E1124 15:42:02.329692 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:42:13 crc kubenswrapper[4790]: I1124 15:42:13.315548 4790 scope.go:117] "RemoveContainer" containerID="c579c182b84b6d5bab1987b612c6237aee4927c7e2c1e98dc1b59eca4477811d" Nov 24 15:42:13 crc kubenswrapper[4790]: E1124 15:42:13.316336 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:42:27 crc kubenswrapper[4790]: I1124 15:42:27.315617 4790 scope.go:117] "RemoveContainer" containerID="c579c182b84b6d5bab1987b612c6237aee4927c7e2c1e98dc1b59eca4477811d" Nov 24 15:42:27 crc kubenswrapper[4790]: E1124 15:42:27.316588 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:42:41 crc kubenswrapper[4790]: I1124 15:42:41.315312 4790 scope.go:117] "RemoveContainer" containerID="c579c182b84b6d5bab1987b612c6237aee4927c7e2c1e98dc1b59eca4477811d" Nov 24 15:42:41 crc kubenswrapper[4790]: E1124 15:42:41.316762 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:42:55 crc kubenswrapper[4790]: I1124 15:42:55.315586 4790 scope.go:117] "RemoveContainer" containerID="c579c182b84b6d5bab1987b612c6237aee4927c7e2c1e98dc1b59eca4477811d" Nov 24 15:42:55 crc kubenswrapper[4790]: E1124 15:42:55.316538 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:43:09 crc kubenswrapper[4790]: I1124 15:43:09.316739 4790 scope.go:117] "RemoveContainer" containerID="c579c182b84b6d5bab1987b612c6237aee4927c7e2c1e98dc1b59eca4477811d" Nov 24 15:43:09 crc kubenswrapper[4790]: E1124 15:43:09.317977 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:43:22 crc kubenswrapper[4790]: I1124 15:43:22.328747 4790 scope.go:117] "RemoveContainer" containerID="c579c182b84b6d5bab1987b612c6237aee4927c7e2c1e98dc1b59eca4477811d" Nov 24 15:43:22 crc kubenswrapper[4790]: E1124 15:43:22.329971 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:43:33 crc kubenswrapper[4790]: I1124 15:43:33.314467 4790 scope.go:117] "RemoveContainer" containerID="c579c182b84b6d5bab1987b612c6237aee4927c7e2c1e98dc1b59eca4477811d" Nov 24 15:43:33 crc kubenswrapper[4790]: E1124 15:43:33.315464 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:43:44 crc kubenswrapper[4790]: I1124 15:43:44.314357 4790 scope.go:117] "RemoveContainer" containerID="c579c182b84b6d5bab1987b612c6237aee4927c7e2c1e98dc1b59eca4477811d" Nov 24 15:43:44 crc kubenswrapper[4790]: E1124 15:43:44.315021 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:43:59 crc kubenswrapper[4790]: I1124 15:43:59.315769 4790 scope.go:117] "RemoveContainer" containerID="c579c182b84b6d5bab1987b612c6237aee4927c7e2c1e98dc1b59eca4477811d" Nov 24 15:43:59 crc kubenswrapper[4790]: E1124 15:43:59.316959 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:44:11 crc kubenswrapper[4790]: I1124 15:44:11.314913 4790 scope.go:117] "RemoveContainer" containerID="c579c182b84b6d5bab1987b612c6237aee4927c7e2c1e98dc1b59eca4477811d" Nov 24 15:44:11 crc kubenswrapper[4790]: E1124 15:44:11.317925 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:44:24 crc kubenswrapper[4790]: I1124 15:44:24.316060 4790 scope.go:117] "RemoveContainer" containerID="c579c182b84b6d5bab1987b612c6237aee4927c7e2c1e98dc1b59eca4477811d" Nov 24 15:44:24 crc kubenswrapper[4790]: E1124 15:44:24.317162 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:44:39 crc kubenswrapper[4790]: I1124 15:44:39.315638 4790 scope.go:117] "RemoveContainer" containerID="c579c182b84b6d5bab1987b612c6237aee4927c7e2c1e98dc1b59eca4477811d" Nov 24 15:44:39 crc kubenswrapper[4790]: E1124 15:44:39.317054 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:44:50 crc kubenswrapper[4790]: I1124 15:44:50.316526 4790 scope.go:117] "RemoveContainer" containerID="c579c182b84b6d5bab1987b612c6237aee4927c7e2c1e98dc1b59eca4477811d" Nov 24 15:44:50 crc kubenswrapper[4790]: E1124 15:44:50.317310 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:44:58 crc kubenswrapper[4790]: I1124 15:44:58.311356 4790 trace.go:236] Trace[997835461]: "Calculate volume metrics of persistence for pod openstack/rabbitmq-server-0" (24-Nov-2025 15:44:57.263) (total time: 1047ms): Nov 24 15:44:58 crc kubenswrapper[4790]: Trace[997835461]: [1.047634224s] [1.047634224s] END Nov 24 15:45:00 crc kubenswrapper[4790]: I1124 15:45:00.344495 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399985-8xpvm"] Nov 24 15:45:00 crc kubenswrapper[4790]: I1124 15:45:00.346983 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399985-8xpvm" Nov 24 15:45:00 crc kubenswrapper[4790]: I1124 15:45:00.348751 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 15:45:00 crc kubenswrapper[4790]: I1124 15:45:00.349411 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 15:45:00 crc kubenswrapper[4790]: I1124 15:45:00.377006 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399985-8xpvm"] Nov 24 15:45:00 crc kubenswrapper[4790]: I1124 15:45:00.409140 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bb0d7854-9d30-4e0a-aefa-983d7a5d505f-config-volume\") pod \"collect-profiles-29399985-8xpvm\" (UID: \"bb0d7854-9d30-4e0a-aefa-983d7a5d505f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399985-8xpvm" Nov 24 15:45:00 crc kubenswrapper[4790]: I1124 15:45:00.409206 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bb0d7854-9d30-4e0a-aefa-983d7a5d505f-secret-volume\") pod \"collect-profiles-29399985-8xpvm\" (UID: \"bb0d7854-9d30-4e0a-aefa-983d7a5d505f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399985-8xpvm" Nov 24 15:45:00 crc kubenswrapper[4790]: I1124 15:45:00.409241 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frbsn\" (UniqueName: \"kubernetes.io/projected/bb0d7854-9d30-4e0a-aefa-983d7a5d505f-kube-api-access-frbsn\") pod \"collect-profiles-29399985-8xpvm\" (UID: \"bb0d7854-9d30-4e0a-aefa-983d7a5d505f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399985-8xpvm" Nov 24 15:45:00 crc kubenswrapper[4790]: I1124 15:45:00.511298 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bb0d7854-9d30-4e0a-aefa-983d7a5d505f-config-volume\") pod \"collect-profiles-29399985-8xpvm\" (UID: \"bb0d7854-9d30-4e0a-aefa-983d7a5d505f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399985-8xpvm" Nov 24 15:45:00 crc kubenswrapper[4790]: I1124 15:45:00.511800 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bb0d7854-9d30-4e0a-aefa-983d7a5d505f-secret-volume\") pod \"collect-profiles-29399985-8xpvm\" (UID: \"bb0d7854-9d30-4e0a-aefa-983d7a5d505f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399985-8xpvm" Nov 24 15:45:00 crc kubenswrapper[4790]: I1124 15:45:00.511869 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frbsn\" (UniqueName: \"kubernetes.io/projected/bb0d7854-9d30-4e0a-aefa-983d7a5d505f-kube-api-access-frbsn\") pod \"collect-profiles-29399985-8xpvm\" (UID: \"bb0d7854-9d30-4e0a-aefa-983d7a5d505f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399985-8xpvm" Nov 24 15:45:00 crc kubenswrapper[4790]: I1124 15:45:00.512521 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bb0d7854-9d30-4e0a-aefa-983d7a5d505f-config-volume\") pod \"collect-profiles-29399985-8xpvm\" (UID: \"bb0d7854-9d30-4e0a-aefa-983d7a5d505f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399985-8xpvm" Nov 24 15:45:00 crc kubenswrapper[4790]: I1124 15:45:00.523388 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bb0d7854-9d30-4e0a-aefa-983d7a5d505f-secret-volume\") pod \"collect-profiles-29399985-8xpvm\" (UID: \"bb0d7854-9d30-4e0a-aefa-983d7a5d505f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399985-8xpvm" Nov 24 15:45:00 crc kubenswrapper[4790]: I1124 15:45:00.530103 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frbsn\" (UniqueName: \"kubernetes.io/projected/bb0d7854-9d30-4e0a-aefa-983d7a5d505f-kube-api-access-frbsn\") pod \"collect-profiles-29399985-8xpvm\" (UID: \"bb0d7854-9d30-4e0a-aefa-983d7a5d505f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399985-8xpvm" Nov 24 15:45:00 crc kubenswrapper[4790]: I1124 15:45:00.670648 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399985-8xpvm" Nov 24 15:45:01 crc kubenswrapper[4790]: I1124 15:45:01.135956 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399985-8xpvm"] Nov 24 15:45:01 crc kubenswrapper[4790]: I1124 15:45:01.316533 4790 scope.go:117] "RemoveContainer" containerID="c579c182b84b6d5bab1987b612c6237aee4927c7e2c1e98dc1b59eca4477811d" Nov 24 15:45:01 crc kubenswrapper[4790]: E1124 15:45:01.316827 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:45:01 crc kubenswrapper[4790]: I1124 15:45:01.537165 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399985-8xpvm" event={"ID":"bb0d7854-9d30-4e0a-aefa-983d7a5d505f","Type":"ContainerStarted","Data":"4fe818aaffb42ca204492f2ffc79bbc57da78e2ca7131f314bdba68c18b6c14b"} Nov 24 15:45:01 crc kubenswrapper[4790]: I1124 15:45:01.537502 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399985-8xpvm" event={"ID":"bb0d7854-9d30-4e0a-aefa-983d7a5d505f","Type":"ContainerStarted","Data":"19a18524817ed86f515e6f9d294977b99938ed7d53af4f93479d101c01b7ab6f"} Nov 24 15:45:01 crc kubenswrapper[4790]: I1124 15:45:01.559496 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29399985-8xpvm" podStartSLOduration=1.5594785519999999 podStartE2EDuration="1.559478552s" podCreationTimestamp="2025-11-24 15:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 15:45:01.554103559 +0000 UTC m=+9149.933997251" watchObservedRunningTime="2025-11-24 15:45:01.559478552 +0000 UTC m=+9149.939372214" Nov 24 15:45:02 crc kubenswrapper[4790]: I1124 15:45:02.565043 4790 generic.go:334] "Generic (PLEG): container finished" podID="bb0d7854-9d30-4e0a-aefa-983d7a5d505f" containerID="4fe818aaffb42ca204492f2ffc79bbc57da78e2ca7131f314bdba68c18b6c14b" exitCode=0 Nov 24 15:45:02 crc kubenswrapper[4790]: I1124 15:45:02.565181 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399985-8xpvm" event={"ID":"bb0d7854-9d30-4e0a-aefa-983d7a5d505f","Type":"ContainerDied","Data":"4fe818aaffb42ca204492f2ffc79bbc57da78e2ca7131f314bdba68c18b6c14b"} Nov 24 15:45:03 crc kubenswrapper[4790]: I1124 15:45:03.998379 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399985-8xpvm" Nov 24 15:45:04 crc kubenswrapper[4790]: I1124 15:45:04.109601 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frbsn\" (UniqueName: \"kubernetes.io/projected/bb0d7854-9d30-4e0a-aefa-983d7a5d505f-kube-api-access-frbsn\") pod \"bb0d7854-9d30-4e0a-aefa-983d7a5d505f\" (UID: \"bb0d7854-9d30-4e0a-aefa-983d7a5d505f\") " Nov 24 15:45:04 crc kubenswrapper[4790]: I1124 15:45:04.109942 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bb0d7854-9d30-4e0a-aefa-983d7a5d505f-secret-volume\") pod \"bb0d7854-9d30-4e0a-aefa-983d7a5d505f\" (UID: \"bb0d7854-9d30-4e0a-aefa-983d7a5d505f\") " Nov 24 15:45:04 crc kubenswrapper[4790]: I1124 15:45:04.109984 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bb0d7854-9d30-4e0a-aefa-983d7a5d505f-config-volume\") pod \"bb0d7854-9d30-4e0a-aefa-983d7a5d505f\" (UID: \"bb0d7854-9d30-4e0a-aefa-983d7a5d505f\") " Nov 24 15:45:04 crc kubenswrapper[4790]: I1124 15:45:04.110872 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb0d7854-9d30-4e0a-aefa-983d7a5d505f-config-volume" (OuterVolumeSpecName: "config-volume") pod "bb0d7854-9d30-4e0a-aefa-983d7a5d505f" (UID: "bb0d7854-9d30-4e0a-aefa-983d7a5d505f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 15:45:04 crc kubenswrapper[4790]: I1124 15:45:04.114933 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb0d7854-9d30-4e0a-aefa-983d7a5d505f-kube-api-access-frbsn" (OuterVolumeSpecName: "kube-api-access-frbsn") pod "bb0d7854-9d30-4e0a-aefa-983d7a5d505f" (UID: "bb0d7854-9d30-4e0a-aefa-983d7a5d505f"). InnerVolumeSpecName "kube-api-access-frbsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:45:04 crc kubenswrapper[4790]: I1124 15:45:04.116571 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb0d7854-9d30-4e0a-aefa-983d7a5d505f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "bb0d7854-9d30-4e0a-aefa-983d7a5d505f" (UID: "bb0d7854-9d30-4e0a-aefa-983d7a5d505f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:45:04 crc kubenswrapper[4790]: I1124 15:45:04.213974 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frbsn\" (UniqueName: \"kubernetes.io/projected/bb0d7854-9d30-4e0a-aefa-983d7a5d505f-kube-api-access-frbsn\") on node \"crc\" DevicePath \"\"" Nov 24 15:45:04 crc kubenswrapper[4790]: I1124 15:45:04.214012 4790 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bb0d7854-9d30-4e0a-aefa-983d7a5d505f-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 15:45:04 crc kubenswrapper[4790]: I1124 15:45:04.214024 4790 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bb0d7854-9d30-4e0a-aefa-983d7a5d505f-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 15:45:04 crc kubenswrapper[4790]: I1124 15:45:04.590661 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399985-8xpvm" event={"ID":"bb0d7854-9d30-4e0a-aefa-983d7a5d505f","Type":"ContainerDied","Data":"19a18524817ed86f515e6f9d294977b99938ed7d53af4f93479d101c01b7ab6f"} Nov 24 15:45:04 crc kubenswrapper[4790]: I1124 15:45:04.590717 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="19a18524817ed86f515e6f9d294977b99938ed7d53af4f93479d101c01b7ab6f" Nov 24 15:45:04 crc kubenswrapper[4790]: I1124 15:45:04.590786 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399985-8xpvm" Nov 24 15:45:04 crc kubenswrapper[4790]: I1124 15:45:04.645247 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399940-48kdm"] Nov 24 15:45:04 crc kubenswrapper[4790]: I1124 15:45:04.653828 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399940-48kdm"] Nov 24 15:45:06 crc kubenswrapper[4790]: I1124 15:45:06.332260 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e688f6b-ed59-42f5-b5fb-3199f6ac7570" path="/var/lib/kubelet/pods/2e688f6b-ed59-42f5-b5fb-3199f6ac7570/volumes" Nov 24 15:45:13 crc kubenswrapper[4790]: I1124 15:45:13.314257 4790 scope.go:117] "RemoveContainer" containerID="c579c182b84b6d5bab1987b612c6237aee4927c7e2c1e98dc1b59eca4477811d" Nov 24 15:45:13 crc kubenswrapper[4790]: E1124 15:45:13.315098 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:45:24 crc kubenswrapper[4790]: I1124 15:45:24.315810 4790 scope.go:117] "RemoveContainer" containerID="c579c182b84b6d5bab1987b612c6237aee4927c7e2c1e98dc1b59eca4477811d" Nov 24 15:45:24 crc kubenswrapper[4790]: E1124 15:45:24.317158 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:45:32 crc kubenswrapper[4790]: I1124 15:45:32.971858 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tlltg"] Nov 24 15:45:32 crc kubenswrapper[4790]: E1124 15:45:32.973431 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb0d7854-9d30-4e0a-aefa-983d7a5d505f" containerName="collect-profiles" Nov 24 15:45:32 crc kubenswrapper[4790]: I1124 15:45:32.973455 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb0d7854-9d30-4e0a-aefa-983d7a5d505f" containerName="collect-profiles" Nov 24 15:45:32 crc kubenswrapper[4790]: I1124 15:45:32.973828 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb0d7854-9d30-4e0a-aefa-983d7a5d505f" containerName="collect-profiles" Nov 24 15:45:32 crc kubenswrapper[4790]: I1124 15:45:32.976594 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tlltg" Nov 24 15:45:32 crc kubenswrapper[4790]: I1124 15:45:32.982716 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tlltg"] Nov 24 15:45:33 crc kubenswrapper[4790]: I1124 15:45:33.041554 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ace68c95-d379-4cc2-a686-ecc89b0b0d5a-utilities\") pod \"certified-operators-tlltg\" (UID: \"ace68c95-d379-4cc2-a686-ecc89b0b0d5a\") " pod="openshift-marketplace/certified-operators-tlltg" Nov 24 15:45:33 crc kubenswrapper[4790]: I1124 15:45:33.041730 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ace68c95-d379-4cc2-a686-ecc89b0b0d5a-catalog-content\") pod \"certified-operators-tlltg\" (UID: \"ace68c95-d379-4cc2-a686-ecc89b0b0d5a\") " pod="openshift-marketplace/certified-operators-tlltg" Nov 24 15:45:33 crc kubenswrapper[4790]: I1124 15:45:33.041797 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km78x\" (UniqueName: \"kubernetes.io/projected/ace68c95-d379-4cc2-a686-ecc89b0b0d5a-kube-api-access-km78x\") pod \"certified-operators-tlltg\" (UID: \"ace68c95-d379-4cc2-a686-ecc89b0b0d5a\") " pod="openshift-marketplace/certified-operators-tlltg" Nov 24 15:45:33 crc kubenswrapper[4790]: I1124 15:45:33.146782 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ace68c95-d379-4cc2-a686-ecc89b0b0d5a-utilities\") pod \"certified-operators-tlltg\" (UID: \"ace68c95-d379-4cc2-a686-ecc89b0b0d5a\") " pod="openshift-marketplace/certified-operators-tlltg" Nov 24 15:45:33 crc kubenswrapper[4790]: I1124 15:45:33.146918 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ace68c95-d379-4cc2-a686-ecc89b0b0d5a-catalog-content\") pod \"certified-operators-tlltg\" (UID: \"ace68c95-d379-4cc2-a686-ecc89b0b0d5a\") " pod="openshift-marketplace/certified-operators-tlltg" Nov 24 15:45:33 crc kubenswrapper[4790]: I1124 15:45:33.146955 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km78x\" (UniqueName: \"kubernetes.io/projected/ace68c95-d379-4cc2-a686-ecc89b0b0d5a-kube-api-access-km78x\") pod \"certified-operators-tlltg\" (UID: \"ace68c95-d379-4cc2-a686-ecc89b0b0d5a\") " pod="openshift-marketplace/certified-operators-tlltg" Nov 24 15:45:33 crc kubenswrapper[4790]: I1124 15:45:33.147441 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ace68c95-d379-4cc2-a686-ecc89b0b0d5a-utilities\") pod \"certified-operators-tlltg\" (UID: \"ace68c95-d379-4cc2-a686-ecc89b0b0d5a\") " pod="openshift-marketplace/certified-operators-tlltg" Nov 24 15:45:33 crc kubenswrapper[4790]: I1124 15:45:33.147676 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ace68c95-d379-4cc2-a686-ecc89b0b0d5a-catalog-content\") pod \"certified-operators-tlltg\" (UID: \"ace68c95-d379-4cc2-a686-ecc89b0b0d5a\") " pod="openshift-marketplace/certified-operators-tlltg" Nov 24 15:45:33 crc kubenswrapper[4790]: I1124 15:45:33.169540 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km78x\" (UniqueName: \"kubernetes.io/projected/ace68c95-d379-4cc2-a686-ecc89b0b0d5a-kube-api-access-km78x\") pod \"certified-operators-tlltg\" (UID: \"ace68c95-d379-4cc2-a686-ecc89b0b0d5a\") " pod="openshift-marketplace/certified-operators-tlltg" Nov 24 15:45:33 crc kubenswrapper[4790]: I1124 15:45:33.336638 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tlltg" Nov 24 15:45:33 crc kubenswrapper[4790]: I1124 15:45:33.855185 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tlltg"] Nov 24 15:45:33 crc kubenswrapper[4790]: I1124 15:45:33.980372 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tlltg" event={"ID":"ace68c95-d379-4cc2-a686-ecc89b0b0d5a","Type":"ContainerStarted","Data":"bb33aa346e1871f3d701e41f5aea35bfb03963196c841ed6984ab6afd6bcbf50"} Nov 24 15:45:34 crc kubenswrapper[4790]: I1124 15:45:34.998753 4790 generic.go:334] "Generic (PLEG): container finished" podID="ace68c95-d379-4cc2-a686-ecc89b0b0d5a" containerID="69e9cbc37733f562d48b704545ad98a2b7bb6ed757ff409ada0a2b6938eb9ea8" exitCode=0 Nov 24 15:45:34 crc kubenswrapper[4790]: I1124 15:45:34.999042 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tlltg" event={"ID":"ace68c95-d379-4cc2-a686-ecc89b0b0d5a","Type":"ContainerDied","Data":"69e9cbc37733f562d48b704545ad98a2b7bb6ed757ff409ada0a2b6938eb9ea8"} Nov 24 15:45:35 crc kubenswrapper[4790]: I1124 15:45:35.003544 4790 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 15:45:36 crc kubenswrapper[4790]: I1124 15:45:36.011668 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tlltg" event={"ID":"ace68c95-d379-4cc2-a686-ecc89b0b0d5a","Type":"ContainerStarted","Data":"1544e9322aba2df1cdce756b3537dd9f4622489ad2fc0a27ea6a712c4bb60c23"} Nov 24 15:45:36 crc kubenswrapper[4790]: I1124 15:45:36.314907 4790 scope.go:117] "RemoveContainer" containerID="c579c182b84b6d5bab1987b612c6237aee4927c7e2c1e98dc1b59eca4477811d" Nov 24 15:45:36 crc kubenswrapper[4790]: E1124 15:45:36.315493 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:45:37 crc kubenswrapper[4790]: I1124 15:45:37.026109 4790 generic.go:334] "Generic (PLEG): container finished" podID="ace68c95-d379-4cc2-a686-ecc89b0b0d5a" containerID="1544e9322aba2df1cdce756b3537dd9f4622489ad2fc0a27ea6a712c4bb60c23" exitCode=0 Nov 24 15:45:37 crc kubenswrapper[4790]: I1124 15:45:37.026161 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tlltg" event={"ID":"ace68c95-d379-4cc2-a686-ecc89b0b0d5a","Type":"ContainerDied","Data":"1544e9322aba2df1cdce756b3537dd9f4622489ad2fc0a27ea6a712c4bb60c23"} Nov 24 15:45:38 crc kubenswrapper[4790]: I1124 15:45:38.036365 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tlltg" event={"ID":"ace68c95-d379-4cc2-a686-ecc89b0b0d5a","Type":"ContainerStarted","Data":"21d3767bc0d8c3ffa09d06fb5623d00ec895a9aeaacb68b0f5b801ac6f3f543b"} Nov 24 15:45:38 crc kubenswrapper[4790]: I1124 15:45:38.064258 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tlltg" podStartSLOduration=3.659599869 podStartE2EDuration="6.064239662s" podCreationTimestamp="2025-11-24 15:45:32 +0000 UTC" firstStartedPulling="2025-11-24 15:45:35.003132542 +0000 UTC m=+9183.383026234" lastFinishedPulling="2025-11-24 15:45:37.407772365 +0000 UTC m=+9185.787666027" observedRunningTime="2025-11-24 15:45:38.054665169 +0000 UTC m=+9186.434558831" watchObservedRunningTime="2025-11-24 15:45:38.064239662 +0000 UTC m=+9186.444133334" Nov 24 15:45:43 crc kubenswrapper[4790]: I1124 15:45:43.337414 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tlltg" Nov 24 15:45:43 crc kubenswrapper[4790]: I1124 15:45:43.338154 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tlltg" Nov 24 15:45:43 crc kubenswrapper[4790]: I1124 15:45:43.391604 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tlltg" Nov 24 15:45:44 crc kubenswrapper[4790]: I1124 15:45:44.198259 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tlltg" Nov 24 15:45:44 crc kubenswrapper[4790]: I1124 15:45:44.258013 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tlltg"] Nov 24 15:45:46 crc kubenswrapper[4790]: I1124 15:45:46.172218 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tlltg" podUID="ace68c95-d379-4cc2-a686-ecc89b0b0d5a" containerName="registry-server" containerID="cri-o://21d3767bc0d8c3ffa09d06fb5623d00ec895a9aeaacb68b0f5b801ac6f3f543b" gracePeriod=2 Nov 24 15:45:46 crc kubenswrapper[4790]: I1124 15:45:46.826149 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tlltg" Nov 24 15:45:46 crc kubenswrapper[4790]: I1124 15:45:46.944930 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-km78x\" (UniqueName: \"kubernetes.io/projected/ace68c95-d379-4cc2-a686-ecc89b0b0d5a-kube-api-access-km78x\") pod \"ace68c95-d379-4cc2-a686-ecc89b0b0d5a\" (UID: \"ace68c95-d379-4cc2-a686-ecc89b0b0d5a\") " Nov 24 15:45:46 crc kubenswrapper[4790]: I1124 15:45:46.945166 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ace68c95-d379-4cc2-a686-ecc89b0b0d5a-utilities\") pod \"ace68c95-d379-4cc2-a686-ecc89b0b0d5a\" (UID: \"ace68c95-d379-4cc2-a686-ecc89b0b0d5a\") " Nov 24 15:45:46 crc kubenswrapper[4790]: I1124 15:45:46.945284 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ace68c95-d379-4cc2-a686-ecc89b0b0d5a-catalog-content\") pod \"ace68c95-d379-4cc2-a686-ecc89b0b0d5a\" (UID: \"ace68c95-d379-4cc2-a686-ecc89b0b0d5a\") " Nov 24 15:45:46 crc kubenswrapper[4790]: I1124 15:45:46.945874 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ace68c95-d379-4cc2-a686-ecc89b0b0d5a-utilities" (OuterVolumeSpecName: "utilities") pod "ace68c95-d379-4cc2-a686-ecc89b0b0d5a" (UID: "ace68c95-d379-4cc2-a686-ecc89b0b0d5a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:45:46 crc kubenswrapper[4790]: I1124 15:45:46.951045 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ace68c95-d379-4cc2-a686-ecc89b0b0d5a-kube-api-access-km78x" (OuterVolumeSpecName: "kube-api-access-km78x") pod "ace68c95-d379-4cc2-a686-ecc89b0b0d5a" (UID: "ace68c95-d379-4cc2-a686-ecc89b0b0d5a"). InnerVolumeSpecName "kube-api-access-km78x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:45:46 crc kubenswrapper[4790]: I1124 15:45:46.987191 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ace68c95-d379-4cc2-a686-ecc89b0b0d5a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ace68c95-d379-4cc2-a686-ecc89b0b0d5a" (UID: "ace68c95-d379-4cc2-a686-ecc89b0b0d5a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:45:47 crc kubenswrapper[4790]: I1124 15:45:47.047735 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ace68c95-d379-4cc2-a686-ecc89b0b0d5a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 15:45:47 crc kubenswrapper[4790]: I1124 15:45:47.047772 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-km78x\" (UniqueName: \"kubernetes.io/projected/ace68c95-d379-4cc2-a686-ecc89b0b0d5a-kube-api-access-km78x\") on node \"crc\" DevicePath \"\"" Nov 24 15:45:47 crc kubenswrapper[4790]: I1124 15:45:47.047784 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ace68c95-d379-4cc2-a686-ecc89b0b0d5a-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 15:45:47 crc kubenswrapper[4790]: I1124 15:45:47.192784 4790 generic.go:334] "Generic (PLEG): container finished" podID="ace68c95-d379-4cc2-a686-ecc89b0b0d5a" containerID="21d3767bc0d8c3ffa09d06fb5623d00ec895a9aeaacb68b0f5b801ac6f3f543b" exitCode=0 Nov 24 15:45:47 crc kubenswrapper[4790]: I1124 15:45:47.192912 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tlltg" Nov 24 15:45:47 crc kubenswrapper[4790]: I1124 15:45:47.192860 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tlltg" event={"ID":"ace68c95-d379-4cc2-a686-ecc89b0b0d5a","Type":"ContainerDied","Data":"21d3767bc0d8c3ffa09d06fb5623d00ec895a9aeaacb68b0f5b801ac6f3f543b"} Nov 24 15:45:47 crc kubenswrapper[4790]: I1124 15:45:47.193295 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tlltg" event={"ID":"ace68c95-d379-4cc2-a686-ecc89b0b0d5a","Type":"ContainerDied","Data":"bb33aa346e1871f3d701e41f5aea35bfb03963196c841ed6984ab6afd6bcbf50"} Nov 24 15:45:47 crc kubenswrapper[4790]: I1124 15:45:47.193417 4790 scope.go:117] "RemoveContainer" containerID="21d3767bc0d8c3ffa09d06fb5623d00ec895a9aeaacb68b0f5b801ac6f3f543b" Nov 24 15:45:47 crc kubenswrapper[4790]: I1124 15:45:47.237108 4790 scope.go:117] "RemoveContainer" containerID="1544e9322aba2df1cdce756b3537dd9f4622489ad2fc0a27ea6a712c4bb60c23" Nov 24 15:45:47 crc kubenswrapper[4790]: I1124 15:45:47.267639 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tlltg"] Nov 24 15:45:47 crc kubenswrapper[4790]: I1124 15:45:47.278878 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tlltg"] Nov 24 15:45:47 crc kubenswrapper[4790]: I1124 15:45:47.310389 4790 scope.go:117] "RemoveContainer" containerID="69e9cbc37733f562d48b704545ad98a2b7bb6ed757ff409ada0a2b6938eb9ea8" Nov 24 15:45:47 crc kubenswrapper[4790]: I1124 15:45:47.349935 4790 scope.go:117] "RemoveContainer" containerID="21d3767bc0d8c3ffa09d06fb5623d00ec895a9aeaacb68b0f5b801ac6f3f543b" Nov 24 15:45:47 crc kubenswrapper[4790]: E1124 15:45:47.350350 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21d3767bc0d8c3ffa09d06fb5623d00ec895a9aeaacb68b0f5b801ac6f3f543b\": container with ID starting with 21d3767bc0d8c3ffa09d06fb5623d00ec895a9aeaacb68b0f5b801ac6f3f543b not found: ID does not exist" containerID="21d3767bc0d8c3ffa09d06fb5623d00ec895a9aeaacb68b0f5b801ac6f3f543b" Nov 24 15:45:47 crc kubenswrapper[4790]: I1124 15:45:47.350397 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21d3767bc0d8c3ffa09d06fb5623d00ec895a9aeaacb68b0f5b801ac6f3f543b"} err="failed to get container status \"21d3767bc0d8c3ffa09d06fb5623d00ec895a9aeaacb68b0f5b801ac6f3f543b\": rpc error: code = NotFound desc = could not find container \"21d3767bc0d8c3ffa09d06fb5623d00ec895a9aeaacb68b0f5b801ac6f3f543b\": container with ID starting with 21d3767bc0d8c3ffa09d06fb5623d00ec895a9aeaacb68b0f5b801ac6f3f543b not found: ID does not exist" Nov 24 15:45:47 crc kubenswrapper[4790]: I1124 15:45:47.350440 4790 scope.go:117] "RemoveContainer" containerID="1544e9322aba2df1cdce756b3537dd9f4622489ad2fc0a27ea6a712c4bb60c23" Nov 24 15:45:47 crc kubenswrapper[4790]: E1124 15:45:47.350935 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1544e9322aba2df1cdce756b3537dd9f4622489ad2fc0a27ea6a712c4bb60c23\": container with ID starting with 1544e9322aba2df1cdce756b3537dd9f4622489ad2fc0a27ea6a712c4bb60c23 not found: ID does not exist" containerID="1544e9322aba2df1cdce756b3537dd9f4622489ad2fc0a27ea6a712c4bb60c23" Nov 24 15:45:47 crc kubenswrapper[4790]: I1124 15:45:47.351015 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1544e9322aba2df1cdce756b3537dd9f4622489ad2fc0a27ea6a712c4bb60c23"} err="failed to get container status \"1544e9322aba2df1cdce756b3537dd9f4622489ad2fc0a27ea6a712c4bb60c23\": rpc error: code = NotFound desc = could not find container \"1544e9322aba2df1cdce756b3537dd9f4622489ad2fc0a27ea6a712c4bb60c23\": container with ID starting with 1544e9322aba2df1cdce756b3537dd9f4622489ad2fc0a27ea6a712c4bb60c23 not found: ID does not exist" Nov 24 15:45:47 crc kubenswrapper[4790]: I1124 15:45:47.351048 4790 scope.go:117] "RemoveContainer" containerID="69e9cbc37733f562d48b704545ad98a2b7bb6ed757ff409ada0a2b6938eb9ea8" Nov 24 15:45:47 crc kubenswrapper[4790]: E1124 15:45:47.351624 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69e9cbc37733f562d48b704545ad98a2b7bb6ed757ff409ada0a2b6938eb9ea8\": container with ID starting with 69e9cbc37733f562d48b704545ad98a2b7bb6ed757ff409ada0a2b6938eb9ea8 not found: ID does not exist" containerID="69e9cbc37733f562d48b704545ad98a2b7bb6ed757ff409ada0a2b6938eb9ea8" Nov 24 15:45:47 crc kubenswrapper[4790]: I1124 15:45:47.351663 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69e9cbc37733f562d48b704545ad98a2b7bb6ed757ff409ada0a2b6938eb9ea8"} err="failed to get container status \"69e9cbc37733f562d48b704545ad98a2b7bb6ed757ff409ada0a2b6938eb9ea8\": rpc error: code = NotFound desc = could not find container \"69e9cbc37733f562d48b704545ad98a2b7bb6ed757ff409ada0a2b6938eb9ea8\": container with ID starting with 69e9cbc37733f562d48b704545ad98a2b7bb6ed757ff409ada0a2b6938eb9ea8 not found: ID does not exist" Nov 24 15:45:48 crc kubenswrapper[4790]: I1124 15:45:48.327580 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ace68c95-d379-4cc2-a686-ecc89b0b0d5a" path="/var/lib/kubelet/pods/ace68c95-d379-4cc2-a686-ecc89b0b0d5a/volumes" Nov 24 15:45:49 crc kubenswrapper[4790]: I1124 15:45:49.315352 4790 scope.go:117] "RemoveContainer" containerID="c579c182b84b6d5bab1987b612c6237aee4927c7e2c1e98dc1b59eca4477811d" Nov 24 15:45:50 crc kubenswrapper[4790]: I1124 15:45:50.242677 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"481705db3593f4e5648f79ea2801d3ecc5478bd6c6a34f70e103d4dccb0eeb10"} Nov 24 15:46:04 crc kubenswrapper[4790]: I1124 15:46:04.357164 4790 scope.go:117] "RemoveContainer" containerID="bca98e9f7a1e74274b9a4a717d4ca591f5ad30f24c0084ec3b64204d853803ab" Nov 24 15:46:27 crc kubenswrapper[4790]: I1124 15:46:27.645838 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2j7zh"] Nov 24 15:46:27 crc kubenswrapper[4790]: E1124 15:46:27.646831 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ace68c95-d379-4cc2-a686-ecc89b0b0d5a" containerName="extract-content" Nov 24 15:46:27 crc kubenswrapper[4790]: I1124 15:46:27.646845 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="ace68c95-d379-4cc2-a686-ecc89b0b0d5a" containerName="extract-content" Nov 24 15:46:27 crc kubenswrapper[4790]: E1124 15:46:27.646872 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ace68c95-d379-4cc2-a686-ecc89b0b0d5a" containerName="extract-utilities" Nov 24 15:46:27 crc kubenswrapper[4790]: I1124 15:46:27.646895 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="ace68c95-d379-4cc2-a686-ecc89b0b0d5a" containerName="extract-utilities" Nov 24 15:46:27 crc kubenswrapper[4790]: E1124 15:46:27.647008 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ace68c95-d379-4cc2-a686-ecc89b0b0d5a" containerName="registry-server" Nov 24 15:46:27 crc kubenswrapper[4790]: I1124 15:46:27.647014 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="ace68c95-d379-4cc2-a686-ecc89b0b0d5a" containerName="registry-server" Nov 24 15:46:27 crc kubenswrapper[4790]: I1124 15:46:27.647225 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="ace68c95-d379-4cc2-a686-ecc89b0b0d5a" containerName="registry-server" Nov 24 15:46:27 crc kubenswrapper[4790]: I1124 15:46:27.648741 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2j7zh" Nov 24 15:46:27 crc kubenswrapper[4790]: I1124 15:46:27.660286 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2j7zh"] Nov 24 15:46:27 crc kubenswrapper[4790]: I1124 15:46:27.747708 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3943108-ee68-4369-b4d0-e80b1ab6aa05-utilities\") pod \"redhat-marketplace-2j7zh\" (UID: \"b3943108-ee68-4369-b4d0-e80b1ab6aa05\") " pod="openshift-marketplace/redhat-marketplace-2j7zh" Nov 24 15:46:27 crc kubenswrapper[4790]: I1124 15:46:27.748068 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3943108-ee68-4369-b4d0-e80b1ab6aa05-catalog-content\") pod \"redhat-marketplace-2j7zh\" (UID: \"b3943108-ee68-4369-b4d0-e80b1ab6aa05\") " pod="openshift-marketplace/redhat-marketplace-2j7zh" Nov 24 15:46:27 crc kubenswrapper[4790]: I1124 15:46:27.748231 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktx5z\" (UniqueName: \"kubernetes.io/projected/b3943108-ee68-4369-b4d0-e80b1ab6aa05-kube-api-access-ktx5z\") pod \"redhat-marketplace-2j7zh\" (UID: \"b3943108-ee68-4369-b4d0-e80b1ab6aa05\") " pod="openshift-marketplace/redhat-marketplace-2j7zh" Nov 24 15:46:27 crc kubenswrapper[4790]: I1124 15:46:27.851615 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3943108-ee68-4369-b4d0-e80b1ab6aa05-utilities\") pod \"redhat-marketplace-2j7zh\" (UID: \"b3943108-ee68-4369-b4d0-e80b1ab6aa05\") " pod="openshift-marketplace/redhat-marketplace-2j7zh" Nov 24 15:46:27 crc kubenswrapper[4790]: I1124 15:46:27.851695 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3943108-ee68-4369-b4d0-e80b1ab6aa05-catalog-content\") pod \"redhat-marketplace-2j7zh\" (UID: \"b3943108-ee68-4369-b4d0-e80b1ab6aa05\") " pod="openshift-marketplace/redhat-marketplace-2j7zh" Nov 24 15:46:27 crc kubenswrapper[4790]: I1124 15:46:27.851764 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktx5z\" (UniqueName: \"kubernetes.io/projected/b3943108-ee68-4369-b4d0-e80b1ab6aa05-kube-api-access-ktx5z\") pod \"redhat-marketplace-2j7zh\" (UID: \"b3943108-ee68-4369-b4d0-e80b1ab6aa05\") " pod="openshift-marketplace/redhat-marketplace-2j7zh" Nov 24 15:46:27 crc kubenswrapper[4790]: I1124 15:46:27.852876 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3943108-ee68-4369-b4d0-e80b1ab6aa05-utilities\") pod \"redhat-marketplace-2j7zh\" (UID: \"b3943108-ee68-4369-b4d0-e80b1ab6aa05\") " pod="openshift-marketplace/redhat-marketplace-2j7zh" Nov 24 15:46:27 crc kubenswrapper[4790]: I1124 15:46:27.853118 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3943108-ee68-4369-b4d0-e80b1ab6aa05-catalog-content\") pod \"redhat-marketplace-2j7zh\" (UID: \"b3943108-ee68-4369-b4d0-e80b1ab6aa05\") " pod="openshift-marketplace/redhat-marketplace-2j7zh" Nov 24 15:46:27 crc kubenswrapper[4790]: I1124 15:46:27.881601 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktx5z\" (UniqueName: \"kubernetes.io/projected/b3943108-ee68-4369-b4d0-e80b1ab6aa05-kube-api-access-ktx5z\") pod \"redhat-marketplace-2j7zh\" (UID: \"b3943108-ee68-4369-b4d0-e80b1ab6aa05\") " pod="openshift-marketplace/redhat-marketplace-2j7zh" Nov 24 15:46:27 crc kubenswrapper[4790]: I1124 15:46:27.985045 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2j7zh" Nov 24 15:46:28 crc kubenswrapper[4790]: I1124 15:46:28.507904 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2j7zh"] Nov 24 15:46:28 crc kubenswrapper[4790]: I1124 15:46:28.747978 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2j7zh" event={"ID":"b3943108-ee68-4369-b4d0-e80b1ab6aa05","Type":"ContainerStarted","Data":"7d89d65885137b7c5e94abe3a72bc1262841c9885a0d9ce7aa966a279339ec7a"} Nov 24 15:46:28 crc kubenswrapper[4790]: I1124 15:46:28.748029 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2j7zh" event={"ID":"b3943108-ee68-4369-b4d0-e80b1ab6aa05","Type":"ContainerStarted","Data":"c488f76796814f08a69f88381a6d0fd6901f59774843f982c63a382f99f021cb"} Nov 24 15:46:29 crc kubenswrapper[4790]: I1124 15:46:29.759901 4790 generic.go:334] "Generic (PLEG): container finished" podID="b3943108-ee68-4369-b4d0-e80b1ab6aa05" containerID="7d89d65885137b7c5e94abe3a72bc1262841c9885a0d9ce7aa966a279339ec7a" exitCode=0 Nov 24 15:46:29 crc kubenswrapper[4790]: I1124 15:46:29.759961 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2j7zh" event={"ID":"b3943108-ee68-4369-b4d0-e80b1ab6aa05","Type":"ContainerDied","Data":"7d89d65885137b7c5e94abe3a72bc1262841c9885a0d9ce7aa966a279339ec7a"} Nov 24 15:46:30 crc kubenswrapper[4790]: I1124 15:46:30.780405 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2j7zh" event={"ID":"b3943108-ee68-4369-b4d0-e80b1ab6aa05","Type":"ContainerStarted","Data":"c999fe5f47bcbcc51ce170c442b312252b0b16d93e7b5dddd9a3c3c27da87f86"} Nov 24 15:46:31 crc kubenswrapper[4790]: I1124 15:46:31.797147 4790 generic.go:334] "Generic (PLEG): container finished" podID="b3943108-ee68-4369-b4d0-e80b1ab6aa05" containerID="c999fe5f47bcbcc51ce170c442b312252b0b16d93e7b5dddd9a3c3c27da87f86" exitCode=0 Nov 24 15:46:31 crc kubenswrapper[4790]: I1124 15:46:31.797236 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2j7zh" event={"ID":"b3943108-ee68-4369-b4d0-e80b1ab6aa05","Type":"ContainerDied","Data":"c999fe5f47bcbcc51ce170c442b312252b0b16d93e7b5dddd9a3c3c27da87f86"} Nov 24 15:46:32 crc kubenswrapper[4790]: I1124 15:46:32.815272 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2j7zh" event={"ID":"b3943108-ee68-4369-b4d0-e80b1ab6aa05","Type":"ContainerStarted","Data":"80c92021418eaba356c3372fc535e9c7f478bd4a606185025b42b04b2b38e3bd"} Nov 24 15:46:32 crc kubenswrapper[4790]: I1124 15:46:32.856944 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2j7zh" podStartSLOduration=3.374471237 podStartE2EDuration="5.856922771s" podCreationTimestamp="2025-11-24 15:46:27 +0000 UTC" firstStartedPulling="2025-11-24 15:46:29.763965212 +0000 UTC m=+9238.143858874" lastFinishedPulling="2025-11-24 15:46:32.246416746 +0000 UTC m=+9240.626310408" observedRunningTime="2025-11-24 15:46:32.840432371 +0000 UTC m=+9241.220326043" watchObservedRunningTime="2025-11-24 15:46:32.856922771 +0000 UTC m=+9241.236816443" Nov 24 15:46:37 crc kubenswrapper[4790]: I1124 15:46:37.986209 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2j7zh" Nov 24 15:46:37 crc kubenswrapper[4790]: I1124 15:46:37.987002 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2j7zh" Nov 24 15:46:38 crc kubenswrapper[4790]: I1124 15:46:38.068821 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2j7zh" Nov 24 15:46:38 crc kubenswrapper[4790]: I1124 15:46:38.961131 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2j7zh" Nov 24 15:46:39 crc kubenswrapper[4790]: I1124 15:46:39.029646 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2j7zh"] Nov 24 15:46:40 crc kubenswrapper[4790]: I1124 15:46:40.908469 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2j7zh" podUID="b3943108-ee68-4369-b4d0-e80b1ab6aa05" containerName="registry-server" containerID="cri-o://80c92021418eaba356c3372fc535e9c7f478bd4a606185025b42b04b2b38e3bd" gracePeriod=2 Nov 24 15:46:41 crc kubenswrapper[4790]: I1124 15:46:41.491317 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2j7zh" Nov 24 15:46:41 crc kubenswrapper[4790]: I1124 15:46:41.600740 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktx5z\" (UniqueName: \"kubernetes.io/projected/b3943108-ee68-4369-b4d0-e80b1ab6aa05-kube-api-access-ktx5z\") pod \"b3943108-ee68-4369-b4d0-e80b1ab6aa05\" (UID: \"b3943108-ee68-4369-b4d0-e80b1ab6aa05\") " Nov 24 15:46:41 crc kubenswrapper[4790]: I1124 15:46:41.601001 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3943108-ee68-4369-b4d0-e80b1ab6aa05-catalog-content\") pod \"b3943108-ee68-4369-b4d0-e80b1ab6aa05\" (UID: \"b3943108-ee68-4369-b4d0-e80b1ab6aa05\") " Nov 24 15:46:41 crc kubenswrapper[4790]: I1124 15:46:41.601092 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3943108-ee68-4369-b4d0-e80b1ab6aa05-utilities\") pod \"b3943108-ee68-4369-b4d0-e80b1ab6aa05\" (UID: \"b3943108-ee68-4369-b4d0-e80b1ab6aa05\") " Nov 24 15:46:41 crc kubenswrapper[4790]: I1124 15:46:41.602875 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3943108-ee68-4369-b4d0-e80b1ab6aa05-utilities" (OuterVolumeSpecName: "utilities") pod "b3943108-ee68-4369-b4d0-e80b1ab6aa05" (UID: "b3943108-ee68-4369-b4d0-e80b1ab6aa05"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:46:41 crc kubenswrapper[4790]: I1124 15:46:41.609504 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3943108-ee68-4369-b4d0-e80b1ab6aa05-kube-api-access-ktx5z" (OuterVolumeSpecName: "kube-api-access-ktx5z") pod "b3943108-ee68-4369-b4d0-e80b1ab6aa05" (UID: "b3943108-ee68-4369-b4d0-e80b1ab6aa05"). InnerVolumeSpecName "kube-api-access-ktx5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:46:41 crc kubenswrapper[4790]: I1124 15:46:41.623959 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3943108-ee68-4369-b4d0-e80b1ab6aa05-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b3943108-ee68-4369-b4d0-e80b1ab6aa05" (UID: "b3943108-ee68-4369-b4d0-e80b1ab6aa05"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:46:41 crc kubenswrapper[4790]: I1124 15:46:41.714708 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktx5z\" (UniqueName: \"kubernetes.io/projected/b3943108-ee68-4369-b4d0-e80b1ab6aa05-kube-api-access-ktx5z\") on node \"crc\" DevicePath \"\"" Nov 24 15:46:41 crc kubenswrapper[4790]: I1124 15:46:41.714785 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3943108-ee68-4369-b4d0-e80b1ab6aa05-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 15:46:41 crc kubenswrapper[4790]: I1124 15:46:41.714806 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3943108-ee68-4369-b4d0-e80b1ab6aa05-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 15:46:41 crc kubenswrapper[4790]: I1124 15:46:41.930330 4790 generic.go:334] "Generic (PLEG): container finished" podID="b3943108-ee68-4369-b4d0-e80b1ab6aa05" containerID="80c92021418eaba356c3372fc535e9c7f478bd4a606185025b42b04b2b38e3bd" exitCode=0 Nov 24 15:46:41 crc kubenswrapper[4790]: I1124 15:46:41.930527 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2j7zh" event={"ID":"b3943108-ee68-4369-b4d0-e80b1ab6aa05","Type":"ContainerDied","Data":"80c92021418eaba356c3372fc535e9c7f478bd4a606185025b42b04b2b38e3bd"} Nov 24 15:46:41 crc kubenswrapper[4790]: I1124 15:46:41.930691 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2j7zh" Nov 24 15:46:41 crc kubenswrapper[4790]: I1124 15:46:41.930853 4790 scope.go:117] "RemoveContainer" containerID="80c92021418eaba356c3372fc535e9c7f478bd4a606185025b42b04b2b38e3bd" Nov 24 15:46:41 crc kubenswrapper[4790]: I1124 15:46:41.930829 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2j7zh" event={"ID":"b3943108-ee68-4369-b4d0-e80b1ab6aa05","Type":"ContainerDied","Data":"c488f76796814f08a69f88381a6d0fd6901f59774843f982c63a382f99f021cb"} Nov 24 15:46:41 crc kubenswrapper[4790]: I1124 15:46:41.968697 4790 scope.go:117] "RemoveContainer" containerID="c999fe5f47bcbcc51ce170c442b312252b0b16d93e7b5dddd9a3c3c27da87f86" Nov 24 15:46:41 crc kubenswrapper[4790]: I1124 15:46:41.982241 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2j7zh"] Nov 24 15:46:41 crc kubenswrapper[4790]: I1124 15:46:41.992438 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2j7zh"] Nov 24 15:46:42 crc kubenswrapper[4790]: I1124 15:46:42.008369 4790 scope.go:117] "RemoveContainer" containerID="7d89d65885137b7c5e94abe3a72bc1262841c9885a0d9ce7aa966a279339ec7a" Nov 24 15:46:42 crc kubenswrapper[4790]: I1124 15:46:42.052528 4790 scope.go:117] "RemoveContainer" containerID="80c92021418eaba356c3372fc535e9c7f478bd4a606185025b42b04b2b38e3bd" Nov 24 15:46:42 crc kubenswrapper[4790]: E1124 15:46:42.053131 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80c92021418eaba356c3372fc535e9c7f478bd4a606185025b42b04b2b38e3bd\": container with ID starting with 80c92021418eaba356c3372fc535e9c7f478bd4a606185025b42b04b2b38e3bd not found: ID does not exist" containerID="80c92021418eaba356c3372fc535e9c7f478bd4a606185025b42b04b2b38e3bd" Nov 24 15:46:42 crc kubenswrapper[4790]: I1124 15:46:42.053194 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80c92021418eaba356c3372fc535e9c7f478bd4a606185025b42b04b2b38e3bd"} err="failed to get container status \"80c92021418eaba356c3372fc535e9c7f478bd4a606185025b42b04b2b38e3bd\": rpc error: code = NotFound desc = could not find container \"80c92021418eaba356c3372fc535e9c7f478bd4a606185025b42b04b2b38e3bd\": container with ID starting with 80c92021418eaba356c3372fc535e9c7f478bd4a606185025b42b04b2b38e3bd not found: ID does not exist" Nov 24 15:46:42 crc kubenswrapper[4790]: I1124 15:46:42.053229 4790 scope.go:117] "RemoveContainer" containerID="c999fe5f47bcbcc51ce170c442b312252b0b16d93e7b5dddd9a3c3c27da87f86" Nov 24 15:46:42 crc kubenswrapper[4790]: E1124 15:46:42.053788 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c999fe5f47bcbcc51ce170c442b312252b0b16d93e7b5dddd9a3c3c27da87f86\": container with ID starting with c999fe5f47bcbcc51ce170c442b312252b0b16d93e7b5dddd9a3c3c27da87f86 not found: ID does not exist" containerID="c999fe5f47bcbcc51ce170c442b312252b0b16d93e7b5dddd9a3c3c27da87f86" Nov 24 15:46:42 crc kubenswrapper[4790]: I1124 15:46:42.053826 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c999fe5f47bcbcc51ce170c442b312252b0b16d93e7b5dddd9a3c3c27da87f86"} err="failed to get container status \"c999fe5f47bcbcc51ce170c442b312252b0b16d93e7b5dddd9a3c3c27da87f86\": rpc error: code = NotFound desc = could not find container \"c999fe5f47bcbcc51ce170c442b312252b0b16d93e7b5dddd9a3c3c27da87f86\": container with ID starting with c999fe5f47bcbcc51ce170c442b312252b0b16d93e7b5dddd9a3c3c27da87f86 not found: ID does not exist" Nov 24 15:46:42 crc kubenswrapper[4790]: I1124 15:46:42.053853 4790 scope.go:117] "RemoveContainer" containerID="7d89d65885137b7c5e94abe3a72bc1262841c9885a0d9ce7aa966a279339ec7a" Nov 24 15:46:42 crc kubenswrapper[4790]: E1124 15:46:42.054122 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d89d65885137b7c5e94abe3a72bc1262841c9885a0d9ce7aa966a279339ec7a\": container with ID starting with 7d89d65885137b7c5e94abe3a72bc1262841c9885a0d9ce7aa966a279339ec7a not found: ID does not exist" containerID="7d89d65885137b7c5e94abe3a72bc1262841c9885a0d9ce7aa966a279339ec7a" Nov 24 15:46:42 crc kubenswrapper[4790]: I1124 15:46:42.054160 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d89d65885137b7c5e94abe3a72bc1262841c9885a0d9ce7aa966a279339ec7a"} err="failed to get container status \"7d89d65885137b7c5e94abe3a72bc1262841c9885a0d9ce7aa966a279339ec7a\": rpc error: code = NotFound desc = could not find container \"7d89d65885137b7c5e94abe3a72bc1262841c9885a0d9ce7aa966a279339ec7a\": container with ID starting with 7d89d65885137b7c5e94abe3a72bc1262841c9885a0d9ce7aa966a279339ec7a not found: ID does not exist" Nov 24 15:46:42 crc kubenswrapper[4790]: I1124 15:46:42.334536 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3943108-ee68-4369-b4d0-e80b1ab6aa05" path="/var/lib/kubelet/pods/b3943108-ee68-4369-b4d0-e80b1ab6aa05/volumes" Nov 24 15:48:13 crc kubenswrapper[4790]: I1124 15:48:13.939216 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:48:13 crc kubenswrapper[4790]: I1124 15:48:13.940010 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:48:28 crc kubenswrapper[4790]: I1124 15:48:28.499492 4790 trace.go:236] Trace[1712391826]: "Calculate volume metrics of persistence for pod openstack/rabbitmq-server-0" (24-Nov-2025 15:48:27.477) (total time: 1022ms): Nov 24 15:48:28 crc kubenswrapper[4790]: Trace[1712391826]: [1.022399998s] [1.022399998s] END Nov 24 15:48:43 crc kubenswrapper[4790]: I1124 15:48:43.939104 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:48:43 crc kubenswrapper[4790]: I1124 15:48:43.940609 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:48:53 crc kubenswrapper[4790]: I1124 15:48:53.771670 4790 generic.go:334] "Generic (PLEG): container finished" podID="515ed57f-53d0-470b-aafe-7deb07d527c9" containerID="522114cd527dfd31f5450ce7f9586288f25d8827aeeb648281ebc806e5bcdc6f" exitCode=0 Nov 24 15:48:53 crc kubenswrapper[4790]: I1124 15:48:53.771776 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-n9962" event={"ID":"515ed57f-53d0-470b-aafe-7deb07d527c9","Type":"ContainerDied","Data":"522114cd527dfd31f5450ce7f9586288f25d8827aeeb648281ebc806e5bcdc6f"} Nov 24 15:48:55 crc kubenswrapper[4790]: I1124 15:48:55.355175 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-n9962" Nov 24 15:48:55 crc kubenswrapper[4790]: I1124 15:48:55.495337 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbl5q\" (UniqueName: \"kubernetes.io/projected/515ed57f-53d0-470b-aafe-7deb07d527c9-kube-api-access-rbl5q\") pod \"515ed57f-53d0-470b-aafe-7deb07d527c9\" (UID: \"515ed57f-53d0-470b-aafe-7deb07d527c9\") " Nov 24 15:48:55 crc kubenswrapper[4790]: I1124 15:48:55.495483 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/515ed57f-53d0-470b-aafe-7deb07d527c9-inventory\") pod \"515ed57f-53d0-470b-aafe-7deb07d527c9\" (UID: \"515ed57f-53d0-470b-aafe-7deb07d527c9\") " Nov 24 15:48:55 crc kubenswrapper[4790]: I1124 15:48:55.495690 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/515ed57f-53d0-470b-aafe-7deb07d527c9-ceph\") pod \"515ed57f-53d0-470b-aafe-7deb07d527c9\" (UID: \"515ed57f-53d0-470b-aafe-7deb07d527c9\") " Nov 24 15:48:55 crc kubenswrapper[4790]: I1124 15:48:55.495802 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/515ed57f-53d0-470b-aafe-7deb07d527c9-ssh-key\") pod \"515ed57f-53d0-470b-aafe-7deb07d527c9\" (UID: \"515ed57f-53d0-470b-aafe-7deb07d527c9\") " Nov 24 15:48:55 crc kubenswrapper[4790]: I1124 15:48:55.495892 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/515ed57f-53d0-470b-aafe-7deb07d527c9-neutron-dhcp-combined-ca-bundle\") pod \"515ed57f-53d0-470b-aafe-7deb07d527c9\" (UID: \"515ed57f-53d0-470b-aafe-7deb07d527c9\") " Nov 24 15:48:55 crc kubenswrapper[4790]: I1124 15:48:55.496138 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/515ed57f-53d0-470b-aafe-7deb07d527c9-neutron-dhcp-agent-neutron-config-0\") pod \"515ed57f-53d0-470b-aafe-7deb07d527c9\" (UID: \"515ed57f-53d0-470b-aafe-7deb07d527c9\") " Nov 24 15:48:55 crc kubenswrapper[4790]: I1124 15:48:55.502231 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/515ed57f-53d0-470b-aafe-7deb07d527c9-kube-api-access-rbl5q" (OuterVolumeSpecName: "kube-api-access-rbl5q") pod "515ed57f-53d0-470b-aafe-7deb07d527c9" (UID: "515ed57f-53d0-470b-aafe-7deb07d527c9"). InnerVolumeSpecName "kube-api-access-rbl5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:48:55 crc kubenswrapper[4790]: I1124 15:48:55.502354 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/515ed57f-53d0-470b-aafe-7deb07d527c9-ceph" (OuterVolumeSpecName: "ceph") pod "515ed57f-53d0-470b-aafe-7deb07d527c9" (UID: "515ed57f-53d0-470b-aafe-7deb07d527c9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:48:55 crc kubenswrapper[4790]: I1124 15:48:55.504047 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/515ed57f-53d0-470b-aafe-7deb07d527c9-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "515ed57f-53d0-470b-aafe-7deb07d527c9" (UID: "515ed57f-53d0-470b-aafe-7deb07d527c9"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:48:55 crc kubenswrapper[4790]: I1124 15:48:55.529491 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/515ed57f-53d0-470b-aafe-7deb07d527c9-inventory" (OuterVolumeSpecName: "inventory") pod "515ed57f-53d0-470b-aafe-7deb07d527c9" (UID: "515ed57f-53d0-470b-aafe-7deb07d527c9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:48:55 crc kubenswrapper[4790]: I1124 15:48:55.558168 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/515ed57f-53d0-470b-aafe-7deb07d527c9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "515ed57f-53d0-470b-aafe-7deb07d527c9" (UID: "515ed57f-53d0-470b-aafe-7deb07d527c9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:48:55 crc kubenswrapper[4790]: I1124 15:48:55.560180 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/515ed57f-53d0-470b-aafe-7deb07d527c9-neutron-dhcp-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-dhcp-agent-neutron-config-0") pod "515ed57f-53d0-470b-aafe-7deb07d527c9" (UID: "515ed57f-53d0-470b-aafe-7deb07d527c9"). InnerVolumeSpecName "neutron-dhcp-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:48:55 crc kubenswrapper[4790]: I1124 15:48:55.600045 4790 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/515ed57f-53d0-470b-aafe-7deb07d527c9-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 15:48:55 crc kubenswrapper[4790]: I1124 15:48:55.600076 4790 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/515ed57f-53d0-470b-aafe-7deb07d527c9-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 15:48:55 crc kubenswrapper[4790]: I1124 15:48:55.600084 4790 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/515ed57f-53d0-470b-aafe-7deb07d527c9-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 15:48:55 crc kubenswrapper[4790]: I1124 15:48:55.600095 4790 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/515ed57f-53d0-470b-aafe-7deb07d527c9-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:48:55 crc kubenswrapper[4790]: I1124 15:48:55.600107 4790 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/515ed57f-53d0-470b-aafe-7deb07d527c9-neutron-dhcp-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 15:48:55 crc kubenswrapper[4790]: I1124 15:48:55.600119 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbl5q\" (UniqueName: \"kubernetes.io/projected/515ed57f-53d0-470b-aafe-7deb07d527c9-kube-api-access-rbl5q\") on node \"crc\" DevicePath \"\"" Nov 24 15:48:55 crc kubenswrapper[4790]: I1124 15:48:55.809916 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-n9962" event={"ID":"515ed57f-53d0-470b-aafe-7deb07d527c9","Type":"ContainerDied","Data":"8ebe7ec7e7e358eac1d12a770ccbf043a78c17d9149b2ac9db7d189c45780a6b"} Nov 24 15:48:55 crc kubenswrapper[4790]: I1124 15:48:55.809989 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ebe7ec7e7e358eac1d12a770ccbf043a78c17d9149b2ac9db7d189c45780a6b" Nov 24 15:48:55 crc kubenswrapper[4790]: I1124 15:48:55.810130 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-n9962" Nov 24 15:49:13 crc kubenswrapper[4790]: I1124 15:49:13.938667 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:49:13 crc kubenswrapper[4790]: I1124 15:49:13.939588 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:49:13 crc kubenswrapper[4790]: I1124 15:49:13.939680 4790 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 15:49:13 crc kubenswrapper[4790]: I1124 15:49:13.941016 4790 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"481705db3593f4e5648f79ea2801d3ecc5478bd6c6a34f70e103d4dccb0eeb10"} pod="openshift-machine-config-operator/machine-config-daemon-xz49t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 15:49:13 crc kubenswrapper[4790]: I1124 15:49:13.941129 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" containerID="cri-o://481705db3593f4e5648f79ea2801d3ecc5478bd6c6a34f70e103d4dccb0eeb10" gracePeriod=600 Nov 24 15:49:14 crc kubenswrapper[4790]: I1124 15:49:14.110118 4790 generic.go:334] "Generic (PLEG): container finished" podID="0d73b133-48f0-455f-8f6a-742e633f631a" containerID="481705db3593f4e5648f79ea2801d3ecc5478bd6c6a34f70e103d4dccb0eeb10" exitCode=0 Nov 24 15:49:14 crc kubenswrapper[4790]: I1124 15:49:14.110177 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerDied","Data":"481705db3593f4e5648f79ea2801d3ecc5478bd6c6a34f70e103d4dccb0eeb10"} Nov 24 15:49:14 crc kubenswrapper[4790]: I1124 15:49:14.110218 4790 scope.go:117] "RemoveContainer" containerID="c579c182b84b6d5bab1987b612c6237aee4927c7e2c1e98dc1b59eca4477811d" Nov 24 15:49:15 crc kubenswrapper[4790]: I1124 15:49:15.132813 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"f1e094cfe52dce391f24557f040b619b252fb0fb27545cacdfae7450eb6c06b2"} Nov 24 15:49:16 crc kubenswrapper[4790]: I1124 15:49:16.437435 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gh2nm"] Nov 24 15:49:16 crc kubenswrapper[4790]: E1124 15:49:16.438546 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3943108-ee68-4369-b4d0-e80b1ab6aa05" containerName="extract-utilities" Nov 24 15:49:16 crc kubenswrapper[4790]: I1124 15:49:16.438570 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3943108-ee68-4369-b4d0-e80b1ab6aa05" containerName="extract-utilities" Nov 24 15:49:16 crc kubenswrapper[4790]: E1124 15:49:16.438601 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3943108-ee68-4369-b4d0-e80b1ab6aa05" containerName="extract-content" Nov 24 15:49:16 crc kubenswrapper[4790]: I1124 15:49:16.438615 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3943108-ee68-4369-b4d0-e80b1ab6aa05" containerName="extract-content" Nov 24 15:49:16 crc kubenswrapper[4790]: E1124 15:49:16.438693 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3943108-ee68-4369-b4d0-e80b1ab6aa05" containerName="registry-server" Nov 24 15:49:16 crc kubenswrapper[4790]: I1124 15:49:16.438705 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3943108-ee68-4369-b4d0-e80b1ab6aa05" containerName="registry-server" Nov 24 15:49:16 crc kubenswrapper[4790]: E1124 15:49:16.438726 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="515ed57f-53d0-470b-aafe-7deb07d527c9" containerName="neutron-dhcp-openstack-openstack-cell1" Nov 24 15:49:16 crc kubenswrapper[4790]: I1124 15:49:16.438739 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="515ed57f-53d0-470b-aafe-7deb07d527c9" containerName="neutron-dhcp-openstack-openstack-cell1" Nov 24 15:49:16 crc kubenswrapper[4790]: I1124 15:49:16.439157 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="515ed57f-53d0-470b-aafe-7deb07d527c9" containerName="neutron-dhcp-openstack-openstack-cell1" Nov 24 15:49:16 crc kubenswrapper[4790]: I1124 15:49:16.439190 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3943108-ee68-4369-b4d0-e80b1ab6aa05" containerName="registry-server" Nov 24 15:49:16 crc kubenswrapper[4790]: I1124 15:49:16.442263 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gh2nm" Nov 24 15:49:16 crc kubenswrapper[4790]: I1124 15:49:16.482376 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gh2nm"] Nov 24 15:49:16 crc kubenswrapper[4790]: I1124 15:49:16.616635 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22c66da7-9f10-44d6-8e01-aab24fe93c78-utilities\") pod \"redhat-operators-gh2nm\" (UID: \"22c66da7-9f10-44d6-8e01-aab24fe93c78\") " pod="openshift-marketplace/redhat-operators-gh2nm" Nov 24 15:49:16 crc kubenswrapper[4790]: I1124 15:49:16.617226 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tll7t\" (UniqueName: \"kubernetes.io/projected/22c66da7-9f10-44d6-8e01-aab24fe93c78-kube-api-access-tll7t\") pod \"redhat-operators-gh2nm\" (UID: \"22c66da7-9f10-44d6-8e01-aab24fe93c78\") " pod="openshift-marketplace/redhat-operators-gh2nm" Nov 24 15:49:16 crc kubenswrapper[4790]: I1124 15:49:16.617557 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22c66da7-9f10-44d6-8e01-aab24fe93c78-catalog-content\") pod \"redhat-operators-gh2nm\" (UID: \"22c66da7-9f10-44d6-8e01-aab24fe93c78\") " pod="openshift-marketplace/redhat-operators-gh2nm" Nov 24 15:49:16 crc kubenswrapper[4790]: I1124 15:49:16.719802 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22c66da7-9f10-44d6-8e01-aab24fe93c78-catalog-content\") pod \"redhat-operators-gh2nm\" (UID: \"22c66da7-9f10-44d6-8e01-aab24fe93c78\") " pod="openshift-marketplace/redhat-operators-gh2nm" Nov 24 15:49:16 crc kubenswrapper[4790]: I1124 15:49:16.719938 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22c66da7-9f10-44d6-8e01-aab24fe93c78-utilities\") pod \"redhat-operators-gh2nm\" (UID: \"22c66da7-9f10-44d6-8e01-aab24fe93c78\") " pod="openshift-marketplace/redhat-operators-gh2nm" Nov 24 15:49:16 crc kubenswrapper[4790]: I1124 15:49:16.720036 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tll7t\" (UniqueName: \"kubernetes.io/projected/22c66da7-9f10-44d6-8e01-aab24fe93c78-kube-api-access-tll7t\") pod \"redhat-operators-gh2nm\" (UID: \"22c66da7-9f10-44d6-8e01-aab24fe93c78\") " pod="openshift-marketplace/redhat-operators-gh2nm" Nov 24 15:49:16 crc kubenswrapper[4790]: I1124 15:49:16.720777 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22c66da7-9f10-44d6-8e01-aab24fe93c78-catalog-content\") pod \"redhat-operators-gh2nm\" (UID: \"22c66da7-9f10-44d6-8e01-aab24fe93c78\") " pod="openshift-marketplace/redhat-operators-gh2nm" Nov 24 15:49:16 crc kubenswrapper[4790]: I1124 15:49:16.721202 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22c66da7-9f10-44d6-8e01-aab24fe93c78-utilities\") pod \"redhat-operators-gh2nm\" (UID: \"22c66da7-9f10-44d6-8e01-aab24fe93c78\") " pod="openshift-marketplace/redhat-operators-gh2nm" Nov 24 15:49:16 crc kubenswrapper[4790]: I1124 15:49:16.743349 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tll7t\" (UniqueName: \"kubernetes.io/projected/22c66da7-9f10-44d6-8e01-aab24fe93c78-kube-api-access-tll7t\") pod \"redhat-operators-gh2nm\" (UID: \"22c66da7-9f10-44d6-8e01-aab24fe93c78\") " pod="openshift-marketplace/redhat-operators-gh2nm" Nov 24 15:49:16 crc kubenswrapper[4790]: I1124 15:49:16.770142 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gh2nm" Nov 24 15:49:17 crc kubenswrapper[4790]: I1124 15:49:17.322758 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gh2nm"] Nov 24 15:49:17 crc kubenswrapper[4790]: W1124 15:49:17.499826 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod22c66da7_9f10_44d6_8e01_aab24fe93c78.slice/crio-5aa209a009dde4964f488d3dc2f660691e4f6949d367688b4e4e509a50fc8114 WatchSource:0}: Error finding container 5aa209a009dde4964f488d3dc2f660691e4f6949d367688b4e4e509a50fc8114: Status 404 returned error can't find the container with id 5aa209a009dde4964f488d3dc2f660691e4f6949d367688b4e4e509a50fc8114 Nov 24 15:49:18 crc kubenswrapper[4790]: I1124 15:49:18.175989 4790 generic.go:334] "Generic (PLEG): container finished" podID="22c66da7-9f10-44d6-8e01-aab24fe93c78" containerID="cb84d2e7ada764bf3206ff30b1d143ba21ccb3595b8cc4622eb3e3ab624e6cb8" exitCode=0 Nov 24 15:49:18 crc kubenswrapper[4790]: I1124 15:49:18.176386 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gh2nm" event={"ID":"22c66da7-9f10-44d6-8e01-aab24fe93c78","Type":"ContainerDied","Data":"cb84d2e7ada764bf3206ff30b1d143ba21ccb3595b8cc4622eb3e3ab624e6cb8"} Nov 24 15:49:18 crc kubenswrapper[4790]: I1124 15:49:18.176457 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gh2nm" event={"ID":"22c66da7-9f10-44d6-8e01-aab24fe93c78","Type":"ContainerStarted","Data":"5aa209a009dde4964f488d3dc2f660691e4f6949d367688b4e4e509a50fc8114"} Nov 24 15:49:20 crc kubenswrapper[4790]: I1124 15:49:20.200969 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gh2nm" event={"ID":"22c66da7-9f10-44d6-8e01-aab24fe93c78","Type":"ContainerStarted","Data":"07a9a4602ba774a72a658d693f4cdb2dd9d2086cc2baabbb8bfefd18e5c96bde"} Nov 24 15:49:21 crc kubenswrapper[4790]: I1124 15:49:21.214547 4790 generic.go:334] "Generic (PLEG): container finished" podID="22c66da7-9f10-44d6-8e01-aab24fe93c78" containerID="07a9a4602ba774a72a658d693f4cdb2dd9d2086cc2baabbb8bfefd18e5c96bde" exitCode=0 Nov 24 15:49:21 crc kubenswrapper[4790]: I1124 15:49:21.214601 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gh2nm" event={"ID":"22c66da7-9f10-44d6-8e01-aab24fe93c78","Type":"ContainerDied","Data":"07a9a4602ba774a72a658d693f4cdb2dd9d2086cc2baabbb8bfefd18e5c96bde"} Nov 24 15:49:22 crc kubenswrapper[4790]: I1124 15:49:22.226558 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gh2nm" event={"ID":"22c66da7-9f10-44d6-8e01-aab24fe93c78","Type":"ContainerStarted","Data":"61b0991ba597d95a922a6d5966b343e4d15528cd4232c199890d392834145037"} Nov 24 15:49:23 crc kubenswrapper[4790]: I1124 15:49:23.257155 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gh2nm" podStartSLOduration=3.7900825769999997 podStartE2EDuration="7.257136897s" podCreationTimestamp="2025-11-24 15:49:16 +0000 UTC" firstStartedPulling="2025-11-24 15:49:18.177815301 +0000 UTC m=+9406.557708963" lastFinishedPulling="2025-11-24 15:49:21.644869621 +0000 UTC m=+9410.024763283" observedRunningTime="2025-11-24 15:49:23.256286632 +0000 UTC m=+9411.636180304" watchObservedRunningTime="2025-11-24 15:49:23.257136897 +0000 UTC m=+9411.637030559" Nov 24 15:49:25 crc kubenswrapper[4790]: I1124 15:49:25.467766 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 15:49:25 crc kubenswrapper[4790]: I1124 15:49:25.468406 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="cda1d09b-2432-4a64-8097-69246eb196a4" containerName="nova-cell0-conductor-conductor" containerID="cri-o://abec8bd8546de72cc3c06177bd23ea100d8a5e09334932e57afe57cd227f16b7" gracePeriod=30 Nov 24 15:49:25 crc kubenswrapper[4790]: I1124 15:49:25.543710 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 15:49:25 crc kubenswrapper[4790]: I1124 15:49:25.544239 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="5cea27eb-31d4-4e3b-a141-d2593bc21427" containerName="nova-cell1-conductor-conductor" containerID="cri-o://47c107e131a16273b3940ec27fea34ff25637218453a77b80c996ad534bf1dca" gracePeriod=30 Nov 24 15:49:26 crc kubenswrapper[4790]: E1124 15:49:26.403586 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="47c107e131a16273b3940ec27fea34ff25637218453a77b80c996ad534bf1dca" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 24 15:49:26 crc kubenswrapper[4790]: E1124 15:49:26.405358 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="47c107e131a16273b3940ec27fea34ff25637218453a77b80c996ad534bf1dca" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 24 15:49:26 crc kubenswrapper[4790]: E1124 15:49:26.406427 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="47c107e131a16273b3940ec27fea34ff25637218453a77b80c996ad534bf1dca" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 24 15:49:26 crc kubenswrapper[4790]: E1124 15:49:26.406582 4790 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="5cea27eb-31d4-4e3b-a141-d2593bc21427" containerName="nova-cell1-conductor-conductor" Nov 24 15:49:26 crc kubenswrapper[4790]: I1124 15:49:26.770599 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gh2nm" Nov 24 15:49:26 crc kubenswrapper[4790]: I1124 15:49:26.770650 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gh2nm" Nov 24 15:49:26 crc kubenswrapper[4790]: I1124 15:49:26.958986 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 15:49:26 crc kubenswrapper[4790]: I1124 15:49:26.959266 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="039efc51-838d-4aac-b0e8-7ba1c9b376ce" containerName="nova-api-log" containerID="cri-o://4a29ff14976dcb69b6759470411ff7beb1d298155f005c3981206b31876014bd" gracePeriod=30 Nov 24 15:49:26 crc kubenswrapper[4790]: I1124 15:49:26.959332 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="039efc51-838d-4aac-b0e8-7ba1c9b376ce" containerName="nova-api-api" containerID="cri-o://58706d8229eea72d6668f75cb9016630db40c6067ad81aa31661875ab841589a" gracePeriod=30 Nov 24 15:49:26 crc kubenswrapper[4790]: I1124 15:49:26.969611 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 15:49:26 crc kubenswrapper[4790]: I1124 15:49:26.969821 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="2a3a068f-92a8-43cf-83dc-8775c1d9ee84" containerName="nova-scheduler-scheduler" containerID="cri-o://56b9940a224d2d309a332cfcbce228cc2bf78759b8ba90c6fba493c1a3781739" gracePeriod=30 Nov 24 15:49:26 crc kubenswrapper[4790]: I1124 15:49:26.990687 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 15:49:26 crc kubenswrapper[4790]: I1124 15:49:26.990947 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="924b2c00-a948-46d2-b103-25f99bc9477e" containerName="nova-metadata-log" containerID="cri-o://5864a737edda91cad9e80b03ac060cd6a2fe7b521eb61e4e209e5c0ea954c4c1" gracePeriod=30 Nov 24 15:49:26 crc kubenswrapper[4790]: I1124 15:49:26.991365 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="924b2c00-a948-46d2-b103-25f99bc9477e" containerName="nova-metadata-metadata" containerID="cri-o://4b805b27d5fe3f30506bbea66532569a1055c70574cd5abc2c7a8a50b9e863b1" gracePeriod=30 Nov 24 15:49:27 crc kubenswrapper[4790]: I1124 15:49:27.913945 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.095563 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cea27eb-31d4-4e3b-a141-d2593bc21427-combined-ca-bundle\") pod \"5cea27eb-31d4-4e3b-a141-d2593bc21427\" (UID: \"5cea27eb-31d4-4e3b-a141-d2593bc21427\") " Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.095701 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cea27eb-31d4-4e3b-a141-d2593bc21427-config-data\") pod \"5cea27eb-31d4-4e3b-a141-d2593bc21427\" (UID: \"5cea27eb-31d4-4e3b-a141-d2593bc21427\") " Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.096369 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpt9m\" (UniqueName: \"kubernetes.io/projected/5cea27eb-31d4-4e3b-a141-d2593bc21427-kube-api-access-zpt9m\") pod \"5cea27eb-31d4-4e3b-a141-d2593bc21427\" (UID: \"5cea27eb-31d4-4e3b-a141-d2593bc21427\") " Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.102066 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cea27eb-31d4-4e3b-a141-d2593bc21427-kube-api-access-zpt9m" (OuterVolumeSpecName: "kube-api-access-zpt9m") pod "5cea27eb-31d4-4e3b-a141-d2593bc21427" (UID: "5cea27eb-31d4-4e3b-a141-d2593bc21427"). InnerVolumeSpecName "kube-api-access-zpt9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.128029 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cea27eb-31d4-4e3b-a141-d2593bc21427-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5cea27eb-31d4-4e3b-a141-d2593bc21427" (UID: "5cea27eb-31d4-4e3b-a141-d2593bc21427"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.128314 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cea27eb-31d4-4e3b-a141-d2593bc21427-config-data" (OuterVolumeSpecName: "config-data") pod "5cea27eb-31d4-4e3b-a141-d2593bc21427" (UID: "5cea27eb-31d4-4e3b-a141-d2593bc21427"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.198521 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cea27eb-31d4-4e3b-a141-d2593bc21427-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.198558 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cea27eb-31d4-4e3b-a141-d2593bc21427-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.198567 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpt9m\" (UniqueName: \"kubernetes.io/projected/5cea27eb-31d4-4e3b-a141-d2593bc21427-kube-api-access-zpt9m\") on node \"crc\" DevicePath \"\"" Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.283082 4790 generic.go:334] "Generic (PLEG): container finished" podID="039efc51-838d-4aac-b0e8-7ba1c9b376ce" containerID="4a29ff14976dcb69b6759470411ff7beb1d298155f005c3981206b31876014bd" exitCode=143 Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.283157 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"039efc51-838d-4aac-b0e8-7ba1c9b376ce","Type":"ContainerDied","Data":"4a29ff14976dcb69b6759470411ff7beb1d298155f005c3981206b31876014bd"} Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.285482 4790 generic.go:334] "Generic (PLEG): container finished" podID="924b2c00-a948-46d2-b103-25f99bc9477e" containerID="5864a737edda91cad9e80b03ac060cd6a2fe7b521eb61e4e209e5c0ea954c4c1" exitCode=143 Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.285542 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"924b2c00-a948-46d2-b103-25f99bc9477e","Type":"ContainerDied","Data":"5864a737edda91cad9e80b03ac060cd6a2fe7b521eb61e4e209e5c0ea954c4c1"} Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.287295 4790 generic.go:334] "Generic (PLEG): container finished" podID="5cea27eb-31d4-4e3b-a141-d2593bc21427" containerID="47c107e131a16273b3940ec27fea34ff25637218453a77b80c996ad534bf1dca" exitCode=0 Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.287329 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"5cea27eb-31d4-4e3b-a141-d2593bc21427","Type":"ContainerDied","Data":"47c107e131a16273b3940ec27fea34ff25637218453a77b80c996ad534bf1dca"} Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.287380 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"5cea27eb-31d4-4e3b-a141-d2593bc21427","Type":"ContainerDied","Data":"021c10b5114e774f39d9ea1db02a0bf29efbc5c35193468f21e6e5e31f7d9e50"} Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.287401 4790 scope.go:117] "RemoveContainer" containerID="47c107e131a16273b3940ec27fea34ff25637218453a77b80c996ad534bf1dca" Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.287633 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.320377 4790 scope.go:117] "RemoveContainer" containerID="47c107e131a16273b3940ec27fea34ff25637218453a77b80c996ad534bf1dca" Nov 24 15:49:28 crc kubenswrapper[4790]: E1124 15:49:28.322110 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47c107e131a16273b3940ec27fea34ff25637218453a77b80c996ad534bf1dca\": container with ID starting with 47c107e131a16273b3940ec27fea34ff25637218453a77b80c996ad534bf1dca not found: ID does not exist" containerID="47c107e131a16273b3940ec27fea34ff25637218453a77b80c996ad534bf1dca" Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.322154 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47c107e131a16273b3940ec27fea34ff25637218453a77b80c996ad534bf1dca"} err="failed to get container status \"47c107e131a16273b3940ec27fea34ff25637218453a77b80c996ad534bf1dca\": rpc error: code = NotFound desc = could not find container \"47c107e131a16273b3940ec27fea34ff25637218453a77b80c996ad534bf1dca\": container with ID starting with 47c107e131a16273b3940ec27fea34ff25637218453a77b80c996ad534bf1dca not found: ID does not exist" Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.368704 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.368744 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.379467 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 15:49:28 crc kubenswrapper[4790]: E1124 15:49:28.380014 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cea27eb-31d4-4e3b-a141-d2593bc21427" containerName="nova-cell1-conductor-conductor" Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.380034 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cea27eb-31d4-4e3b-a141-d2593bc21427" containerName="nova-cell1-conductor-conductor" Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.380254 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cea27eb-31d4-4e3b-a141-d2593bc21427" containerName="nova-cell1-conductor-conductor" Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.380988 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.383540 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.412771 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 15:49:28 crc kubenswrapper[4790]: E1124 15:49:28.472024 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="56b9940a224d2d309a332cfcbce228cc2bf78759b8ba90c6fba493c1a3781739" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 15:49:28 crc kubenswrapper[4790]: E1124 15:49:28.498003 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="56b9940a224d2d309a332cfcbce228cc2bf78759b8ba90c6fba493c1a3781739" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 15:49:28 crc kubenswrapper[4790]: E1124 15:49:28.501998 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="56b9940a224d2d309a332cfcbce228cc2bf78759b8ba90c6fba493c1a3781739" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 15:49:28 crc kubenswrapper[4790]: E1124 15:49:28.502059 4790 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="2a3a068f-92a8-43cf-83dc-8775c1d9ee84" containerName="nova-scheduler-scheduler" Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.517236 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17309035-deb1-4185-80ac-b4483116301f-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"17309035-deb1-4185-80ac-b4483116301f\") " pod="openstack/nova-cell1-conductor-0" Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.517472 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17309035-deb1-4185-80ac-b4483116301f-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"17309035-deb1-4185-80ac-b4483116301f\") " pod="openstack/nova-cell1-conductor-0" Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.517496 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wc2t\" (UniqueName: \"kubernetes.io/projected/17309035-deb1-4185-80ac-b4483116301f-kube-api-access-2wc2t\") pod \"nova-cell1-conductor-0\" (UID: \"17309035-deb1-4185-80ac-b4483116301f\") " pod="openstack/nova-cell1-conductor-0" Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.537325 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gh2nm" podUID="22c66da7-9f10-44d6-8e01-aab24fe93c78" containerName="registry-server" probeResult="failure" output=< Nov 24 15:49:28 crc kubenswrapper[4790]: timeout: failed to connect service ":50051" within 1s Nov 24 15:49:28 crc kubenswrapper[4790]: > Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.621152 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17309035-deb1-4185-80ac-b4483116301f-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"17309035-deb1-4185-80ac-b4483116301f\") " pod="openstack/nova-cell1-conductor-0" Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.621238 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wc2t\" (UniqueName: \"kubernetes.io/projected/17309035-deb1-4185-80ac-b4483116301f-kube-api-access-2wc2t\") pod \"nova-cell1-conductor-0\" (UID: \"17309035-deb1-4185-80ac-b4483116301f\") " pod="openstack/nova-cell1-conductor-0" Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.621325 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17309035-deb1-4185-80ac-b4483116301f-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"17309035-deb1-4185-80ac-b4483116301f\") " pod="openstack/nova-cell1-conductor-0" Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.629761 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17309035-deb1-4185-80ac-b4483116301f-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"17309035-deb1-4185-80ac-b4483116301f\") " pod="openstack/nova-cell1-conductor-0" Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.641511 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17309035-deb1-4185-80ac-b4483116301f-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"17309035-deb1-4185-80ac-b4483116301f\") " pod="openstack/nova-cell1-conductor-0" Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.642418 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wc2t\" (UniqueName: \"kubernetes.io/projected/17309035-deb1-4185-80ac-b4483116301f-kube-api-access-2wc2t\") pod \"nova-cell1-conductor-0\" (UID: \"17309035-deb1-4185-80ac-b4483116301f\") " pod="openstack/nova-cell1-conductor-0" Nov 24 15:49:28 crc kubenswrapper[4790]: I1124 15:49:28.749299 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 15:49:29 crc kubenswrapper[4790]: I1124 15:49:29.275942 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 15:49:29 crc kubenswrapper[4790]: I1124 15:49:29.302522 4790 generic.go:334] "Generic (PLEG): container finished" podID="cda1d09b-2432-4a64-8097-69246eb196a4" containerID="abec8bd8546de72cc3c06177bd23ea100d8a5e09334932e57afe57cd227f16b7" exitCode=0 Nov 24 15:49:29 crc kubenswrapper[4790]: I1124 15:49:29.302567 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"cda1d09b-2432-4a64-8097-69246eb196a4","Type":"ContainerDied","Data":"abec8bd8546de72cc3c06177bd23ea100d8a5e09334932e57afe57cd227f16b7"} Nov 24 15:49:29 crc kubenswrapper[4790]: I1124 15:49:29.570298 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-swvz2"] Nov 24 15:49:29 crc kubenswrapper[4790]: I1124 15:49:29.573251 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-swvz2" Nov 24 15:49:29 crc kubenswrapper[4790]: I1124 15:49:29.597026 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-swvz2"] Nov 24 15:49:29 crc kubenswrapper[4790]: E1124 15:49:29.604347 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of abec8bd8546de72cc3c06177bd23ea100d8a5e09334932e57afe57cd227f16b7 is running failed: container process not found" containerID="abec8bd8546de72cc3c06177bd23ea100d8a5e09334932e57afe57cd227f16b7" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 24 15:49:29 crc kubenswrapper[4790]: E1124 15:49:29.604704 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of abec8bd8546de72cc3c06177bd23ea100d8a5e09334932e57afe57cd227f16b7 is running failed: container process not found" containerID="abec8bd8546de72cc3c06177bd23ea100d8a5e09334932e57afe57cd227f16b7" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 24 15:49:29 crc kubenswrapper[4790]: E1124 15:49:29.605069 4790 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of abec8bd8546de72cc3c06177bd23ea100d8a5e09334932e57afe57cd227f16b7 is running failed: container process not found" containerID="abec8bd8546de72cc3c06177bd23ea100d8a5e09334932e57afe57cd227f16b7" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 24 15:49:29 crc kubenswrapper[4790]: E1124 15:49:29.605114 4790 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of abec8bd8546de72cc3c06177bd23ea100d8a5e09334932e57afe57cd227f16b7 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="cda1d09b-2432-4a64-8097-69246eb196a4" containerName="nova-cell0-conductor-conductor" Nov 24 15:49:29 crc kubenswrapper[4790]: I1124 15:49:29.744495 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s958k\" (UniqueName: \"kubernetes.io/projected/1f246eee-cba0-4bc5-9d93-bb318e763cc1-kube-api-access-s958k\") pod \"community-operators-swvz2\" (UID: \"1f246eee-cba0-4bc5-9d93-bb318e763cc1\") " pod="openshift-marketplace/community-operators-swvz2" Nov 24 15:49:29 crc kubenswrapper[4790]: I1124 15:49:29.744558 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f246eee-cba0-4bc5-9d93-bb318e763cc1-catalog-content\") pod \"community-operators-swvz2\" (UID: \"1f246eee-cba0-4bc5-9d93-bb318e763cc1\") " pod="openshift-marketplace/community-operators-swvz2" Nov 24 15:49:29 crc kubenswrapper[4790]: I1124 15:49:29.744585 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f246eee-cba0-4bc5-9d93-bb318e763cc1-utilities\") pod \"community-operators-swvz2\" (UID: \"1f246eee-cba0-4bc5-9d93-bb318e763cc1\") " pod="openshift-marketplace/community-operators-swvz2" Nov 24 15:49:29 crc kubenswrapper[4790]: I1124 15:49:29.797131 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 15:49:29 crc kubenswrapper[4790]: I1124 15:49:29.846427 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f246eee-cba0-4bc5-9d93-bb318e763cc1-catalog-content\") pod \"community-operators-swvz2\" (UID: \"1f246eee-cba0-4bc5-9d93-bb318e763cc1\") " pod="openshift-marketplace/community-operators-swvz2" Nov 24 15:49:29 crc kubenswrapper[4790]: I1124 15:49:29.846477 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f246eee-cba0-4bc5-9d93-bb318e763cc1-utilities\") pod \"community-operators-swvz2\" (UID: \"1f246eee-cba0-4bc5-9d93-bb318e763cc1\") " pod="openshift-marketplace/community-operators-swvz2" Nov 24 15:49:29 crc kubenswrapper[4790]: I1124 15:49:29.846649 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s958k\" (UniqueName: \"kubernetes.io/projected/1f246eee-cba0-4bc5-9d93-bb318e763cc1-kube-api-access-s958k\") pod \"community-operators-swvz2\" (UID: \"1f246eee-cba0-4bc5-9d93-bb318e763cc1\") " pod="openshift-marketplace/community-operators-swvz2" Nov 24 15:49:29 crc kubenswrapper[4790]: I1124 15:49:29.847279 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f246eee-cba0-4bc5-9d93-bb318e763cc1-catalog-content\") pod \"community-operators-swvz2\" (UID: \"1f246eee-cba0-4bc5-9d93-bb318e763cc1\") " pod="openshift-marketplace/community-operators-swvz2" Nov 24 15:49:29 crc kubenswrapper[4790]: I1124 15:49:29.847482 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f246eee-cba0-4bc5-9d93-bb318e763cc1-utilities\") pod \"community-operators-swvz2\" (UID: \"1f246eee-cba0-4bc5-9d93-bb318e763cc1\") " pod="openshift-marketplace/community-operators-swvz2" Nov 24 15:49:29 crc kubenswrapper[4790]: I1124 15:49:29.867587 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s958k\" (UniqueName: \"kubernetes.io/projected/1f246eee-cba0-4bc5-9d93-bb318e763cc1-kube-api-access-s958k\") pod \"community-operators-swvz2\" (UID: \"1f246eee-cba0-4bc5-9d93-bb318e763cc1\") " pod="openshift-marketplace/community-operators-swvz2" Nov 24 15:49:29 crc kubenswrapper[4790]: I1124 15:49:29.950551 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cda1d09b-2432-4a64-8097-69246eb196a4-config-data\") pod \"cda1d09b-2432-4a64-8097-69246eb196a4\" (UID: \"cda1d09b-2432-4a64-8097-69246eb196a4\") " Nov 24 15:49:29 crc kubenswrapper[4790]: I1124 15:49:29.951084 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cda1d09b-2432-4a64-8097-69246eb196a4-combined-ca-bundle\") pod \"cda1d09b-2432-4a64-8097-69246eb196a4\" (UID: \"cda1d09b-2432-4a64-8097-69246eb196a4\") " Nov 24 15:49:29 crc kubenswrapper[4790]: I1124 15:49:29.951172 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mlpq\" (UniqueName: \"kubernetes.io/projected/cda1d09b-2432-4a64-8097-69246eb196a4-kube-api-access-2mlpq\") pod \"cda1d09b-2432-4a64-8097-69246eb196a4\" (UID: \"cda1d09b-2432-4a64-8097-69246eb196a4\") " Nov 24 15:49:29 crc kubenswrapper[4790]: I1124 15:49:29.981363 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cda1d09b-2432-4a64-8097-69246eb196a4-kube-api-access-2mlpq" (OuterVolumeSpecName: "kube-api-access-2mlpq") pod "cda1d09b-2432-4a64-8097-69246eb196a4" (UID: "cda1d09b-2432-4a64-8097-69246eb196a4"). InnerVolumeSpecName "kube-api-access-2mlpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:49:29 crc kubenswrapper[4790]: I1124 15:49:29.995719 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cda1d09b-2432-4a64-8097-69246eb196a4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cda1d09b-2432-4a64-8097-69246eb196a4" (UID: "cda1d09b-2432-4a64-8097-69246eb196a4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:49:29 crc kubenswrapper[4790]: I1124 15:49:29.997208 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cda1d09b-2432-4a64-8097-69246eb196a4-config-data" (OuterVolumeSpecName: "config-data") pod "cda1d09b-2432-4a64-8097-69246eb196a4" (UID: "cda1d09b-2432-4a64-8097-69246eb196a4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.058474 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cda1d09b-2432-4a64-8097-69246eb196a4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.058518 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mlpq\" (UniqueName: \"kubernetes.io/projected/cda1d09b-2432-4a64-8097-69246eb196a4-kube-api-access-2mlpq\") on node \"crc\" DevicePath \"\"" Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.058532 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cda1d09b-2432-4a64-8097-69246eb196a4-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.111341 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-swvz2" Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.202409 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="924b2c00-a948-46d2-b103-25f99bc9477e" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.85:8775/\": read tcp 10.217.0.2:59552->10.217.1.85:8775: read: connection reset by peer" Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.202676 4790 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="924b2c00-a948-46d2-b103-25f99bc9477e" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.85:8775/\": read tcp 10.217.0.2:59554->10.217.1.85:8775: read: connection reset by peer" Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.327366 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.328374 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cea27eb-31d4-4e3b-a141-d2593bc21427" path="/var/lib/kubelet/pods/5cea27eb-31d4-4e3b-a141-d2593bc21427/volumes" Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.340711 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"cda1d09b-2432-4a64-8097-69246eb196a4","Type":"ContainerDied","Data":"1f6ae3c97f494a3db8e448cd3592e357186cd2813c61bf0619565836d892037e"} Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.340752 4790 scope.go:117] "RemoveContainer" containerID="abec8bd8546de72cc3c06177bd23ea100d8a5e09334932e57afe57cd227f16b7" Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.363671 4790 generic.go:334] "Generic (PLEG): container finished" podID="039efc51-838d-4aac-b0e8-7ba1c9b376ce" containerID="58706d8229eea72d6668f75cb9016630db40c6067ad81aa31661875ab841589a" exitCode=0 Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.363750 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"039efc51-838d-4aac-b0e8-7ba1c9b376ce","Type":"ContainerDied","Data":"58706d8229eea72d6668f75cb9016630db40c6067ad81aa31661875ab841589a"} Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.382449 4790 generic.go:334] "Generic (PLEG): container finished" podID="924b2c00-a948-46d2-b103-25f99bc9477e" containerID="4b805b27d5fe3f30506bbea66532569a1055c70574cd5abc2c7a8a50b9e863b1" exitCode=0 Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.382512 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"924b2c00-a948-46d2-b103-25f99bc9477e","Type":"ContainerDied","Data":"4b805b27d5fe3f30506bbea66532569a1055c70574cd5abc2c7a8a50b9e863b1"} Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.386589 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"17309035-deb1-4185-80ac-b4483116301f","Type":"ContainerStarted","Data":"e6fadec1bcceed24fba587f267852968446be1cfb11cbf6454adc50e75e46000"} Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.386752 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"17309035-deb1-4185-80ac-b4483116301f","Type":"ContainerStarted","Data":"e479b403ece99fceab330567ad842e02b7450c96b94df146afc656c718e2caac"} Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.387828 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.407202 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.424164 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.434488 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 15:49:30 crc kubenswrapper[4790]: E1124 15:49:30.434976 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cda1d09b-2432-4a64-8097-69246eb196a4" containerName="nova-cell0-conductor-conductor" Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.434992 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="cda1d09b-2432-4a64-8097-69246eb196a4" containerName="nova-cell0-conductor-conductor" Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.435216 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="cda1d09b-2432-4a64-8097-69246eb196a4" containerName="nova-cell0-conductor-conductor" Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.435996 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.437660 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.437644647 podStartE2EDuration="2.437644647s" podCreationTimestamp="2025-11-24 15:49:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 15:49:30.407182528 +0000 UTC m=+9418.787076190" watchObservedRunningTime="2025-11-24 15:49:30.437644647 +0000 UTC m=+9418.817538309" Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.438598 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.451715 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.571905 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwp4r\" (UniqueName: \"kubernetes.io/projected/3df0c98f-9380-488b-a72e-4111d10941d9-kube-api-access-mwp4r\") pod \"nova-cell0-conductor-0\" (UID: \"3df0c98f-9380-488b-a72e-4111d10941d9\") " pod="openstack/nova-cell0-conductor-0" Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.572071 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3df0c98f-9380-488b-a72e-4111d10941d9-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"3df0c98f-9380-488b-a72e-4111d10941d9\") " pod="openstack/nova-cell0-conductor-0" Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.572109 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3df0c98f-9380-488b-a72e-4111d10941d9-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"3df0c98f-9380-488b-a72e-4111d10941d9\") " pod="openstack/nova-cell0-conductor-0" Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.641658 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-swvz2"] Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.676740 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3df0c98f-9380-488b-a72e-4111d10941d9-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"3df0c98f-9380-488b-a72e-4111d10941d9\") " pod="openstack/nova-cell0-conductor-0" Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.677125 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3df0c98f-9380-488b-a72e-4111d10941d9-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"3df0c98f-9380-488b-a72e-4111d10941d9\") " pod="openstack/nova-cell0-conductor-0" Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.677593 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwp4r\" (UniqueName: \"kubernetes.io/projected/3df0c98f-9380-488b-a72e-4111d10941d9-kube-api-access-mwp4r\") pod \"nova-cell0-conductor-0\" (UID: \"3df0c98f-9380-488b-a72e-4111d10941d9\") " pod="openstack/nova-cell0-conductor-0" Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.686725 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3df0c98f-9380-488b-a72e-4111d10941d9-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"3df0c98f-9380-488b-a72e-4111d10941d9\") " pod="openstack/nova-cell0-conductor-0" Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.698467 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwp4r\" (UniqueName: \"kubernetes.io/projected/3df0c98f-9380-488b-a72e-4111d10941d9-kube-api-access-mwp4r\") pod \"nova-cell0-conductor-0\" (UID: \"3df0c98f-9380-488b-a72e-4111d10941d9\") " pod="openstack/nova-cell0-conductor-0" Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.706411 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3df0c98f-9380-488b-a72e-4111d10941d9-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"3df0c98f-9380-488b-a72e-4111d10941d9\") " pod="openstack/nova-cell0-conductor-0" Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.758361 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 15:49:30 crc kubenswrapper[4790]: I1124 15:49:30.961923 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:30.981868 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.096380 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/924b2c00-a948-46d2-b103-25f99bc9477e-logs\") pod \"924b2c00-a948-46d2-b103-25f99bc9477e\" (UID: \"924b2c00-a948-46d2-b103-25f99bc9477e\") " Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.096422 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4lpw\" (UniqueName: \"kubernetes.io/projected/039efc51-838d-4aac-b0e8-7ba1c9b376ce-kube-api-access-t4lpw\") pod \"039efc51-838d-4aac-b0e8-7ba1c9b376ce\" (UID: \"039efc51-838d-4aac-b0e8-7ba1c9b376ce\") " Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.096443 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/039efc51-838d-4aac-b0e8-7ba1c9b376ce-combined-ca-bundle\") pod \"039efc51-838d-4aac-b0e8-7ba1c9b376ce\" (UID: \"039efc51-838d-4aac-b0e8-7ba1c9b376ce\") " Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.096530 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/924b2c00-a948-46d2-b103-25f99bc9477e-combined-ca-bundle\") pod \"924b2c00-a948-46d2-b103-25f99bc9477e\" (UID: \"924b2c00-a948-46d2-b103-25f99bc9477e\") " Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.096549 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4z2t\" (UniqueName: \"kubernetes.io/projected/924b2c00-a948-46d2-b103-25f99bc9477e-kube-api-access-b4z2t\") pod \"924b2c00-a948-46d2-b103-25f99bc9477e\" (UID: \"924b2c00-a948-46d2-b103-25f99bc9477e\") " Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.096675 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/924b2c00-a948-46d2-b103-25f99bc9477e-config-data\") pod \"924b2c00-a948-46d2-b103-25f99bc9477e\" (UID: \"924b2c00-a948-46d2-b103-25f99bc9477e\") " Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.096713 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/039efc51-838d-4aac-b0e8-7ba1c9b376ce-logs\") pod \"039efc51-838d-4aac-b0e8-7ba1c9b376ce\" (UID: \"039efc51-838d-4aac-b0e8-7ba1c9b376ce\") " Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.096808 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/039efc51-838d-4aac-b0e8-7ba1c9b376ce-config-data\") pod \"039efc51-838d-4aac-b0e8-7ba1c9b376ce\" (UID: \"039efc51-838d-4aac-b0e8-7ba1c9b376ce\") " Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.097696 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/039efc51-838d-4aac-b0e8-7ba1c9b376ce-logs" (OuterVolumeSpecName: "logs") pod "039efc51-838d-4aac-b0e8-7ba1c9b376ce" (UID: "039efc51-838d-4aac-b0e8-7ba1c9b376ce"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.098374 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/924b2c00-a948-46d2-b103-25f99bc9477e-logs" (OuterVolumeSpecName: "logs") pod "924b2c00-a948-46d2-b103-25f99bc9477e" (UID: "924b2c00-a948-46d2-b103-25f99bc9477e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.102348 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/924b2c00-a948-46d2-b103-25f99bc9477e-kube-api-access-b4z2t" (OuterVolumeSpecName: "kube-api-access-b4z2t") pod "924b2c00-a948-46d2-b103-25f99bc9477e" (UID: "924b2c00-a948-46d2-b103-25f99bc9477e"). InnerVolumeSpecName "kube-api-access-b4z2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.103657 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/039efc51-838d-4aac-b0e8-7ba1c9b376ce-kube-api-access-t4lpw" (OuterVolumeSpecName: "kube-api-access-t4lpw") pod "039efc51-838d-4aac-b0e8-7ba1c9b376ce" (UID: "039efc51-838d-4aac-b0e8-7ba1c9b376ce"). InnerVolumeSpecName "kube-api-access-t4lpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.139800 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/924b2c00-a948-46d2-b103-25f99bc9477e-config-data" (OuterVolumeSpecName: "config-data") pod "924b2c00-a948-46d2-b103-25f99bc9477e" (UID: "924b2c00-a948-46d2-b103-25f99bc9477e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.139969 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/924b2c00-a948-46d2-b103-25f99bc9477e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "924b2c00-a948-46d2-b103-25f99bc9477e" (UID: "924b2c00-a948-46d2-b103-25f99bc9477e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.149027 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/039efc51-838d-4aac-b0e8-7ba1c9b376ce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "039efc51-838d-4aac-b0e8-7ba1c9b376ce" (UID: "039efc51-838d-4aac-b0e8-7ba1c9b376ce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.151650 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/039efc51-838d-4aac-b0e8-7ba1c9b376ce-config-data" (OuterVolumeSpecName: "config-data") pod "039efc51-838d-4aac-b0e8-7ba1c9b376ce" (UID: "039efc51-838d-4aac-b0e8-7ba1c9b376ce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.199313 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/924b2c00-a948-46d2-b103-25f99bc9477e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.199539 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4z2t\" (UniqueName: \"kubernetes.io/projected/924b2c00-a948-46d2-b103-25f99bc9477e-kube-api-access-b4z2t\") on node \"crc\" DevicePath \"\"" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.199664 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/924b2c00-a948-46d2-b103-25f99bc9477e-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.199748 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/039efc51-838d-4aac-b0e8-7ba1c9b376ce-logs\") on node \"crc\" DevicePath \"\"" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.199822 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/039efc51-838d-4aac-b0e8-7ba1c9b376ce-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.199912 4790 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/924b2c00-a948-46d2-b103-25f99bc9477e-logs\") on node \"crc\" DevicePath \"\"" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.200003 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4lpw\" (UniqueName: \"kubernetes.io/projected/039efc51-838d-4aac-b0e8-7ba1c9b376ce-kube-api-access-t4lpw\") on node \"crc\" DevicePath \"\"" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.200081 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/039efc51-838d-4aac-b0e8-7ba1c9b376ce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.347694 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 15:49:31 crc kubenswrapper[4790]: W1124 15:49:31.348955 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3df0c98f_9380_488b_a72e_4111d10941d9.slice/crio-d1027b4756f69b62e154d2a2b57d4bfbae503cbf9316b7c69e2e92d143f9e115 WatchSource:0}: Error finding container d1027b4756f69b62e154d2a2b57d4bfbae503cbf9316b7c69e2e92d143f9e115: Status 404 returned error can't find the container with id d1027b4756f69b62e154d2a2b57d4bfbae503cbf9316b7c69e2e92d143f9e115 Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.397068 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"3df0c98f-9380-488b-a72e-4111d10941d9","Type":"ContainerStarted","Data":"d1027b4756f69b62e154d2a2b57d4bfbae503cbf9316b7c69e2e92d143f9e115"} Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.400272 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.400273 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"039efc51-838d-4aac-b0e8-7ba1c9b376ce","Type":"ContainerDied","Data":"41bb5a692a4693b864f8feb5f54d078f84409c52facf3089106c328c4c61a7ed"} Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.400404 4790 scope.go:117] "RemoveContainer" containerID="58706d8229eea72d6668f75cb9016630db40c6067ad81aa31661875ab841589a" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.414591 4790 generic.go:334] "Generic (PLEG): container finished" podID="1f246eee-cba0-4bc5-9d93-bb318e763cc1" containerID="0b8e941eb590d225ba2bacda6ee356b06b2afcebe831e7eddb16a05bd8770b0d" exitCode=0 Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.414700 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swvz2" event={"ID":"1f246eee-cba0-4bc5-9d93-bb318e763cc1","Type":"ContainerDied","Data":"0b8e941eb590d225ba2bacda6ee356b06b2afcebe831e7eddb16a05bd8770b0d"} Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.414731 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swvz2" event={"ID":"1f246eee-cba0-4bc5-9d93-bb318e763cc1","Type":"ContainerStarted","Data":"c70518138d727dc1e39d95d267d35d39708f64cc2df3bd41a8670190175bfdf5"} Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.418538 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.418568 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"924b2c00-a948-46d2-b103-25f99bc9477e","Type":"ContainerDied","Data":"a70f11282a58121dc1bb2e63baa459ff35f4c79eb897248fb2f4d504419180ee"} Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.440273 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.443773 4790 scope.go:117] "RemoveContainer" containerID="4a29ff14976dcb69b6759470411ff7beb1d298155f005c3981206b31876014bd" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.478207 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.493146 4790 scope.go:117] "RemoveContainer" containerID="4b805b27d5fe3f30506bbea66532569a1055c70574cd5abc2c7a8a50b9e863b1" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.510939 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 15:49:31 crc kubenswrapper[4790]: E1124 15:49:31.511397 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="039efc51-838d-4aac-b0e8-7ba1c9b376ce" containerName="nova-api-log" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.511414 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="039efc51-838d-4aac-b0e8-7ba1c9b376ce" containerName="nova-api-log" Nov 24 15:49:31 crc kubenswrapper[4790]: E1124 15:49:31.511432 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="924b2c00-a948-46d2-b103-25f99bc9477e" containerName="nova-metadata-metadata" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.511440 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="924b2c00-a948-46d2-b103-25f99bc9477e" containerName="nova-metadata-metadata" Nov 24 15:49:31 crc kubenswrapper[4790]: E1124 15:49:31.511454 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="924b2c00-a948-46d2-b103-25f99bc9477e" containerName="nova-metadata-log" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.511460 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="924b2c00-a948-46d2-b103-25f99bc9477e" containerName="nova-metadata-log" Nov 24 15:49:31 crc kubenswrapper[4790]: E1124 15:49:31.511481 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="039efc51-838d-4aac-b0e8-7ba1c9b376ce" containerName="nova-api-api" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.511486 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="039efc51-838d-4aac-b0e8-7ba1c9b376ce" containerName="nova-api-api" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.511698 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="924b2c00-a948-46d2-b103-25f99bc9477e" containerName="nova-metadata-metadata" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.511714 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="924b2c00-a948-46d2-b103-25f99bc9477e" containerName="nova-metadata-log" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.511723 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="039efc51-838d-4aac-b0e8-7ba1c9b376ce" containerName="nova-api-log" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.511734 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="039efc51-838d-4aac-b0e8-7ba1c9b376ce" containerName="nova-api-api" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.512824 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.550450 4790 scope.go:117] "RemoveContainer" containerID="5864a737edda91cad9e80b03ac060cd6a2fe7b521eb61e4e209e5c0ea954c4c1" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.552046 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.589331 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.601207 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.612411 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.621766 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.623553 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.626114 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.641453 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.645529 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d36ae33-dbe8-4323-8bc6-83ca5dcf40ea-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3d36ae33-dbe8-4323-8bc6-83ca5dcf40ea\") " pod="openstack/nova-api-0" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.645592 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d36ae33-dbe8-4323-8bc6-83ca5dcf40ea-logs\") pod \"nova-api-0\" (UID: \"3d36ae33-dbe8-4323-8bc6-83ca5dcf40ea\") " pod="openstack/nova-api-0" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.645619 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztw6c\" (UniqueName: \"kubernetes.io/projected/3d36ae33-dbe8-4323-8bc6-83ca5dcf40ea-kube-api-access-ztw6c\") pod \"nova-api-0\" (UID: \"3d36ae33-dbe8-4323-8bc6-83ca5dcf40ea\") " pod="openstack/nova-api-0" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.645703 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d36ae33-dbe8-4323-8bc6-83ca5dcf40ea-config-data\") pod \"nova-api-0\" (UID: \"3d36ae33-dbe8-4323-8bc6-83ca5dcf40ea\") " pod="openstack/nova-api-0" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.747004 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d36ae33-dbe8-4323-8bc6-83ca5dcf40ea-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3d36ae33-dbe8-4323-8bc6-83ca5dcf40ea\") " pod="openstack/nova-api-0" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.747111 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d36ae33-dbe8-4323-8bc6-83ca5dcf40ea-logs\") pod \"nova-api-0\" (UID: \"3d36ae33-dbe8-4323-8bc6-83ca5dcf40ea\") " pod="openstack/nova-api-0" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.747156 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztw6c\" (UniqueName: \"kubernetes.io/projected/3d36ae33-dbe8-4323-8bc6-83ca5dcf40ea-kube-api-access-ztw6c\") pod \"nova-api-0\" (UID: \"3d36ae33-dbe8-4323-8bc6-83ca5dcf40ea\") " pod="openstack/nova-api-0" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.747220 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d36ae33-dbe8-4323-8bc6-83ca5dcf40ea-config-data\") pod \"nova-api-0\" (UID: \"3d36ae33-dbe8-4323-8bc6-83ca5dcf40ea\") " pod="openstack/nova-api-0" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.747274 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvltz\" (UniqueName: \"kubernetes.io/projected/9a7f156b-e253-4f3b-b4ce-5c3c226ccfcd-kube-api-access-zvltz\") pod \"nova-metadata-0\" (UID: \"9a7f156b-e253-4f3b-b4ce-5c3c226ccfcd\") " pod="openstack/nova-metadata-0" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.747300 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a7f156b-e253-4f3b-b4ce-5c3c226ccfcd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9a7f156b-e253-4f3b-b4ce-5c3c226ccfcd\") " pod="openstack/nova-metadata-0" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.747335 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a7f156b-e253-4f3b-b4ce-5c3c226ccfcd-config-data\") pod \"nova-metadata-0\" (UID: \"9a7f156b-e253-4f3b-b4ce-5c3c226ccfcd\") " pod="openstack/nova-metadata-0" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.747370 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a7f156b-e253-4f3b-b4ce-5c3c226ccfcd-logs\") pod \"nova-metadata-0\" (UID: \"9a7f156b-e253-4f3b-b4ce-5c3c226ccfcd\") " pod="openstack/nova-metadata-0" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.748027 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d36ae33-dbe8-4323-8bc6-83ca5dcf40ea-logs\") pod \"nova-api-0\" (UID: \"3d36ae33-dbe8-4323-8bc6-83ca5dcf40ea\") " pod="openstack/nova-api-0" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.751592 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d36ae33-dbe8-4323-8bc6-83ca5dcf40ea-config-data\") pod \"nova-api-0\" (UID: \"3d36ae33-dbe8-4323-8bc6-83ca5dcf40ea\") " pod="openstack/nova-api-0" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.752740 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d36ae33-dbe8-4323-8bc6-83ca5dcf40ea-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3d36ae33-dbe8-4323-8bc6-83ca5dcf40ea\") " pod="openstack/nova-api-0" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.768389 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztw6c\" (UniqueName: \"kubernetes.io/projected/3d36ae33-dbe8-4323-8bc6-83ca5dcf40ea-kube-api-access-ztw6c\") pod \"nova-api-0\" (UID: \"3d36ae33-dbe8-4323-8bc6-83ca5dcf40ea\") " pod="openstack/nova-api-0" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.849369 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvltz\" (UniqueName: \"kubernetes.io/projected/9a7f156b-e253-4f3b-b4ce-5c3c226ccfcd-kube-api-access-zvltz\") pod \"nova-metadata-0\" (UID: \"9a7f156b-e253-4f3b-b4ce-5c3c226ccfcd\") " pod="openstack/nova-metadata-0" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.849448 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a7f156b-e253-4f3b-b4ce-5c3c226ccfcd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9a7f156b-e253-4f3b-b4ce-5c3c226ccfcd\") " pod="openstack/nova-metadata-0" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.849488 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a7f156b-e253-4f3b-b4ce-5c3c226ccfcd-config-data\") pod \"nova-metadata-0\" (UID: \"9a7f156b-e253-4f3b-b4ce-5c3c226ccfcd\") " pod="openstack/nova-metadata-0" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.849542 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a7f156b-e253-4f3b-b4ce-5c3c226ccfcd-logs\") pod \"nova-metadata-0\" (UID: \"9a7f156b-e253-4f3b-b4ce-5c3c226ccfcd\") " pod="openstack/nova-metadata-0" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.850073 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a7f156b-e253-4f3b-b4ce-5c3c226ccfcd-logs\") pod \"nova-metadata-0\" (UID: \"9a7f156b-e253-4f3b-b4ce-5c3c226ccfcd\") " pod="openstack/nova-metadata-0" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.853240 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a7f156b-e253-4f3b-b4ce-5c3c226ccfcd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9a7f156b-e253-4f3b-b4ce-5c3c226ccfcd\") " pod="openstack/nova-metadata-0" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.854505 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a7f156b-e253-4f3b-b4ce-5c3c226ccfcd-config-data\") pod \"nova-metadata-0\" (UID: \"9a7f156b-e253-4f3b-b4ce-5c3c226ccfcd\") " pod="openstack/nova-metadata-0" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.870170 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvltz\" (UniqueName: \"kubernetes.io/projected/9a7f156b-e253-4f3b-b4ce-5c3c226ccfcd-kube-api-access-zvltz\") pod \"nova-metadata-0\" (UID: \"9a7f156b-e253-4f3b-b4ce-5c3c226ccfcd\") " pod="openstack/nova-metadata-0" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.883763 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 15:49:31 crc kubenswrapper[4790]: I1124 15:49:31.945720 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 15:49:32 crc kubenswrapper[4790]: I1124 15:49:32.331762 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="039efc51-838d-4aac-b0e8-7ba1c9b376ce" path="/var/lib/kubelet/pods/039efc51-838d-4aac-b0e8-7ba1c9b376ce/volumes" Nov 24 15:49:32 crc kubenswrapper[4790]: I1124 15:49:32.332857 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="924b2c00-a948-46d2-b103-25f99bc9477e" path="/var/lib/kubelet/pods/924b2c00-a948-46d2-b103-25f99bc9477e/volumes" Nov 24 15:49:32 crc kubenswrapper[4790]: I1124 15:49:32.333424 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cda1d09b-2432-4a64-8097-69246eb196a4" path="/var/lib/kubelet/pods/cda1d09b-2432-4a64-8097-69246eb196a4/volumes" Nov 24 15:49:32 crc kubenswrapper[4790]: I1124 15:49:32.458483 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"3df0c98f-9380-488b-a72e-4111d10941d9","Type":"ContainerStarted","Data":"e39b059780b732b1f2d189e08c9f2ec33fe35c7984acd0faa378e5386b97679c"} Nov 24 15:49:32 crc kubenswrapper[4790]: I1124 15:49:32.459950 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 24 15:49:32 crc kubenswrapper[4790]: I1124 15:49:32.468835 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 15:49:32 crc kubenswrapper[4790]: I1124 15:49:32.488037 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 15:49:32 crc kubenswrapper[4790]: W1124 15:49:32.490784 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a7f156b_e253_4f3b_b4ce_5c3c226ccfcd.slice/crio-de3bbe01046dacae6ca609dd8c4b817f764965763b971c05bc997421447c13d9 WatchSource:0}: Error finding container de3bbe01046dacae6ca609dd8c4b817f764965763b971c05bc997421447c13d9: Status 404 returned error can't find the container with id de3bbe01046dacae6ca609dd8c4b817f764965763b971c05bc997421447c13d9 Nov 24 15:49:32 crc kubenswrapper[4790]: I1124 15:49:32.492690 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swvz2" event={"ID":"1f246eee-cba0-4bc5-9d93-bb318e763cc1","Type":"ContainerStarted","Data":"005b8586bcee4e28a61fb6aa607fb5a0099d605d0635aa8013821f11f59baf16"} Nov 24 15:49:32 crc kubenswrapper[4790]: I1124 15:49:32.501961 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.501938689 podStartE2EDuration="2.501938689s" podCreationTimestamp="2025-11-24 15:49:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 15:49:32.480223949 +0000 UTC m=+9420.860117621" watchObservedRunningTime="2025-11-24 15:49:32.501938689 +0000 UTC m=+9420.881832351" Nov 24 15:49:32 crc kubenswrapper[4790]: I1124 15:49:32.923032 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.082316 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nk4ts\" (UniqueName: \"kubernetes.io/projected/2a3a068f-92a8-43cf-83dc-8775c1d9ee84-kube-api-access-nk4ts\") pod \"2a3a068f-92a8-43cf-83dc-8775c1d9ee84\" (UID: \"2a3a068f-92a8-43cf-83dc-8775c1d9ee84\") " Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.082757 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a3a068f-92a8-43cf-83dc-8775c1d9ee84-combined-ca-bundle\") pod \"2a3a068f-92a8-43cf-83dc-8775c1d9ee84\" (UID: \"2a3a068f-92a8-43cf-83dc-8775c1d9ee84\") " Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.083145 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a3a068f-92a8-43cf-83dc-8775c1d9ee84-config-data\") pod \"2a3a068f-92a8-43cf-83dc-8775c1d9ee84\" (UID: \"2a3a068f-92a8-43cf-83dc-8775c1d9ee84\") " Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.090961 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a3a068f-92a8-43cf-83dc-8775c1d9ee84-kube-api-access-nk4ts" (OuterVolumeSpecName: "kube-api-access-nk4ts") pod "2a3a068f-92a8-43cf-83dc-8775c1d9ee84" (UID: "2a3a068f-92a8-43cf-83dc-8775c1d9ee84"). InnerVolumeSpecName "kube-api-access-nk4ts". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.127071 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a3a068f-92a8-43cf-83dc-8775c1d9ee84-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2a3a068f-92a8-43cf-83dc-8775c1d9ee84" (UID: "2a3a068f-92a8-43cf-83dc-8775c1d9ee84"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.141961 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a3a068f-92a8-43cf-83dc-8775c1d9ee84-config-data" (OuterVolumeSpecName: "config-data") pod "2a3a068f-92a8-43cf-83dc-8775c1d9ee84" (UID: "2a3a068f-92a8-43cf-83dc-8775c1d9ee84"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.187267 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a3a068f-92a8-43cf-83dc-8775c1d9ee84-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.187492 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nk4ts\" (UniqueName: \"kubernetes.io/projected/2a3a068f-92a8-43cf-83dc-8775c1d9ee84-kube-api-access-nk4ts\") on node \"crc\" DevicePath \"\"" Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.187553 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a3a068f-92a8-43cf-83dc-8775c1d9ee84-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.519055 4790 generic.go:334] "Generic (PLEG): container finished" podID="1f246eee-cba0-4bc5-9d93-bb318e763cc1" containerID="005b8586bcee4e28a61fb6aa607fb5a0099d605d0635aa8013821f11f59baf16" exitCode=0 Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.519134 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swvz2" event={"ID":"1f246eee-cba0-4bc5-9d93-bb318e763cc1","Type":"ContainerDied","Data":"005b8586bcee4e28a61fb6aa607fb5a0099d605d0635aa8013821f11f59baf16"} Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.521432 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9a7f156b-e253-4f3b-b4ce-5c3c226ccfcd","Type":"ContainerStarted","Data":"614fee5b2e69ec6a446fbc5b368660d774d7a8c3df78345a161a5f00ad6fc2b5"} Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.521472 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9a7f156b-e253-4f3b-b4ce-5c3c226ccfcd","Type":"ContainerStarted","Data":"0dc6a2c6093bbde3b93d4b392a4f6d71774fd8574e51bae32bfe7ed3e7b3a863"} Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.521484 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9a7f156b-e253-4f3b-b4ce-5c3c226ccfcd","Type":"ContainerStarted","Data":"de3bbe01046dacae6ca609dd8c4b817f764965763b971c05bc997421447c13d9"} Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.527310 4790 generic.go:334] "Generic (PLEG): container finished" podID="2a3a068f-92a8-43cf-83dc-8775c1d9ee84" containerID="56b9940a224d2d309a332cfcbce228cc2bf78759b8ba90c6fba493c1a3781739" exitCode=0 Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.527371 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2a3a068f-92a8-43cf-83dc-8775c1d9ee84","Type":"ContainerDied","Data":"56b9940a224d2d309a332cfcbce228cc2bf78759b8ba90c6fba493c1a3781739"} Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.527396 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2a3a068f-92a8-43cf-83dc-8775c1d9ee84","Type":"ContainerDied","Data":"9adeb24326bccef100166923b4275ac81338d33224168a41977c2e353450b62c"} Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.527413 4790 scope.go:117] "RemoveContainer" containerID="56b9940a224d2d309a332cfcbce228cc2bf78759b8ba90c6fba493c1a3781739" Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.527449 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.542897 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3d36ae33-dbe8-4323-8bc6-83ca5dcf40ea","Type":"ContainerStarted","Data":"6c3a6d2612fa69bc6af52e6237bb8e3376435f7596109f8f07d30cbdf0af9de7"} Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.542975 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3d36ae33-dbe8-4323-8bc6-83ca5dcf40ea","Type":"ContainerStarted","Data":"422f4d34d6cac8769bff61523b520b3ba6c60c9dfc5a948d4a8fad0fbd3f14a6"} Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.542989 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3d36ae33-dbe8-4323-8bc6-83ca5dcf40ea","Type":"ContainerStarted","Data":"e8ae134333333cdb473bccd24dc08abd6123551f65a01988fabab22758311f47"} Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.563245 4790 scope.go:117] "RemoveContainer" containerID="56b9940a224d2d309a332cfcbce228cc2bf78759b8ba90c6fba493c1a3781739" Nov 24 15:49:33 crc kubenswrapper[4790]: E1124 15:49:33.571123 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56b9940a224d2d309a332cfcbce228cc2bf78759b8ba90c6fba493c1a3781739\": container with ID starting with 56b9940a224d2d309a332cfcbce228cc2bf78759b8ba90c6fba493c1a3781739 not found: ID does not exist" containerID="56b9940a224d2d309a332cfcbce228cc2bf78759b8ba90c6fba493c1a3781739" Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.571196 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56b9940a224d2d309a332cfcbce228cc2bf78759b8ba90c6fba493c1a3781739"} err="failed to get container status \"56b9940a224d2d309a332cfcbce228cc2bf78759b8ba90c6fba493c1a3781739\": rpc error: code = NotFound desc = could not find container \"56b9940a224d2d309a332cfcbce228cc2bf78759b8ba90c6fba493c1a3781739\": container with ID starting with 56b9940a224d2d309a332cfcbce228cc2bf78759b8ba90c6fba493c1a3781739 not found: ID does not exist" Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.598721 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.598699097 podStartE2EDuration="2.598699097s" podCreationTimestamp="2025-11-24 15:49:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 15:49:33.573435636 +0000 UTC m=+9421.953329298" watchObservedRunningTime="2025-11-24 15:49:33.598699097 +0000 UTC m=+9421.978592759" Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.605549 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.605531082 podStartE2EDuration="2.605531082s" podCreationTimestamp="2025-11-24 15:49:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 15:49:33.593373425 +0000 UTC m=+9421.973267087" watchObservedRunningTime="2025-11-24 15:49:33.605531082 +0000 UTC m=+9421.985424744" Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.616439 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.635704 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.646918 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 15:49:33 crc kubenswrapper[4790]: E1124 15:49:33.647429 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a3a068f-92a8-43cf-83dc-8775c1d9ee84" containerName="nova-scheduler-scheduler" Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.647442 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a3a068f-92a8-43cf-83dc-8775c1d9ee84" containerName="nova-scheduler-scheduler" Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.647667 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a3a068f-92a8-43cf-83dc-8775c1d9ee84" containerName="nova-scheduler-scheduler" Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.648449 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.650818 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.656544 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.805838 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c19692e0-e586-4506-a9e7-28195138b290-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c19692e0-e586-4506-a9e7-28195138b290\") " pod="openstack/nova-scheduler-0" Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.806022 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c19692e0-e586-4506-a9e7-28195138b290-config-data\") pod \"nova-scheduler-0\" (UID: \"c19692e0-e586-4506-a9e7-28195138b290\") " pod="openstack/nova-scheduler-0" Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.806271 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9thn\" (UniqueName: \"kubernetes.io/projected/c19692e0-e586-4506-a9e7-28195138b290-kube-api-access-s9thn\") pod \"nova-scheduler-0\" (UID: \"c19692e0-e586-4506-a9e7-28195138b290\") " pod="openstack/nova-scheduler-0" Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.908498 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c19692e0-e586-4506-a9e7-28195138b290-config-data\") pod \"nova-scheduler-0\" (UID: \"c19692e0-e586-4506-a9e7-28195138b290\") " pod="openstack/nova-scheduler-0" Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.908711 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9thn\" (UniqueName: \"kubernetes.io/projected/c19692e0-e586-4506-a9e7-28195138b290-kube-api-access-s9thn\") pod \"nova-scheduler-0\" (UID: \"c19692e0-e586-4506-a9e7-28195138b290\") " pod="openstack/nova-scheduler-0" Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.909047 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c19692e0-e586-4506-a9e7-28195138b290-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c19692e0-e586-4506-a9e7-28195138b290\") " pod="openstack/nova-scheduler-0" Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.913564 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c19692e0-e586-4506-a9e7-28195138b290-config-data\") pod \"nova-scheduler-0\" (UID: \"c19692e0-e586-4506-a9e7-28195138b290\") " pod="openstack/nova-scheduler-0" Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.918116 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c19692e0-e586-4506-a9e7-28195138b290-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c19692e0-e586-4506-a9e7-28195138b290\") " pod="openstack/nova-scheduler-0" Nov 24 15:49:33 crc kubenswrapper[4790]: I1124 15:49:33.925186 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9thn\" (UniqueName: \"kubernetes.io/projected/c19692e0-e586-4506-a9e7-28195138b290-kube-api-access-s9thn\") pod \"nova-scheduler-0\" (UID: \"c19692e0-e586-4506-a9e7-28195138b290\") " pod="openstack/nova-scheduler-0" Nov 24 15:49:34 crc kubenswrapper[4790]: I1124 15:49:34.026795 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 15:49:34 crc kubenswrapper[4790]: I1124 15:49:34.333012 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a3a068f-92a8-43cf-83dc-8775c1d9ee84" path="/var/lib/kubelet/pods/2a3a068f-92a8-43cf-83dc-8775c1d9ee84/volumes" Nov 24 15:49:34 crc kubenswrapper[4790]: I1124 15:49:34.529202 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 15:49:34 crc kubenswrapper[4790]: W1124 15:49:34.530903 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc19692e0_e586_4506_a9e7_28195138b290.slice/crio-d95bdf9eb12d2484e17a6dda19281eb8c155a9dadff1e4c76adf9d9fca3d1c07 WatchSource:0}: Error finding container d95bdf9eb12d2484e17a6dda19281eb8c155a9dadff1e4c76adf9d9fca3d1c07: Status 404 returned error can't find the container with id d95bdf9eb12d2484e17a6dda19281eb8c155a9dadff1e4c76adf9d9fca3d1c07 Nov 24 15:49:34 crc kubenswrapper[4790]: I1124 15:49:34.558539 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c19692e0-e586-4506-a9e7-28195138b290","Type":"ContainerStarted","Data":"d95bdf9eb12d2484e17a6dda19281eb8c155a9dadff1e4c76adf9d9fca3d1c07"} Nov 24 15:49:34 crc kubenswrapper[4790]: I1124 15:49:34.566986 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swvz2" event={"ID":"1f246eee-cba0-4bc5-9d93-bb318e763cc1","Type":"ContainerStarted","Data":"703b983ce194b0f5172b0eca1b7e8d8cccb49b9f1d371ac95521ff76bf19730c"} Nov 24 15:49:34 crc kubenswrapper[4790]: I1124 15:49:34.592560 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-swvz2" podStartSLOduration=3.074452053 podStartE2EDuration="5.59254106s" podCreationTimestamp="2025-11-24 15:49:29 +0000 UTC" firstStartedPulling="2025-11-24 15:49:31.415924896 +0000 UTC m=+9419.795818558" lastFinishedPulling="2025-11-24 15:49:33.934013893 +0000 UTC m=+9422.313907565" observedRunningTime="2025-11-24 15:49:34.587707652 +0000 UTC m=+9422.967601314" watchObservedRunningTime="2025-11-24 15:49:34.59254106 +0000 UTC m=+9422.972434722" Nov 24 15:49:35 crc kubenswrapper[4790]: I1124 15:49:35.575129 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c19692e0-e586-4506-a9e7-28195138b290","Type":"ContainerStarted","Data":"31f7a17887c9b278bb731693fda959f6c3256bf297d364a2961d9a5adcae4524"} Nov 24 15:49:36 crc kubenswrapper[4790]: I1124 15:49:36.838780 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gh2nm" Nov 24 15:49:36 crc kubenswrapper[4790]: I1124 15:49:36.879991 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.879959228 podStartE2EDuration="3.879959228s" podCreationTimestamp="2025-11-24 15:49:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 15:49:35.602090842 +0000 UTC m=+9423.981984504" watchObservedRunningTime="2025-11-24 15:49:36.879959228 +0000 UTC m=+9425.259852920" Nov 24 15:49:36 crc kubenswrapper[4790]: I1124 15:49:36.906919 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gh2nm" Nov 24 15:49:36 crc kubenswrapper[4790]: I1124 15:49:36.946744 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 15:49:36 crc kubenswrapper[4790]: I1124 15:49:36.946793 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 15:49:37 crc kubenswrapper[4790]: I1124 15:49:37.973289 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gh2nm"] Nov 24 15:49:38 crc kubenswrapper[4790]: I1124 15:49:38.613137 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gh2nm" podUID="22c66da7-9f10-44d6-8e01-aab24fe93c78" containerName="registry-server" containerID="cri-o://61b0991ba597d95a922a6d5966b343e4d15528cd4232c199890d392834145037" gracePeriod=2 Nov 24 15:49:38 crc kubenswrapper[4790]: I1124 15:49:38.822089 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 24 15:49:39 crc kubenswrapper[4790]: I1124 15:49:39.027827 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 15:49:39 crc kubenswrapper[4790]: I1124 15:49:39.250262 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gh2nm" Nov 24 15:49:39 crc kubenswrapper[4790]: I1124 15:49:39.347611 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tll7t\" (UniqueName: \"kubernetes.io/projected/22c66da7-9f10-44d6-8e01-aab24fe93c78-kube-api-access-tll7t\") pod \"22c66da7-9f10-44d6-8e01-aab24fe93c78\" (UID: \"22c66da7-9f10-44d6-8e01-aab24fe93c78\") " Nov 24 15:49:39 crc kubenswrapper[4790]: I1124 15:49:39.347793 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22c66da7-9f10-44d6-8e01-aab24fe93c78-utilities\") pod \"22c66da7-9f10-44d6-8e01-aab24fe93c78\" (UID: \"22c66da7-9f10-44d6-8e01-aab24fe93c78\") " Nov 24 15:49:39 crc kubenswrapper[4790]: I1124 15:49:39.347854 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22c66da7-9f10-44d6-8e01-aab24fe93c78-catalog-content\") pod \"22c66da7-9f10-44d6-8e01-aab24fe93c78\" (UID: \"22c66da7-9f10-44d6-8e01-aab24fe93c78\") " Nov 24 15:49:39 crc kubenswrapper[4790]: I1124 15:49:39.348988 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22c66da7-9f10-44d6-8e01-aab24fe93c78-utilities" (OuterVolumeSpecName: "utilities") pod "22c66da7-9f10-44d6-8e01-aab24fe93c78" (UID: "22c66da7-9f10-44d6-8e01-aab24fe93c78"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:49:39 crc kubenswrapper[4790]: I1124 15:49:39.354147 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c66da7-9f10-44d6-8e01-aab24fe93c78-kube-api-access-tll7t" (OuterVolumeSpecName: "kube-api-access-tll7t") pod "22c66da7-9f10-44d6-8e01-aab24fe93c78" (UID: "22c66da7-9f10-44d6-8e01-aab24fe93c78"). InnerVolumeSpecName "kube-api-access-tll7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:49:39 crc kubenswrapper[4790]: I1124 15:49:39.443281 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22c66da7-9f10-44d6-8e01-aab24fe93c78-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "22c66da7-9f10-44d6-8e01-aab24fe93c78" (UID: "22c66da7-9f10-44d6-8e01-aab24fe93c78"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:49:39 crc kubenswrapper[4790]: I1124 15:49:39.450711 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tll7t\" (UniqueName: \"kubernetes.io/projected/22c66da7-9f10-44d6-8e01-aab24fe93c78-kube-api-access-tll7t\") on node \"crc\" DevicePath \"\"" Nov 24 15:49:39 crc kubenswrapper[4790]: I1124 15:49:39.450745 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22c66da7-9f10-44d6-8e01-aab24fe93c78-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 15:49:39 crc kubenswrapper[4790]: I1124 15:49:39.450758 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22c66da7-9f10-44d6-8e01-aab24fe93c78-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 15:49:39 crc kubenswrapper[4790]: I1124 15:49:39.627377 4790 generic.go:334] "Generic (PLEG): container finished" podID="22c66da7-9f10-44d6-8e01-aab24fe93c78" containerID="61b0991ba597d95a922a6d5966b343e4d15528cd4232c199890d392834145037" exitCode=0 Nov 24 15:49:39 crc kubenswrapper[4790]: I1124 15:49:39.627423 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gh2nm" event={"ID":"22c66da7-9f10-44d6-8e01-aab24fe93c78","Type":"ContainerDied","Data":"61b0991ba597d95a922a6d5966b343e4d15528cd4232c199890d392834145037"} Nov 24 15:49:39 crc kubenswrapper[4790]: I1124 15:49:39.627455 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gh2nm" event={"ID":"22c66da7-9f10-44d6-8e01-aab24fe93c78","Type":"ContainerDied","Data":"5aa209a009dde4964f488d3dc2f660691e4f6949d367688b4e4e509a50fc8114"} Nov 24 15:49:39 crc kubenswrapper[4790]: I1124 15:49:39.627483 4790 scope.go:117] "RemoveContainer" containerID="61b0991ba597d95a922a6d5966b343e4d15528cd4232c199890d392834145037" Nov 24 15:49:39 crc kubenswrapper[4790]: I1124 15:49:39.627582 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gh2nm" Nov 24 15:49:39 crc kubenswrapper[4790]: I1124 15:49:39.650111 4790 scope.go:117] "RemoveContainer" containerID="07a9a4602ba774a72a658d693f4cdb2dd9d2086cc2baabbb8bfefd18e5c96bde" Nov 24 15:49:39 crc kubenswrapper[4790]: I1124 15:49:39.678695 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gh2nm"] Nov 24 15:49:39 crc kubenswrapper[4790]: I1124 15:49:39.690143 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gh2nm"] Nov 24 15:49:40 crc kubenswrapper[4790]: I1124 15:49:40.112208 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-swvz2" Nov 24 15:49:40 crc kubenswrapper[4790]: I1124 15:49:40.112323 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-swvz2" Nov 24 15:49:40 crc kubenswrapper[4790]: I1124 15:49:40.339809 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c66da7-9f10-44d6-8e01-aab24fe93c78" path="/var/lib/kubelet/pods/22c66da7-9f10-44d6-8e01-aab24fe93c78/volumes" Nov 24 15:49:40 crc kubenswrapper[4790]: I1124 15:49:40.417344 4790 scope.go:117] "RemoveContainer" containerID="cb84d2e7ada764bf3206ff30b1d143ba21ccb3595b8cc4622eb3e3ab624e6cb8" Nov 24 15:49:40 crc kubenswrapper[4790]: I1124 15:49:40.519545 4790 scope.go:117] "RemoveContainer" containerID="61b0991ba597d95a922a6d5966b343e4d15528cd4232c199890d392834145037" Nov 24 15:49:40 crc kubenswrapper[4790]: E1124 15:49:40.520303 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61b0991ba597d95a922a6d5966b343e4d15528cd4232c199890d392834145037\": container with ID starting with 61b0991ba597d95a922a6d5966b343e4d15528cd4232c199890d392834145037 not found: ID does not exist" containerID="61b0991ba597d95a922a6d5966b343e4d15528cd4232c199890d392834145037" Nov 24 15:49:40 crc kubenswrapper[4790]: I1124 15:49:40.520387 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61b0991ba597d95a922a6d5966b343e4d15528cd4232c199890d392834145037"} err="failed to get container status \"61b0991ba597d95a922a6d5966b343e4d15528cd4232c199890d392834145037\": rpc error: code = NotFound desc = could not find container \"61b0991ba597d95a922a6d5966b343e4d15528cd4232c199890d392834145037\": container with ID starting with 61b0991ba597d95a922a6d5966b343e4d15528cd4232c199890d392834145037 not found: ID does not exist" Nov 24 15:49:40 crc kubenswrapper[4790]: I1124 15:49:40.520524 4790 scope.go:117] "RemoveContainer" containerID="07a9a4602ba774a72a658d693f4cdb2dd9d2086cc2baabbb8bfefd18e5c96bde" Nov 24 15:49:40 crc kubenswrapper[4790]: E1124 15:49:40.522279 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07a9a4602ba774a72a658d693f4cdb2dd9d2086cc2baabbb8bfefd18e5c96bde\": container with ID starting with 07a9a4602ba774a72a658d693f4cdb2dd9d2086cc2baabbb8bfefd18e5c96bde not found: ID does not exist" containerID="07a9a4602ba774a72a658d693f4cdb2dd9d2086cc2baabbb8bfefd18e5c96bde" Nov 24 15:49:40 crc kubenswrapper[4790]: I1124 15:49:40.524212 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07a9a4602ba774a72a658d693f4cdb2dd9d2086cc2baabbb8bfefd18e5c96bde"} err="failed to get container status \"07a9a4602ba774a72a658d693f4cdb2dd9d2086cc2baabbb8bfefd18e5c96bde\": rpc error: code = NotFound desc = could not find container \"07a9a4602ba774a72a658d693f4cdb2dd9d2086cc2baabbb8bfefd18e5c96bde\": container with ID starting with 07a9a4602ba774a72a658d693f4cdb2dd9d2086cc2baabbb8bfefd18e5c96bde not found: ID does not exist" Nov 24 15:49:40 crc kubenswrapper[4790]: I1124 15:49:40.524261 4790 scope.go:117] "RemoveContainer" containerID="cb84d2e7ada764bf3206ff30b1d143ba21ccb3595b8cc4622eb3e3ab624e6cb8" Nov 24 15:49:40 crc kubenswrapper[4790]: E1124 15:49:40.525478 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb84d2e7ada764bf3206ff30b1d143ba21ccb3595b8cc4622eb3e3ab624e6cb8\": container with ID starting with cb84d2e7ada764bf3206ff30b1d143ba21ccb3595b8cc4622eb3e3ab624e6cb8 not found: ID does not exist" containerID="cb84d2e7ada764bf3206ff30b1d143ba21ccb3595b8cc4622eb3e3ab624e6cb8" Nov 24 15:49:40 crc kubenswrapper[4790]: I1124 15:49:40.525548 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb84d2e7ada764bf3206ff30b1d143ba21ccb3595b8cc4622eb3e3ab624e6cb8"} err="failed to get container status \"cb84d2e7ada764bf3206ff30b1d143ba21ccb3595b8cc4622eb3e3ab624e6cb8\": rpc error: code = NotFound desc = could not find container \"cb84d2e7ada764bf3206ff30b1d143ba21ccb3595b8cc4622eb3e3ab624e6cb8\": container with ID starting with cb84d2e7ada764bf3206ff30b1d143ba21ccb3595b8cc4622eb3e3ab624e6cb8 not found: ID does not exist" Nov 24 15:49:40 crc kubenswrapper[4790]: I1124 15:49:40.661476 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-swvz2" Nov 24 15:49:40 crc kubenswrapper[4790]: I1124 15:49:40.711688 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-swvz2" Nov 24 15:49:40 crc kubenswrapper[4790]: I1124 15:49:40.798676 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 24 15:49:41 crc kubenswrapper[4790]: I1124 15:49:41.884371 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 15:49:41 crc kubenswrapper[4790]: I1124 15:49:41.884855 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 15:49:41 crc kubenswrapper[4790]: I1124 15:49:41.947590 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 15:49:41 crc kubenswrapper[4790]: I1124 15:49:41.947660 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 15:49:42 crc kubenswrapper[4790]: I1124 15:49:42.373801 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-swvz2"] Nov 24 15:49:42 crc kubenswrapper[4790]: I1124 15:49:42.677970 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-swvz2" podUID="1f246eee-cba0-4bc5-9d93-bb318e763cc1" containerName="registry-server" containerID="cri-o://703b983ce194b0f5172b0eca1b7e8d8cccb49b9f1d371ac95521ff76bf19730c" gracePeriod=2 Nov 24 15:49:42 crc kubenswrapper[4790]: I1124 15:49:42.967119 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3d36ae33-dbe8-4323-8bc6-83ca5dcf40ea" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.196:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 15:49:42 crc kubenswrapper[4790]: I1124 15:49:42.967230 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3d36ae33-dbe8-4323-8bc6-83ca5dcf40ea" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.196:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 15:49:43 crc kubenswrapper[4790]: I1124 15:49:43.052166 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="9a7f156b-e253-4f3b-b4ce-5c3c226ccfcd" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.197:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 15:49:43 crc kubenswrapper[4790]: I1124 15:49:43.053576 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="9a7f156b-e253-4f3b-b4ce-5c3c226ccfcd" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.197:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 15:49:43 crc kubenswrapper[4790]: I1124 15:49:43.227190 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-swvz2" Nov 24 15:49:43 crc kubenswrapper[4790]: I1124 15:49:43.372270 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f246eee-cba0-4bc5-9d93-bb318e763cc1-utilities\") pod \"1f246eee-cba0-4bc5-9d93-bb318e763cc1\" (UID: \"1f246eee-cba0-4bc5-9d93-bb318e763cc1\") " Nov 24 15:49:43 crc kubenswrapper[4790]: I1124 15:49:43.375221 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f246eee-cba0-4bc5-9d93-bb318e763cc1-utilities" (OuterVolumeSpecName: "utilities") pod "1f246eee-cba0-4bc5-9d93-bb318e763cc1" (UID: "1f246eee-cba0-4bc5-9d93-bb318e763cc1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:49:43 crc kubenswrapper[4790]: I1124 15:49:43.375593 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f246eee-cba0-4bc5-9d93-bb318e763cc1-catalog-content\") pod \"1f246eee-cba0-4bc5-9d93-bb318e763cc1\" (UID: \"1f246eee-cba0-4bc5-9d93-bb318e763cc1\") " Nov 24 15:49:43 crc kubenswrapper[4790]: I1124 15:49:43.379442 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s958k\" (UniqueName: \"kubernetes.io/projected/1f246eee-cba0-4bc5-9d93-bb318e763cc1-kube-api-access-s958k\") pod \"1f246eee-cba0-4bc5-9d93-bb318e763cc1\" (UID: \"1f246eee-cba0-4bc5-9d93-bb318e763cc1\") " Nov 24 15:49:43 crc kubenswrapper[4790]: I1124 15:49:43.384596 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f246eee-cba0-4bc5-9d93-bb318e763cc1-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 15:49:43 crc kubenswrapper[4790]: I1124 15:49:43.389065 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f246eee-cba0-4bc5-9d93-bb318e763cc1-kube-api-access-s958k" (OuterVolumeSpecName: "kube-api-access-s958k") pod "1f246eee-cba0-4bc5-9d93-bb318e763cc1" (UID: "1f246eee-cba0-4bc5-9d93-bb318e763cc1"). InnerVolumeSpecName "kube-api-access-s958k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:49:43 crc kubenswrapper[4790]: I1124 15:49:43.462098 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f246eee-cba0-4bc5-9d93-bb318e763cc1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1f246eee-cba0-4bc5-9d93-bb318e763cc1" (UID: "1f246eee-cba0-4bc5-9d93-bb318e763cc1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:49:43 crc kubenswrapper[4790]: I1124 15:49:43.487667 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f246eee-cba0-4bc5-9d93-bb318e763cc1-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 15:49:43 crc kubenswrapper[4790]: I1124 15:49:43.487696 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s958k\" (UniqueName: \"kubernetes.io/projected/1f246eee-cba0-4bc5-9d93-bb318e763cc1-kube-api-access-s958k\") on node \"crc\" DevicePath \"\"" Nov 24 15:49:43 crc kubenswrapper[4790]: I1124 15:49:43.691696 4790 generic.go:334] "Generic (PLEG): container finished" podID="1f246eee-cba0-4bc5-9d93-bb318e763cc1" containerID="703b983ce194b0f5172b0eca1b7e8d8cccb49b9f1d371ac95521ff76bf19730c" exitCode=0 Nov 24 15:49:43 crc kubenswrapper[4790]: I1124 15:49:43.691939 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swvz2" event={"ID":"1f246eee-cba0-4bc5-9d93-bb318e763cc1","Type":"ContainerDied","Data":"703b983ce194b0f5172b0eca1b7e8d8cccb49b9f1d371ac95521ff76bf19730c"} Nov 24 15:49:43 crc kubenswrapper[4790]: I1124 15:49:43.691966 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swvz2" event={"ID":"1f246eee-cba0-4bc5-9d93-bb318e763cc1","Type":"ContainerDied","Data":"c70518138d727dc1e39d95d267d35d39708f64cc2df3bd41a8670190175bfdf5"} Nov 24 15:49:43 crc kubenswrapper[4790]: I1124 15:49:43.691983 4790 scope.go:117] "RemoveContainer" containerID="703b983ce194b0f5172b0eca1b7e8d8cccb49b9f1d371ac95521ff76bf19730c" Nov 24 15:49:43 crc kubenswrapper[4790]: I1124 15:49:43.692117 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-swvz2" Nov 24 15:49:43 crc kubenswrapper[4790]: I1124 15:49:43.725010 4790 scope.go:117] "RemoveContainer" containerID="005b8586bcee4e28a61fb6aa607fb5a0099d605d0635aa8013821f11f59baf16" Nov 24 15:49:43 crc kubenswrapper[4790]: I1124 15:49:43.734059 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-swvz2"] Nov 24 15:49:43 crc kubenswrapper[4790]: I1124 15:49:43.745842 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-swvz2"] Nov 24 15:49:43 crc kubenswrapper[4790]: I1124 15:49:43.756687 4790 scope.go:117] "RemoveContainer" containerID="0b8e941eb590d225ba2bacda6ee356b06b2afcebe831e7eddb16a05bd8770b0d" Nov 24 15:49:43 crc kubenswrapper[4790]: I1124 15:49:43.806919 4790 scope.go:117] "RemoveContainer" containerID="703b983ce194b0f5172b0eca1b7e8d8cccb49b9f1d371ac95521ff76bf19730c" Nov 24 15:49:43 crc kubenswrapper[4790]: E1124 15:49:43.807813 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"703b983ce194b0f5172b0eca1b7e8d8cccb49b9f1d371ac95521ff76bf19730c\": container with ID starting with 703b983ce194b0f5172b0eca1b7e8d8cccb49b9f1d371ac95521ff76bf19730c not found: ID does not exist" containerID="703b983ce194b0f5172b0eca1b7e8d8cccb49b9f1d371ac95521ff76bf19730c" Nov 24 15:49:43 crc kubenswrapper[4790]: I1124 15:49:43.807868 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"703b983ce194b0f5172b0eca1b7e8d8cccb49b9f1d371ac95521ff76bf19730c"} err="failed to get container status \"703b983ce194b0f5172b0eca1b7e8d8cccb49b9f1d371ac95521ff76bf19730c\": rpc error: code = NotFound desc = could not find container \"703b983ce194b0f5172b0eca1b7e8d8cccb49b9f1d371ac95521ff76bf19730c\": container with ID starting with 703b983ce194b0f5172b0eca1b7e8d8cccb49b9f1d371ac95521ff76bf19730c not found: ID does not exist" Nov 24 15:49:43 crc kubenswrapper[4790]: I1124 15:49:43.807905 4790 scope.go:117] "RemoveContainer" containerID="005b8586bcee4e28a61fb6aa607fb5a0099d605d0635aa8013821f11f59baf16" Nov 24 15:49:43 crc kubenswrapper[4790]: E1124 15:49:43.808505 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"005b8586bcee4e28a61fb6aa607fb5a0099d605d0635aa8013821f11f59baf16\": container with ID starting with 005b8586bcee4e28a61fb6aa607fb5a0099d605d0635aa8013821f11f59baf16 not found: ID does not exist" containerID="005b8586bcee4e28a61fb6aa607fb5a0099d605d0635aa8013821f11f59baf16" Nov 24 15:49:43 crc kubenswrapper[4790]: I1124 15:49:43.808535 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"005b8586bcee4e28a61fb6aa607fb5a0099d605d0635aa8013821f11f59baf16"} err="failed to get container status \"005b8586bcee4e28a61fb6aa607fb5a0099d605d0635aa8013821f11f59baf16\": rpc error: code = NotFound desc = could not find container \"005b8586bcee4e28a61fb6aa607fb5a0099d605d0635aa8013821f11f59baf16\": container with ID starting with 005b8586bcee4e28a61fb6aa607fb5a0099d605d0635aa8013821f11f59baf16 not found: ID does not exist" Nov 24 15:49:43 crc kubenswrapper[4790]: I1124 15:49:43.808550 4790 scope.go:117] "RemoveContainer" containerID="0b8e941eb590d225ba2bacda6ee356b06b2afcebe831e7eddb16a05bd8770b0d" Nov 24 15:49:43 crc kubenswrapper[4790]: E1124 15:49:43.808971 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b8e941eb590d225ba2bacda6ee356b06b2afcebe831e7eddb16a05bd8770b0d\": container with ID starting with 0b8e941eb590d225ba2bacda6ee356b06b2afcebe831e7eddb16a05bd8770b0d not found: ID does not exist" containerID="0b8e941eb590d225ba2bacda6ee356b06b2afcebe831e7eddb16a05bd8770b0d" Nov 24 15:49:43 crc kubenswrapper[4790]: I1124 15:49:43.809018 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b8e941eb590d225ba2bacda6ee356b06b2afcebe831e7eddb16a05bd8770b0d"} err="failed to get container status \"0b8e941eb590d225ba2bacda6ee356b06b2afcebe831e7eddb16a05bd8770b0d\": rpc error: code = NotFound desc = could not find container \"0b8e941eb590d225ba2bacda6ee356b06b2afcebe831e7eddb16a05bd8770b0d\": container with ID starting with 0b8e941eb590d225ba2bacda6ee356b06b2afcebe831e7eddb16a05bd8770b0d not found: ID does not exist" Nov 24 15:49:44 crc kubenswrapper[4790]: I1124 15:49:44.027191 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 15:49:44 crc kubenswrapper[4790]: I1124 15:49:44.087818 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 15:49:44 crc kubenswrapper[4790]: I1124 15:49:44.334331 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f246eee-cba0-4bc5-9d93-bb318e763cc1" path="/var/lib/kubelet/pods/1f246eee-cba0-4bc5-9d93-bb318e763cc1/volumes" Nov 24 15:49:44 crc kubenswrapper[4790]: I1124 15:49:44.756524 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 15:49:51 crc kubenswrapper[4790]: I1124 15:49:51.889104 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 15:49:51 crc kubenswrapper[4790]: I1124 15:49:51.891648 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 15:49:51 crc kubenswrapper[4790]: I1124 15:49:51.895713 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 15:49:51 crc kubenswrapper[4790]: I1124 15:49:51.897965 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 15:49:51 crc kubenswrapper[4790]: I1124 15:49:51.948914 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 15:49:51 crc kubenswrapper[4790]: I1124 15:49:51.949708 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 15:49:51 crc kubenswrapper[4790]: I1124 15:49:51.951958 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 15:49:52 crc kubenswrapper[4790]: I1124 15:49:52.809468 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 15:49:52 crc kubenswrapper[4790]: I1124 15:49:52.812877 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 15:49:52 crc kubenswrapper[4790]: I1124 15:49:52.814617 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 15:49:53 crc kubenswrapper[4790]: I1124 15:49:53.900870 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf"] Nov 24 15:49:53 crc kubenswrapper[4790]: E1124 15:49:53.901634 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22c66da7-9f10-44d6-8e01-aab24fe93c78" containerName="registry-server" Nov 24 15:49:53 crc kubenswrapper[4790]: I1124 15:49:53.901649 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="22c66da7-9f10-44d6-8e01-aab24fe93c78" containerName="registry-server" Nov 24 15:49:53 crc kubenswrapper[4790]: E1124 15:49:53.901660 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f246eee-cba0-4bc5-9d93-bb318e763cc1" containerName="extract-content" Nov 24 15:49:53 crc kubenswrapper[4790]: I1124 15:49:53.901668 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f246eee-cba0-4bc5-9d93-bb318e763cc1" containerName="extract-content" Nov 24 15:49:53 crc kubenswrapper[4790]: E1124 15:49:53.901701 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f246eee-cba0-4bc5-9d93-bb318e763cc1" containerName="registry-server" Nov 24 15:49:53 crc kubenswrapper[4790]: I1124 15:49:53.901709 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f246eee-cba0-4bc5-9d93-bb318e763cc1" containerName="registry-server" Nov 24 15:49:53 crc kubenswrapper[4790]: E1124 15:49:53.901733 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22c66da7-9f10-44d6-8e01-aab24fe93c78" containerName="extract-content" Nov 24 15:49:53 crc kubenswrapper[4790]: I1124 15:49:53.901740 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="22c66da7-9f10-44d6-8e01-aab24fe93c78" containerName="extract-content" Nov 24 15:49:53 crc kubenswrapper[4790]: E1124 15:49:53.901767 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22c66da7-9f10-44d6-8e01-aab24fe93c78" containerName="extract-utilities" Nov 24 15:49:53 crc kubenswrapper[4790]: I1124 15:49:53.901775 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="22c66da7-9f10-44d6-8e01-aab24fe93c78" containerName="extract-utilities" Nov 24 15:49:53 crc kubenswrapper[4790]: E1124 15:49:53.901790 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f246eee-cba0-4bc5-9d93-bb318e763cc1" containerName="extract-utilities" Nov 24 15:49:53 crc kubenswrapper[4790]: I1124 15:49:53.901798 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f246eee-cba0-4bc5-9d93-bb318e763cc1" containerName="extract-utilities" Nov 24 15:49:53 crc kubenswrapper[4790]: I1124 15:49:53.902055 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f246eee-cba0-4bc5-9d93-bb318e763cc1" containerName="registry-server" Nov 24 15:49:53 crc kubenswrapper[4790]: I1124 15:49:53.902089 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="22c66da7-9f10-44d6-8e01-aab24fe93c78" containerName="registry-server" Nov 24 15:49:53 crc kubenswrapper[4790]: I1124 15:49:53.903070 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:53 crc kubenswrapper[4790]: I1124 15:49:53.908241 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 15:49:53 crc kubenswrapper[4790]: I1124 15:49:53.908933 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 24 15:49:53 crc kubenswrapper[4790]: I1124 15:49:53.909153 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fc497" Nov 24 15:49:53 crc kubenswrapper[4790]: I1124 15:49:53.910101 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 24 15:49:53 crc kubenswrapper[4790]: I1124 15:49:53.910233 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 15:49:53 crc kubenswrapper[4790]: I1124 15:49:53.910364 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Nov 24 15:49:53 crc kubenswrapper[4790]: I1124 15:49:53.911231 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 15:49:53 crc kubenswrapper[4790]: I1124 15:49:53.920732 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf"] Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.067237 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.067297 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.067332 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.067409 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.067543 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.067612 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxghh\" (UniqueName: \"kubernetes.io/projected/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-kube-api-access-dxghh\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.067633 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.067778 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.067862 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.068012 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.068098 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.170320 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.170373 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxghh\" (UniqueName: \"kubernetes.io/projected/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-kube-api-access-dxghh\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.170402 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.170491 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.170540 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.170762 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.171298 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.171360 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.171393 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.171426 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.171772 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.172563 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.172838 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.175574 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.176411 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.176497 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.176760 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.177054 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.177529 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.179438 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.183815 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.189415 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxghh\" (UniqueName: \"kubernetes.io/projected/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-kube-api-access-dxghh\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.263437 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:49:54 crc kubenswrapper[4790]: I1124 15:49:54.824878 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf"] Nov 24 15:49:55 crc kubenswrapper[4790]: I1124 15:49:55.849997 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" event={"ID":"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4","Type":"ContainerStarted","Data":"8170bfa938899480f3600d7d108b0b972f18c16920d7d850e0cd9706106f4d66"} Nov 24 15:49:55 crc kubenswrapper[4790]: I1124 15:49:55.850457 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" event={"ID":"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4","Type":"ContainerStarted","Data":"e1625ade383c73f8a5bf0f1bd7d600635a01cada557b61315f1e423a28e6578a"} Nov 24 15:49:55 crc kubenswrapper[4790]: I1124 15:49:55.874598 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" podStartSLOduration=2.71402077 podStartE2EDuration="2.87457898s" podCreationTimestamp="2025-11-24 15:49:53 +0000 UTC" firstStartedPulling="2025-11-24 15:49:54.833242643 +0000 UTC m=+9443.213136315" lastFinishedPulling="2025-11-24 15:49:54.993800823 +0000 UTC m=+9443.373694525" observedRunningTime="2025-11-24 15:49:55.868574109 +0000 UTC m=+9444.248467771" watchObservedRunningTime="2025-11-24 15:49:55.87457898 +0000 UTC m=+9444.254472642" Nov 24 15:51:43 crc kubenswrapper[4790]: I1124 15:51:43.938953 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:51:43 crc kubenswrapper[4790]: I1124 15:51:43.939677 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:52:13 crc kubenswrapper[4790]: I1124 15:52:13.938795 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:52:13 crc kubenswrapper[4790]: I1124 15:52:13.939413 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:52:43 crc kubenswrapper[4790]: I1124 15:52:43.938655 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 15:52:43 crc kubenswrapper[4790]: I1124 15:52:43.939347 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 15:52:43 crc kubenswrapper[4790]: I1124 15:52:43.939417 4790 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 15:52:43 crc kubenswrapper[4790]: I1124 15:52:43.940390 4790 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f1e094cfe52dce391f24557f040b619b252fb0fb27545cacdfae7450eb6c06b2"} pod="openshift-machine-config-operator/machine-config-daemon-xz49t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 15:52:43 crc kubenswrapper[4790]: I1124 15:52:43.940457 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" containerID="cri-o://f1e094cfe52dce391f24557f040b619b252fb0fb27545cacdfae7450eb6c06b2" gracePeriod=600 Nov 24 15:52:44 crc kubenswrapper[4790]: E1124 15:52:44.077460 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:52:44 crc kubenswrapper[4790]: I1124 15:52:44.382073 4790 generic.go:334] "Generic (PLEG): container finished" podID="0d73b133-48f0-455f-8f6a-742e633f631a" containerID="f1e094cfe52dce391f24557f040b619b252fb0fb27545cacdfae7450eb6c06b2" exitCode=0 Nov 24 15:52:44 crc kubenswrapper[4790]: I1124 15:52:44.382142 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerDied","Data":"f1e094cfe52dce391f24557f040b619b252fb0fb27545cacdfae7450eb6c06b2"} Nov 24 15:52:44 crc kubenswrapper[4790]: I1124 15:52:44.382409 4790 scope.go:117] "RemoveContainer" containerID="481705db3593f4e5648f79ea2801d3ecc5478bd6c6a34f70e103d4dccb0eeb10" Nov 24 15:52:44 crc kubenswrapper[4790]: I1124 15:52:44.382801 4790 scope.go:117] "RemoveContainer" containerID="f1e094cfe52dce391f24557f040b619b252fb0fb27545cacdfae7450eb6c06b2" Nov 24 15:52:44 crc kubenswrapper[4790]: E1124 15:52:44.383172 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:52:55 crc kubenswrapper[4790]: I1124 15:52:55.314589 4790 scope.go:117] "RemoveContainer" containerID="f1e094cfe52dce391f24557f040b619b252fb0fb27545cacdfae7450eb6c06b2" Nov 24 15:52:55 crc kubenswrapper[4790]: E1124 15:52:55.315463 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:52:59 crc kubenswrapper[4790]: I1124 15:52:59.213142 4790 trace.go:236] Trace[1810487361]: "Calculate volume metrics of ovndbcluster-sb-etc-ovn for pod openstack/ovsdbserver-sb-0" (24-Nov-2025 15:52:58.083) (total time: 1129ms): Nov 24 15:52:59 crc kubenswrapper[4790]: Trace[1810487361]: [1.129662998s] [1.129662998s] END Nov 24 15:53:10 crc kubenswrapper[4790]: I1124 15:53:10.315238 4790 scope.go:117] "RemoveContainer" containerID="f1e094cfe52dce391f24557f040b619b252fb0fb27545cacdfae7450eb6c06b2" Nov 24 15:53:10 crc kubenswrapper[4790]: E1124 15:53:10.316337 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:53:24 crc kubenswrapper[4790]: I1124 15:53:24.314967 4790 scope.go:117] "RemoveContainer" containerID="f1e094cfe52dce391f24557f040b619b252fb0fb27545cacdfae7450eb6c06b2" Nov 24 15:53:24 crc kubenswrapper[4790]: E1124 15:53:24.316268 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:53:36 crc kubenswrapper[4790]: I1124 15:53:36.314873 4790 scope.go:117] "RemoveContainer" containerID="f1e094cfe52dce391f24557f040b619b252fb0fb27545cacdfae7450eb6c06b2" Nov 24 15:53:36 crc kubenswrapper[4790]: E1124 15:53:36.317287 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:53:47 crc kubenswrapper[4790]: I1124 15:53:47.316498 4790 scope.go:117] "RemoveContainer" containerID="f1e094cfe52dce391f24557f040b619b252fb0fb27545cacdfae7450eb6c06b2" Nov 24 15:53:47 crc kubenswrapper[4790]: E1124 15:53:47.318168 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:53:59 crc kubenswrapper[4790]: I1124 15:53:59.365462 4790 scope.go:117] "RemoveContainer" containerID="f1e094cfe52dce391f24557f040b619b252fb0fb27545cacdfae7450eb6c06b2" Nov 24 15:53:59 crc kubenswrapper[4790]: E1124 15:53:59.366422 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:54:13 crc kubenswrapper[4790]: I1124 15:54:13.314924 4790 scope.go:117] "RemoveContainer" containerID="f1e094cfe52dce391f24557f040b619b252fb0fb27545cacdfae7450eb6c06b2" Nov 24 15:54:13 crc kubenswrapper[4790]: E1124 15:54:13.316282 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:54:28 crc kubenswrapper[4790]: I1124 15:54:28.315399 4790 scope.go:117] "RemoveContainer" containerID="f1e094cfe52dce391f24557f040b619b252fb0fb27545cacdfae7450eb6c06b2" Nov 24 15:54:28 crc kubenswrapper[4790]: E1124 15:54:28.318991 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:54:34 crc kubenswrapper[4790]: I1124 15:54:34.112017 4790 trace.go:236] Trace[1524509742]: "Calculate volume metrics of ovndbcluster-sb-etc-ovn for pod openstack/ovsdbserver-sb-2" (24-Nov-2025 15:54:33.099) (total time: 1012ms): Nov 24 15:54:34 crc kubenswrapper[4790]: Trace[1524509742]: [1.01265435s] [1.01265435s] END Nov 24 15:54:42 crc kubenswrapper[4790]: I1124 15:54:42.323648 4790 scope.go:117] "RemoveContainer" containerID="f1e094cfe52dce391f24557f040b619b252fb0fb27545cacdfae7450eb6c06b2" Nov 24 15:54:42 crc kubenswrapper[4790]: E1124 15:54:42.324420 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:54:56 crc kubenswrapper[4790]: I1124 15:54:56.315184 4790 scope.go:117] "RemoveContainer" containerID="f1e094cfe52dce391f24557f040b619b252fb0fb27545cacdfae7450eb6c06b2" Nov 24 15:54:56 crc kubenswrapper[4790]: E1124 15:54:56.316216 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:55:07 crc kubenswrapper[4790]: I1124 15:55:07.315368 4790 scope.go:117] "RemoveContainer" containerID="f1e094cfe52dce391f24557f040b619b252fb0fb27545cacdfae7450eb6c06b2" Nov 24 15:55:07 crc kubenswrapper[4790]: E1124 15:55:07.316455 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:55:20 crc kubenswrapper[4790]: I1124 15:55:20.315317 4790 scope.go:117] "RemoveContainer" containerID="f1e094cfe52dce391f24557f040b619b252fb0fb27545cacdfae7450eb6c06b2" Nov 24 15:55:20 crc kubenswrapper[4790]: E1124 15:55:20.315979 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:55:32 crc kubenswrapper[4790]: I1124 15:55:32.335991 4790 scope.go:117] "RemoveContainer" containerID="f1e094cfe52dce391f24557f040b619b252fb0fb27545cacdfae7450eb6c06b2" Nov 24 15:55:32 crc kubenswrapper[4790]: E1124 15:55:32.337219 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:55:38 crc kubenswrapper[4790]: I1124 15:55:38.675267 4790 generic.go:334] "Generic (PLEG): container finished" podID="9e561ed1-5cd7-4a50-8629-bf48c3a21ac4" containerID="8170bfa938899480f3600d7d108b0b972f18c16920d7d850e0cd9706106f4d66" exitCode=0 Nov 24 15:55:38 crc kubenswrapper[4790]: I1124 15:55:38.675347 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" event={"ID":"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4","Type":"ContainerDied","Data":"8170bfa938899480f3600d7d108b0b972f18c16920d7d850e0cd9706106f4d66"} Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.350803 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.371102 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-q7fss"] Nov 24 15:55:40 crc kubenswrapper[4790]: E1124 15:55:40.371645 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e561ed1-5cd7-4a50-8629-bf48c3a21ac4" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.371661 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e561ed1-5cd7-4a50-8629-bf48c3a21ac4" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.371976 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e561ed1-5cd7-4a50-8629-bf48c3a21ac4" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.373744 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q7fss" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.385127 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q7fss"] Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.521520 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-cell1-compute-config-0\") pod \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.521602 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxghh\" (UniqueName: \"kubernetes.io/projected/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-kube-api-access-dxghh\") pod \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.521647 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-cell1-combined-ca-bundle\") pod \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.521696 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-cells-global-config-0\") pod \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.521741 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-inventory\") pod \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.521773 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-ssh-key\") pod \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.521817 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-cell1-compute-config-1\") pod \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.521921 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-migration-ssh-key-0\") pod \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.521995 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-migration-ssh-key-1\") pod \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.522041 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-cells-global-config-1\") pod \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.522068 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-ceph\") pod \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\" (UID: \"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4\") " Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.523211 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5e92738-452f-4a85-a6a7-278241c373f6-utilities\") pod \"certified-operators-q7fss\" (UID: \"e5e92738-452f-4a85-a6a7-278241c373f6\") " pod="openshift-marketplace/certified-operators-q7fss" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.523274 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5e92738-452f-4a85-a6a7-278241c373f6-catalog-content\") pod \"certified-operators-q7fss\" (UID: \"e5e92738-452f-4a85-a6a7-278241c373f6\") " pod="openshift-marketplace/certified-operators-q7fss" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.523335 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clmnh\" (UniqueName: \"kubernetes.io/projected/e5e92738-452f-4a85-a6a7-278241c373f6-kube-api-access-clmnh\") pod \"certified-operators-q7fss\" (UID: \"e5e92738-452f-4a85-a6a7-278241c373f6\") " pod="openshift-marketplace/certified-operators-q7fss" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.531324 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-ceph" (OuterVolumeSpecName: "ceph") pod "9e561ed1-5cd7-4a50-8629-bf48c3a21ac4" (UID: "9e561ed1-5cd7-4a50-8629-bf48c3a21ac4"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.531924 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-kube-api-access-dxghh" (OuterVolumeSpecName: "kube-api-access-dxghh") pod "9e561ed1-5cd7-4a50-8629-bf48c3a21ac4" (UID: "9e561ed1-5cd7-4a50-8629-bf48c3a21ac4"). InnerVolumeSpecName "kube-api-access-dxghh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.534218 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "9e561ed1-5cd7-4a50-8629-bf48c3a21ac4" (UID: "9e561ed1-5cd7-4a50-8629-bf48c3a21ac4"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.558319 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "9e561ed1-5cd7-4a50-8629-bf48c3a21ac4" (UID: "9e561ed1-5cd7-4a50-8629-bf48c3a21ac4"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.561662 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "9e561ed1-5cd7-4a50-8629-bf48c3a21ac4" (UID: "9e561ed1-5cd7-4a50-8629-bf48c3a21ac4"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.561843 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "9e561ed1-5cd7-4a50-8629-bf48c3a21ac4" (UID: "9e561ed1-5cd7-4a50-8629-bf48c3a21ac4"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.563826 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "9e561ed1-5cd7-4a50-8629-bf48c3a21ac4" (UID: "9e561ed1-5cd7-4a50-8629-bf48c3a21ac4"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.581019 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "9e561ed1-5cd7-4a50-8629-bf48c3a21ac4" (UID: "9e561ed1-5cd7-4a50-8629-bf48c3a21ac4"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.581283 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-inventory" (OuterVolumeSpecName: "inventory") pod "9e561ed1-5cd7-4a50-8629-bf48c3a21ac4" (UID: "9e561ed1-5cd7-4a50-8629-bf48c3a21ac4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.588405 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9e561ed1-5cd7-4a50-8629-bf48c3a21ac4" (UID: "9e561ed1-5cd7-4a50-8629-bf48c3a21ac4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.593697 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "9e561ed1-5cd7-4a50-8629-bf48c3a21ac4" (UID: "9e561ed1-5cd7-4a50-8629-bf48c3a21ac4"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.625609 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5e92738-452f-4a85-a6a7-278241c373f6-utilities\") pod \"certified-operators-q7fss\" (UID: \"e5e92738-452f-4a85-a6a7-278241c373f6\") " pod="openshift-marketplace/certified-operators-q7fss" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.625680 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5e92738-452f-4a85-a6a7-278241c373f6-catalog-content\") pod \"certified-operators-q7fss\" (UID: \"e5e92738-452f-4a85-a6a7-278241c373f6\") " pod="openshift-marketplace/certified-operators-q7fss" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.625732 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clmnh\" (UniqueName: \"kubernetes.io/projected/e5e92738-452f-4a85-a6a7-278241c373f6-kube-api-access-clmnh\") pod \"certified-operators-q7fss\" (UID: \"e5e92738-452f-4a85-a6a7-278241c373f6\") " pod="openshift-marketplace/certified-operators-q7fss" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.625922 4790 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.625938 4790 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.625952 4790 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.625968 4790 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.625983 4790 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.625996 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxghh\" (UniqueName: \"kubernetes.io/projected/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-kube-api-access-dxghh\") on node \"crc\" DevicePath \"\"" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.626008 4790 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.626021 4790 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.626034 4790 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.626045 4790 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.626056 4790 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9e561ed1-5cd7-4a50-8629-bf48c3a21ac4-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.626197 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5e92738-452f-4a85-a6a7-278241c373f6-utilities\") pod \"certified-operators-q7fss\" (UID: \"e5e92738-452f-4a85-a6a7-278241c373f6\") " pod="openshift-marketplace/certified-operators-q7fss" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.626494 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5e92738-452f-4a85-a6a7-278241c373f6-catalog-content\") pod \"certified-operators-q7fss\" (UID: \"e5e92738-452f-4a85-a6a7-278241c373f6\") " pod="openshift-marketplace/certified-operators-q7fss" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.642587 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clmnh\" (UniqueName: \"kubernetes.io/projected/e5e92738-452f-4a85-a6a7-278241c373f6-kube-api-access-clmnh\") pod \"certified-operators-q7fss\" (UID: \"e5e92738-452f-4a85-a6a7-278241c373f6\") " pod="openshift-marketplace/certified-operators-q7fss" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.700768 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" event={"ID":"9e561ed1-5cd7-4a50-8629-bf48c3a21ac4","Type":"ContainerDied","Data":"e1625ade383c73f8a5bf0f1bd7d600635a01cada557b61315f1e423a28e6578a"} Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.700814 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1625ade383c73f8a5bf0f1bd7d600635a01cada557b61315f1e423a28e6578a" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.700917 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf" Nov 24 15:55:40 crc kubenswrapper[4790]: I1124 15:55:40.704505 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q7fss" Nov 24 15:55:41 crc kubenswrapper[4790]: I1124 15:55:41.263419 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q7fss"] Nov 24 15:55:41 crc kubenswrapper[4790]: I1124 15:55:41.713727 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q7fss" event={"ID":"e5e92738-452f-4a85-a6a7-278241c373f6","Type":"ContainerStarted","Data":"4dc6485f2e90e77cc83a10be60989334ef2edbde098cd86cb48d166b598cbba4"} Nov 24 15:55:42 crc kubenswrapper[4790]: I1124 15:55:42.728067 4790 generic.go:334] "Generic (PLEG): container finished" podID="e5e92738-452f-4a85-a6a7-278241c373f6" containerID="79260f300b90c210ec961b7692d25cf825bb39a1e8b589f218397bc3863034e3" exitCode=0 Nov 24 15:55:42 crc kubenswrapper[4790]: I1124 15:55:42.728191 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q7fss" event={"ID":"e5e92738-452f-4a85-a6a7-278241c373f6","Type":"ContainerDied","Data":"79260f300b90c210ec961b7692d25cf825bb39a1e8b589f218397bc3863034e3"} Nov 24 15:55:42 crc kubenswrapper[4790]: I1124 15:55:42.731229 4790 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 15:55:44 crc kubenswrapper[4790]: I1124 15:55:44.319708 4790 scope.go:117] "RemoveContainer" containerID="f1e094cfe52dce391f24557f040b619b252fb0fb27545cacdfae7450eb6c06b2" Nov 24 15:55:44 crc kubenswrapper[4790]: E1124 15:55:44.320551 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:55:44 crc kubenswrapper[4790]: I1124 15:55:44.753648 4790 generic.go:334] "Generic (PLEG): container finished" podID="e5e92738-452f-4a85-a6a7-278241c373f6" containerID="ec9565c90541433b8ab9d95295050a50f09a35906d2acbaa697c1fd5358ea754" exitCode=0 Nov 24 15:55:44 crc kubenswrapper[4790]: I1124 15:55:44.753737 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q7fss" event={"ID":"e5e92738-452f-4a85-a6a7-278241c373f6","Type":"ContainerDied","Data":"ec9565c90541433b8ab9d95295050a50f09a35906d2acbaa697c1fd5358ea754"} Nov 24 15:55:45 crc kubenswrapper[4790]: I1124 15:55:45.775823 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q7fss" event={"ID":"e5e92738-452f-4a85-a6a7-278241c373f6","Type":"ContainerStarted","Data":"0f47cc27e95a2c1e328bbd37d3ba3535c2db4507f407867f29f2b391abf8b39a"} Nov 24 15:55:45 crc kubenswrapper[4790]: I1124 15:55:45.809327 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-q7fss" podStartSLOduration=3.35888424 podStartE2EDuration="5.809310058s" podCreationTimestamp="2025-11-24 15:55:40 +0000 UTC" firstStartedPulling="2025-11-24 15:55:42.730869004 +0000 UTC m=+9791.110762686" lastFinishedPulling="2025-11-24 15:55:45.181294832 +0000 UTC m=+9793.561188504" observedRunningTime="2025-11-24 15:55:45.800378183 +0000 UTC m=+9794.180271845" watchObservedRunningTime="2025-11-24 15:55:45.809310058 +0000 UTC m=+9794.189203720" Nov 24 15:55:50 crc kubenswrapper[4790]: I1124 15:55:50.705697 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-q7fss" Nov 24 15:55:50 crc kubenswrapper[4790]: I1124 15:55:50.706139 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-q7fss" Nov 24 15:55:50 crc kubenswrapper[4790]: I1124 15:55:50.783045 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-q7fss" Nov 24 15:55:50 crc kubenswrapper[4790]: I1124 15:55:50.899187 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-q7fss" Nov 24 15:55:51 crc kubenswrapper[4790]: I1124 15:55:51.025175 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q7fss"] Nov 24 15:55:52 crc kubenswrapper[4790]: I1124 15:55:52.866614 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-q7fss" podUID="e5e92738-452f-4a85-a6a7-278241c373f6" containerName="registry-server" containerID="cri-o://0f47cc27e95a2c1e328bbd37d3ba3535c2db4507f407867f29f2b391abf8b39a" gracePeriod=2 Nov 24 15:55:53 crc kubenswrapper[4790]: I1124 15:55:53.501398 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q7fss" Nov 24 15:55:53 crc kubenswrapper[4790]: I1124 15:55:53.646905 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5e92738-452f-4a85-a6a7-278241c373f6-utilities\") pod \"e5e92738-452f-4a85-a6a7-278241c373f6\" (UID: \"e5e92738-452f-4a85-a6a7-278241c373f6\") " Nov 24 15:55:53 crc kubenswrapper[4790]: I1124 15:55:53.646984 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clmnh\" (UniqueName: \"kubernetes.io/projected/e5e92738-452f-4a85-a6a7-278241c373f6-kube-api-access-clmnh\") pod \"e5e92738-452f-4a85-a6a7-278241c373f6\" (UID: \"e5e92738-452f-4a85-a6a7-278241c373f6\") " Nov 24 15:55:53 crc kubenswrapper[4790]: I1124 15:55:53.647014 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5e92738-452f-4a85-a6a7-278241c373f6-catalog-content\") pod \"e5e92738-452f-4a85-a6a7-278241c373f6\" (UID: \"e5e92738-452f-4a85-a6a7-278241c373f6\") " Nov 24 15:55:53 crc kubenswrapper[4790]: I1124 15:55:53.648289 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5e92738-452f-4a85-a6a7-278241c373f6-utilities" (OuterVolumeSpecName: "utilities") pod "e5e92738-452f-4a85-a6a7-278241c373f6" (UID: "e5e92738-452f-4a85-a6a7-278241c373f6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:55:53 crc kubenswrapper[4790]: I1124 15:55:53.657482 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5e92738-452f-4a85-a6a7-278241c373f6-kube-api-access-clmnh" (OuterVolumeSpecName: "kube-api-access-clmnh") pod "e5e92738-452f-4a85-a6a7-278241c373f6" (UID: "e5e92738-452f-4a85-a6a7-278241c373f6"). InnerVolumeSpecName "kube-api-access-clmnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:55:53 crc kubenswrapper[4790]: I1124 15:55:53.698741 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5e92738-452f-4a85-a6a7-278241c373f6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e5e92738-452f-4a85-a6a7-278241c373f6" (UID: "e5e92738-452f-4a85-a6a7-278241c373f6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:55:53 crc kubenswrapper[4790]: I1124 15:55:53.749383 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5e92738-452f-4a85-a6a7-278241c373f6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 15:55:53 crc kubenswrapper[4790]: I1124 15:55:53.749421 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5e92738-452f-4a85-a6a7-278241c373f6-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 15:55:53 crc kubenswrapper[4790]: I1124 15:55:53.749434 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clmnh\" (UniqueName: \"kubernetes.io/projected/e5e92738-452f-4a85-a6a7-278241c373f6-kube-api-access-clmnh\") on node \"crc\" DevicePath \"\"" Nov 24 15:55:53 crc kubenswrapper[4790]: I1124 15:55:53.885051 4790 generic.go:334] "Generic (PLEG): container finished" podID="e5e92738-452f-4a85-a6a7-278241c373f6" containerID="0f47cc27e95a2c1e328bbd37d3ba3535c2db4507f407867f29f2b391abf8b39a" exitCode=0 Nov 24 15:55:53 crc kubenswrapper[4790]: I1124 15:55:53.885147 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q7fss" event={"ID":"e5e92738-452f-4a85-a6a7-278241c373f6","Type":"ContainerDied","Data":"0f47cc27e95a2c1e328bbd37d3ba3535c2db4507f407867f29f2b391abf8b39a"} Nov 24 15:55:53 crc kubenswrapper[4790]: I1124 15:55:53.885175 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q7fss" Nov 24 15:55:53 crc kubenswrapper[4790]: I1124 15:55:53.885211 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q7fss" event={"ID":"e5e92738-452f-4a85-a6a7-278241c373f6","Type":"ContainerDied","Data":"4dc6485f2e90e77cc83a10be60989334ef2edbde098cd86cb48d166b598cbba4"} Nov 24 15:55:53 crc kubenswrapper[4790]: I1124 15:55:53.885245 4790 scope.go:117] "RemoveContainer" containerID="0f47cc27e95a2c1e328bbd37d3ba3535c2db4507f407867f29f2b391abf8b39a" Nov 24 15:55:53 crc kubenswrapper[4790]: I1124 15:55:53.922714 4790 scope.go:117] "RemoveContainer" containerID="ec9565c90541433b8ab9d95295050a50f09a35906d2acbaa697c1fd5358ea754" Nov 24 15:55:53 crc kubenswrapper[4790]: I1124 15:55:53.953218 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q7fss"] Nov 24 15:55:53 crc kubenswrapper[4790]: I1124 15:55:53.960149 4790 scope.go:117] "RemoveContainer" containerID="79260f300b90c210ec961b7692d25cf825bb39a1e8b589f218397bc3863034e3" Nov 24 15:55:53 crc kubenswrapper[4790]: I1124 15:55:53.973048 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-q7fss"] Nov 24 15:55:54 crc kubenswrapper[4790]: I1124 15:55:54.039019 4790 scope.go:117] "RemoveContainer" containerID="0f47cc27e95a2c1e328bbd37d3ba3535c2db4507f407867f29f2b391abf8b39a" Nov 24 15:55:54 crc kubenswrapper[4790]: E1124 15:55:54.039581 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f47cc27e95a2c1e328bbd37d3ba3535c2db4507f407867f29f2b391abf8b39a\": container with ID starting with 0f47cc27e95a2c1e328bbd37d3ba3535c2db4507f407867f29f2b391abf8b39a not found: ID does not exist" containerID="0f47cc27e95a2c1e328bbd37d3ba3535c2db4507f407867f29f2b391abf8b39a" Nov 24 15:55:54 crc kubenswrapper[4790]: I1124 15:55:54.039631 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f47cc27e95a2c1e328bbd37d3ba3535c2db4507f407867f29f2b391abf8b39a"} err="failed to get container status \"0f47cc27e95a2c1e328bbd37d3ba3535c2db4507f407867f29f2b391abf8b39a\": rpc error: code = NotFound desc = could not find container \"0f47cc27e95a2c1e328bbd37d3ba3535c2db4507f407867f29f2b391abf8b39a\": container with ID starting with 0f47cc27e95a2c1e328bbd37d3ba3535c2db4507f407867f29f2b391abf8b39a not found: ID does not exist" Nov 24 15:55:54 crc kubenswrapper[4790]: I1124 15:55:54.039666 4790 scope.go:117] "RemoveContainer" containerID="ec9565c90541433b8ab9d95295050a50f09a35906d2acbaa697c1fd5358ea754" Nov 24 15:55:54 crc kubenswrapper[4790]: E1124 15:55:54.040390 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec9565c90541433b8ab9d95295050a50f09a35906d2acbaa697c1fd5358ea754\": container with ID starting with ec9565c90541433b8ab9d95295050a50f09a35906d2acbaa697c1fd5358ea754 not found: ID does not exist" containerID="ec9565c90541433b8ab9d95295050a50f09a35906d2acbaa697c1fd5358ea754" Nov 24 15:55:54 crc kubenswrapper[4790]: I1124 15:55:54.040428 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec9565c90541433b8ab9d95295050a50f09a35906d2acbaa697c1fd5358ea754"} err="failed to get container status \"ec9565c90541433b8ab9d95295050a50f09a35906d2acbaa697c1fd5358ea754\": rpc error: code = NotFound desc = could not find container \"ec9565c90541433b8ab9d95295050a50f09a35906d2acbaa697c1fd5358ea754\": container with ID starting with ec9565c90541433b8ab9d95295050a50f09a35906d2acbaa697c1fd5358ea754 not found: ID does not exist" Nov 24 15:55:54 crc kubenswrapper[4790]: I1124 15:55:54.040487 4790 scope.go:117] "RemoveContainer" containerID="79260f300b90c210ec961b7692d25cf825bb39a1e8b589f218397bc3863034e3" Nov 24 15:55:54 crc kubenswrapper[4790]: E1124 15:55:54.040921 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79260f300b90c210ec961b7692d25cf825bb39a1e8b589f218397bc3863034e3\": container with ID starting with 79260f300b90c210ec961b7692d25cf825bb39a1e8b589f218397bc3863034e3 not found: ID does not exist" containerID="79260f300b90c210ec961b7692d25cf825bb39a1e8b589f218397bc3863034e3" Nov 24 15:55:54 crc kubenswrapper[4790]: I1124 15:55:54.040970 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79260f300b90c210ec961b7692d25cf825bb39a1e8b589f218397bc3863034e3"} err="failed to get container status \"79260f300b90c210ec961b7692d25cf825bb39a1e8b589f218397bc3863034e3\": rpc error: code = NotFound desc = could not find container \"79260f300b90c210ec961b7692d25cf825bb39a1e8b589f218397bc3863034e3\": container with ID starting with 79260f300b90c210ec961b7692d25cf825bb39a1e8b589f218397bc3863034e3 not found: ID does not exist" Nov 24 15:55:54 crc kubenswrapper[4790]: I1124 15:55:54.332940 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5e92738-452f-4a85-a6a7-278241c373f6" path="/var/lib/kubelet/pods/e5e92738-452f-4a85-a6a7-278241c373f6/volumes" Nov 24 15:55:56 crc kubenswrapper[4790]: I1124 15:55:56.314510 4790 scope.go:117] "RemoveContainer" containerID="f1e094cfe52dce391f24557f040b619b252fb0fb27545cacdfae7450eb6c06b2" Nov 24 15:55:56 crc kubenswrapper[4790]: E1124 15:55:56.316015 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:56:09 crc kubenswrapper[4790]: I1124 15:56:09.314978 4790 scope.go:117] "RemoveContainer" containerID="f1e094cfe52dce391f24557f040b619b252fb0fb27545cacdfae7450eb6c06b2" Nov 24 15:56:09 crc kubenswrapper[4790]: E1124 15:56:09.315995 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:56:24 crc kubenswrapper[4790]: I1124 15:56:24.315485 4790 scope.go:117] "RemoveContainer" containerID="f1e094cfe52dce391f24557f040b619b252fb0fb27545cacdfae7450eb6c06b2" Nov 24 15:56:24 crc kubenswrapper[4790]: E1124 15:56:24.316442 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:56:39 crc kubenswrapper[4790]: I1124 15:56:39.321997 4790 scope.go:117] "RemoveContainer" containerID="f1e094cfe52dce391f24557f040b619b252fb0fb27545cacdfae7450eb6c06b2" Nov 24 15:56:39 crc kubenswrapper[4790]: E1124 15:56:39.322894 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:56:50 crc kubenswrapper[4790]: I1124 15:56:50.316210 4790 scope.go:117] "RemoveContainer" containerID="f1e094cfe52dce391f24557f040b619b252fb0fb27545cacdfae7450eb6c06b2" Nov 24 15:56:50 crc kubenswrapper[4790]: E1124 15:56:50.317105 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:57:05 crc kubenswrapper[4790]: I1124 15:57:05.315936 4790 scope.go:117] "RemoveContainer" containerID="f1e094cfe52dce391f24557f040b619b252fb0fb27545cacdfae7450eb6c06b2" Nov 24 15:57:05 crc kubenswrapper[4790]: E1124 15:57:05.317487 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:57:14 crc kubenswrapper[4790]: E1124 15:57:14.694162 4790 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.129.56.69:35866->38.129.56.69:36625: read tcp 38.129.56.69:35866->38.129.56.69:36625: read: connection reset by peer Nov 24 15:57:16 crc kubenswrapper[4790]: I1124 15:57:16.316315 4790 scope.go:117] "RemoveContainer" containerID="f1e094cfe52dce391f24557f040b619b252fb0fb27545cacdfae7450eb6c06b2" Nov 24 15:57:16 crc kubenswrapper[4790]: E1124 15:57:16.316822 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:57:27 crc kubenswrapper[4790]: I1124 15:57:27.343036 4790 scope.go:117] "RemoveContainer" containerID="f1e094cfe52dce391f24557f040b619b252fb0fb27545cacdfae7450eb6c06b2" Nov 24 15:57:27 crc kubenswrapper[4790]: E1124 15:57:27.347373 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:57:41 crc kubenswrapper[4790]: I1124 15:57:41.316929 4790 scope.go:117] "RemoveContainer" containerID="f1e094cfe52dce391f24557f040b619b252fb0fb27545cacdfae7450eb6c06b2" Nov 24 15:57:41 crc kubenswrapper[4790]: E1124 15:57:41.318463 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 15:57:44 crc kubenswrapper[4790]: I1124 15:57:44.482945 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fscrk"] Nov 24 15:57:44 crc kubenswrapper[4790]: E1124 15:57:44.492408 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5e92738-452f-4a85-a6a7-278241c373f6" containerName="extract-utilities" Nov 24 15:57:44 crc kubenswrapper[4790]: I1124 15:57:44.492459 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5e92738-452f-4a85-a6a7-278241c373f6" containerName="extract-utilities" Nov 24 15:57:44 crc kubenswrapper[4790]: E1124 15:57:44.492487 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5e92738-452f-4a85-a6a7-278241c373f6" containerName="extract-content" Nov 24 15:57:44 crc kubenswrapper[4790]: I1124 15:57:44.492506 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5e92738-452f-4a85-a6a7-278241c373f6" containerName="extract-content" Nov 24 15:57:44 crc kubenswrapper[4790]: E1124 15:57:44.492605 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5e92738-452f-4a85-a6a7-278241c373f6" containerName="registry-server" Nov 24 15:57:44 crc kubenswrapper[4790]: I1124 15:57:44.492617 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5e92738-452f-4a85-a6a7-278241c373f6" containerName="registry-server" Nov 24 15:57:44 crc kubenswrapper[4790]: I1124 15:57:44.493154 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5e92738-452f-4a85-a6a7-278241c373f6" containerName="registry-server" Nov 24 15:57:44 crc kubenswrapper[4790]: I1124 15:57:44.495970 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fscrk"] Nov 24 15:57:44 crc kubenswrapper[4790]: I1124 15:57:44.496134 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fscrk" Nov 24 15:57:44 crc kubenswrapper[4790]: I1124 15:57:44.521038 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6e1eef8-0365-42fd-bbd9-444096c27c62-utilities\") pod \"redhat-marketplace-fscrk\" (UID: \"b6e1eef8-0365-42fd-bbd9-444096c27c62\") " pod="openshift-marketplace/redhat-marketplace-fscrk" Nov 24 15:57:44 crc kubenswrapper[4790]: I1124 15:57:44.521081 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45fx4\" (UniqueName: \"kubernetes.io/projected/b6e1eef8-0365-42fd-bbd9-444096c27c62-kube-api-access-45fx4\") pod \"redhat-marketplace-fscrk\" (UID: \"b6e1eef8-0365-42fd-bbd9-444096c27c62\") " pod="openshift-marketplace/redhat-marketplace-fscrk" Nov 24 15:57:44 crc kubenswrapper[4790]: I1124 15:57:44.521114 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6e1eef8-0365-42fd-bbd9-444096c27c62-catalog-content\") pod \"redhat-marketplace-fscrk\" (UID: \"b6e1eef8-0365-42fd-bbd9-444096c27c62\") " pod="openshift-marketplace/redhat-marketplace-fscrk" Nov 24 15:57:44 crc kubenswrapper[4790]: I1124 15:57:44.622817 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6e1eef8-0365-42fd-bbd9-444096c27c62-utilities\") pod \"redhat-marketplace-fscrk\" (UID: \"b6e1eef8-0365-42fd-bbd9-444096c27c62\") " pod="openshift-marketplace/redhat-marketplace-fscrk" Nov 24 15:57:44 crc kubenswrapper[4790]: I1124 15:57:44.622873 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45fx4\" (UniqueName: \"kubernetes.io/projected/b6e1eef8-0365-42fd-bbd9-444096c27c62-kube-api-access-45fx4\") pod \"redhat-marketplace-fscrk\" (UID: \"b6e1eef8-0365-42fd-bbd9-444096c27c62\") " pod="openshift-marketplace/redhat-marketplace-fscrk" Nov 24 15:57:44 crc kubenswrapper[4790]: I1124 15:57:44.622933 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6e1eef8-0365-42fd-bbd9-444096c27c62-catalog-content\") pod \"redhat-marketplace-fscrk\" (UID: \"b6e1eef8-0365-42fd-bbd9-444096c27c62\") " pod="openshift-marketplace/redhat-marketplace-fscrk" Nov 24 15:57:44 crc kubenswrapper[4790]: I1124 15:57:44.623312 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6e1eef8-0365-42fd-bbd9-444096c27c62-utilities\") pod \"redhat-marketplace-fscrk\" (UID: \"b6e1eef8-0365-42fd-bbd9-444096c27c62\") " pod="openshift-marketplace/redhat-marketplace-fscrk" Nov 24 15:57:44 crc kubenswrapper[4790]: I1124 15:57:44.623377 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6e1eef8-0365-42fd-bbd9-444096c27c62-catalog-content\") pod \"redhat-marketplace-fscrk\" (UID: \"b6e1eef8-0365-42fd-bbd9-444096c27c62\") " pod="openshift-marketplace/redhat-marketplace-fscrk" Nov 24 15:57:44 crc kubenswrapper[4790]: I1124 15:57:44.649092 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45fx4\" (UniqueName: \"kubernetes.io/projected/b6e1eef8-0365-42fd-bbd9-444096c27c62-kube-api-access-45fx4\") pod \"redhat-marketplace-fscrk\" (UID: \"b6e1eef8-0365-42fd-bbd9-444096c27c62\") " pod="openshift-marketplace/redhat-marketplace-fscrk" Nov 24 15:57:44 crc kubenswrapper[4790]: I1124 15:57:44.830339 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fscrk" Nov 24 15:57:45 crc kubenswrapper[4790]: I1124 15:57:45.371151 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fscrk"] Nov 24 15:57:45 crc kubenswrapper[4790]: W1124 15:57:45.374749 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb6e1eef8_0365_42fd_bbd9_444096c27c62.slice/crio-d4a454acc4d664680e8782e642b6e62514401c14294656b010022f2fc2fa7050 WatchSource:0}: Error finding container d4a454acc4d664680e8782e642b6e62514401c14294656b010022f2fc2fa7050: Status 404 returned error can't find the container with id d4a454acc4d664680e8782e642b6e62514401c14294656b010022f2fc2fa7050 Nov 24 15:57:45 crc kubenswrapper[4790]: I1124 15:57:45.460090 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fscrk" event={"ID":"b6e1eef8-0365-42fd-bbd9-444096c27c62","Type":"ContainerStarted","Data":"d4a454acc4d664680e8782e642b6e62514401c14294656b010022f2fc2fa7050"} Nov 24 15:57:46 crc kubenswrapper[4790]: I1124 15:57:46.472788 4790 generic.go:334] "Generic (PLEG): container finished" podID="b6e1eef8-0365-42fd-bbd9-444096c27c62" containerID="da668ce73bc7aaba66e61ed82b3d5603f452e672a271005a5feea687a5045edb" exitCode=0 Nov 24 15:57:46 crc kubenswrapper[4790]: I1124 15:57:46.472970 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fscrk" event={"ID":"b6e1eef8-0365-42fd-bbd9-444096c27c62","Type":"ContainerDied","Data":"da668ce73bc7aaba66e61ed82b3d5603f452e672a271005a5feea687a5045edb"} Nov 24 15:57:47 crc kubenswrapper[4790]: I1124 15:57:47.486461 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fscrk" event={"ID":"b6e1eef8-0365-42fd-bbd9-444096c27c62","Type":"ContainerStarted","Data":"0cb468739476d63e606f3c307a49cd580d033ee8628a79ea33135b848afc8541"} Nov 24 15:57:48 crc kubenswrapper[4790]: I1124 15:57:48.505646 4790 generic.go:334] "Generic (PLEG): container finished" podID="b6e1eef8-0365-42fd-bbd9-444096c27c62" containerID="0cb468739476d63e606f3c307a49cd580d033ee8628a79ea33135b848afc8541" exitCode=0 Nov 24 15:57:48 crc kubenswrapper[4790]: I1124 15:57:48.505721 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fscrk" event={"ID":"b6e1eef8-0365-42fd-bbd9-444096c27c62","Type":"ContainerDied","Data":"0cb468739476d63e606f3c307a49cd580d033ee8628a79ea33135b848afc8541"} Nov 24 15:57:49 crc kubenswrapper[4790]: I1124 15:57:49.518895 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fscrk" event={"ID":"b6e1eef8-0365-42fd-bbd9-444096c27c62","Type":"ContainerStarted","Data":"e8f9b05277abd13405c156a72dc4f18b2ca72afc1f26b6d409eaace5234b0cdf"} Nov 24 15:57:49 crc kubenswrapper[4790]: I1124 15:57:49.560736 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fscrk" podStartSLOduration=3.1225927430000002 podStartE2EDuration="5.560712429s" podCreationTimestamp="2025-11-24 15:57:44 +0000 UTC" firstStartedPulling="2025-11-24 15:57:46.474989868 +0000 UTC m=+9914.854883540" lastFinishedPulling="2025-11-24 15:57:48.913109554 +0000 UTC m=+9917.293003226" observedRunningTime="2025-11-24 15:57:49.54670241 +0000 UTC m=+9917.926596082" watchObservedRunningTime="2025-11-24 15:57:49.560712429 +0000 UTC m=+9917.940606101" Nov 24 15:57:50 crc kubenswrapper[4790]: I1124 15:57:50.717098 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Nov 24 15:57:50 crc kubenswrapper[4790]: I1124 15:57:50.717769 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mariadb-copy-data" podUID="185395e2-de39-45a6-acd6-44f6a5405018" containerName="adoption" containerID="cri-o://972bb850604ae326cd1ba67e236b9e051e64189a7b2711475af7417ec61cdd17" gracePeriod=30 Nov 24 15:57:54 crc kubenswrapper[4790]: I1124 15:57:54.831185 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fscrk" Nov 24 15:57:54 crc kubenswrapper[4790]: I1124 15:57:54.832011 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fscrk" Nov 24 15:57:54 crc kubenswrapper[4790]: I1124 15:57:54.921355 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fscrk" Nov 24 15:57:55 crc kubenswrapper[4790]: I1124 15:57:55.684241 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fscrk" Nov 24 15:57:55 crc kubenswrapper[4790]: I1124 15:57:55.773786 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fscrk"] Nov 24 15:57:56 crc kubenswrapper[4790]: I1124 15:57:56.315289 4790 scope.go:117] "RemoveContainer" containerID="f1e094cfe52dce391f24557f040b619b252fb0fb27545cacdfae7450eb6c06b2" Nov 24 15:57:57 crc kubenswrapper[4790]: I1124 15:57:57.625379 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"25209c017376e3ac46b1a33e19b0b7fe5acafd9ba6b9ced78aacd10037991a2f"} Nov 24 15:57:57 crc kubenswrapper[4790]: I1124 15:57:57.625580 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fscrk" podUID="b6e1eef8-0365-42fd-bbd9-444096c27c62" containerName="registry-server" containerID="cri-o://e8f9b05277abd13405c156a72dc4f18b2ca72afc1f26b6d409eaace5234b0cdf" gracePeriod=2 Nov 24 15:57:58 crc kubenswrapper[4790]: I1124 15:57:58.142229 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fscrk" Nov 24 15:57:58 crc kubenswrapper[4790]: I1124 15:57:58.251441 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6e1eef8-0365-42fd-bbd9-444096c27c62-utilities\") pod \"b6e1eef8-0365-42fd-bbd9-444096c27c62\" (UID: \"b6e1eef8-0365-42fd-bbd9-444096c27c62\") " Nov 24 15:57:58 crc kubenswrapper[4790]: I1124 15:57:58.251502 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45fx4\" (UniqueName: \"kubernetes.io/projected/b6e1eef8-0365-42fd-bbd9-444096c27c62-kube-api-access-45fx4\") pod \"b6e1eef8-0365-42fd-bbd9-444096c27c62\" (UID: \"b6e1eef8-0365-42fd-bbd9-444096c27c62\") " Nov 24 15:57:58 crc kubenswrapper[4790]: I1124 15:57:58.251537 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6e1eef8-0365-42fd-bbd9-444096c27c62-catalog-content\") pod \"b6e1eef8-0365-42fd-bbd9-444096c27c62\" (UID: \"b6e1eef8-0365-42fd-bbd9-444096c27c62\") " Nov 24 15:57:58 crc kubenswrapper[4790]: I1124 15:57:58.252491 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6e1eef8-0365-42fd-bbd9-444096c27c62-utilities" (OuterVolumeSpecName: "utilities") pod "b6e1eef8-0365-42fd-bbd9-444096c27c62" (UID: "b6e1eef8-0365-42fd-bbd9-444096c27c62"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:57:58 crc kubenswrapper[4790]: I1124 15:57:58.260087 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6e1eef8-0365-42fd-bbd9-444096c27c62-kube-api-access-45fx4" (OuterVolumeSpecName: "kube-api-access-45fx4") pod "b6e1eef8-0365-42fd-bbd9-444096c27c62" (UID: "b6e1eef8-0365-42fd-bbd9-444096c27c62"). InnerVolumeSpecName "kube-api-access-45fx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:57:58 crc kubenswrapper[4790]: I1124 15:57:58.274329 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6e1eef8-0365-42fd-bbd9-444096c27c62-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b6e1eef8-0365-42fd-bbd9-444096c27c62" (UID: "b6e1eef8-0365-42fd-bbd9-444096c27c62"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 15:57:58 crc kubenswrapper[4790]: I1124 15:57:58.354414 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6e1eef8-0365-42fd-bbd9-444096c27c62-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 15:57:58 crc kubenswrapper[4790]: I1124 15:57:58.354454 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45fx4\" (UniqueName: \"kubernetes.io/projected/b6e1eef8-0365-42fd-bbd9-444096c27c62-kube-api-access-45fx4\") on node \"crc\" DevicePath \"\"" Nov 24 15:57:58 crc kubenswrapper[4790]: I1124 15:57:58.354475 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6e1eef8-0365-42fd-bbd9-444096c27c62-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 15:57:58 crc kubenswrapper[4790]: I1124 15:57:58.639229 4790 generic.go:334] "Generic (PLEG): container finished" podID="b6e1eef8-0365-42fd-bbd9-444096c27c62" containerID="e8f9b05277abd13405c156a72dc4f18b2ca72afc1f26b6d409eaace5234b0cdf" exitCode=0 Nov 24 15:57:58 crc kubenswrapper[4790]: I1124 15:57:58.639277 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fscrk" Nov 24 15:57:58 crc kubenswrapper[4790]: I1124 15:57:58.639295 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fscrk" event={"ID":"b6e1eef8-0365-42fd-bbd9-444096c27c62","Type":"ContainerDied","Data":"e8f9b05277abd13405c156a72dc4f18b2ca72afc1f26b6d409eaace5234b0cdf"} Nov 24 15:57:58 crc kubenswrapper[4790]: I1124 15:57:58.639748 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fscrk" event={"ID":"b6e1eef8-0365-42fd-bbd9-444096c27c62","Type":"ContainerDied","Data":"d4a454acc4d664680e8782e642b6e62514401c14294656b010022f2fc2fa7050"} Nov 24 15:57:58 crc kubenswrapper[4790]: I1124 15:57:58.639798 4790 scope.go:117] "RemoveContainer" containerID="e8f9b05277abd13405c156a72dc4f18b2ca72afc1f26b6d409eaace5234b0cdf" Nov 24 15:57:58 crc kubenswrapper[4790]: I1124 15:57:58.682599 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fscrk"] Nov 24 15:57:58 crc kubenswrapper[4790]: I1124 15:57:58.691785 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fscrk"] Nov 24 15:57:58 crc kubenswrapper[4790]: I1124 15:57:58.704057 4790 scope.go:117] "RemoveContainer" containerID="0cb468739476d63e606f3c307a49cd580d033ee8628a79ea33135b848afc8541" Nov 24 15:57:58 crc kubenswrapper[4790]: I1124 15:57:58.730043 4790 scope.go:117] "RemoveContainer" containerID="da668ce73bc7aaba66e61ed82b3d5603f452e672a271005a5feea687a5045edb" Nov 24 15:57:58 crc kubenswrapper[4790]: I1124 15:57:58.776082 4790 scope.go:117] "RemoveContainer" containerID="e8f9b05277abd13405c156a72dc4f18b2ca72afc1f26b6d409eaace5234b0cdf" Nov 24 15:57:58 crc kubenswrapper[4790]: E1124 15:57:58.776955 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8f9b05277abd13405c156a72dc4f18b2ca72afc1f26b6d409eaace5234b0cdf\": container with ID starting with e8f9b05277abd13405c156a72dc4f18b2ca72afc1f26b6d409eaace5234b0cdf not found: ID does not exist" containerID="e8f9b05277abd13405c156a72dc4f18b2ca72afc1f26b6d409eaace5234b0cdf" Nov 24 15:57:58 crc kubenswrapper[4790]: I1124 15:57:58.777052 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8f9b05277abd13405c156a72dc4f18b2ca72afc1f26b6d409eaace5234b0cdf"} err="failed to get container status \"e8f9b05277abd13405c156a72dc4f18b2ca72afc1f26b6d409eaace5234b0cdf\": rpc error: code = NotFound desc = could not find container \"e8f9b05277abd13405c156a72dc4f18b2ca72afc1f26b6d409eaace5234b0cdf\": container with ID starting with e8f9b05277abd13405c156a72dc4f18b2ca72afc1f26b6d409eaace5234b0cdf not found: ID does not exist" Nov 24 15:57:58 crc kubenswrapper[4790]: I1124 15:57:58.777135 4790 scope.go:117] "RemoveContainer" containerID="0cb468739476d63e606f3c307a49cd580d033ee8628a79ea33135b848afc8541" Nov 24 15:57:58 crc kubenswrapper[4790]: E1124 15:57:58.777590 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0cb468739476d63e606f3c307a49cd580d033ee8628a79ea33135b848afc8541\": container with ID starting with 0cb468739476d63e606f3c307a49cd580d033ee8628a79ea33135b848afc8541 not found: ID does not exist" containerID="0cb468739476d63e606f3c307a49cd580d033ee8628a79ea33135b848afc8541" Nov 24 15:57:58 crc kubenswrapper[4790]: I1124 15:57:58.777656 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0cb468739476d63e606f3c307a49cd580d033ee8628a79ea33135b848afc8541"} err="failed to get container status \"0cb468739476d63e606f3c307a49cd580d033ee8628a79ea33135b848afc8541\": rpc error: code = NotFound desc = could not find container \"0cb468739476d63e606f3c307a49cd580d033ee8628a79ea33135b848afc8541\": container with ID starting with 0cb468739476d63e606f3c307a49cd580d033ee8628a79ea33135b848afc8541 not found: ID does not exist" Nov 24 15:57:58 crc kubenswrapper[4790]: I1124 15:57:58.777695 4790 scope.go:117] "RemoveContainer" containerID="da668ce73bc7aaba66e61ed82b3d5603f452e672a271005a5feea687a5045edb" Nov 24 15:57:58 crc kubenswrapper[4790]: E1124 15:57:58.778049 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da668ce73bc7aaba66e61ed82b3d5603f452e672a271005a5feea687a5045edb\": container with ID starting with da668ce73bc7aaba66e61ed82b3d5603f452e672a271005a5feea687a5045edb not found: ID does not exist" containerID="da668ce73bc7aaba66e61ed82b3d5603f452e672a271005a5feea687a5045edb" Nov 24 15:57:58 crc kubenswrapper[4790]: I1124 15:57:58.778148 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da668ce73bc7aaba66e61ed82b3d5603f452e672a271005a5feea687a5045edb"} err="failed to get container status \"da668ce73bc7aaba66e61ed82b3d5603f452e672a271005a5feea687a5045edb\": rpc error: code = NotFound desc = could not find container \"da668ce73bc7aaba66e61ed82b3d5603f452e672a271005a5feea687a5045edb\": container with ID starting with da668ce73bc7aaba66e61ed82b3d5603f452e672a271005a5feea687a5045edb not found: ID does not exist" Nov 24 15:58:00 crc kubenswrapper[4790]: I1124 15:58:00.333523 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6e1eef8-0365-42fd-bbd9-444096c27c62" path="/var/lib/kubelet/pods/b6e1eef8-0365-42fd-bbd9-444096c27c62/volumes" Nov 24 15:58:20 crc kubenswrapper[4790]: I1124 15:58:20.935533 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"185395e2-de39-45a6-acd6-44f6a5405018","Type":"ContainerDied","Data":"972bb850604ae326cd1ba67e236b9e051e64189a7b2711475af7417ec61cdd17"} Nov 24 15:58:20 crc kubenswrapper[4790]: I1124 15:58:20.935613 4790 generic.go:334] "Generic (PLEG): container finished" podID="185395e2-de39-45a6-acd6-44f6a5405018" containerID="972bb850604ae326cd1ba67e236b9e051e64189a7b2711475af7417ec61cdd17" exitCode=137 Nov 24 15:58:21 crc kubenswrapper[4790]: I1124 15:58:21.244224 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Nov 24 15:58:21 crc kubenswrapper[4790]: I1124 15:58:21.320534 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mariadb-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f5e64ede-e98c-4007-9957-6a9735fe71e4\") pod \"185395e2-de39-45a6-acd6-44f6a5405018\" (UID: \"185395e2-de39-45a6-acd6-44f6a5405018\") " Nov 24 15:58:21 crc kubenswrapper[4790]: I1124 15:58:21.320615 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-845wh\" (UniqueName: \"kubernetes.io/projected/185395e2-de39-45a6-acd6-44f6a5405018-kube-api-access-845wh\") pod \"185395e2-de39-45a6-acd6-44f6a5405018\" (UID: \"185395e2-de39-45a6-acd6-44f6a5405018\") " Nov 24 15:58:21 crc kubenswrapper[4790]: I1124 15:58:21.329162 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/185395e2-de39-45a6-acd6-44f6a5405018-kube-api-access-845wh" (OuterVolumeSpecName: "kube-api-access-845wh") pod "185395e2-de39-45a6-acd6-44f6a5405018" (UID: "185395e2-de39-45a6-acd6-44f6a5405018"). InnerVolumeSpecName "kube-api-access-845wh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:58:21 crc kubenswrapper[4790]: I1124 15:58:21.351940 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f5e64ede-e98c-4007-9957-6a9735fe71e4" (OuterVolumeSpecName: "mariadb-data") pod "185395e2-de39-45a6-acd6-44f6a5405018" (UID: "185395e2-de39-45a6-acd6-44f6a5405018"). InnerVolumeSpecName "pvc-f5e64ede-e98c-4007-9957-6a9735fe71e4". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 15:58:21 crc kubenswrapper[4790]: I1124 15:58:21.422947 4790 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-f5e64ede-e98c-4007-9957-6a9735fe71e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f5e64ede-e98c-4007-9957-6a9735fe71e4\") on node \"crc\" " Nov 24 15:58:21 crc kubenswrapper[4790]: I1124 15:58:21.423014 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-845wh\" (UniqueName: \"kubernetes.io/projected/185395e2-de39-45a6-acd6-44f6a5405018-kube-api-access-845wh\") on node \"crc\" DevicePath \"\"" Nov 24 15:58:21 crc kubenswrapper[4790]: I1124 15:58:21.460842 4790 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 24 15:58:21 crc kubenswrapper[4790]: I1124 15:58:21.461005 4790 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-f5e64ede-e98c-4007-9957-6a9735fe71e4" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f5e64ede-e98c-4007-9957-6a9735fe71e4") on node "crc" Nov 24 15:58:21 crc kubenswrapper[4790]: I1124 15:58:21.525998 4790 reconciler_common.go:293] "Volume detached for volume \"pvc-f5e64ede-e98c-4007-9957-6a9735fe71e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f5e64ede-e98c-4007-9957-6a9735fe71e4\") on node \"crc\" DevicePath \"\"" Nov 24 15:58:21 crc kubenswrapper[4790]: I1124 15:58:21.947274 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"185395e2-de39-45a6-acd6-44f6a5405018","Type":"ContainerDied","Data":"6f5bba76d4cf33212a2645d0cc6576e9f1373420e461875fa85d1bf278a95db9"} Nov 24 15:58:21 crc kubenswrapper[4790]: I1124 15:58:21.947323 4790 scope.go:117] "RemoveContainer" containerID="972bb850604ae326cd1ba67e236b9e051e64189a7b2711475af7417ec61cdd17" Nov 24 15:58:21 crc kubenswrapper[4790]: I1124 15:58:21.947339 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Nov 24 15:58:21 crc kubenswrapper[4790]: I1124 15:58:21.981636 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Nov 24 15:58:21 crc kubenswrapper[4790]: I1124 15:58:21.991395 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-copy-data"] Nov 24 15:58:22 crc kubenswrapper[4790]: I1124 15:58:22.335323 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="185395e2-de39-45a6-acd6-44f6a5405018" path="/var/lib/kubelet/pods/185395e2-de39-45a6-acd6-44f6a5405018/volumes" Nov 24 15:58:22 crc kubenswrapper[4790]: I1124 15:58:22.568209 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Nov 24 15:58:22 crc kubenswrapper[4790]: I1124 15:58:22.568607 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-copy-data" podUID="67fabd6a-992b-49ab-b90f-98fddba09ce0" containerName="adoption" containerID="cri-o://ec835f53de660a2c046a34875abb04980f7575d9c55dcb8fc367cafb608533c0" gracePeriod=30 Nov 24 15:58:53 crc kubenswrapper[4790]: I1124 15:58:53.189084 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Nov 24 15:58:53 crc kubenswrapper[4790]: I1124 15:58:53.267520 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmmp9\" (UniqueName: \"kubernetes.io/projected/67fabd6a-992b-49ab-b90f-98fddba09ce0-kube-api-access-bmmp9\") pod \"67fabd6a-992b-49ab-b90f-98fddba09ce0\" (UID: \"67fabd6a-992b-49ab-b90f-98fddba09ce0\") " Nov 24 15:58:53 crc kubenswrapper[4790]: I1124 15:58:53.268563 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-012d64e2-c94b-454b-b5bb-6a6a047812a5\") pod \"67fabd6a-992b-49ab-b90f-98fddba09ce0\" (UID: \"67fabd6a-992b-49ab-b90f-98fddba09ce0\") " Nov 24 15:58:53 crc kubenswrapper[4790]: I1124 15:58:53.268599 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/67fabd6a-992b-49ab-b90f-98fddba09ce0-ovn-data-cert\") pod \"67fabd6a-992b-49ab-b90f-98fddba09ce0\" (UID: \"67fabd6a-992b-49ab-b90f-98fddba09ce0\") " Nov 24 15:58:53 crc kubenswrapper[4790]: I1124 15:58:53.275335 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67fabd6a-992b-49ab-b90f-98fddba09ce0-ovn-data-cert" (OuterVolumeSpecName: "ovn-data-cert") pod "67fabd6a-992b-49ab-b90f-98fddba09ce0" (UID: "67fabd6a-992b-49ab-b90f-98fddba09ce0"). InnerVolumeSpecName "ovn-data-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 15:58:53 crc kubenswrapper[4790]: I1124 15:58:53.277311 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67fabd6a-992b-49ab-b90f-98fddba09ce0-kube-api-access-bmmp9" (OuterVolumeSpecName: "kube-api-access-bmmp9") pod "67fabd6a-992b-49ab-b90f-98fddba09ce0" (UID: "67fabd6a-992b-49ab-b90f-98fddba09ce0"). InnerVolumeSpecName "kube-api-access-bmmp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 15:58:53 crc kubenswrapper[4790]: I1124 15:58:53.293512 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-012d64e2-c94b-454b-b5bb-6a6a047812a5" (OuterVolumeSpecName: "ovn-data") pod "67fabd6a-992b-49ab-b90f-98fddba09ce0" (UID: "67fabd6a-992b-49ab-b90f-98fddba09ce0"). InnerVolumeSpecName "pvc-012d64e2-c94b-454b-b5bb-6a6a047812a5". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 15:58:53 crc kubenswrapper[4790]: I1124 15:58:53.371356 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmmp9\" (UniqueName: \"kubernetes.io/projected/67fabd6a-992b-49ab-b90f-98fddba09ce0-kube-api-access-bmmp9\") on node \"crc\" DevicePath \"\"" Nov 24 15:58:53 crc kubenswrapper[4790]: I1124 15:58:53.371410 4790 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-012d64e2-c94b-454b-b5bb-6a6a047812a5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-012d64e2-c94b-454b-b5bb-6a6a047812a5\") on node \"crc\" " Nov 24 15:58:53 crc kubenswrapper[4790]: I1124 15:58:53.371427 4790 reconciler_common.go:293] "Volume detached for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/67fabd6a-992b-49ab-b90f-98fddba09ce0-ovn-data-cert\") on node \"crc\" DevicePath \"\"" Nov 24 15:58:53 crc kubenswrapper[4790]: I1124 15:58:53.378004 4790 generic.go:334] "Generic (PLEG): container finished" podID="67fabd6a-992b-49ab-b90f-98fddba09ce0" containerID="ec835f53de660a2c046a34875abb04980f7575d9c55dcb8fc367cafb608533c0" exitCode=137 Nov 24 15:58:53 crc kubenswrapper[4790]: I1124 15:58:53.378057 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"67fabd6a-992b-49ab-b90f-98fddba09ce0","Type":"ContainerDied","Data":"ec835f53de660a2c046a34875abb04980f7575d9c55dcb8fc367cafb608533c0"} Nov 24 15:58:53 crc kubenswrapper[4790]: I1124 15:58:53.378088 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"67fabd6a-992b-49ab-b90f-98fddba09ce0","Type":"ContainerDied","Data":"6064a7e91df03fe9ebf63359ffb45c3b8aafeb51c330ddea8d62ad2a96147db8"} Nov 24 15:58:53 crc kubenswrapper[4790]: I1124 15:58:53.378061 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Nov 24 15:58:53 crc kubenswrapper[4790]: I1124 15:58:53.378110 4790 scope.go:117] "RemoveContainer" containerID="ec835f53de660a2c046a34875abb04980f7575d9c55dcb8fc367cafb608533c0" Nov 24 15:58:53 crc kubenswrapper[4790]: I1124 15:58:53.400662 4790 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 24 15:58:53 crc kubenswrapper[4790]: I1124 15:58:53.400938 4790 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-012d64e2-c94b-454b-b5bb-6a6a047812a5" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-012d64e2-c94b-454b-b5bb-6a6a047812a5") on node "crc" Nov 24 15:58:53 crc kubenswrapper[4790]: I1124 15:58:53.401983 4790 scope.go:117] "RemoveContainer" containerID="ec835f53de660a2c046a34875abb04980f7575d9c55dcb8fc367cafb608533c0" Nov 24 15:58:53 crc kubenswrapper[4790]: E1124 15:58:53.402530 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec835f53de660a2c046a34875abb04980f7575d9c55dcb8fc367cafb608533c0\": container with ID starting with ec835f53de660a2c046a34875abb04980f7575d9c55dcb8fc367cafb608533c0 not found: ID does not exist" containerID="ec835f53de660a2c046a34875abb04980f7575d9c55dcb8fc367cafb608533c0" Nov 24 15:58:53 crc kubenswrapper[4790]: I1124 15:58:53.405977 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec835f53de660a2c046a34875abb04980f7575d9c55dcb8fc367cafb608533c0"} err="failed to get container status \"ec835f53de660a2c046a34875abb04980f7575d9c55dcb8fc367cafb608533c0\": rpc error: code = NotFound desc = could not find container \"ec835f53de660a2c046a34875abb04980f7575d9c55dcb8fc367cafb608533c0\": container with ID starting with ec835f53de660a2c046a34875abb04980f7575d9c55dcb8fc367cafb608533c0 not found: ID does not exist" Nov 24 15:58:53 crc kubenswrapper[4790]: I1124 15:58:53.419601 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Nov 24 15:58:53 crc kubenswrapper[4790]: I1124 15:58:53.434229 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-copy-data"] Nov 24 15:58:53 crc kubenswrapper[4790]: I1124 15:58:53.473519 4790 reconciler_common.go:293] "Volume detached for volume \"pvc-012d64e2-c94b-454b-b5bb-6a6a047812a5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-012d64e2-c94b-454b-b5bb-6a6a047812a5\") on node \"crc\" DevicePath \"\"" Nov 24 15:58:54 crc kubenswrapper[4790]: I1124 15:58:54.326094 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67fabd6a-992b-49ab-b90f-98fddba09ce0" path="/var/lib/kubelet/pods/67fabd6a-992b-49ab-b90f-98fddba09ce0/volumes" Nov 24 15:59:39 crc kubenswrapper[4790]: I1124 15:59:39.088628 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kb6g6"] Nov 24 15:59:39 crc kubenswrapper[4790]: E1124 15:59:39.090044 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6e1eef8-0365-42fd-bbd9-444096c27c62" containerName="extract-utilities" Nov 24 15:59:39 crc kubenswrapper[4790]: I1124 15:59:39.090067 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6e1eef8-0365-42fd-bbd9-444096c27c62" containerName="extract-utilities" Nov 24 15:59:39 crc kubenswrapper[4790]: E1124 15:59:39.090107 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67fabd6a-992b-49ab-b90f-98fddba09ce0" containerName="adoption" Nov 24 15:59:39 crc kubenswrapper[4790]: I1124 15:59:39.090121 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="67fabd6a-992b-49ab-b90f-98fddba09ce0" containerName="adoption" Nov 24 15:59:39 crc kubenswrapper[4790]: E1124 15:59:39.090146 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6e1eef8-0365-42fd-bbd9-444096c27c62" containerName="extract-content" Nov 24 15:59:39 crc kubenswrapper[4790]: I1124 15:59:39.090188 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6e1eef8-0365-42fd-bbd9-444096c27c62" containerName="extract-content" Nov 24 15:59:39 crc kubenswrapper[4790]: E1124 15:59:39.090214 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6e1eef8-0365-42fd-bbd9-444096c27c62" containerName="registry-server" Nov 24 15:59:39 crc kubenswrapper[4790]: I1124 15:59:39.090225 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6e1eef8-0365-42fd-bbd9-444096c27c62" containerName="registry-server" Nov 24 15:59:39 crc kubenswrapper[4790]: E1124 15:59:39.090260 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="185395e2-de39-45a6-acd6-44f6a5405018" containerName="adoption" Nov 24 15:59:39 crc kubenswrapper[4790]: I1124 15:59:39.090273 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="185395e2-de39-45a6-acd6-44f6a5405018" containerName="adoption" Nov 24 15:59:39 crc kubenswrapper[4790]: I1124 15:59:39.090658 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="185395e2-de39-45a6-acd6-44f6a5405018" containerName="adoption" Nov 24 15:59:39 crc kubenswrapper[4790]: I1124 15:59:39.090691 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="67fabd6a-992b-49ab-b90f-98fddba09ce0" containerName="adoption" Nov 24 15:59:39 crc kubenswrapper[4790]: I1124 15:59:39.090720 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6e1eef8-0365-42fd-bbd9-444096c27c62" containerName="registry-server" Nov 24 15:59:39 crc kubenswrapper[4790]: I1124 15:59:39.093392 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kb6g6" Nov 24 15:59:39 crc kubenswrapper[4790]: I1124 15:59:39.113868 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kb6g6"] Nov 24 15:59:39 crc kubenswrapper[4790]: I1124 15:59:39.220785 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/445af5b5-5fa0-4e0d-b760-524775e030fd-catalog-content\") pod \"redhat-operators-kb6g6\" (UID: \"445af5b5-5fa0-4e0d-b760-524775e030fd\") " pod="openshift-marketplace/redhat-operators-kb6g6" Nov 24 15:59:39 crc kubenswrapper[4790]: I1124 15:59:39.221127 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/445af5b5-5fa0-4e0d-b760-524775e030fd-utilities\") pod \"redhat-operators-kb6g6\" (UID: \"445af5b5-5fa0-4e0d-b760-524775e030fd\") " pod="openshift-marketplace/redhat-operators-kb6g6" Nov 24 15:59:39 crc kubenswrapper[4790]: I1124 15:59:39.221442 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z95ll\" (UniqueName: \"kubernetes.io/projected/445af5b5-5fa0-4e0d-b760-524775e030fd-kube-api-access-z95ll\") pod \"redhat-operators-kb6g6\" (UID: \"445af5b5-5fa0-4e0d-b760-524775e030fd\") " pod="openshift-marketplace/redhat-operators-kb6g6" Nov 24 15:59:39 crc kubenswrapper[4790]: I1124 15:59:39.323790 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/445af5b5-5fa0-4e0d-b760-524775e030fd-catalog-content\") pod \"redhat-operators-kb6g6\" (UID: \"445af5b5-5fa0-4e0d-b760-524775e030fd\") " pod="openshift-marketplace/redhat-operators-kb6g6" Nov 24 15:59:39 crc kubenswrapper[4790]: I1124 15:59:39.323965 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/445af5b5-5fa0-4e0d-b760-524775e030fd-utilities\") pod \"redhat-operators-kb6g6\" (UID: \"445af5b5-5fa0-4e0d-b760-524775e030fd\") " pod="openshift-marketplace/redhat-operators-kb6g6" Nov 24 15:59:39 crc kubenswrapper[4790]: I1124 15:59:39.324043 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z95ll\" (UniqueName: \"kubernetes.io/projected/445af5b5-5fa0-4e0d-b760-524775e030fd-kube-api-access-z95ll\") pod \"redhat-operators-kb6g6\" (UID: \"445af5b5-5fa0-4e0d-b760-524775e030fd\") " pod="openshift-marketplace/redhat-operators-kb6g6" Nov 24 15:59:39 crc kubenswrapper[4790]: I1124 15:59:39.324734 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/445af5b5-5fa0-4e0d-b760-524775e030fd-utilities\") pod \"redhat-operators-kb6g6\" (UID: \"445af5b5-5fa0-4e0d-b760-524775e030fd\") " pod="openshift-marketplace/redhat-operators-kb6g6" Nov 24 15:59:39 crc kubenswrapper[4790]: I1124 15:59:39.324821 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/445af5b5-5fa0-4e0d-b760-524775e030fd-catalog-content\") pod \"redhat-operators-kb6g6\" (UID: \"445af5b5-5fa0-4e0d-b760-524775e030fd\") " pod="openshift-marketplace/redhat-operators-kb6g6" Nov 24 15:59:39 crc kubenswrapper[4790]: I1124 15:59:39.353177 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z95ll\" (UniqueName: \"kubernetes.io/projected/445af5b5-5fa0-4e0d-b760-524775e030fd-kube-api-access-z95ll\") pod \"redhat-operators-kb6g6\" (UID: \"445af5b5-5fa0-4e0d-b760-524775e030fd\") " pod="openshift-marketplace/redhat-operators-kb6g6" Nov 24 15:59:39 crc kubenswrapper[4790]: I1124 15:59:39.428126 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kb6g6" Nov 24 15:59:39 crc kubenswrapper[4790]: I1124 15:59:39.986825 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kb6g6"] Nov 24 15:59:40 crc kubenswrapper[4790]: I1124 15:59:40.951166 4790 generic.go:334] "Generic (PLEG): container finished" podID="445af5b5-5fa0-4e0d-b760-524775e030fd" containerID="556229960b0ccb5628d9ddb94b5461bcec8f5319f3762a5b9bdf7f0b2cfc6889" exitCode=0 Nov 24 15:59:40 crc kubenswrapper[4790]: I1124 15:59:40.951241 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kb6g6" event={"ID":"445af5b5-5fa0-4e0d-b760-524775e030fd","Type":"ContainerDied","Data":"556229960b0ccb5628d9ddb94b5461bcec8f5319f3762a5b9bdf7f0b2cfc6889"} Nov 24 15:59:40 crc kubenswrapper[4790]: I1124 15:59:40.951478 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kb6g6" event={"ID":"445af5b5-5fa0-4e0d-b760-524775e030fd","Type":"ContainerStarted","Data":"a5eb1d539ed8ca681dc1927b808c5139020d2a32c69829355033444ddf353746"} Nov 24 15:59:44 crc kubenswrapper[4790]: I1124 15:59:44.994600 4790 generic.go:334] "Generic (PLEG): container finished" podID="445af5b5-5fa0-4e0d-b760-524775e030fd" containerID="fe3432b1f09e0f41b020d28b779ec96d4436f56a479bdfb4f71bfe2a8ea4ad4b" exitCode=0 Nov 24 15:59:44 crc kubenswrapper[4790]: I1124 15:59:44.995024 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kb6g6" event={"ID":"445af5b5-5fa0-4e0d-b760-524775e030fd","Type":"ContainerDied","Data":"fe3432b1f09e0f41b020d28b779ec96d4436f56a479bdfb4f71bfe2a8ea4ad4b"} Nov 24 15:59:52 crc kubenswrapper[4790]: I1124 15:59:52.021575 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-cd6sz/must-gather-99rhs"] Nov 24 15:59:52 crc kubenswrapper[4790]: I1124 15:59:52.025411 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cd6sz/must-gather-99rhs" Nov 24 15:59:52 crc kubenswrapper[4790]: I1124 15:59:52.028338 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-cd6sz"/"kube-root-ca.crt" Nov 24 15:59:52 crc kubenswrapper[4790]: I1124 15:59:52.028499 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-cd6sz"/"default-dockercfg-zcsfq" Nov 24 15:59:52 crc kubenswrapper[4790]: I1124 15:59:52.029436 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-cd6sz"/"openshift-service-ca.crt" Nov 24 15:59:52 crc kubenswrapper[4790]: I1124 15:59:52.038485 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-cd6sz/must-gather-99rhs"] Nov 24 15:59:52 crc kubenswrapper[4790]: I1124 15:59:52.151350 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7ac3f1cd-6b96-4378-949e-fc4ece0e6a09-must-gather-output\") pod \"must-gather-99rhs\" (UID: \"7ac3f1cd-6b96-4378-949e-fc4ece0e6a09\") " pod="openshift-must-gather-cd6sz/must-gather-99rhs" Nov 24 15:59:52 crc kubenswrapper[4790]: I1124 15:59:52.151422 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnnh5\" (UniqueName: \"kubernetes.io/projected/7ac3f1cd-6b96-4378-949e-fc4ece0e6a09-kube-api-access-hnnh5\") pod \"must-gather-99rhs\" (UID: \"7ac3f1cd-6b96-4378-949e-fc4ece0e6a09\") " pod="openshift-must-gather-cd6sz/must-gather-99rhs" Nov 24 15:59:52 crc kubenswrapper[4790]: I1124 15:59:52.252658 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7ac3f1cd-6b96-4378-949e-fc4ece0e6a09-must-gather-output\") pod \"must-gather-99rhs\" (UID: \"7ac3f1cd-6b96-4378-949e-fc4ece0e6a09\") " pod="openshift-must-gather-cd6sz/must-gather-99rhs" Nov 24 15:59:52 crc kubenswrapper[4790]: I1124 15:59:52.252716 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnnh5\" (UniqueName: \"kubernetes.io/projected/7ac3f1cd-6b96-4378-949e-fc4ece0e6a09-kube-api-access-hnnh5\") pod \"must-gather-99rhs\" (UID: \"7ac3f1cd-6b96-4378-949e-fc4ece0e6a09\") " pod="openshift-must-gather-cd6sz/must-gather-99rhs" Nov 24 15:59:52 crc kubenswrapper[4790]: I1124 15:59:52.253284 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7ac3f1cd-6b96-4378-949e-fc4ece0e6a09-must-gather-output\") pod \"must-gather-99rhs\" (UID: \"7ac3f1cd-6b96-4378-949e-fc4ece0e6a09\") " pod="openshift-must-gather-cd6sz/must-gather-99rhs" Nov 24 15:59:52 crc kubenswrapper[4790]: I1124 15:59:52.277097 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnnh5\" (UniqueName: \"kubernetes.io/projected/7ac3f1cd-6b96-4378-949e-fc4ece0e6a09-kube-api-access-hnnh5\") pod \"must-gather-99rhs\" (UID: \"7ac3f1cd-6b96-4378-949e-fc4ece0e6a09\") " pod="openshift-must-gather-cd6sz/must-gather-99rhs" Nov 24 15:59:52 crc kubenswrapper[4790]: I1124 15:59:52.346447 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cd6sz/must-gather-99rhs" Nov 24 15:59:54 crc kubenswrapper[4790]: I1124 15:59:54.138470 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-cd6sz/must-gather-99rhs"] Nov 24 15:59:55 crc kubenswrapper[4790]: I1124 15:59:55.101009 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kb6g6" event={"ID":"445af5b5-5fa0-4e0d-b760-524775e030fd","Type":"ContainerStarted","Data":"6d2469efb1324ac75244bf3f9b55e86e6a1c897ec87430db044498345201a644"} Nov 24 15:59:55 crc kubenswrapper[4790]: I1124 15:59:55.102414 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cd6sz/must-gather-99rhs" event={"ID":"7ac3f1cd-6b96-4378-949e-fc4ece0e6a09","Type":"ContainerStarted","Data":"0ab461042f23a21f50a6131629d4380c893ab00c94e1d6577e62dd3014129a1a"} Nov 24 15:59:55 crc kubenswrapper[4790]: I1124 15:59:55.125984 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kb6g6" podStartSLOduration=3.488571641 podStartE2EDuration="16.12596532s" podCreationTimestamp="2025-11-24 15:59:39 +0000 UTC" firstStartedPulling="2025-11-24 15:59:40.954024702 +0000 UTC m=+10029.333918374" lastFinishedPulling="2025-11-24 15:59:53.591418381 +0000 UTC m=+10041.971312053" observedRunningTime="2025-11-24 15:59:55.122037938 +0000 UTC m=+10043.501931640" watchObservedRunningTime="2025-11-24 15:59:55.12596532 +0000 UTC m=+10043.505858982" Nov 24 15:59:59 crc kubenswrapper[4790]: I1124 15:59:59.429221 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kb6g6" Nov 24 15:59:59 crc kubenswrapper[4790]: I1124 15:59:59.429771 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kb6g6" Nov 24 16:00:00 crc kubenswrapper[4790]: I1124 16:00:00.175854 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400000-69vs5"] Nov 24 16:00:00 crc kubenswrapper[4790]: I1124 16:00:00.177913 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400000-69vs5" Nov 24 16:00:00 crc kubenswrapper[4790]: I1124 16:00:00.179764 4790 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 16:00:00 crc kubenswrapper[4790]: I1124 16:00:00.179992 4790 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 16:00:00 crc kubenswrapper[4790]: I1124 16:00:00.203389 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400000-69vs5"] Nov 24 16:00:00 crc kubenswrapper[4790]: I1124 16:00:00.269936 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pl9vx\" (UniqueName: \"kubernetes.io/projected/41eac02b-794f-4700-9ba5-ef9b50f4424b-kube-api-access-pl9vx\") pod \"collect-profiles-29400000-69vs5\" (UID: \"41eac02b-794f-4700-9ba5-ef9b50f4424b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400000-69vs5" Nov 24 16:00:00 crc kubenswrapper[4790]: I1124 16:00:00.270340 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/41eac02b-794f-4700-9ba5-ef9b50f4424b-config-volume\") pod \"collect-profiles-29400000-69vs5\" (UID: \"41eac02b-794f-4700-9ba5-ef9b50f4424b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400000-69vs5" Nov 24 16:00:00 crc kubenswrapper[4790]: I1124 16:00:00.270569 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/41eac02b-794f-4700-9ba5-ef9b50f4424b-secret-volume\") pod \"collect-profiles-29400000-69vs5\" (UID: \"41eac02b-794f-4700-9ba5-ef9b50f4424b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400000-69vs5" Nov 24 16:00:00 crc kubenswrapper[4790]: I1124 16:00:00.373648 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/41eac02b-794f-4700-9ba5-ef9b50f4424b-config-volume\") pod \"collect-profiles-29400000-69vs5\" (UID: \"41eac02b-794f-4700-9ba5-ef9b50f4424b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400000-69vs5" Nov 24 16:00:00 crc kubenswrapper[4790]: I1124 16:00:00.373818 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/41eac02b-794f-4700-9ba5-ef9b50f4424b-secret-volume\") pod \"collect-profiles-29400000-69vs5\" (UID: \"41eac02b-794f-4700-9ba5-ef9b50f4424b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400000-69vs5" Nov 24 16:00:00 crc kubenswrapper[4790]: I1124 16:00:00.373951 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pl9vx\" (UniqueName: \"kubernetes.io/projected/41eac02b-794f-4700-9ba5-ef9b50f4424b-kube-api-access-pl9vx\") pod \"collect-profiles-29400000-69vs5\" (UID: \"41eac02b-794f-4700-9ba5-ef9b50f4424b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400000-69vs5" Nov 24 16:00:00 crc kubenswrapper[4790]: I1124 16:00:00.375402 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/41eac02b-794f-4700-9ba5-ef9b50f4424b-config-volume\") pod \"collect-profiles-29400000-69vs5\" (UID: \"41eac02b-794f-4700-9ba5-ef9b50f4424b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400000-69vs5" Nov 24 16:00:00 crc kubenswrapper[4790]: I1124 16:00:00.382924 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/41eac02b-794f-4700-9ba5-ef9b50f4424b-secret-volume\") pod \"collect-profiles-29400000-69vs5\" (UID: \"41eac02b-794f-4700-9ba5-ef9b50f4424b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400000-69vs5" Nov 24 16:00:00 crc kubenswrapper[4790]: I1124 16:00:00.394948 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pl9vx\" (UniqueName: \"kubernetes.io/projected/41eac02b-794f-4700-9ba5-ef9b50f4424b-kube-api-access-pl9vx\") pod \"collect-profiles-29400000-69vs5\" (UID: \"41eac02b-794f-4700-9ba5-ef9b50f4424b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400000-69vs5" Nov 24 16:00:00 crc kubenswrapper[4790]: I1124 16:00:00.503676 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400000-69vs5" Nov 24 16:00:00 crc kubenswrapper[4790]: I1124 16:00:00.505036 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kb6g6" podUID="445af5b5-5fa0-4e0d-b760-524775e030fd" containerName="registry-server" probeResult="failure" output=< Nov 24 16:00:00 crc kubenswrapper[4790]: timeout: failed to connect service ":50051" within 1s Nov 24 16:00:00 crc kubenswrapper[4790]: > Nov 24 16:00:02 crc kubenswrapper[4790]: I1124 16:00:02.387204 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400000-69vs5"] Nov 24 16:00:03 crc kubenswrapper[4790]: I1124 16:00:03.199568 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400000-69vs5" event={"ID":"41eac02b-794f-4700-9ba5-ef9b50f4424b","Type":"ContainerStarted","Data":"888d0e451c47a9d0986938e86346efc71e91a4fa1513cfa332efdfee15a52f52"} Nov 24 16:00:03 crc kubenswrapper[4790]: I1124 16:00:03.200144 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400000-69vs5" event={"ID":"41eac02b-794f-4700-9ba5-ef9b50f4424b","Type":"ContainerStarted","Data":"af870fb8670781b59ed8faa712f4509bd8aa5583a42f3dc99c4d5668692206de"} Nov 24 16:00:03 crc kubenswrapper[4790]: I1124 16:00:03.217380 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29400000-69vs5" podStartSLOduration=3.217362957 podStartE2EDuration="3.217362957s" podCreationTimestamp="2025-11-24 16:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 16:00:03.212618671 +0000 UTC m=+10051.592512333" watchObservedRunningTime="2025-11-24 16:00:03.217362957 +0000 UTC m=+10051.597256619" Nov 24 16:00:04 crc kubenswrapper[4790]: I1124 16:00:04.213783 4790 generic.go:334] "Generic (PLEG): container finished" podID="41eac02b-794f-4700-9ba5-ef9b50f4424b" containerID="888d0e451c47a9d0986938e86346efc71e91a4fa1513cfa332efdfee15a52f52" exitCode=0 Nov 24 16:00:04 crc kubenswrapper[4790]: I1124 16:00:04.214128 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400000-69vs5" event={"ID":"41eac02b-794f-4700-9ba5-ef9b50f4424b","Type":"ContainerDied","Data":"888d0e451c47a9d0986938e86346efc71e91a4fa1513cfa332efdfee15a52f52"} Nov 24 16:00:06 crc kubenswrapper[4790]: I1124 16:00:06.383608 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400000-69vs5" Nov 24 16:00:06 crc kubenswrapper[4790]: I1124 16:00:06.556835 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/41eac02b-794f-4700-9ba5-ef9b50f4424b-config-volume\") pod \"41eac02b-794f-4700-9ba5-ef9b50f4424b\" (UID: \"41eac02b-794f-4700-9ba5-ef9b50f4424b\") " Nov 24 16:00:06 crc kubenswrapper[4790]: I1124 16:00:06.556977 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/41eac02b-794f-4700-9ba5-ef9b50f4424b-secret-volume\") pod \"41eac02b-794f-4700-9ba5-ef9b50f4424b\" (UID: \"41eac02b-794f-4700-9ba5-ef9b50f4424b\") " Nov 24 16:00:06 crc kubenswrapper[4790]: I1124 16:00:06.557024 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pl9vx\" (UniqueName: \"kubernetes.io/projected/41eac02b-794f-4700-9ba5-ef9b50f4424b-kube-api-access-pl9vx\") pod \"41eac02b-794f-4700-9ba5-ef9b50f4424b\" (UID: \"41eac02b-794f-4700-9ba5-ef9b50f4424b\") " Nov 24 16:00:06 crc kubenswrapper[4790]: I1124 16:00:06.557729 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41eac02b-794f-4700-9ba5-ef9b50f4424b-config-volume" (OuterVolumeSpecName: "config-volume") pod "41eac02b-794f-4700-9ba5-ef9b50f4424b" (UID: "41eac02b-794f-4700-9ba5-ef9b50f4424b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 16:00:06 crc kubenswrapper[4790]: I1124 16:00:06.564452 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41eac02b-794f-4700-9ba5-ef9b50f4424b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "41eac02b-794f-4700-9ba5-ef9b50f4424b" (UID: "41eac02b-794f-4700-9ba5-ef9b50f4424b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 16:00:06 crc kubenswrapper[4790]: I1124 16:00:06.565332 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41eac02b-794f-4700-9ba5-ef9b50f4424b-kube-api-access-pl9vx" (OuterVolumeSpecName: "kube-api-access-pl9vx") pod "41eac02b-794f-4700-9ba5-ef9b50f4424b" (UID: "41eac02b-794f-4700-9ba5-ef9b50f4424b"). InnerVolumeSpecName "kube-api-access-pl9vx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 16:00:06 crc kubenswrapper[4790]: I1124 16:00:06.660966 4790 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/41eac02b-794f-4700-9ba5-ef9b50f4424b-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 16:00:06 crc kubenswrapper[4790]: I1124 16:00:06.660994 4790 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/41eac02b-794f-4700-9ba5-ef9b50f4424b-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 16:00:06 crc kubenswrapper[4790]: I1124 16:00:06.661004 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pl9vx\" (UniqueName: \"kubernetes.io/projected/41eac02b-794f-4700-9ba5-ef9b50f4424b-kube-api-access-pl9vx\") on node \"crc\" DevicePath \"\"" Nov 24 16:00:07 crc kubenswrapper[4790]: I1124 16:00:07.245600 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400000-69vs5" event={"ID":"41eac02b-794f-4700-9ba5-ef9b50f4424b","Type":"ContainerDied","Data":"af870fb8670781b59ed8faa712f4509bd8aa5583a42f3dc99c4d5668692206de"} Nov 24 16:00:07 crc kubenswrapper[4790]: I1124 16:00:07.245641 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af870fb8670781b59ed8faa712f4509bd8aa5583a42f3dc99c4d5668692206de" Nov 24 16:00:07 crc kubenswrapper[4790]: I1124 16:00:07.245694 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400000-69vs5" Nov 24 16:00:07 crc kubenswrapper[4790]: I1124 16:00:07.534342 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399955-nvl5w"] Nov 24 16:00:07 crc kubenswrapper[4790]: I1124 16:00:07.547608 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399955-nvl5w"] Nov 24 16:00:08 crc kubenswrapper[4790]: I1124 16:00:08.257286 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cd6sz/must-gather-99rhs" event={"ID":"7ac3f1cd-6b96-4378-949e-fc4ece0e6a09","Type":"ContainerStarted","Data":"bb6a206675c4846e0b28f4384fde2dd72e125d332d379af1ef22271b9e9ceb6a"} Nov 24 16:00:08 crc kubenswrapper[4790]: I1124 16:00:08.338688 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="833d9b1e-3d1a-4269-8dbb-26091e6479be" path="/var/lib/kubelet/pods/833d9b1e-3d1a-4269-8dbb-26091e6479be/volumes" Nov 24 16:00:09 crc kubenswrapper[4790]: I1124 16:00:09.280903 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cd6sz/must-gather-99rhs" event={"ID":"7ac3f1cd-6b96-4378-949e-fc4ece0e6a09","Type":"ContainerStarted","Data":"dedf4c468ab3948b7ae8302d054d763bef0474421cc2dcb428bb0320d7418a04"} Nov 24 16:00:09 crc kubenswrapper[4790]: I1124 16:00:09.318400 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-cd6sz/must-gather-99rhs" podStartSLOduration=4.238838089 podStartE2EDuration="17.3183827s" podCreationTimestamp="2025-11-24 15:59:52 +0000 UTC" firstStartedPulling="2025-11-24 15:59:54.203141243 +0000 UTC m=+10042.583034905" lastFinishedPulling="2025-11-24 16:00:07.282685854 +0000 UTC m=+10055.662579516" observedRunningTime="2025-11-24 16:00:09.310388312 +0000 UTC m=+10057.690281974" watchObservedRunningTime="2025-11-24 16:00:09.3183827 +0000 UTC m=+10057.698276362" Nov 24 16:00:10 crc kubenswrapper[4790]: I1124 16:00:10.507933 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kb6g6" podUID="445af5b5-5fa0-4e0d-b760-524775e030fd" containerName="registry-server" probeResult="failure" output=< Nov 24 16:00:10 crc kubenswrapper[4790]: timeout: failed to connect service ":50051" within 1s Nov 24 16:00:10 crc kubenswrapper[4790]: > Nov 24 16:00:12 crc kubenswrapper[4790]: I1124 16:00:12.193515 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-cd6sz/crc-debug-wgfrf"] Nov 24 16:00:12 crc kubenswrapper[4790]: E1124 16:00:12.194728 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41eac02b-794f-4700-9ba5-ef9b50f4424b" containerName="collect-profiles" Nov 24 16:00:12 crc kubenswrapper[4790]: I1124 16:00:12.194745 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="41eac02b-794f-4700-9ba5-ef9b50f4424b" containerName="collect-profiles" Nov 24 16:00:12 crc kubenswrapper[4790]: I1124 16:00:12.195063 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="41eac02b-794f-4700-9ba5-ef9b50f4424b" containerName="collect-profiles" Nov 24 16:00:12 crc kubenswrapper[4790]: I1124 16:00:12.196125 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cd6sz/crc-debug-wgfrf" Nov 24 16:00:12 crc kubenswrapper[4790]: I1124 16:00:12.287323 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fa001a3d-6aae-4cb7-9f48-e1adc714ad8f-host\") pod \"crc-debug-wgfrf\" (UID: \"fa001a3d-6aae-4cb7-9f48-e1adc714ad8f\") " pod="openshift-must-gather-cd6sz/crc-debug-wgfrf" Nov 24 16:00:12 crc kubenswrapper[4790]: I1124 16:00:12.287748 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bngcp\" (UniqueName: \"kubernetes.io/projected/fa001a3d-6aae-4cb7-9f48-e1adc714ad8f-kube-api-access-bngcp\") pod \"crc-debug-wgfrf\" (UID: \"fa001a3d-6aae-4cb7-9f48-e1adc714ad8f\") " pod="openshift-must-gather-cd6sz/crc-debug-wgfrf" Nov 24 16:00:12 crc kubenswrapper[4790]: I1124 16:00:12.390978 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bngcp\" (UniqueName: \"kubernetes.io/projected/fa001a3d-6aae-4cb7-9f48-e1adc714ad8f-kube-api-access-bngcp\") pod \"crc-debug-wgfrf\" (UID: \"fa001a3d-6aae-4cb7-9f48-e1adc714ad8f\") " pod="openshift-must-gather-cd6sz/crc-debug-wgfrf" Nov 24 16:00:12 crc kubenswrapper[4790]: I1124 16:00:12.391188 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fa001a3d-6aae-4cb7-9f48-e1adc714ad8f-host\") pod \"crc-debug-wgfrf\" (UID: \"fa001a3d-6aae-4cb7-9f48-e1adc714ad8f\") " pod="openshift-must-gather-cd6sz/crc-debug-wgfrf" Nov 24 16:00:12 crc kubenswrapper[4790]: I1124 16:00:12.391480 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fa001a3d-6aae-4cb7-9f48-e1adc714ad8f-host\") pod \"crc-debug-wgfrf\" (UID: \"fa001a3d-6aae-4cb7-9f48-e1adc714ad8f\") " pod="openshift-must-gather-cd6sz/crc-debug-wgfrf" Nov 24 16:00:12 crc kubenswrapper[4790]: I1124 16:00:12.418410 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bngcp\" (UniqueName: \"kubernetes.io/projected/fa001a3d-6aae-4cb7-9f48-e1adc714ad8f-kube-api-access-bngcp\") pod \"crc-debug-wgfrf\" (UID: \"fa001a3d-6aae-4cb7-9f48-e1adc714ad8f\") " pod="openshift-must-gather-cd6sz/crc-debug-wgfrf" Nov 24 16:00:12 crc kubenswrapper[4790]: I1124 16:00:12.531983 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cd6sz/crc-debug-wgfrf" Nov 24 16:00:12 crc kubenswrapper[4790]: W1124 16:00:12.579251 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfa001a3d_6aae_4cb7_9f48_e1adc714ad8f.slice/crio-304a48c3b5ec183a8a53c52e1d469b244eb5f3624b2435f9fff6bcdf55af721c WatchSource:0}: Error finding container 304a48c3b5ec183a8a53c52e1d469b244eb5f3624b2435f9fff6bcdf55af721c: Status 404 returned error can't find the container with id 304a48c3b5ec183a8a53c52e1d469b244eb5f3624b2435f9fff6bcdf55af721c Nov 24 16:00:13 crc kubenswrapper[4790]: I1124 16:00:13.324711 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cd6sz/crc-debug-wgfrf" event={"ID":"fa001a3d-6aae-4cb7-9f48-e1adc714ad8f","Type":"ContainerStarted","Data":"304a48c3b5ec183a8a53c52e1d469b244eb5f3624b2435f9fff6bcdf55af721c"} Nov 24 16:00:13 crc kubenswrapper[4790]: I1124 16:00:13.938422 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 16:00:13 crc kubenswrapper[4790]: I1124 16:00:13.938538 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 16:00:20 crc kubenswrapper[4790]: I1124 16:00:20.476594 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kb6g6" podUID="445af5b5-5fa0-4e0d-b760-524775e030fd" containerName="registry-server" probeResult="failure" output=< Nov 24 16:00:20 crc kubenswrapper[4790]: timeout: failed to connect service ":50051" within 1s Nov 24 16:00:20 crc kubenswrapper[4790]: > Nov 24 16:00:30 crc kubenswrapper[4790]: I1124 16:00:30.482550 4790 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kb6g6" podUID="445af5b5-5fa0-4e0d-b760-524775e030fd" containerName="registry-server" probeResult="failure" output=< Nov 24 16:00:30 crc kubenswrapper[4790]: timeout: failed to connect service ":50051" within 1s Nov 24 16:00:30 crc kubenswrapper[4790]: > Nov 24 16:00:31 crc kubenswrapper[4790]: E1124 16:00:31.623740 4790 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296" Nov 24 16:00:31 crc kubenswrapper[4790]: E1124 16:00:31.624340 4790 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:container-00,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296,Command:[chroot /host bash -c echo 'TOOLBOX_NAME=toolbox-osp' > /root/.toolboxrc ; rm -rf \"/var/tmp/sos-osp\" && mkdir -p \"/var/tmp/sos-osp\" && sudo podman rm --force toolbox-osp; sudo --preserve-env podman pull --authfile /var/lib/kubelet/config.json registry.redhat.io/rhel9/support-tools && toolbox sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir=\"/var/tmp/sos-osp\" && if [[ \"$(ls /var/log/pods/*/{*.log.*,*/*.log.*} 2>/dev/null)\" != '' ]]; then tar --ignore-failed-read --warning=no-file-changed -cJf \"/var/tmp/sos-osp/podlogs.tar.xz\" --transform 's,^,podlogs/,' /var/log/pods/*/{*.log.*,*/*.log.*} || true; fi],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:TMOUT,Value:900,ValueFrom:nil,},EnvVar{Name:HOST,Value:/host,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host,ReadOnly:false,MountPath:/host,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bngcp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod crc-debug-wgfrf_openshift-must-gather-cd6sz(fa001a3d-6aae-4cb7-9f48-e1adc714ad8f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 16:00:31 crc kubenswrapper[4790]: E1124 16:00:31.625915 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openshift-must-gather-cd6sz/crc-debug-wgfrf" podUID="fa001a3d-6aae-4cb7-9f48-e1adc714ad8f" Nov 24 16:00:32 crc kubenswrapper[4790]: E1124 16:00:32.515438 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296\\\"\"" pod="openshift-must-gather-cd6sz/crc-debug-wgfrf" podUID="fa001a3d-6aae-4cb7-9f48-e1adc714ad8f" Nov 24 16:00:33 crc kubenswrapper[4790]: I1124 16:00:33.745935 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qs924"] Nov 24 16:00:33 crc kubenswrapper[4790]: I1124 16:00:33.748822 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qs924" Nov 24 16:00:33 crc kubenswrapper[4790]: I1124 16:00:33.762850 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qs924"] Nov 24 16:00:33 crc kubenswrapper[4790]: I1124 16:00:33.884398 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f64ns\" (UniqueName: \"kubernetes.io/projected/017da2fa-87fc-44a6-9eb0-e0e1d4601633-kube-api-access-f64ns\") pod \"community-operators-qs924\" (UID: \"017da2fa-87fc-44a6-9eb0-e0e1d4601633\") " pod="openshift-marketplace/community-operators-qs924" Nov 24 16:00:33 crc kubenswrapper[4790]: I1124 16:00:33.884749 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/017da2fa-87fc-44a6-9eb0-e0e1d4601633-utilities\") pod \"community-operators-qs924\" (UID: \"017da2fa-87fc-44a6-9eb0-e0e1d4601633\") " pod="openshift-marketplace/community-operators-qs924" Nov 24 16:00:33 crc kubenswrapper[4790]: I1124 16:00:33.884778 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/017da2fa-87fc-44a6-9eb0-e0e1d4601633-catalog-content\") pod \"community-operators-qs924\" (UID: \"017da2fa-87fc-44a6-9eb0-e0e1d4601633\") " pod="openshift-marketplace/community-operators-qs924" Nov 24 16:00:33 crc kubenswrapper[4790]: I1124 16:00:33.986163 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/017da2fa-87fc-44a6-9eb0-e0e1d4601633-utilities\") pod \"community-operators-qs924\" (UID: \"017da2fa-87fc-44a6-9eb0-e0e1d4601633\") " pod="openshift-marketplace/community-operators-qs924" Nov 24 16:00:33 crc kubenswrapper[4790]: I1124 16:00:33.986213 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/017da2fa-87fc-44a6-9eb0-e0e1d4601633-catalog-content\") pod \"community-operators-qs924\" (UID: \"017da2fa-87fc-44a6-9eb0-e0e1d4601633\") " pod="openshift-marketplace/community-operators-qs924" Nov 24 16:00:33 crc kubenswrapper[4790]: I1124 16:00:33.986303 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f64ns\" (UniqueName: \"kubernetes.io/projected/017da2fa-87fc-44a6-9eb0-e0e1d4601633-kube-api-access-f64ns\") pod \"community-operators-qs924\" (UID: \"017da2fa-87fc-44a6-9eb0-e0e1d4601633\") " pod="openshift-marketplace/community-operators-qs924" Nov 24 16:00:33 crc kubenswrapper[4790]: I1124 16:00:33.994760 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/017da2fa-87fc-44a6-9eb0-e0e1d4601633-utilities\") pod \"community-operators-qs924\" (UID: \"017da2fa-87fc-44a6-9eb0-e0e1d4601633\") " pod="openshift-marketplace/community-operators-qs924" Nov 24 16:00:33 crc kubenswrapper[4790]: I1124 16:00:33.994788 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/017da2fa-87fc-44a6-9eb0-e0e1d4601633-catalog-content\") pod \"community-operators-qs924\" (UID: \"017da2fa-87fc-44a6-9eb0-e0e1d4601633\") " pod="openshift-marketplace/community-operators-qs924" Nov 24 16:00:34 crc kubenswrapper[4790]: I1124 16:00:34.013441 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f64ns\" (UniqueName: \"kubernetes.io/projected/017da2fa-87fc-44a6-9eb0-e0e1d4601633-kube-api-access-f64ns\") pod \"community-operators-qs924\" (UID: \"017da2fa-87fc-44a6-9eb0-e0e1d4601633\") " pod="openshift-marketplace/community-operators-qs924" Nov 24 16:00:34 crc kubenswrapper[4790]: I1124 16:00:34.082173 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qs924" Nov 24 16:00:34 crc kubenswrapper[4790]: I1124 16:00:34.752751 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qs924"] Nov 24 16:00:35 crc kubenswrapper[4790]: E1124 16:00:35.148704 4790 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod017da2fa_87fc_44a6_9eb0_e0e1d4601633.slice/crio-conmon-96dadac1271f149ce3b7e2c86772507b7c2da99219c7fb71eb4091defabc5ac1.scope\": RecentStats: unable to find data in memory cache]" Nov 24 16:00:35 crc kubenswrapper[4790]: I1124 16:00:35.549833 4790 generic.go:334] "Generic (PLEG): container finished" podID="017da2fa-87fc-44a6-9eb0-e0e1d4601633" containerID="96dadac1271f149ce3b7e2c86772507b7c2da99219c7fb71eb4091defabc5ac1" exitCode=0 Nov 24 16:00:35 crc kubenswrapper[4790]: I1124 16:00:35.549909 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qs924" event={"ID":"017da2fa-87fc-44a6-9eb0-e0e1d4601633","Type":"ContainerDied","Data":"96dadac1271f149ce3b7e2c86772507b7c2da99219c7fb71eb4091defabc5ac1"} Nov 24 16:00:35 crc kubenswrapper[4790]: I1124 16:00:35.550167 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qs924" event={"ID":"017da2fa-87fc-44a6-9eb0-e0e1d4601633","Type":"ContainerStarted","Data":"caf01b7c87154ed96b7e0601bcc8a3b550fe0397d89dd5a881cb530780a2f053"} Nov 24 16:00:37 crc kubenswrapper[4790]: I1124 16:00:37.570635 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qs924" event={"ID":"017da2fa-87fc-44a6-9eb0-e0e1d4601633","Type":"ContainerStarted","Data":"c7312c7ab3927d281e85c6afb31e3ca0d4ab4a664c11f8af03ee1b935357d11c"} Nov 24 16:00:39 crc kubenswrapper[4790]: I1124 16:00:39.498246 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kb6g6" Nov 24 16:00:39 crc kubenswrapper[4790]: I1124 16:00:39.556311 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kb6g6" Nov 24 16:00:40 crc kubenswrapper[4790]: I1124 16:00:40.129684 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kb6g6"] Nov 24 16:00:40 crc kubenswrapper[4790]: I1124 16:00:40.602608 4790 generic.go:334] "Generic (PLEG): container finished" podID="017da2fa-87fc-44a6-9eb0-e0e1d4601633" containerID="c7312c7ab3927d281e85c6afb31e3ca0d4ab4a664c11f8af03ee1b935357d11c" exitCode=0 Nov 24 16:00:40 crc kubenswrapper[4790]: I1124 16:00:40.602687 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qs924" event={"ID":"017da2fa-87fc-44a6-9eb0-e0e1d4601633","Type":"ContainerDied","Data":"c7312c7ab3927d281e85c6afb31e3ca0d4ab4a664c11f8af03ee1b935357d11c"} Nov 24 16:00:40 crc kubenswrapper[4790]: I1124 16:00:40.604227 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kb6g6" podUID="445af5b5-5fa0-4e0d-b760-524775e030fd" containerName="registry-server" containerID="cri-o://6d2469efb1324ac75244bf3f9b55e86e6a1c897ec87430db044498345201a644" gracePeriod=2 Nov 24 16:00:41 crc kubenswrapper[4790]: I1124 16:00:41.224537 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kb6g6" Nov 24 16:00:41 crc kubenswrapper[4790]: I1124 16:00:41.346519 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/445af5b5-5fa0-4e0d-b760-524775e030fd-catalog-content\") pod \"445af5b5-5fa0-4e0d-b760-524775e030fd\" (UID: \"445af5b5-5fa0-4e0d-b760-524775e030fd\") " Nov 24 16:00:41 crc kubenswrapper[4790]: I1124 16:00:41.346709 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z95ll\" (UniqueName: \"kubernetes.io/projected/445af5b5-5fa0-4e0d-b760-524775e030fd-kube-api-access-z95ll\") pod \"445af5b5-5fa0-4e0d-b760-524775e030fd\" (UID: \"445af5b5-5fa0-4e0d-b760-524775e030fd\") " Nov 24 16:00:41 crc kubenswrapper[4790]: I1124 16:00:41.346923 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/445af5b5-5fa0-4e0d-b760-524775e030fd-utilities\") pod \"445af5b5-5fa0-4e0d-b760-524775e030fd\" (UID: \"445af5b5-5fa0-4e0d-b760-524775e030fd\") " Nov 24 16:00:41 crc kubenswrapper[4790]: I1124 16:00:41.347436 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/445af5b5-5fa0-4e0d-b760-524775e030fd-utilities" (OuterVolumeSpecName: "utilities") pod "445af5b5-5fa0-4e0d-b760-524775e030fd" (UID: "445af5b5-5fa0-4e0d-b760-524775e030fd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 16:00:41 crc kubenswrapper[4790]: I1124 16:00:41.359571 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/445af5b5-5fa0-4e0d-b760-524775e030fd-kube-api-access-z95ll" (OuterVolumeSpecName: "kube-api-access-z95ll") pod "445af5b5-5fa0-4e0d-b760-524775e030fd" (UID: "445af5b5-5fa0-4e0d-b760-524775e030fd"). InnerVolumeSpecName "kube-api-access-z95ll". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 16:00:41 crc kubenswrapper[4790]: I1124 16:00:41.438662 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/445af5b5-5fa0-4e0d-b760-524775e030fd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "445af5b5-5fa0-4e0d-b760-524775e030fd" (UID: "445af5b5-5fa0-4e0d-b760-524775e030fd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 16:00:41 crc kubenswrapper[4790]: I1124 16:00:41.449805 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/445af5b5-5fa0-4e0d-b760-524775e030fd-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 16:00:41 crc kubenswrapper[4790]: I1124 16:00:41.449853 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z95ll\" (UniqueName: \"kubernetes.io/projected/445af5b5-5fa0-4e0d-b760-524775e030fd-kube-api-access-z95ll\") on node \"crc\" DevicePath \"\"" Nov 24 16:00:41 crc kubenswrapper[4790]: I1124 16:00:41.449869 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/445af5b5-5fa0-4e0d-b760-524775e030fd-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 16:00:41 crc kubenswrapper[4790]: I1124 16:00:41.617760 4790 generic.go:334] "Generic (PLEG): container finished" podID="445af5b5-5fa0-4e0d-b760-524775e030fd" containerID="6d2469efb1324ac75244bf3f9b55e86e6a1c897ec87430db044498345201a644" exitCode=0 Nov 24 16:00:41 crc kubenswrapper[4790]: I1124 16:00:41.617811 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kb6g6" Nov 24 16:00:41 crc kubenswrapper[4790]: I1124 16:00:41.617831 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kb6g6" event={"ID":"445af5b5-5fa0-4e0d-b760-524775e030fd","Type":"ContainerDied","Data":"6d2469efb1324ac75244bf3f9b55e86e6a1c897ec87430db044498345201a644"} Nov 24 16:00:41 crc kubenswrapper[4790]: I1124 16:00:41.617901 4790 scope.go:117] "RemoveContainer" containerID="6d2469efb1324ac75244bf3f9b55e86e6a1c897ec87430db044498345201a644" Nov 24 16:00:41 crc kubenswrapper[4790]: I1124 16:00:41.617925 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kb6g6" event={"ID":"445af5b5-5fa0-4e0d-b760-524775e030fd","Type":"ContainerDied","Data":"a5eb1d539ed8ca681dc1927b808c5139020d2a32c69829355033444ddf353746"} Nov 24 16:00:41 crc kubenswrapper[4790]: I1124 16:00:41.620380 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qs924" event={"ID":"017da2fa-87fc-44a6-9eb0-e0e1d4601633","Type":"ContainerStarted","Data":"36ca0862a5f273c118cdceeb669d389587372b66bee52d9c890b68afefaace97"} Nov 24 16:00:41 crc kubenswrapper[4790]: I1124 16:00:41.641233 4790 scope.go:117] "RemoveContainer" containerID="fe3432b1f09e0f41b020d28b779ec96d4436f56a479bdfb4f71bfe2a8ea4ad4b" Nov 24 16:00:41 crc kubenswrapper[4790]: I1124 16:00:41.644785 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qs924" podStartSLOduration=3.021175715 podStartE2EDuration="8.644762277s" podCreationTimestamp="2025-11-24 16:00:33 +0000 UTC" firstStartedPulling="2025-11-24 16:00:35.551716922 +0000 UTC m=+10083.931610584" lastFinishedPulling="2025-11-24 16:00:41.175303484 +0000 UTC m=+10089.555197146" observedRunningTime="2025-11-24 16:00:41.637265524 +0000 UTC m=+10090.017159186" watchObservedRunningTime="2025-11-24 16:00:41.644762277 +0000 UTC m=+10090.024655959" Nov 24 16:00:41 crc kubenswrapper[4790]: I1124 16:00:41.662984 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kb6g6"] Nov 24 16:00:41 crc kubenswrapper[4790]: I1124 16:00:41.673812 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kb6g6"] Nov 24 16:00:41 crc kubenswrapper[4790]: I1124 16:00:41.674411 4790 scope.go:117] "RemoveContainer" containerID="556229960b0ccb5628d9ddb94b5461bcec8f5319f3762a5b9bdf7f0b2cfc6889" Nov 24 16:00:41 crc kubenswrapper[4790]: I1124 16:00:41.695341 4790 scope.go:117] "RemoveContainer" containerID="6d2469efb1324ac75244bf3f9b55e86e6a1c897ec87430db044498345201a644" Nov 24 16:00:41 crc kubenswrapper[4790]: E1124 16:00:41.695706 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d2469efb1324ac75244bf3f9b55e86e6a1c897ec87430db044498345201a644\": container with ID starting with 6d2469efb1324ac75244bf3f9b55e86e6a1c897ec87430db044498345201a644 not found: ID does not exist" containerID="6d2469efb1324ac75244bf3f9b55e86e6a1c897ec87430db044498345201a644" Nov 24 16:00:41 crc kubenswrapper[4790]: I1124 16:00:41.695737 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d2469efb1324ac75244bf3f9b55e86e6a1c897ec87430db044498345201a644"} err="failed to get container status \"6d2469efb1324ac75244bf3f9b55e86e6a1c897ec87430db044498345201a644\": rpc error: code = NotFound desc = could not find container \"6d2469efb1324ac75244bf3f9b55e86e6a1c897ec87430db044498345201a644\": container with ID starting with 6d2469efb1324ac75244bf3f9b55e86e6a1c897ec87430db044498345201a644 not found: ID does not exist" Nov 24 16:00:41 crc kubenswrapper[4790]: I1124 16:00:41.695758 4790 scope.go:117] "RemoveContainer" containerID="fe3432b1f09e0f41b020d28b779ec96d4436f56a479bdfb4f71bfe2a8ea4ad4b" Nov 24 16:00:41 crc kubenswrapper[4790]: E1124 16:00:41.695972 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe3432b1f09e0f41b020d28b779ec96d4436f56a479bdfb4f71bfe2a8ea4ad4b\": container with ID starting with fe3432b1f09e0f41b020d28b779ec96d4436f56a479bdfb4f71bfe2a8ea4ad4b not found: ID does not exist" containerID="fe3432b1f09e0f41b020d28b779ec96d4436f56a479bdfb4f71bfe2a8ea4ad4b" Nov 24 16:00:41 crc kubenswrapper[4790]: I1124 16:00:41.696011 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe3432b1f09e0f41b020d28b779ec96d4436f56a479bdfb4f71bfe2a8ea4ad4b"} err="failed to get container status \"fe3432b1f09e0f41b020d28b779ec96d4436f56a479bdfb4f71bfe2a8ea4ad4b\": rpc error: code = NotFound desc = could not find container \"fe3432b1f09e0f41b020d28b779ec96d4436f56a479bdfb4f71bfe2a8ea4ad4b\": container with ID starting with fe3432b1f09e0f41b020d28b779ec96d4436f56a479bdfb4f71bfe2a8ea4ad4b not found: ID does not exist" Nov 24 16:00:41 crc kubenswrapper[4790]: I1124 16:00:41.696026 4790 scope.go:117] "RemoveContainer" containerID="556229960b0ccb5628d9ddb94b5461bcec8f5319f3762a5b9bdf7f0b2cfc6889" Nov 24 16:00:41 crc kubenswrapper[4790]: E1124 16:00:41.696894 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"556229960b0ccb5628d9ddb94b5461bcec8f5319f3762a5b9bdf7f0b2cfc6889\": container with ID starting with 556229960b0ccb5628d9ddb94b5461bcec8f5319f3762a5b9bdf7f0b2cfc6889 not found: ID does not exist" containerID="556229960b0ccb5628d9ddb94b5461bcec8f5319f3762a5b9bdf7f0b2cfc6889" Nov 24 16:00:41 crc kubenswrapper[4790]: I1124 16:00:41.696917 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"556229960b0ccb5628d9ddb94b5461bcec8f5319f3762a5b9bdf7f0b2cfc6889"} err="failed to get container status \"556229960b0ccb5628d9ddb94b5461bcec8f5319f3762a5b9bdf7f0b2cfc6889\": rpc error: code = NotFound desc = could not find container \"556229960b0ccb5628d9ddb94b5461bcec8f5319f3762a5b9bdf7f0b2cfc6889\": container with ID starting with 556229960b0ccb5628d9ddb94b5461bcec8f5319f3762a5b9bdf7f0b2cfc6889 not found: ID does not exist" Nov 24 16:00:42 crc kubenswrapper[4790]: I1124 16:00:42.347557 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="445af5b5-5fa0-4e0d-b760-524775e030fd" path="/var/lib/kubelet/pods/445af5b5-5fa0-4e0d-b760-524775e030fd/volumes" Nov 24 16:00:43 crc kubenswrapper[4790]: I1124 16:00:43.938754 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 16:00:43 crc kubenswrapper[4790]: I1124 16:00:43.939257 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 16:00:44 crc kubenswrapper[4790]: I1124 16:00:44.083154 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qs924" Nov 24 16:00:44 crc kubenswrapper[4790]: I1124 16:00:44.083218 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qs924" Nov 24 16:00:44 crc kubenswrapper[4790]: I1124 16:00:44.138355 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qs924" Nov 24 16:00:45 crc kubenswrapper[4790]: I1124 16:00:45.319356 4790 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 16:00:46 crc kubenswrapper[4790]: I1124 16:00:46.673706 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cd6sz/crc-debug-wgfrf" event={"ID":"fa001a3d-6aae-4cb7-9f48-e1adc714ad8f","Type":"ContainerStarted","Data":"3645ccf35e68622a20b515b8ff900fd24e114c48a1c96f5d6ae57ea7d2ae8843"} Nov 24 16:00:46 crc kubenswrapper[4790]: I1124 16:00:46.688116 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-cd6sz/crc-debug-wgfrf" podStartSLOduration=1.1594965 podStartE2EDuration="34.688101386s" podCreationTimestamp="2025-11-24 16:00:12 +0000 UTC" firstStartedPulling="2025-11-24 16:00:12.581470801 +0000 UTC m=+10060.961364463" lastFinishedPulling="2025-11-24 16:00:46.110075677 +0000 UTC m=+10094.489969349" observedRunningTime="2025-11-24 16:00:46.68612899 +0000 UTC m=+10095.066022652" watchObservedRunningTime="2025-11-24 16:00:46.688101386 +0000 UTC m=+10095.067995048" Nov 24 16:00:54 crc kubenswrapper[4790]: I1124 16:00:54.140477 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qs924" Nov 24 16:00:54 crc kubenswrapper[4790]: I1124 16:00:54.197979 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qs924"] Nov 24 16:00:54 crc kubenswrapper[4790]: I1124 16:00:54.750137 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qs924" podUID="017da2fa-87fc-44a6-9eb0-e0e1d4601633" containerName="registry-server" containerID="cri-o://36ca0862a5f273c118cdceeb669d389587372b66bee52d9c890b68afefaace97" gracePeriod=2 Nov 24 16:00:55 crc kubenswrapper[4790]: I1124 16:00:55.790942 4790 generic.go:334] "Generic (PLEG): container finished" podID="017da2fa-87fc-44a6-9eb0-e0e1d4601633" containerID="36ca0862a5f273c118cdceeb669d389587372b66bee52d9c890b68afefaace97" exitCode=0 Nov 24 16:00:55 crc kubenswrapper[4790]: I1124 16:00:55.791516 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qs924" event={"ID":"017da2fa-87fc-44a6-9eb0-e0e1d4601633","Type":"ContainerDied","Data":"36ca0862a5f273c118cdceeb669d389587372b66bee52d9c890b68afefaace97"} Nov 24 16:01:01 crc kubenswrapper[4790]: I1124 16:01:01.338751 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29400001-dvmpw"] Nov 24 16:01:01 crc kubenswrapper[4790]: E1124 16:01:01.339557 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="445af5b5-5fa0-4e0d-b760-524775e030fd" containerName="extract-content" Nov 24 16:01:01 crc kubenswrapper[4790]: I1124 16:01:01.339570 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="445af5b5-5fa0-4e0d-b760-524775e030fd" containerName="extract-content" Nov 24 16:01:01 crc kubenswrapper[4790]: E1124 16:01:01.339591 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="445af5b5-5fa0-4e0d-b760-524775e030fd" containerName="extract-utilities" Nov 24 16:01:01 crc kubenswrapper[4790]: I1124 16:01:01.339599 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="445af5b5-5fa0-4e0d-b760-524775e030fd" containerName="extract-utilities" Nov 24 16:01:01 crc kubenswrapper[4790]: E1124 16:01:01.339610 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="445af5b5-5fa0-4e0d-b760-524775e030fd" containerName="registry-server" Nov 24 16:01:01 crc kubenswrapper[4790]: I1124 16:01:01.339616 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="445af5b5-5fa0-4e0d-b760-524775e030fd" containerName="registry-server" Nov 24 16:01:01 crc kubenswrapper[4790]: I1124 16:01:01.340627 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="445af5b5-5fa0-4e0d-b760-524775e030fd" containerName="registry-server" Nov 24 16:01:01 crc kubenswrapper[4790]: I1124 16:01:01.341374 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29400001-dvmpw" Nov 24 16:01:01 crc kubenswrapper[4790]: I1124 16:01:01.356601 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29400001-dvmpw"] Nov 24 16:01:01 crc kubenswrapper[4790]: I1124 16:01:01.486168 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jw45k\" (UniqueName: \"kubernetes.io/projected/a0b00350-df63-40ed-995b-6c2a9178f7d7-kube-api-access-jw45k\") pod \"keystone-cron-29400001-dvmpw\" (UID: \"a0b00350-df63-40ed-995b-6c2a9178f7d7\") " pod="openstack/keystone-cron-29400001-dvmpw" Nov 24 16:01:01 crc kubenswrapper[4790]: I1124 16:01:01.486235 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a0b00350-df63-40ed-995b-6c2a9178f7d7-fernet-keys\") pod \"keystone-cron-29400001-dvmpw\" (UID: \"a0b00350-df63-40ed-995b-6c2a9178f7d7\") " pod="openstack/keystone-cron-29400001-dvmpw" Nov 24 16:01:01 crc kubenswrapper[4790]: I1124 16:01:01.486479 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0b00350-df63-40ed-995b-6c2a9178f7d7-combined-ca-bundle\") pod \"keystone-cron-29400001-dvmpw\" (UID: \"a0b00350-df63-40ed-995b-6c2a9178f7d7\") " pod="openstack/keystone-cron-29400001-dvmpw" Nov 24 16:01:01 crc kubenswrapper[4790]: I1124 16:01:01.486571 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0b00350-df63-40ed-995b-6c2a9178f7d7-config-data\") pod \"keystone-cron-29400001-dvmpw\" (UID: \"a0b00350-df63-40ed-995b-6c2a9178f7d7\") " pod="openstack/keystone-cron-29400001-dvmpw" Nov 24 16:01:01 crc kubenswrapper[4790]: I1124 16:01:01.498492 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qs924" Nov 24 16:01:01 crc kubenswrapper[4790]: I1124 16:01:01.589144 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/017da2fa-87fc-44a6-9eb0-e0e1d4601633-catalog-content\") pod \"017da2fa-87fc-44a6-9eb0-e0e1d4601633\" (UID: \"017da2fa-87fc-44a6-9eb0-e0e1d4601633\") " Nov 24 16:01:01 crc kubenswrapper[4790]: I1124 16:01:01.589257 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/017da2fa-87fc-44a6-9eb0-e0e1d4601633-utilities\") pod \"017da2fa-87fc-44a6-9eb0-e0e1d4601633\" (UID: \"017da2fa-87fc-44a6-9eb0-e0e1d4601633\") " Nov 24 16:01:01 crc kubenswrapper[4790]: I1124 16:01:01.589329 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f64ns\" (UniqueName: \"kubernetes.io/projected/017da2fa-87fc-44a6-9eb0-e0e1d4601633-kube-api-access-f64ns\") pod \"017da2fa-87fc-44a6-9eb0-e0e1d4601633\" (UID: \"017da2fa-87fc-44a6-9eb0-e0e1d4601633\") " Nov 24 16:01:01 crc kubenswrapper[4790]: I1124 16:01:01.589785 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0b00350-df63-40ed-995b-6c2a9178f7d7-combined-ca-bundle\") pod \"keystone-cron-29400001-dvmpw\" (UID: \"a0b00350-df63-40ed-995b-6c2a9178f7d7\") " pod="openstack/keystone-cron-29400001-dvmpw" Nov 24 16:01:01 crc kubenswrapper[4790]: I1124 16:01:01.589838 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0b00350-df63-40ed-995b-6c2a9178f7d7-config-data\") pod \"keystone-cron-29400001-dvmpw\" (UID: \"a0b00350-df63-40ed-995b-6c2a9178f7d7\") " pod="openstack/keystone-cron-29400001-dvmpw" Nov 24 16:01:01 crc kubenswrapper[4790]: I1124 16:01:01.589976 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jw45k\" (UniqueName: \"kubernetes.io/projected/a0b00350-df63-40ed-995b-6c2a9178f7d7-kube-api-access-jw45k\") pod \"keystone-cron-29400001-dvmpw\" (UID: \"a0b00350-df63-40ed-995b-6c2a9178f7d7\") " pod="openstack/keystone-cron-29400001-dvmpw" Nov 24 16:01:01 crc kubenswrapper[4790]: I1124 16:01:01.590029 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a0b00350-df63-40ed-995b-6c2a9178f7d7-fernet-keys\") pod \"keystone-cron-29400001-dvmpw\" (UID: \"a0b00350-df63-40ed-995b-6c2a9178f7d7\") " pod="openstack/keystone-cron-29400001-dvmpw" Nov 24 16:01:01 crc kubenswrapper[4790]: I1124 16:01:01.590305 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/017da2fa-87fc-44a6-9eb0-e0e1d4601633-utilities" (OuterVolumeSpecName: "utilities") pod "017da2fa-87fc-44a6-9eb0-e0e1d4601633" (UID: "017da2fa-87fc-44a6-9eb0-e0e1d4601633"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 16:01:01 crc kubenswrapper[4790]: I1124 16:01:01.648271 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/017da2fa-87fc-44a6-9eb0-e0e1d4601633-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "017da2fa-87fc-44a6-9eb0-e0e1d4601633" (UID: "017da2fa-87fc-44a6-9eb0-e0e1d4601633"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 16:01:01 crc kubenswrapper[4790]: I1124 16:01:01.692643 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/017da2fa-87fc-44a6-9eb0-e0e1d4601633-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 16:01:01 crc kubenswrapper[4790]: I1124 16:01:01.692687 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/017da2fa-87fc-44a6-9eb0-e0e1d4601633-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 16:01:02 crc kubenswrapper[4790]: I1124 16:01:02.193545 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/017da2fa-87fc-44a6-9eb0-e0e1d4601633-kube-api-access-f64ns" (OuterVolumeSpecName: "kube-api-access-f64ns") pod "017da2fa-87fc-44a6-9eb0-e0e1d4601633" (UID: "017da2fa-87fc-44a6-9eb0-e0e1d4601633"). InnerVolumeSpecName "kube-api-access-f64ns". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 16:01:02 crc kubenswrapper[4790]: I1124 16:01:02.193723 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0b00350-df63-40ed-995b-6c2a9178f7d7-combined-ca-bundle\") pod \"keystone-cron-29400001-dvmpw\" (UID: \"a0b00350-df63-40ed-995b-6c2a9178f7d7\") " pod="openstack/keystone-cron-29400001-dvmpw" Nov 24 16:01:02 crc kubenswrapper[4790]: I1124 16:01:02.193912 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a0b00350-df63-40ed-995b-6c2a9178f7d7-fernet-keys\") pod \"keystone-cron-29400001-dvmpw\" (UID: \"a0b00350-df63-40ed-995b-6c2a9178f7d7\") " pod="openstack/keystone-cron-29400001-dvmpw" Nov 24 16:01:02 crc kubenswrapper[4790]: I1124 16:01:02.194068 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jw45k\" (UniqueName: \"kubernetes.io/projected/a0b00350-df63-40ed-995b-6c2a9178f7d7-kube-api-access-jw45k\") pod \"keystone-cron-29400001-dvmpw\" (UID: \"a0b00350-df63-40ed-995b-6c2a9178f7d7\") " pod="openstack/keystone-cron-29400001-dvmpw" Nov 24 16:01:02 crc kubenswrapper[4790]: I1124 16:01:02.194949 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0b00350-df63-40ed-995b-6c2a9178f7d7-config-data\") pod \"keystone-cron-29400001-dvmpw\" (UID: \"a0b00350-df63-40ed-995b-6c2a9178f7d7\") " pod="openstack/keystone-cron-29400001-dvmpw" Nov 24 16:01:02 crc kubenswrapper[4790]: I1124 16:01:02.202850 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f64ns\" (UniqueName: \"kubernetes.io/projected/017da2fa-87fc-44a6-9eb0-e0e1d4601633-kube-api-access-f64ns\") on node \"crc\" DevicePath \"\"" Nov 24 16:01:02 crc kubenswrapper[4790]: I1124 16:01:02.297288 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qs924" event={"ID":"017da2fa-87fc-44a6-9eb0-e0e1d4601633","Type":"ContainerDied","Data":"caf01b7c87154ed96b7e0601bcc8a3b550fe0397d89dd5a881cb530780a2f053"} Nov 24 16:01:02 crc kubenswrapper[4790]: I1124 16:01:02.297343 4790 scope.go:117] "RemoveContainer" containerID="36ca0862a5f273c118cdceeb669d389587372b66bee52d9c890b68afefaace97" Nov 24 16:01:02 crc kubenswrapper[4790]: I1124 16:01:02.297413 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qs924" Nov 24 16:01:02 crc kubenswrapper[4790]: I1124 16:01:02.352186 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qs924"] Nov 24 16:01:02 crc kubenswrapper[4790]: I1124 16:01:02.363330 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qs924"] Nov 24 16:01:02 crc kubenswrapper[4790]: I1124 16:01:02.395187 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29400001-dvmpw" Nov 24 16:01:02 crc kubenswrapper[4790]: I1124 16:01:02.427337 4790 scope.go:117] "RemoveContainer" containerID="c7312c7ab3927d281e85c6afb31e3ca0d4ab4a664c11f8af03ee1b935357d11c" Nov 24 16:01:02 crc kubenswrapper[4790]: I1124 16:01:02.490501 4790 scope.go:117] "RemoveContainer" containerID="96dadac1271f149ce3b7e2c86772507b7c2da99219c7fb71eb4091defabc5ac1" Nov 24 16:01:03 crc kubenswrapper[4790]: I1124 16:01:03.019295 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29400001-dvmpw"] Nov 24 16:01:03 crc kubenswrapper[4790]: W1124 16:01:03.021348 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0b00350_df63_40ed_995b_6c2a9178f7d7.slice/crio-45c2369a29a6515ae4152b07a33a2d17aa65251fec7dc500c06b9fe2d4cf60c7 WatchSource:0}: Error finding container 45c2369a29a6515ae4152b07a33a2d17aa65251fec7dc500c06b9fe2d4cf60c7: Status 404 returned error can't find the container with id 45c2369a29a6515ae4152b07a33a2d17aa65251fec7dc500c06b9fe2d4cf60c7 Nov 24 16:01:03 crc kubenswrapper[4790]: I1124 16:01:03.311365 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29400001-dvmpw" event={"ID":"a0b00350-df63-40ed-995b-6c2a9178f7d7","Type":"ContainerStarted","Data":"246b663e58c9efdc9c9a754b342f119b1a0127feceb065b9304b73e241c43e69"} Nov 24 16:01:03 crc kubenswrapper[4790]: I1124 16:01:03.311687 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29400001-dvmpw" event={"ID":"a0b00350-df63-40ed-995b-6c2a9178f7d7","Type":"ContainerStarted","Data":"45c2369a29a6515ae4152b07a33a2d17aa65251fec7dc500c06b9fe2d4cf60c7"} Nov 24 16:01:03 crc kubenswrapper[4790]: I1124 16:01:03.331089 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29400001-dvmpw" podStartSLOduration=2.331070838 podStartE2EDuration="2.331070838s" podCreationTimestamp="2025-11-24 16:01:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 16:01:03.325708315 +0000 UTC m=+10111.705601977" watchObservedRunningTime="2025-11-24 16:01:03.331070838 +0000 UTC m=+10111.710964500" Nov 24 16:01:04 crc kubenswrapper[4790]: I1124 16:01:04.326317 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="017da2fa-87fc-44a6-9eb0-e0e1d4601633" path="/var/lib/kubelet/pods/017da2fa-87fc-44a6-9eb0-e0e1d4601633/volumes" Nov 24 16:01:05 crc kubenswrapper[4790]: I1124 16:01:05.063124 4790 scope.go:117] "RemoveContainer" containerID="f6629e9e00132a19cdb1e8cb2f35e186041658cfbc9bb94d2c71f1b5f442ca0d" Nov 24 16:01:07 crc kubenswrapper[4790]: I1124 16:01:07.376340 4790 generic.go:334] "Generic (PLEG): container finished" podID="a0b00350-df63-40ed-995b-6c2a9178f7d7" containerID="246b663e58c9efdc9c9a754b342f119b1a0127feceb065b9304b73e241c43e69" exitCode=0 Nov 24 16:01:07 crc kubenswrapper[4790]: I1124 16:01:07.376481 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29400001-dvmpw" event={"ID":"a0b00350-df63-40ed-995b-6c2a9178f7d7","Type":"ContainerDied","Data":"246b663e58c9efdc9c9a754b342f119b1a0127feceb065b9304b73e241c43e69"} Nov 24 16:01:09 crc kubenswrapper[4790]: I1124 16:01:09.161689 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29400001-dvmpw" Nov 24 16:01:09 crc kubenswrapper[4790]: I1124 16:01:09.269374 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jw45k\" (UniqueName: \"kubernetes.io/projected/a0b00350-df63-40ed-995b-6c2a9178f7d7-kube-api-access-jw45k\") pod \"a0b00350-df63-40ed-995b-6c2a9178f7d7\" (UID: \"a0b00350-df63-40ed-995b-6c2a9178f7d7\") " Nov 24 16:01:09 crc kubenswrapper[4790]: I1124 16:01:09.269591 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a0b00350-df63-40ed-995b-6c2a9178f7d7-fernet-keys\") pod \"a0b00350-df63-40ed-995b-6c2a9178f7d7\" (UID: \"a0b00350-df63-40ed-995b-6c2a9178f7d7\") " Nov 24 16:01:09 crc kubenswrapper[4790]: I1124 16:01:09.269621 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0b00350-df63-40ed-995b-6c2a9178f7d7-combined-ca-bundle\") pod \"a0b00350-df63-40ed-995b-6c2a9178f7d7\" (UID: \"a0b00350-df63-40ed-995b-6c2a9178f7d7\") " Nov 24 16:01:09 crc kubenswrapper[4790]: I1124 16:01:09.269749 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0b00350-df63-40ed-995b-6c2a9178f7d7-config-data\") pod \"a0b00350-df63-40ed-995b-6c2a9178f7d7\" (UID: \"a0b00350-df63-40ed-995b-6c2a9178f7d7\") " Nov 24 16:01:09 crc kubenswrapper[4790]: I1124 16:01:09.291944 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0b00350-df63-40ed-995b-6c2a9178f7d7-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "a0b00350-df63-40ed-995b-6c2a9178f7d7" (UID: "a0b00350-df63-40ed-995b-6c2a9178f7d7"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 16:01:09 crc kubenswrapper[4790]: I1124 16:01:09.294035 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0b00350-df63-40ed-995b-6c2a9178f7d7-kube-api-access-jw45k" (OuterVolumeSpecName: "kube-api-access-jw45k") pod "a0b00350-df63-40ed-995b-6c2a9178f7d7" (UID: "a0b00350-df63-40ed-995b-6c2a9178f7d7"). InnerVolumeSpecName "kube-api-access-jw45k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 16:01:09 crc kubenswrapper[4790]: I1124 16:01:09.341201 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0b00350-df63-40ed-995b-6c2a9178f7d7-config-data" (OuterVolumeSpecName: "config-data") pod "a0b00350-df63-40ed-995b-6c2a9178f7d7" (UID: "a0b00350-df63-40ed-995b-6c2a9178f7d7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 16:01:09 crc kubenswrapper[4790]: I1124 16:01:09.372943 4790 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a0b00350-df63-40ed-995b-6c2a9178f7d7-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 16:01:09 crc kubenswrapper[4790]: I1124 16:01:09.374353 4790 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0b00350-df63-40ed-995b-6c2a9178f7d7-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 16:01:09 crc kubenswrapper[4790]: I1124 16:01:09.374466 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jw45k\" (UniqueName: \"kubernetes.io/projected/a0b00350-df63-40ed-995b-6c2a9178f7d7-kube-api-access-jw45k\") on node \"crc\" DevicePath \"\"" Nov 24 16:01:09 crc kubenswrapper[4790]: I1124 16:01:09.380587 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0b00350-df63-40ed-995b-6c2a9178f7d7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a0b00350-df63-40ed-995b-6c2a9178f7d7" (UID: "a0b00350-df63-40ed-995b-6c2a9178f7d7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 16:01:09 crc kubenswrapper[4790]: I1124 16:01:09.405596 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29400001-dvmpw" event={"ID":"a0b00350-df63-40ed-995b-6c2a9178f7d7","Type":"ContainerDied","Data":"45c2369a29a6515ae4152b07a33a2d17aa65251fec7dc500c06b9fe2d4cf60c7"} Nov 24 16:01:09 crc kubenswrapper[4790]: I1124 16:01:09.405631 4790 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45c2369a29a6515ae4152b07a33a2d17aa65251fec7dc500c06b9fe2d4cf60c7" Nov 24 16:01:09 crc kubenswrapper[4790]: I1124 16:01:09.405717 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29400001-dvmpw" Nov 24 16:01:09 crc kubenswrapper[4790]: I1124 16:01:09.475990 4790 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0b00350-df63-40ed-995b-6c2a9178f7d7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 16:01:12 crc kubenswrapper[4790]: I1124 16:01:12.431163 4790 generic.go:334] "Generic (PLEG): container finished" podID="fa001a3d-6aae-4cb7-9f48-e1adc714ad8f" containerID="3645ccf35e68622a20b515b8ff900fd24e114c48a1c96f5d6ae57ea7d2ae8843" exitCode=0 Nov 24 16:01:12 crc kubenswrapper[4790]: I1124 16:01:12.431627 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cd6sz/crc-debug-wgfrf" event={"ID":"fa001a3d-6aae-4cb7-9f48-e1adc714ad8f","Type":"ContainerDied","Data":"3645ccf35e68622a20b515b8ff900fd24e114c48a1c96f5d6ae57ea7d2ae8843"} Nov 24 16:01:13 crc kubenswrapper[4790]: I1124 16:01:13.614429 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cd6sz/crc-debug-wgfrf" Nov 24 16:01:13 crc kubenswrapper[4790]: I1124 16:01:13.661358 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-cd6sz/crc-debug-wgfrf"] Nov 24 16:01:13 crc kubenswrapper[4790]: I1124 16:01:13.664126 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fa001a3d-6aae-4cb7-9f48-e1adc714ad8f-host\") pod \"fa001a3d-6aae-4cb7-9f48-e1adc714ad8f\" (UID: \"fa001a3d-6aae-4cb7-9f48-e1adc714ad8f\") " Nov 24 16:01:13 crc kubenswrapper[4790]: I1124 16:01:13.664278 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa001a3d-6aae-4cb7-9f48-e1adc714ad8f-host" (OuterVolumeSpecName: "host") pod "fa001a3d-6aae-4cb7-9f48-e1adc714ad8f" (UID: "fa001a3d-6aae-4cb7-9f48-e1adc714ad8f"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 16:01:13 crc kubenswrapper[4790]: I1124 16:01:13.664569 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bngcp\" (UniqueName: \"kubernetes.io/projected/fa001a3d-6aae-4cb7-9f48-e1adc714ad8f-kube-api-access-bngcp\") pod \"fa001a3d-6aae-4cb7-9f48-e1adc714ad8f\" (UID: \"fa001a3d-6aae-4cb7-9f48-e1adc714ad8f\") " Nov 24 16:01:13 crc kubenswrapper[4790]: I1124 16:01:13.665078 4790 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fa001a3d-6aae-4cb7-9f48-e1adc714ad8f-host\") on node \"crc\" DevicePath \"\"" Nov 24 16:01:13 crc kubenswrapper[4790]: I1124 16:01:13.670539 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa001a3d-6aae-4cb7-9f48-e1adc714ad8f-kube-api-access-bngcp" (OuterVolumeSpecName: "kube-api-access-bngcp") pod "fa001a3d-6aae-4cb7-9f48-e1adc714ad8f" (UID: "fa001a3d-6aae-4cb7-9f48-e1adc714ad8f"). InnerVolumeSpecName "kube-api-access-bngcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 16:01:13 crc kubenswrapper[4790]: I1124 16:01:13.691525 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-cd6sz/crc-debug-wgfrf"] Nov 24 16:01:13 crc kubenswrapper[4790]: I1124 16:01:13.767831 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bngcp\" (UniqueName: \"kubernetes.io/projected/fa001a3d-6aae-4cb7-9f48-e1adc714ad8f-kube-api-access-bngcp\") on node \"crc\" DevicePath \"\"" Nov 24 16:01:13 crc kubenswrapper[4790]: I1124 16:01:13.938576 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 16:01:13 crc kubenswrapper[4790]: I1124 16:01:13.938638 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 16:01:13 crc kubenswrapper[4790]: I1124 16:01:13.938677 4790 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 16:01:13 crc kubenswrapper[4790]: I1124 16:01:13.939519 4790 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"25209c017376e3ac46b1a33e19b0b7fe5acafd9ba6b9ced78aacd10037991a2f"} pod="openshift-machine-config-operator/machine-config-daemon-xz49t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 16:01:13 crc kubenswrapper[4790]: I1124 16:01:13.939578 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" containerID="cri-o://25209c017376e3ac46b1a33e19b0b7fe5acafd9ba6b9ced78aacd10037991a2f" gracePeriod=600 Nov 24 16:01:14 crc kubenswrapper[4790]: I1124 16:01:14.326685 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa001a3d-6aae-4cb7-9f48-e1adc714ad8f" path="/var/lib/kubelet/pods/fa001a3d-6aae-4cb7-9f48-e1adc714ad8f/volumes" Nov 24 16:01:14 crc kubenswrapper[4790]: I1124 16:01:14.455505 4790 generic.go:334] "Generic (PLEG): container finished" podID="0d73b133-48f0-455f-8f6a-742e633f631a" containerID="25209c017376e3ac46b1a33e19b0b7fe5acafd9ba6b9ced78aacd10037991a2f" exitCode=0 Nov 24 16:01:14 crc kubenswrapper[4790]: I1124 16:01:14.455558 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerDied","Data":"25209c017376e3ac46b1a33e19b0b7fe5acafd9ba6b9ced78aacd10037991a2f"} Nov 24 16:01:14 crc kubenswrapper[4790]: I1124 16:01:14.455598 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"04f0f5e9e4cbc0de998b73d51f7a42d9e018f4e80bf0907bef135486c6f1967b"} Nov 24 16:01:14 crc kubenswrapper[4790]: I1124 16:01:14.455618 4790 scope.go:117] "RemoveContainer" containerID="f1e094cfe52dce391f24557f040b619b252fb0fb27545cacdfae7450eb6c06b2" Nov 24 16:01:14 crc kubenswrapper[4790]: I1124 16:01:14.461121 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cd6sz/crc-debug-wgfrf" Nov 24 16:01:14 crc kubenswrapper[4790]: I1124 16:01:14.483872 4790 scope.go:117] "RemoveContainer" containerID="3645ccf35e68622a20b515b8ff900fd24e114c48a1c96f5d6ae57ea7d2ae8843" Nov 24 16:01:14 crc kubenswrapper[4790]: I1124 16:01:14.882580 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-cd6sz/crc-debug-sxpkv"] Nov 24 16:01:14 crc kubenswrapper[4790]: E1124 16:01:14.883263 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="017da2fa-87fc-44a6-9eb0-e0e1d4601633" containerName="extract-utilities" Nov 24 16:01:14 crc kubenswrapper[4790]: I1124 16:01:14.883277 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="017da2fa-87fc-44a6-9eb0-e0e1d4601633" containerName="extract-utilities" Nov 24 16:01:14 crc kubenswrapper[4790]: E1124 16:01:14.883290 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0b00350-df63-40ed-995b-6c2a9178f7d7" containerName="keystone-cron" Nov 24 16:01:14 crc kubenswrapper[4790]: I1124 16:01:14.883296 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b00350-df63-40ed-995b-6c2a9178f7d7" containerName="keystone-cron" Nov 24 16:01:14 crc kubenswrapper[4790]: E1124 16:01:14.883310 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="017da2fa-87fc-44a6-9eb0-e0e1d4601633" containerName="extract-content" Nov 24 16:01:14 crc kubenswrapper[4790]: I1124 16:01:14.883316 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="017da2fa-87fc-44a6-9eb0-e0e1d4601633" containerName="extract-content" Nov 24 16:01:14 crc kubenswrapper[4790]: E1124 16:01:14.883328 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="017da2fa-87fc-44a6-9eb0-e0e1d4601633" containerName="registry-server" Nov 24 16:01:14 crc kubenswrapper[4790]: I1124 16:01:14.883334 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="017da2fa-87fc-44a6-9eb0-e0e1d4601633" containerName="registry-server" Nov 24 16:01:14 crc kubenswrapper[4790]: E1124 16:01:14.883351 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa001a3d-6aae-4cb7-9f48-e1adc714ad8f" containerName="container-00" Nov 24 16:01:14 crc kubenswrapper[4790]: I1124 16:01:14.883358 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa001a3d-6aae-4cb7-9f48-e1adc714ad8f" containerName="container-00" Nov 24 16:01:14 crc kubenswrapper[4790]: I1124 16:01:14.883559 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0b00350-df63-40ed-995b-6c2a9178f7d7" containerName="keystone-cron" Nov 24 16:01:14 crc kubenswrapper[4790]: I1124 16:01:14.883583 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="017da2fa-87fc-44a6-9eb0-e0e1d4601633" containerName="registry-server" Nov 24 16:01:14 crc kubenswrapper[4790]: I1124 16:01:14.883600 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa001a3d-6aae-4cb7-9f48-e1adc714ad8f" containerName="container-00" Nov 24 16:01:14 crc kubenswrapper[4790]: I1124 16:01:14.884393 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cd6sz/crc-debug-sxpkv" Nov 24 16:01:14 crc kubenswrapper[4790]: I1124 16:01:14.997809 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7d4b26c0-b10e-43d4-8203-0f7244e1325f-host\") pod \"crc-debug-sxpkv\" (UID: \"7d4b26c0-b10e-43d4-8203-0f7244e1325f\") " pod="openshift-must-gather-cd6sz/crc-debug-sxpkv" Nov 24 16:01:14 crc kubenswrapper[4790]: I1124 16:01:14.997995 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7thnk\" (UniqueName: \"kubernetes.io/projected/7d4b26c0-b10e-43d4-8203-0f7244e1325f-kube-api-access-7thnk\") pod \"crc-debug-sxpkv\" (UID: \"7d4b26c0-b10e-43d4-8203-0f7244e1325f\") " pod="openshift-must-gather-cd6sz/crc-debug-sxpkv" Nov 24 16:01:15 crc kubenswrapper[4790]: I1124 16:01:15.100263 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7thnk\" (UniqueName: \"kubernetes.io/projected/7d4b26c0-b10e-43d4-8203-0f7244e1325f-kube-api-access-7thnk\") pod \"crc-debug-sxpkv\" (UID: \"7d4b26c0-b10e-43d4-8203-0f7244e1325f\") " pod="openshift-must-gather-cd6sz/crc-debug-sxpkv" Nov 24 16:01:15 crc kubenswrapper[4790]: I1124 16:01:15.100465 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7d4b26c0-b10e-43d4-8203-0f7244e1325f-host\") pod \"crc-debug-sxpkv\" (UID: \"7d4b26c0-b10e-43d4-8203-0f7244e1325f\") " pod="openshift-must-gather-cd6sz/crc-debug-sxpkv" Nov 24 16:01:15 crc kubenswrapper[4790]: I1124 16:01:15.100578 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7d4b26c0-b10e-43d4-8203-0f7244e1325f-host\") pod \"crc-debug-sxpkv\" (UID: \"7d4b26c0-b10e-43d4-8203-0f7244e1325f\") " pod="openshift-must-gather-cd6sz/crc-debug-sxpkv" Nov 24 16:01:15 crc kubenswrapper[4790]: I1124 16:01:15.121789 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7thnk\" (UniqueName: \"kubernetes.io/projected/7d4b26c0-b10e-43d4-8203-0f7244e1325f-kube-api-access-7thnk\") pod \"crc-debug-sxpkv\" (UID: \"7d4b26c0-b10e-43d4-8203-0f7244e1325f\") " pod="openshift-must-gather-cd6sz/crc-debug-sxpkv" Nov 24 16:01:15 crc kubenswrapper[4790]: I1124 16:01:15.204544 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cd6sz/crc-debug-sxpkv" Nov 24 16:01:15 crc kubenswrapper[4790]: W1124 16:01:15.236834 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d4b26c0_b10e_43d4_8203_0f7244e1325f.slice/crio-ded119c6f8db821323b18264eae65046c22bf00b06e451465c2d3b90461afb43 WatchSource:0}: Error finding container ded119c6f8db821323b18264eae65046c22bf00b06e451465c2d3b90461afb43: Status 404 returned error can't find the container with id ded119c6f8db821323b18264eae65046c22bf00b06e451465c2d3b90461afb43 Nov 24 16:01:15 crc kubenswrapper[4790]: I1124 16:01:15.475736 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cd6sz/crc-debug-sxpkv" event={"ID":"7d4b26c0-b10e-43d4-8203-0f7244e1325f","Type":"ContainerStarted","Data":"ded119c6f8db821323b18264eae65046c22bf00b06e451465c2d3b90461afb43"} Nov 24 16:01:16 crc kubenswrapper[4790]: I1124 16:01:16.486654 4790 generic.go:334] "Generic (PLEG): container finished" podID="7d4b26c0-b10e-43d4-8203-0f7244e1325f" containerID="d45b10db3b0312279b24850f04c5f808cd5b1e2d48d492896680f7633ce1252b" exitCode=1 Nov 24 16:01:16 crc kubenswrapper[4790]: I1124 16:01:16.487295 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cd6sz/crc-debug-sxpkv" event={"ID":"7d4b26c0-b10e-43d4-8203-0f7244e1325f","Type":"ContainerDied","Data":"d45b10db3b0312279b24850f04c5f808cd5b1e2d48d492896680f7633ce1252b"} Nov 24 16:01:16 crc kubenswrapper[4790]: I1124 16:01:16.549849 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-cd6sz/crc-debug-sxpkv"] Nov 24 16:01:16 crc kubenswrapper[4790]: I1124 16:01:16.559626 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-cd6sz/crc-debug-sxpkv"] Nov 24 16:01:17 crc kubenswrapper[4790]: I1124 16:01:17.638467 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cd6sz/crc-debug-sxpkv" Nov 24 16:01:17 crc kubenswrapper[4790]: I1124 16:01:17.655330 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7thnk\" (UniqueName: \"kubernetes.io/projected/7d4b26c0-b10e-43d4-8203-0f7244e1325f-kube-api-access-7thnk\") pod \"7d4b26c0-b10e-43d4-8203-0f7244e1325f\" (UID: \"7d4b26c0-b10e-43d4-8203-0f7244e1325f\") " Nov 24 16:01:17 crc kubenswrapper[4790]: I1124 16:01:17.655460 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7d4b26c0-b10e-43d4-8203-0f7244e1325f-host\") pod \"7d4b26c0-b10e-43d4-8203-0f7244e1325f\" (UID: \"7d4b26c0-b10e-43d4-8203-0f7244e1325f\") " Nov 24 16:01:17 crc kubenswrapper[4790]: I1124 16:01:17.656283 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7d4b26c0-b10e-43d4-8203-0f7244e1325f-host" (OuterVolumeSpecName: "host") pod "7d4b26c0-b10e-43d4-8203-0f7244e1325f" (UID: "7d4b26c0-b10e-43d4-8203-0f7244e1325f"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 16:01:17 crc kubenswrapper[4790]: I1124 16:01:17.662319 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d4b26c0-b10e-43d4-8203-0f7244e1325f-kube-api-access-7thnk" (OuterVolumeSpecName: "kube-api-access-7thnk") pod "7d4b26c0-b10e-43d4-8203-0f7244e1325f" (UID: "7d4b26c0-b10e-43d4-8203-0f7244e1325f"). InnerVolumeSpecName "kube-api-access-7thnk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 16:01:17 crc kubenswrapper[4790]: I1124 16:01:17.757936 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7thnk\" (UniqueName: \"kubernetes.io/projected/7d4b26c0-b10e-43d4-8203-0f7244e1325f-kube-api-access-7thnk\") on node \"crc\" DevicePath \"\"" Nov 24 16:01:17 crc kubenswrapper[4790]: I1124 16:01:17.757989 4790 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7d4b26c0-b10e-43d4-8203-0f7244e1325f-host\") on node \"crc\" DevicePath \"\"" Nov 24 16:01:18 crc kubenswrapper[4790]: I1124 16:01:18.328100 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d4b26c0-b10e-43d4-8203-0f7244e1325f" path="/var/lib/kubelet/pods/7d4b26c0-b10e-43d4-8203-0f7244e1325f/volumes" Nov 24 16:01:18 crc kubenswrapper[4790]: I1124 16:01:18.516738 4790 scope.go:117] "RemoveContainer" containerID="d45b10db3b0312279b24850f04c5f808cd5b1e2d48d492896680f7633ce1252b" Nov 24 16:01:18 crc kubenswrapper[4790]: I1124 16:01:18.516769 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cd6sz/crc-debug-sxpkv" Nov 24 16:03:43 crc kubenswrapper[4790]: I1124 16:03:43.938629 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 16:03:43 crc kubenswrapper[4790]: I1124 16:03:43.939290 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 16:03:56 crc kubenswrapper[4790]: I1124 16:03:56.518211 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_35216e6d-676a-48f1-aeda-e379d7f6ddb4/init-config-reloader/0.log" Nov 24 16:03:56 crc kubenswrapper[4790]: I1124 16:03:56.760160 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_35216e6d-676a-48f1-aeda-e379d7f6ddb4/init-config-reloader/0.log" Nov 24 16:03:56 crc kubenswrapper[4790]: I1124 16:03:56.780713 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_35216e6d-676a-48f1-aeda-e379d7f6ddb4/alertmanager/0.log" Nov 24 16:03:56 crc kubenswrapper[4790]: I1124 16:03:56.814866 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_35216e6d-676a-48f1-aeda-e379d7f6ddb4/config-reloader/0.log" Nov 24 16:03:56 crc kubenswrapper[4790]: I1124 16:03:56.949393 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_d0c3fa90-a58b-4a63-836a-35744ec234df/aodh-api/0.log" Nov 24 16:03:57 crc kubenswrapper[4790]: I1124 16:03:57.060182 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_d0c3fa90-a58b-4a63-836a-35744ec234df/aodh-listener/0.log" Nov 24 16:03:57 crc kubenswrapper[4790]: I1124 16:03:57.077017 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_d0c3fa90-a58b-4a63-836a-35744ec234df/aodh-evaluator/0.log" Nov 24 16:03:57 crc kubenswrapper[4790]: I1124 16:03:57.713228 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_d0c3fa90-a58b-4a63-836a-35744ec234df/aodh-notifier/0.log" Nov 24 16:03:57 crc kubenswrapper[4790]: I1124 16:03:57.735215 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6c6d75b7d-nbj94_b0e05357-2caf-4e9c-b18d-d0fc75a067d7/barbican-api-log/0.log" Nov 24 16:03:57 crc kubenswrapper[4790]: I1124 16:03:57.746540 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6c6d75b7d-nbj94_b0e05357-2caf-4e9c-b18d-d0fc75a067d7/barbican-api/0.log" Nov 24 16:03:57 crc kubenswrapper[4790]: I1124 16:03:57.985824 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-846d64984d-vjkzr_18179e0b-5942-4780-8240-ea2496142b5c/barbican-keystone-listener/0.log" Nov 24 16:03:58 crc kubenswrapper[4790]: I1124 16:03:58.111220 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-846d64984d-vjkzr_18179e0b-5942-4780-8240-ea2496142b5c/barbican-keystone-listener-log/0.log" Nov 24 16:03:58 crc kubenswrapper[4790]: I1124 16:03:58.230477 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6f48884455-sk4db_f00df2c8-01c0-4cd2-a699-97fadd168c78/barbican-worker/0.log" Nov 24 16:03:58 crc kubenswrapper[4790]: I1124 16:03:58.353854 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6f48884455-sk4db_f00df2c8-01c0-4cd2-a699-97fadd168c78/barbican-worker-log/0.log" Nov 24 16:03:58 crc kubenswrapper[4790]: I1124 16:03:58.407424 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-openstack-openstack-cell1-9cmwr_5991caca-bb27-439c-b5f8-4137b17639d9/bootstrap-openstack-openstack-cell1/0.log" Nov 24 16:03:58 crc kubenswrapper[4790]: I1124 16:03:58.832140 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_01048e56-f62e-4007-b717-ac76629de1ed/ceilometer-notification-agent/0.log" Nov 24 16:03:58 crc kubenswrapper[4790]: I1124 16:03:58.862199 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_01048e56-f62e-4007-b717-ac76629de1ed/ceilometer-central-agent/0.log" Nov 24 16:03:58 crc kubenswrapper[4790]: I1124 16:03:58.985798 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_01048e56-f62e-4007-b717-ac76629de1ed/proxy-httpd/0.log" Nov 24 16:03:59 crc kubenswrapper[4790]: I1124 16:03:59.006560 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_01048e56-f62e-4007-b717-ac76629de1ed/sg-core/0.log" Nov 24 16:03:59 crc kubenswrapper[4790]: I1124 16:03:59.444905 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-openstack-openstack-cell1-4mkwj_31455824-5727-44a5-b2c9-9eb68e20c305/ceph-client-openstack-openstack-cell1/0.log" Nov 24 16:03:59 crc kubenswrapper[4790]: I1124 16:03:59.501164 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_70481c96-3eef-4652-8918-9558f9ec74ae/cinder-api/0.log" Nov 24 16:04:00 crc kubenswrapper[4790]: I1124 16:04:00.053598 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_70481c96-3eef-4652-8918-9558f9ec74ae/cinder-api-log/0.log" Nov 24 16:04:00 crc kubenswrapper[4790]: I1124 16:04:00.126130 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_b54f7c8e-31ff-4565-bebe-663f7de8b110/probe/0.log" Nov 24 16:04:00 crc kubenswrapper[4790]: I1124 16:04:00.259347 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_b54f7c8e-31ff-4565-bebe-663f7de8b110/cinder-backup/0.log" Nov 24 16:04:00 crc kubenswrapper[4790]: I1124 16:04:00.335382 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_73610ce7-d271-4573-a14b-fd4563757182/cinder-scheduler/0.log" Nov 24 16:04:00 crc kubenswrapper[4790]: I1124 16:04:00.402407 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_73610ce7-d271-4573-a14b-fd4563757182/probe/0.log" Nov 24 16:04:00 crc kubenswrapper[4790]: I1124 16:04:00.618258 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_777b5526-f7f6-4ab3-9399-7aa596ff36a0/cinder-volume/0.log" Nov 24 16:04:00 crc kubenswrapper[4790]: I1124 16:04:00.622644 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_777b5526-f7f6-4ab3-9399-7aa596ff36a0/probe/0.log" Nov 24 16:04:00 crc kubenswrapper[4790]: I1124 16:04:00.710138 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-openstack-openstack-cell1-cfxxv_81c6adc7-3f2c-471d-a3ea-f2a947deb94a/configure-network-openstack-openstack-cell1/0.log" Nov 24 16:04:00 crc kubenswrapper[4790]: I1124 16:04:00.824413 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-cell1-xfkr7_8aebdda2-47dc-4465-b3b1-6de4cb11c07c/configure-os-openstack-openstack-cell1/0.log" Nov 24 16:04:00 crc kubenswrapper[4790]: I1124 16:04:00.909593 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6b9bbc5-vrcjr_35ee9564-c213-4eba-b011-467edd892cd0/init/0.log" Nov 24 16:04:01 crc kubenswrapper[4790]: I1124 16:04:01.086373 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6b9bbc5-vrcjr_35ee9564-c213-4eba-b011-467edd892cd0/init/0.log" Nov 24 16:04:01 crc kubenswrapper[4790]: I1124 16:04:01.128949 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-openstack-openstack-cell1-d8www_e3c33642-ddec-46a5-b86d-220f086ed2cf/download-cache-openstack-openstack-cell1/0.log" Nov 24 16:04:01 crc kubenswrapper[4790]: I1124 16:04:01.140600 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6b9bbc5-vrcjr_35ee9564-c213-4eba-b011-467edd892cd0/dnsmasq-dns/0.log" Nov 24 16:04:01 crc kubenswrapper[4790]: I1124 16:04:01.272928 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_f53e305e-ffe9-4c30-90bd-c3b4a1134e39/glance-httpd/0.log" Nov 24 16:04:01 crc kubenswrapper[4790]: I1124 16:04:01.328851 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_f53e305e-ffe9-4c30-90bd-c3b4a1134e39/glance-log/0.log" Nov 24 16:04:01 crc kubenswrapper[4790]: I1124 16:04:01.436245 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_130ad045-32d0-460f-83e8-a357594f53e9/glance-httpd/0.log" Nov 24 16:04:01 crc kubenswrapper[4790]: I1124 16:04:01.464864 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_130ad045-32d0-460f-83e8-a357594f53e9/glance-log/0.log" Nov 24 16:04:01 crc kubenswrapper[4790]: I1124 16:04:01.653176 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-746456fccd-k95mp_c9d135e7-5d36-4403-826e-e75f7d654ce9/heat-api/0.log" Nov 24 16:04:01 crc kubenswrapper[4790]: I1124 16:04:01.754689 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-688db64b5b-xg9f7_3a294679-0487-4813-8948-97732fe25595/heat-cfnapi/0.log" Nov 24 16:04:01 crc kubenswrapper[4790]: I1124 16:04:01.864599 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-76c78d699-44kq8_f5f0775f-7428-4cda-aa40-8e8b50df6c81/heat-engine/0.log" Nov 24 16:04:02 crc kubenswrapper[4790]: I1124 16:04:02.076190 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-8487684449-pkmn4_a96b1a40-c526-4796-9023-fd8f4f699323/horizon/0.log" Nov 24 16:04:02 crc kubenswrapper[4790]: I1124 16:04:02.091087 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-openstack-openstack-cell1-xrpsr_08f58547-542f-4a8f-a1a5-c433c2747d70/install-certs-openstack-openstack-cell1/0.log" Nov 24 16:04:02 crc kubenswrapper[4790]: I1124 16:04:02.094074 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-8487684449-pkmn4_a96b1a40-c526-4796-9023-fd8f4f699323/horizon-log/0.log" Nov 24 16:04:02 crc kubenswrapper[4790]: I1124 16:04:02.519118 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-openstack-openstack-cell1-cb4w5_f039968e-b352-4952-8420-71a78ab6cc49/install-os-openstack-openstack-cell1/0.log" Nov 24 16:04:02 crc kubenswrapper[4790]: I1124 16:04:02.651034 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29399941-28pp2_292b9a6b-b644-46a2-b4ba-2ba30d46e345/keystone-cron/0.log" Nov 24 16:04:02 crc kubenswrapper[4790]: I1124 16:04:02.688957 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-787bf45df4-blpgw_7dbb6d63-d31f-4951-ac0c-5ec7309c129a/keystone-api/0.log" Nov 24 16:04:02 crc kubenswrapper[4790]: I1124 16:04:02.785760 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29400001-dvmpw_a0b00350-df63-40ed-995b-6c2a9178f7d7/keystone-cron/0.log" Nov 24 16:04:02 crc kubenswrapper[4790]: I1124 16:04:02.873970 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_582a457e-86b9-42e9-8a6a-99598f5e85c9/kube-state-metrics/0.log" Nov 24 16:04:03 crc kubenswrapper[4790]: I1124 16:04:03.055474 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-openstack-openstack-cell1-zgjq6_abbde07e-75cb-4db6-9f73-726b7fe5f03c/libvirt-openstack-openstack-cell1/0.log" Nov 24 16:04:03 crc kubenswrapper[4790]: I1124 16:04:03.158381 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_d42b6e7d-ef2f-4d91-b9d0-138efb9549eb/manila-api-log/0.log" Nov 24 16:04:03 crc kubenswrapper[4790]: I1124 16:04:03.189641 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_d42b6e7d-ef2f-4d91-b9d0-138efb9549eb/manila-api/0.log" Nov 24 16:04:03 crc kubenswrapper[4790]: I1124 16:04:03.336611 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_b75fcca8-34a3-42d4-bfca-43c21b63227b/probe/0.log" Nov 24 16:04:03 crc kubenswrapper[4790]: I1124 16:04:03.342369 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_b75fcca8-34a3-42d4-bfca-43c21b63227b/manila-scheduler/0.log" Nov 24 16:04:03 crc kubenswrapper[4790]: I1124 16:04:03.467682 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_6cacb59b-e19b-44b9-b4c4-3a12559e64ed/manila-share/0.log" Nov 24 16:04:03 crc kubenswrapper[4790]: I1124 16:04:03.475432 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_6cacb59b-e19b-44b9-b4c4-3a12559e64ed/probe/0.log" Nov 24 16:04:03 crc kubenswrapper[4790]: I1124 16:04:03.816136 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-bc5df678c-79bx5_15b5bece-4f0a-41fb-8473-44d31fd383b6/neutron-api/0.log" Nov 24 16:04:03 crc kubenswrapper[4790]: I1124 16:04:03.853819 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-bc5df678c-79bx5_15b5bece-4f0a-41fb-8473-44d31fd383b6/neutron-httpd/0.log" Nov 24 16:04:04 crc kubenswrapper[4790]: I1124 16:04:04.103068 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dhcp-openstack-openstack-cell1-n9962_515ed57f-53d0-470b-aafe-7deb07d527c9/neutron-dhcp-openstack-openstack-cell1/0.log" Nov 24 16:04:04 crc kubenswrapper[4790]: I1124 16:04:04.204388 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-openstack-openstack-cell1-68gxl_052884d0-d104-44a6-bcf8-86ac5a67b6ff/neutron-metadata-openstack-openstack-cell1/0.log" Nov 24 16:04:04 crc kubenswrapper[4790]: I1124 16:04:04.436068 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-sriov-openstack-openstack-cell1-d9tc9_367b480c-25ab-497e-aac9-26a29735867c/neutron-sriov-openstack-openstack-cell1/0.log" Nov 24 16:04:04 crc kubenswrapper[4790]: I1124 16:04:04.494017 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_3d36ae33-dbe8-4323-8bc6-83ca5dcf40ea/nova-api-api/0.log" Nov 24 16:04:04 crc kubenswrapper[4790]: I1124 16:04:04.590690 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_3d36ae33-dbe8-4323-8bc6-83ca5dcf40ea/nova-api-log/0.log" Nov 24 16:04:04 crc kubenswrapper[4790]: I1124 16:04:04.810529 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_3df0c98f-9380-488b-a72e-4111d10941d9/nova-cell0-conductor-conductor/0.log" Nov 24 16:04:04 crc kubenswrapper[4790]: I1124 16:04:04.902585 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_17309035-deb1-4185-80ac-b4483116301f/nova-cell1-conductor-conductor/0.log" Nov 24 16:04:05 crc kubenswrapper[4790]: I1124 16:04:05.133848 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_06c8ec77-c053-44cc-b394-e3170e011b3d/nova-cell1-novncproxy-novncproxy/0.log" Nov 24 16:04:05 crc kubenswrapper[4790]: I1124 16:04:05.411555 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell9qcrf_9e561ed1-5cd7-4a50-8629-bf48c3a21ac4/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1/0.log" Nov 24 16:04:05 crc kubenswrapper[4790]: I1124 16:04:05.460646 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-openstack-cell1-k6gv2_1feac502-08da-479d-8efe-2c84747f9d0a/nova-cell1-openstack-openstack-cell1/0.log" Nov 24 16:04:05 crc kubenswrapper[4790]: I1124 16:04:05.641055 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_9a7f156b-e253-4f3b-b4ce-5c3c226ccfcd/nova-metadata-log/0.log" Nov 24 16:04:06 crc kubenswrapper[4790]: I1124 16:04:06.062004 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_9a7f156b-e253-4f3b-b4ce-5c3c226ccfcd/nova-metadata-metadata/0.log" Nov 24 16:04:06 crc kubenswrapper[4790]: I1124 16:04:06.094815 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_c19692e0-e586-4506-a9e7-28195138b290/nova-scheduler-scheduler/0.log" Nov 24 16:04:06 crc kubenswrapper[4790]: I1124 16:04:06.270570 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-5b99f4ff4b-5rkfz_008c0092-bd88-4696-91a3-3f124d181264/init/0.log" Nov 24 16:04:06 crc kubenswrapper[4790]: I1124 16:04:06.494139 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-5b99f4ff4b-5rkfz_008c0092-bd88-4696-91a3-3f124d181264/init/0.log" Nov 24 16:04:06 crc kubenswrapper[4790]: I1124 16:04:06.520737 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-5b99f4ff4b-5rkfz_008c0092-bd88-4696-91a3-3f124d181264/octavia-api-provider-agent/0.log" Nov 24 16:04:06 crc kubenswrapper[4790]: I1124 16:04:06.783716 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-5b99f4ff4b-5rkfz_008c0092-bd88-4696-91a3-3f124d181264/octavia-api/0.log" Nov 24 16:04:07 crc kubenswrapper[4790]: I1124 16:04:07.102826 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-n974m_617d4157-f445-4d62-ac72-9d16b5fd52be/init/0.log" Nov 24 16:04:07 crc kubenswrapper[4790]: I1124 16:04:07.357684 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-n974m_617d4157-f445-4d62-ac72-9d16b5fd52be/octavia-healthmanager/0.log" Nov 24 16:04:07 crc kubenswrapper[4790]: I1124 16:04:07.365265 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-kfk5m_2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427/init/0.log" Nov 24 16:04:07 crc kubenswrapper[4790]: I1124 16:04:07.411986 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-n974m_617d4157-f445-4d62-ac72-9d16b5fd52be/init/0.log" Nov 24 16:04:07 crc kubenswrapper[4790]: I1124 16:04:07.656225 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-kfk5m_2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427/init/0.log" Nov 24 16:04:07 crc kubenswrapper[4790]: I1124 16:04:07.666108 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-kfk5m_2999c2ac-3bd1-4a5f-b6d5-ef70fd44d427/octavia-housekeeping/0.log" Nov 24 16:04:07 crc kubenswrapper[4790]: I1124 16:04:07.713809 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-59f8cff499-6pxxp_e7f3d559-a9e1-4cb6-bc14-e85e8df77929/init/0.log" Nov 24 16:04:07 crc kubenswrapper[4790]: I1124 16:04:07.862135 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-59f8cff499-6pxxp_e7f3d559-a9e1-4cb6-bc14-e85e8df77929/octavia-amphora-httpd/0.log" Nov 24 16:04:07 crc kubenswrapper[4790]: I1124 16:04:07.876158 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-59f8cff499-6pxxp_e7f3d559-a9e1-4cb6-bc14-e85e8df77929/init/0.log" Nov 24 16:04:07 crc kubenswrapper[4790]: I1124 16:04:07.966020 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-7rfsl_d749d6ea-42cc-47d0-880e-d079f59b3ad6/init/0.log" Nov 24 16:04:08 crc kubenswrapper[4790]: I1124 16:04:08.194458 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-7rfsl_d749d6ea-42cc-47d0-880e-d079f59b3ad6/octavia-rsyslog/0.log" Nov 24 16:04:08 crc kubenswrapper[4790]: I1124 16:04:08.227631 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-7rfsl_d749d6ea-42cc-47d0-880e-d079f59b3ad6/init/0.log" Nov 24 16:04:08 crc kubenswrapper[4790]: I1124 16:04:08.268225 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-fkc29_3aa341b2-1c7e-4827-9d80-9a26d5b46fa9/init/0.log" Nov 24 16:04:09 crc kubenswrapper[4790]: I1124 16:04:09.383990 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_09cadf79-68ab-4fc3-b391-4a7d8318a787/mysql-bootstrap/0.log" Nov 24 16:04:09 crc kubenswrapper[4790]: I1124 16:04:09.392898 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-fkc29_3aa341b2-1c7e-4827-9d80-9a26d5b46fa9/init/0.log" Nov 24 16:04:09 crc kubenswrapper[4790]: I1124 16:04:09.581190 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-fkc29_3aa341b2-1c7e-4827-9d80-9a26d5b46fa9/octavia-worker/0.log" Nov 24 16:04:09 crc kubenswrapper[4790]: I1124 16:04:09.749773 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_09cadf79-68ab-4fc3-b391-4a7d8318a787/galera/0.log" Nov 24 16:04:09 crc kubenswrapper[4790]: I1124 16:04:09.755357 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_09cadf79-68ab-4fc3-b391-4a7d8318a787/mysql-bootstrap/0.log" Nov 24 16:04:09 crc kubenswrapper[4790]: I1124 16:04:09.862337 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_95752c01-3e10-44b4-8af5-7b74622dbae3/mysql-bootstrap/0.log" Nov 24 16:04:10 crc kubenswrapper[4790]: I1124 16:04:10.034584 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_95752c01-3e10-44b4-8af5-7b74622dbae3/mysql-bootstrap/0.log" Nov 24 16:04:10 crc kubenswrapper[4790]: I1124 16:04:10.111749 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_a6d18574-10ed-4b11-b5a9-eeed45e66cb2/openstackclient/0.log" Nov 24 16:04:10 crc kubenswrapper[4790]: I1124 16:04:10.121341 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_95752c01-3e10-44b4-8af5-7b74622dbae3/galera/0.log" Nov 24 16:04:10 crc kubenswrapper[4790]: I1124 16:04:10.335689 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-6p54v_ec4e9e28-345c-4cc4-acba-7a06cac52117/ovn-controller/0.log" Nov 24 16:04:10 crc kubenswrapper[4790]: I1124 16:04:10.383967 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-9x8wb_5f1ba98b-2cf8-471f-94c6-ce7a5699d421/openstack-network-exporter/0.log" Nov 24 16:04:10 crc kubenswrapper[4790]: I1124 16:04:10.785581 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-l46w2_d9800a51-bf87-4786-8cef-c8fcafdc72d9/ovsdb-server-init/0.log" Nov 24 16:04:10 crc kubenswrapper[4790]: I1124 16:04:10.997756 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-l46w2_d9800a51-bf87-4786-8cef-c8fcafdc72d9/ovsdb-server-init/0.log" Nov 24 16:04:10 crc kubenswrapper[4790]: I1124 16:04:10.999417 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-l46w2_d9800a51-bf87-4786-8cef-c8fcafdc72d9/ovsdb-server/0.log" Nov 24 16:04:11 crc kubenswrapper[4790]: I1124 16:04:11.020228 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-l46w2_d9800a51-bf87-4786-8cef-c8fcafdc72d9/ovs-vswitchd/0.log" Nov 24 16:04:11 crc kubenswrapper[4790]: I1124 16:04:11.200140 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_5364dcd2-a902-4a92-a96c-a7ff91d0b50e/ovn-northd/0.log" Nov 24 16:04:11 crc kubenswrapper[4790]: I1124 16:04:11.244390 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_5364dcd2-a902-4a92-a96c-a7ff91d0b50e/openstack-network-exporter/0.log" Nov 24 16:04:11 crc kubenswrapper[4790]: I1124 16:04:11.464020 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_35ca55b2-63a0-4c8e-b3f5-f4b645692137/openstack-network-exporter/0.log" Nov 24 16:04:11 crc kubenswrapper[4790]: I1124 16:04:11.598785 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-openstack-openstack-cell1-znrmt_6036a536-d292-481d-b3f3-22b8ca768c8e/ovn-openstack-openstack-cell1/0.log" Nov 24 16:04:11 crc kubenswrapper[4790]: I1124 16:04:11.599927 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_35ca55b2-63a0-4c8e-b3f5-f4b645692137/ovsdbserver-nb/0.log" Nov 24 16:04:11 crc kubenswrapper[4790]: I1124 16:04:11.747954 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_361b98d3-3b9b-4671-85b9-7b9c62ac72d7/openstack-network-exporter/0.log" Nov 24 16:04:11 crc kubenswrapper[4790]: I1124 16:04:11.916358 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_361b98d3-3b9b-4671-85b9-7b9c62ac72d7/ovsdbserver-nb/0.log" Nov 24 16:04:11 crc kubenswrapper[4790]: I1124 16:04:11.918066 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_d069694e-57dd-4c26-9dca-b063d74fb7ea/openstack-network-exporter/0.log" Nov 24 16:04:12 crc kubenswrapper[4790]: I1124 16:04:12.014444 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_d069694e-57dd-4c26-9dca-b063d74fb7ea/ovsdbserver-nb/0.log" Nov 24 16:04:12 crc kubenswrapper[4790]: I1124 16:04:12.158830 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_04996e8f-e013-4af2-8890-ca8bd3c21974/ovsdbserver-sb/0.log" Nov 24 16:04:12 crc kubenswrapper[4790]: I1124 16:04:12.190454 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_04996e8f-e013-4af2-8890-ca8bd3c21974/openstack-network-exporter/0.log" Nov 24 16:04:12 crc kubenswrapper[4790]: I1124 16:04:12.363291 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_bb538d03-ca9f-4629-935e-f586455cd4e5/ovsdbserver-sb/0.log" Nov 24 16:04:12 crc kubenswrapper[4790]: I1124 16:04:12.427970 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_bb538d03-ca9f-4629-935e-f586455cd4e5/openstack-network-exporter/0.log" Nov 24 16:04:12 crc kubenswrapper[4790]: I1124 16:04:12.540466 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_b3d12b48-59a8-45df-8ff3-7850796b5163/openstack-network-exporter/0.log" Nov 24 16:04:12 crc kubenswrapper[4790]: I1124 16:04:12.599241 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_b3d12b48-59a8-45df-8ff3-7850796b5163/ovsdbserver-sb/0.log" Nov 24 16:04:12 crc kubenswrapper[4790]: I1124 16:04:12.745129 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-59d7f799d8-grzhn_ca49278e-82e5-42f8-9b83-a7f44e4e21e6/placement-api/0.log" Nov 24 16:04:12 crc kubenswrapper[4790]: I1124 16:04:12.950353 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-59d7f799d8-grzhn_ca49278e-82e5-42f8-9b83-a7f44e4e21e6/placement-log/0.log" Nov 24 16:04:13 crc kubenswrapper[4790]: I1124 16:04:13.000517 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_pre-adoption-validation-openstack-pre-adoption-openstack-cxqdsk_ba0f2f40-136f-402f-b25c-4fdf2236b7d3/pre-adoption-validation-openstack-pre-adoption-openstack-cell1/0.log" Nov 24 16:04:13 crc kubenswrapper[4790]: I1124 16:04:13.056233 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_09da3379-5d69-4618-ad5c-baa983f57405/init-config-reloader/0.log" Nov 24 16:04:13 crc kubenswrapper[4790]: I1124 16:04:13.301216 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_09da3379-5d69-4618-ad5c-baa983f57405/init-config-reloader/0.log" Nov 24 16:04:13 crc kubenswrapper[4790]: I1124 16:04:13.332282 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_09da3379-5d69-4618-ad5c-baa983f57405/thanos-sidecar/0.log" Nov 24 16:04:13 crc kubenswrapper[4790]: I1124 16:04:13.333091 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_09da3379-5d69-4618-ad5c-baa983f57405/prometheus/0.log" Nov 24 16:04:13 crc kubenswrapper[4790]: I1124 16:04:13.345552 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_09da3379-5d69-4618-ad5c-baa983f57405/config-reloader/0.log" Nov 24 16:04:13 crc kubenswrapper[4790]: I1124 16:04:13.541964 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_260598aa-c7fe-453b-8f55-c40dd2037e8d/setup-container/0.log" Nov 24 16:04:13 crc kubenswrapper[4790]: I1124 16:04:13.765929 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_260598aa-c7fe-453b-8f55-c40dd2037e8d/setup-container/0.log" Nov 24 16:04:13 crc kubenswrapper[4790]: I1124 16:04:13.801718 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_260598aa-c7fe-453b-8f55-c40dd2037e8d/rabbitmq/0.log" Nov 24 16:04:13 crc kubenswrapper[4790]: I1124 16:04:13.856171 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_5f59a746-4906-45aa-b6ee-7feb652fa17a/memcached/0.log" Nov 24 16:04:13 crc kubenswrapper[4790]: I1124 16:04:13.857546 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_c1043d54-fc1a-4758-abaa-c53a371cebb8/setup-container/0.log" Nov 24 16:04:13 crc kubenswrapper[4790]: I1124 16:04:13.938457 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 16:04:13 crc kubenswrapper[4790]: I1124 16:04:13.938518 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 16:04:14 crc kubenswrapper[4790]: I1124 16:04:14.001749 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_c1043d54-fc1a-4758-abaa-c53a371cebb8/setup-container/0.log" Nov 24 16:04:14 crc kubenswrapper[4790]: I1124 16:04:14.021698 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_c1043d54-fc1a-4758-abaa-c53a371cebb8/rabbitmq/0.log" Nov 24 16:04:14 crc kubenswrapper[4790]: I1124 16:04:14.108896 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-openstack-openstack-cell1-2wk4p_f6689103-8cb5-494e-995f-36f3a9e1f83c/reboot-os-openstack-openstack-cell1/0.log" Nov 24 16:04:14 crc kubenswrapper[4790]: I1124 16:04:14.147956 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-openstack-openstack-cell1-v7fx7_2c4b8d2c-3996-4673-9a36-b21c45d58dec/run-os-openstack-openstack-cell1/0.log" Nov 24 16:04:14 crc kubenswrapper[4790]: I1124 16:04:14.320804 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-openstack-h9qxq_9b974f0b-e04c-4394-85fa-14b57cff7504/ssh-known-hosts-openstack/0.log" Nov 24 16:04:14 crc kubenswrapper[4790]: I1124 16:04:14.389117 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-openstack-openstack-cell1-ncldn_d286ee05-e6fc-4945-b7a9-87d56a06486c/telemetry-openstack-openstack-cell1/0.log" Nov 24 16:04:14 crc kubenswrapper[4790]: I1124 16:04:14.501372 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tripleo-cleanup-tripleo-cleanup-openstack-cell1-httz6_0effec18-dddb-4967-a5f8-1a98979e3338/tripleo-cleanup-tripleo-cleanup-openstack-cell1/0.log" Nov 24 16:04:14 crc kubenswrapper[4790]: I1124 16:04:14.521252 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-openstack-openstack-cell1-jswk7_abfebf54-171c-46e3-bbb3-882266bca305/validate-network-openstack-openstack-cell1/0.log" Nov 24 16:04:34 crc kubenswrapper[4790]: I1124 16:04:34.921710 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-75fb479bcc-9wl9p_b55ba4e8-3d0a-4746-817c-feb1a210b01e/kube-rbac-proxy/0.log" Nov 24 16:04:35 crc kubenswrapper[4790]: I1124 16:04:35.013378 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-75fb479bcc-9wl9p_b55ba4e8-3d0a-4746-817c-feb1a210b01e/manager/0.log" Nov 24 16:04:35 crc kubenswrapper[4790]: I1124 16:04:35.154044 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6498cbf48f-44wlb_09e60927-110e-4cda-b7e3-6f47e70f2388/kube-rbac-proxy/0.log" Nov 24 16:04:35 crc kubenswrapper[4790]: I1124 16:04:35.206853 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6498cbf48f-44wlb_09e60927-110e-4cda-b7e3-6f47e70f2388/manager/0.log" Nov 24 16:04:35 crc kubenswrapper[4790]: I1124 16:04:35.305111 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-767ccfd65f-h8nkv_b2b9b9e9-800e-4213-8605-058d91412276/kube-rbac-proxy/0.log" Nov 24 16:04:35 crc kubenswrapper[4790]: I1124 16:04:35.343604 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-767ccfd65f-h8nkv_b2b9b9e9-800e-4213-8605-058d91412276/manager/0.log" Nov 24 16:04:35 crc kubenswrapper[4790]: I1124 16:04:35.421597 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch_8c370b97-cb4b-410e-b8cb-fd5b807de698/util/0.log" Nov 24 16:04:35 crc kubenswrapper[4790]: I1124 16:04:35.638521 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch_8c370b97-cb4b-410e-b8cb-fd5b807de698/pull/0.log" Nov 24 16:04:35 crc kubenswrapper[4790]: I1124 16:04:35.643605 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch_8c370b97-cb4b-410e-b8cb-fd5b807de698/pull/0.log" Nov 24 16:04:35 crc kubenswrapper[4790]: I1124 16:04:35.645192 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch_8c370b97-cb4b-410e-b8cb-fd5b807de698/util/0.log" Nov 24 16:04:35 crc kubenswrapper[4790]: I1124 16:04:35.784378 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch_8c370b97-cb4b-410e-b8cb-fd5b807de698/pull/0.log" Nov 24 16:04:35 crc kubenswrapper[4790]: I1124 16:04:35.786304 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch_8c370b97-cb4b-410e-b8cb-fd5b807de698/util/0.log" Nov 24 16:04:35 crc kubenswrapper[4790]: I1124 16:04:35.790179 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db66052f29ch_8c370b97-cb4b-410e-b8cb-fd5b807de698/extract/0.log" Nov 24 16:04:35 crc kubenswrapper[4790]: I1124 16:04:35.958995 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7969689c84-wq7t5_d39e8176-e052-4a0e-9d16-e66b6312fbfc/kube-rbac-proxy/0.log" Nov 24 16:04:36 crc kubenswrapper[4790]: I1124 16:04:36.063786 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-56f54d6746-bzh6h_49de2fb5-2dde-400e-9ae4-d36dfe196ff1/kube-rbac-proxy/0.log" Nov 24 16:04:36 crc kubenswrapper[4790]: I1124 16:04:36.097775 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7969689c84-wq7t5_d39e8176-e052-4a0e-9d16-e66b6312fbfc/manager/0.log" Nov 24 16:04:36 crc kubenswrapper[4790]: I1124 16:04:36.181817 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-56f54d6746-bzh6h_49de2fb5-2dde-400e-9ae4-d36dfe196ff1/manager/0.log" Nov 24 16:04:36 crc kubenswrapper[4790]: I1124 16:04:36.265718 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-598f69df5d-95l27_8b819ac2-ef31-494d-b4cf-882b602feb70/kube-rbac-proxy/0.log" Nov 24 16:04:36 crc kubenswrapper[4790]: I1124 16:04:36.312697 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-598f69df5d-95l27_8b819ac2-ef31-494d-b4cf-882b602feb70/manager/0.log" Nov 24 16:04:36 crc kubenswrapper[4790]: I1124 16:04:36.350905 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6dd8864d7c-fxplj_7000614b-01cb-4878-b85f-b5a1f8ef91b2/kube-rbac-proxy/0.log" Nov 24 16:04:36 crc kubenswrapper[4790]: I1124 16:04:36.556988 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-99b499f4-ntfjb_52519a2b-7c33-425f-9115-6fd4876ba60d/kube-rbac-proxy/0.log" Nov 24 16:04:36 crc kubenswrapper[4790]: I1124 16:04:36.616408 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-99b499f4-ntfjb_52519a2b-7c33-425f-9115-6fd4876ba60d/manager/0.log" Nov 24 16:04:36 crc kubenswrapper[4790]: I1124 16:04:36.717758 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6dd8864d7c-fxplj_7000614b-01cb-4878-b85f-b5a1f8ef91b2/manager/0.log" Nov 24 16:04:36 crc kubenswrapper[4790]: I1124 16:04:36.826651 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7454b96578-4bndq_2c64049e-cf92-40ac-84d9-e2724c334c3f/kube-rbac-proxy/0.log" Nov 24 16:04:36 crc kubenswrapper[4790]: I1124 16:04:36.918719 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7454b96578-4bndq_2c64049e-cf92-40ac-84d9-e2724c334c3f/manager/0.log" Nov 24 16:04:37 crc kubenswrapper[4790]: I1124 16:04:37.040030 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58f887965d-qvbnj_3dc6a5ba-d447-40fc-bd8b-640e4b49990b/kube-rbac-proxy/0.log" Nov 24 16:04:37 crc kubenswrapper[4790]: I1124 16:04:37.106534 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58f887965d-qvbnj_3dc6a5ba-d447-40fc-bd8b-640e4b49990b/manager/0.log" Nov 24 16:04:37 crc kubenswrapper[4790]: I1124 16:04:37.135675 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-54b5986bb8-wzvsq_4c1c13a1-3285-45f5-888b-d9abb0008661/kube-rbac-proxy/0.log" Nov 24 16:04:37 crc kubenswrapper[4790]: I1124 16:04:37.308351 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-54b5986bb8-wzvsq_4c1c13a1-3285-45f5-888b-d9abb0008661/manager/0.log" Nov 24 16:04:37 crc kubenswrapper[4790]: I1124 16:04:37.344366 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78bd47f458-d8gzf_0c8ec99a-dfcc-414d-85c9-f005c476919d/kube-rbac-proxy/0.log" Nov 24 16:04:37 crc kubenswrapper[4790]: I1124 16:04:37.433957 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78bd47f458-d8gzf_0c8ec99a-dfcc-414d-85c9-f005c476919d/manager/0.log" Nov 24 16:04:37 crc kubenswrapper[4790]: I1124 16:04:37.497749 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-cfbb9c588-gkdz7_e25e339c-c102-44f0-b4a4-15f0ce87b38c/kube-rbac-proxy/0.log" Nov 24 16:04:37 crc kubenswrapper[4790]: I1124 16:04:37.668618 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-54cfbf4c7d-9gnsv_dc60f94f-01f9-4179-a0ce-67112913da75/kube-rbac-proxy/0.log" Nov 24 16:04:37 crc kubenswrapper[4790]: I1124 16:04:37.857582 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-8c7444f48-jg2n4_c861ea1f-e001-4f01-8e0d-12e97b7628cc/kube-rbac-proxy/0.log" Nov 24 16:04:37 crc kubenswrapper[4790]: I1124 16:04:37.871368 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-54cfbf4c7d-9gnsv_dc60f94f-01f9-4179-a0ce-67112913da75/manager/0.log" Nov 24 16:04:37 crc kubenswrapper[4790]: I1124 16:04:37.965123 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-cfbb9c588-gkdz7_e25e339c-c102-44f0-b4a4-15f0ce87b38c/manager/0.log" Nov 24 16:04:38 crc kubenswrapper[4790]: I1124 16:04:38.051162 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-8c7444f48-jg2n4_c861ea1f-e001-4f01-8e0d-12e97b7628cc/manager/0.log" Nov 24 16:04:38 crc kubenswrapper[4790]: I1124 16:04:38.241332 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7547bd9b9-skhbm_7c72ca69-273b-4c92-a5f7-54c2b968621b/kube-rbac-proxy/0.log" Nov 24 16:04:38 crc kubenswrapper[4790]: I1124 16:04:38.322392 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-866b74d8b7-t2h2s_123e2f50-7f37-422b-a3ca-91252d609541/kube-rbac-proxy/0.log" Nov 24 16:04:38 crc kubenswrapper[4790]: I1124 16:04:38.647018 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-vkxbs_382216c5-b05f-4e86-b5c2-eb93a948c4e5/registry-server/0.log" Nov 24 16:04:38 crc kubenswrapper[4790]: I1124 16:04:38.660978 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-866b74d8b7-t2h2s_123e2f50-7f37-422b-a3ca-91252d609541/operator/0.log" Nov 24 16:04:38 crc kubenswrapper[4790]: I1124 16:04:38.836893 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54fc5f65b7-ld798_84281187-e5c6-4b57-807a-ae5637942260/kube-rbac-proxy/0.log" Nov 24 16:04:39 crc kubenswrapper[4790]: I1124 16:04:39.014989 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b797b8dff-c8xtz_cb16570b-3e90-4b2e-996b-d0cd5c0a0259/kube-rbac-proxy/0.log" Nov 24 16:04:39 crc kubenswrapper[4790]: I1124 16:04:39.037816 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54fc5f65b7-ld798_84281187-e5c6-4b57-807a-ae5637942260/manager/0.log" Nov 24 16:04:39 crc kubenswrapper[4790]: I1124 16:04:39.172965 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b797b8dff-c8xtz_cb16570b-3e90-4b2e-996b-d0cd5c0a0259/manager/0.log" Nov 24 16:04:39 crc kubenswrapper[4790]: I1124 16:04:39.284010 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-6dwzt_8e1bedf8-78bf-4558-91e6-4228e9765356/operator/0.log" Nov 24 16:04:39 crc kubenswrapper[4790]: I1124 16:04:39.391198 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d656998f4-lfvdf_69644679-f0fb-4c09-abec-8e4084140f05/kube-rbac-proxy/0.log" Nov 24 16:04:39 crc kubenswrapper[4790]: I1124 16:04:39.519856 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d656998f4-lfvdf_69644679-f0fb-4c09-abec-8e4084140f05/manager/0.log" Nov 24 16:04:39 crc kubenswrapper[4790]: I1124 16:04:39.531320 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6d4bf84b58-fg9nc_1e565309-31eb-4278-98fd-5a4321d5eb10/kube-rbac-proxy/0.log" Nov 24 16:04:39 crc kubenswrapper[4790]: I1124 16:04:39.792298 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-b4c496f69-8k88p_5e93acb0-367e-4f6c-b13e-530d41db7851/manager/0.log" Nov 24 16:04:39 crc kubenswrapper[4790]: I1124 16:04:39.812541 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-b4c496f69-8k88p_5e93acb0-367e-4f6c-b13e-530d41db7851/kube-rbac-proxy/0.log" Nov 24 16:04:39 crc kubenswrapper[4790]: I1124 16:04:39.864605 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6d4bf84b58-fg9nc_1e565309-31eb-4278-98fd-5a4321d5eb10/manager/0.log" Nov 24 16:04:40 crc kubenswrapper[4790]: I1124 16:04:40.029520 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-8c6448b9f-cddlm_e85f1bfb-0153-40e0-b54b-728171cb39f0/kube-rbac-proxy/0.log" Nov 24 16:04:40 crc kubenswrapper[4790]: I1124 16:04:40.050365 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-8c6448b9f-cddlm_e85f1bfb-0153-40e0-b54b-728171cb39f0/manager/0.log" Nov 24 16:04:40 crc kubenswrapper[4790]: I1124 16:04:40.658703 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7547bd9b9-skhbm_7c72ca69-273b-4c92-a5f7-54c2b968621b/manager/0.log" Nov 24 16:04:43 crc kubenswrapper[4790]: I1124 16:04:43.938762 4790 patch_prober.go:28] interesting pod/machine-config-daemon-xz49t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 16:04:43 crc kubenswrapper[4790]: I1124 16:04:43.939341 4790 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 16:04:43 crc kubenswrapper[4790]: I1124 16:04:43.939396 4790 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" Nov 24 16:04:43 crc kubenswrapper[4790]: I1124 16:04:43.940279 4790 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"04f0f5e9e4cbc0de998b73d51f7a42d9e018f4e80bf0907bef135486c6f1967b"} pod="openshift-machine-config-operator/machine-config-daemon-xz49t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 16:04:43 crc kubenswrapper[4790]: I1124 16:04:43.940370 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" containerName="machine-config-daemon" containerID="cri-o://04f0f5e9e4cbc0de998b73d51f7a42d9e018f4e80bf0907bef135486c6f1967b" gracePeriod=600 Nov 24 16:04:44 crc kubenswrapper[4790]: E1124 16:04:44.073671 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 16:04:44 crc kubenswrapper[4790]: I1124 16:04:44.395322 4790 generic.go:334] "Generic (PLEG): container finished" podID="0d73b133-48f0-455f-8f6a-742e633f631a" containerID="04f0f5e9e4cbc0de998b73d51f7a42d9e018f4e80bf0907bef135486c6f1967b" exitCode=0 Nov 24 16:04:44 crc kubenswrapper[4790]: I1124 16:04:44.395363 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerDied","Data":"04f0f5e9e4cbc0de998b73d51f7a42d9e018f4e80bf0907bef135486c6f1967b"} Nov 24 16:04:44 crc kubenswrapper[4790]: I1124 16:04:44.395395 4790 scope.go:117] "RemoveContainer" containerID="25209c017376e3ac46b1a33e19b0b7fe5acafd9ba6b9ced78aacd10037991a2f" Nov 24 16:04:44 crc kubenswrapper[4790]: I1124 16:04:44.396214 4790 scope.go:117] "RemoveContainer" containerID="04f0f5e9e4cbc0de998b73d51f7a42d9e018f4e80bf0907bef135486c6f1967b" Nov 24 16:04:44 crc kubenswrapper[4790]: E1124 16:04:44.396702 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 16:04:56 crc kubenswrapper[4790]: I1124 16:04:56.315470 4790 scope.go:117] "RemoveContainer" containerID="04f0f5e9e4cbc0de998b73d51f7a42d9e018f4e80bf0907bef135486c6f1967b" Nov 24 16:04:56 crc kubenswrapper[4790]: E1124 16:04:56.316557 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 16:04:58 crc kubenswrapper[4790]: I1124 16:04:58.568325 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-96268_acb2b70f-4a66-4290-aa8a-a13cb1b4b2a9/control-plane-machine-set-operator/0.log" Nov 24 16:04:58 crc kubenswrapper[4790]: I1124 16:04:58.701396 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-slkpm_17718a81-aa70-4f5a-90bd-06e9251f4ef8/kube-rbac-proxy/0.log" Nov 24 16:04:58 crc kubenswrapper[4790]: I1124 16:04:58.757405 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-slkpm_17718a81-aa70-4f5a-90bd-06e9251f4ef8/machine-api-operator/0.log" Nov 24 16:05:10 crc kubenswrapper[4790]: I1124 16:05:10.315550 4790 scope.go:117] "RemoveContainer" containerID="04f0f5e9e4cbc0de998b73d51f7a42d9e018f4e80bf0907bef135486c6f1967b" Nov 24 16:05:10 crc kubenswrapper[4790]: E1124 16:05:10.316508 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 16:05:13 crc kubenswrapper[4790]: I1124 16:05:13.508166 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-q2m9q_4da891b8-efa8-4c31-921a-878c9b78c8a6/cert-manager-controller/0.log" Nov 24 16:05:13 crc kubenswrapper[4790]: I1124 16:05:13.689945 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-gz7ct_462f7ce3-f481-4d1b-b466-ffdbc9038f3a/cert-manager-cainjector/0.log" Nov 24 16:05:13 crc kubenswrapper[4790]: I1124 16:05:13.776570 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-dm8pw_cdbadd01-32e5-48d1-bf16-68939d60bd02/cert-manager-webhook/0.log" Nov 24 16:05:22 crc kubenswrapper[4790]: I1124 16:05:22.344720 4790 scope.go:117] "RemoveContainer" containerID="04f0f5e9e4cbc0de998b73d51f7a42d9e018f4e80bf0907bef135486c6f1967b" Nov 24 16:05:22 crc kubenswrapper[4790]: E1124 16:05:22.346018 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 16:05:28 crc kubenswrapper[4790]: I1124 16:05:28.211893 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-l2ltp_e6ee737d-a089-4ecb-a3a8-34ce580b84aa/nmstate-console-plugin/0.log" Nov 24 16:05:28 crc kubenswrapper[4790]: I1124 16:05:28.328349 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-kbn22_da4996e7-f19c-47e9-a75c-000e4248dc2d/nmstate-handler/0.log" Nov 24 16:05:28 crc kubenswrapper[4790]: I1124 16:05:28.395125 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-xg8tt_d9e69cb8-ffab-4a0d-be5d-e7d683a44c87/kube-rbac-proxy/0.log" Nov 24 16:05:28 crc kubenswrapper[4790]: I1124 16:05:28.473028 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-xg8tt_d9e69cb8-ffab-4a0d-be5d-e7d683a44c87/nmstate-metrics/0.log" Nov 24 16:05:28 crc kubenswrapper[4790]: I1124 16:05:28.578669 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-kshj8_9f0af7f5-b65a-47dc-8c85-c452cb493912/nmstate-operator/0.log" Nov 24 16:05:28 crc kubenswrapper[4790]: I1124 16:05:28.793684 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-ggp8t_c71e7468-3526-41e1-87bd-09809fe6ae07/nmstate-webhook/0.log" Nov 24 16:05:35 crc kubenswrapper[4790]: I1124 16:05:35.314623 4790 scope.go:117] "RemoveContainer" containerID="04f0f5e9e4cbc0de998b73d51f7a42d9e018f4e80bf0907bef135486c6f1967b" Nov 24 16:05:35 crc kubenswrapper[4790]: E1124 16:05:35.315386 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 16:05:46 crc kubenswrapper[4790]: I1124 16:05:46.243560 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-fx9l5_20b558ce-130d-4100-af6e-d2f4d58cafe3/kube-rbac-proxy/0.log" Nov 24 16:05:46 crc kubenswrapper[4790]: I1124 16:05:46.601693 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-fx9l5_20b558ce-130d-4100-af6e-d2f4d58cafe3/controller/0.log" Nov 24 16:05:46 crc kubenswrapper[4790]: I1124 16:05:46.873498 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bvwnn_8ddb656f-ec21-4a37-a89c-3c02fdda39e2/cp-frr-files/0.log" Nov 24 16:05:47 crc kubenswrapper[4790]: I1124 16:05:47.059622 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bvwnn_8ddb656f-ec21-4a37-a89c-3c02fdda39e2/cp-metrics/0.log" Nov 24 16:05:47 crc kubenswrapper[4790]: I1124 16:05:47.061181 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bvwnn_8ddb656f-ec21-4a37-a89c-3c02fdda39e2/cp-reloader/0.log" Nov 24 16:05:47 crc kubenswrapper[4790]: I1124 16:05:47.069674 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bvwnn_8ddb656f-ec21-4a37-a89c-3c02fdda39e2/cp-reloader/0.log" Nov 24 16:05:47 crc kubenswrapper[4790]: I1124 16:05:47.077713 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bvwnn_8ddb656f-ec21-4a37-a89c-3c02fdda39e2/cp-frr-files/0.log" Nov 24 16:05:47 crc kubenswrapper[4790]: I1124 16:05:47.229470 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bvwnn_8ddb656f-ec21-4a37-a89c-3c02fdda39e2/cp-frr-files/0.log" Nov 24 16:05:47 crc kubenswrapper[4790]: I1124 16:05:47.276202 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bvwnn_8ddb656f-ec21-4a37-a89c-3c02fdda39e2/cp-reloader/0.log" Nov 24 16:05:47 crc kubenswrapper[4790]: I1124 16:05:47.277711 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bvwnn_8ddb656f-ec21-4a37-a89c-3c02fdda39e2/cp-metrics/0.log" Nov 24 16:05:47 crc kubenswrapper[4790]: I1124 16:05:47.314901 4790 scope.go:117] "RemoveContainer" containerID="04f0f5e9e4cbc0de998b73d51f7a42d9e018f4e80bf0907bef135486c6f1967b" Nov 24 16:05:47 crc kubenswrapper[4790]: E1124 16:05:47.315152 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 16:05:47 crc kubenswrapper[4790]: I1124 16:05:47.336539 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bvwnn_8ddb656f-ec21-4a37-a89c-3c02fdda39e2/cp-metrics/0.log" Nov 24 16:05:47 crc kubenswrapper[4790]: I1124 16:05:47.417439 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bvwnn_8ddb656f-ec21-4a37-a89c-3c02fdda39e2/cp-frr-files/0.log" Nov 24 16:05:47 crc kubenswrapper[4790]: I1124 16:05:47.469590 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bvwnn_8ddb656f-ec21-4a37-a89c-3c02fdda39e2/cp-metrics/0.log" Nov 24 16:05:47 crc kubenswrapper[4790]: I1124 16:05:47.485280 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bvwnn_8ddb656f-ec21-4a37-a89c-3c02fdda39e2/cp-reloader/0.log" Nov 24 16:05:47 crc kubenswrapper[4790]: I1124 16:05:47.547053 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bvwnn_8ddb656f-ec21-4a37-a89c-3c02fdda39e2/controller/0.log" Nov 24 16:05:47 crc kubenswrapper[4790]: I1124 16:05:47.668717 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bvwnn_8ddb656f-ec21-4a37-a89c-3c02fdda39e2/frr-metrics/0.log" Nov 24 16:05:47 crc kubenswrapper[4790]: I1124 16:05:47.732491 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bvwnn_8ddb656f-ec21-4a37-a89c-3c02fdda39e2/kube-rbac-proxy/0.log" Nov 24 16:05:47 crc kubenswrapper[4790]: I1124 16:05:47.759014 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bvwnn_8ddb656f-ec21-4a37-a89c-3c02fdda39e2/kube-rbac-proxy-frr/0.log" Nov 24 16:05:47 crc kubenswrapper[4790]: I1124 16:05:47.896711 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bvwnn_8ddb656f-ec21-4a37-a89c-3c02fdda39e2/reloader/0.log" Nov 24 16:05:47 crc kubenswrapper[4790]: I1124 16:05:47.949591 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-sdghv_47d9b626-df79-437b-a1af-5c584ff33f11/frr-k8s-webhook-server/0.log" Nov 24 16:05:48 crc kubenswrapper[4790]: I1124 16:05:48.236802 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-768d985d78-np9x5_79948d0b-b705-4c1b-8d13-836359599770/manager/0.log" Nov 24 16:05:48 crc kubenswrapper[4790]: I1124 16:05:48.300760 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-78c98cf574-zvq6f_1bdc2295-5ae4-4d34-969d-f0a5251ab904/webhook-server/0.log" Nov 24 16:05:48 crc kubenswrapper[4790]: I1124 16:05:48.480760 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-nglb4_ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d/kube-rbac-proxy/0.log" Nov 24 16:05:49 crc kubenswrapper[4790]: I1124 16:05:49.425272 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-nglb4_ddc6a4d3-6419-4d05-a3a1-4dc3b3f79e5d/speaker/0.log" Nov 24 16:05:50 crc kubenswrapper[4790]: I1124 16:05:50.942660 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bvwnn_8ddb656f-ec21-4a37-a89c-3c02fdda39e2/frr/0.log" Nov 24 16:05:58 crc kubenswrapper[4790]: I1124 16:05:58.313950 4790 scope.go:117] "RemoveContainer" containerID="04f0f5e9e4cbc0de998b73d51f7a42d9e018f4e80bf0907bef135486c6f1967b" Nov 24 16:05:58 crc kubenswrapper[4790]: E1124 16:05:58.314739 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 16:06:00 crc kubenswrapper[4790]: I1124 16:06:00.195543 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qw9kh"] Nov 24 16:06:00 crc kubenswrapper[4790]: E1124 16:06:00.196468 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d4b26c0-b10e-43d4-8203-0f7244e1325f" containerName="container-00" Nov 24 16:06:00 crc kubenswrapper[4790]: I1124 16:06:00.196484 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d4b26c0-b10e-43d4-8203-0f7244e1325f" containerName="container-00" Nov 24 16:06:00 crc kubenswrapper[4790]: I1124 16:06:00.196734 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d4b26c0-b10e-43d4-8203-0f7244e1325f" containerName="container-00" Nov 24 16:06:00 crc kubenswrapper[4790]: I1124 16:06:00.198920 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qw9kh" Nov 24 16:06:00 crc kubenswrapper[4790]: I1124 16:06:00.215058 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qw9kh"] Nov 24 16:06:00 crc kubenswrapper[4790]: I1124 16:06:00.320396 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a49772bf-6801-49ef-931e-8c40e4a759c3-utilities\") pod \"certified-operators-qw9kh\" (UID: \"a49772bf-6801-49ef-931e-8c40e4a759c3\") " pod="openshift-marketplace/certified-operators-qw9kh" Nov 24 16:06:00 crc kubenswrapper[4790]: I1124 16:06:00.320550 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a49772bf-6801-49ef-931e-8c40e4a759c3-catalog-content\") pod \"certified-operators-qw9kh\" (UID: \"a49772bf-6801-49ef-931e-8c40e4a759c3\") " pod="openshift-marketplace/certified-operators-qw9kh" Nov 24 16:06:00 crc kubenswrapper[4790]: I1124 16:06:00.320588 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdbcv\" (UniqueName: \"kubernetes.io/projected/a49772bf-6801-49ef-931e-8c40e4a759c3-kube-api-access-vdbcv\") pod \"certified-operators-qw9kh\" (UID: \"a49772bf-6801-49ef-931e-8c40e4a759c3\") " pod="openshift-marketplace/certified-operators-qw9kh" Nov 24 16:06:00 crc kubenswrapper[4790]: I1124 16:06:00.423238 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a49772bf-6801-49ef-931e-8c40e4a759c3-utilities\") pod \"certified-operators-qw9kh\" (UID: \"a49772bf-6801-49ef-931e-8c40e4a759c3\") " pod="openshift-marketplace/certified-operators-qw9kh" Nov 24 16:06:00 crc kubenswrapper[4790]: I1124 16:06:00.423665 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a49772bf-6801-49ef-931e-8c40e4a759c3-catalog-content\") pod \"certified-operators-qw9kh\" (UID: \"a49772bf-6801-49ef-931e-8c40e4a759c3\") " pod="openshift-marketplace/certified-operators-qw9kh" Nov 24 16:06:00 crc kubenswrapper[4790]: I1124 16:06:00.423807 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdbcv\" (UniqueName: \"kubernetes.io/projected/a49772bf-6801-49ef-931e-8c40e4a759c3-kube-api-access-vdbcv\") pod \"certified-operators-qw9kh\" (UID: \"a49772bf-6801-49ef-931e-8c40e4a759c3\") " pod="openshift-marketplace/certified-operators-qw9kh" Nov 24 16:06:00 crc kubenswrapper[4790]: I1124 16:06:00.423820 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a49772bf-6801-49ef-931e-8c40e4a759c3-utilities\") pod \"certified-operators-qw9kh\" (UID: \"a49772bf-6801-49ef-931e-8c40e4a759c3\") " pod="openshift-marketplace/certified-operators-qw9kh" Nov 24 16:06:00 crc kubenswrapper[4790]: I1124 16:06:00.424174 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a49772bf-6801-49ef-931e-8c40e4a759c3-catalog-content\") pod \"certified-operators-qw9kh\" (UID: \"a49772bf-6801-49ef-931e-8c40e4a759c3\") " pod="openshift-marketplace/certified-operators-qw9kh" Nov 24 16:06:00 crc kubenswrapper[4790]: I1124 16:06:00.452641 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdbcv\" (UniqueName: \"kubernetes.io/projected/a49772bf-6801-49ef-931e-8c40e4a759c3-kube-api-access-vdbcv\") pod \"certified-operators-qw9kh\" (UID: \"a49772bf-6801-49ef-931e-8c40e4a759c3\") " pod="openshift-marketplace/certified-operators-qw9kh" Nov 24 16:06:00 crc kubenswrapper[4790]: I1124 16:06:00.563507 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qw9kh" Nov 24 16:06:01 crc kubenswrapper[4790]: W1124 16:06:01.117859 4790 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda49772bf_6801_49ef_931e_8c40e4a759c3.slice/crio-2eb549ff410fb2e05509b845cc0d3e2be6183fa3a72e5b468781be172bcc6f87 WatchSource:0}: Error finding container 2eb549ff410fb2e05509b845cc0d3e2be6183fa3a72e5b468781be172bcc6f87: Status 404 returned error can't find the container with id 2eb549ff410fb2e05509b845cc0d3e2be6183fa3a72e5b468781be172bcc6f87 Nov 24 16:06:01 crc kubenswrapper[4790]: I1124 16:06:01.127730 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qw9kh"] Nov 24 16:06:01 crc kubenswrapper[4790]: I1124 16:06:01.259929 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qw9kh" event={"ID":"a49772bf-6801-49ef-931e-8c40e4a759c3","Type":"ContainerStarted","Data":"2eb549ff410fb2e05509b845cc0d3e2be6183fa3a72e5b468781be172bcc6f87"} Nov 24 16:06:02 crc kubenswrapper[4790]: I1124 16:06:02.275144 4790 generic.go:334] "Generic (PLEG): container finished" podID="a49772bf-6801-49ef-931e-8c40e4a759c3" containerID="ada5896fbc3132da70993b7960c0dd6bae357cb225206bd5e6cb0f4ad285335f" exitCode=0 Nov 24 16:06:02 crc kubenswrapper[4790]: I1124 16:06:02.275276 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qw9kh" event={"ID":"a49772bf-6801-49ef-931e-8c40e4a759c3","Type":"ContainerDied","Data":"ada5896fbc3132da70993b7960c0dd6bae357cb225206bd5e6cb0f4ad285335f"} Nov 24 16:06:02 crc kubenswrapper[4790]: I1124 16:06:02.278305 4790 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 16:06:04 crc kubenswrapper[4790]: I1124 16:06:04.297261 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qw9kh" event={"ID":"a49772bf-6801-49ef-931e-8c40e4a759c3","Type":"ContainerStarted","Data":"0821fc3c9e1b9acbb0dff07bc36f5a18dfb0679dff51438dff6ab3a85af9c01d"} Nov 24 16:06:04 crc kubenswrapper[4790]: I1124 16:06:04.775050 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9_47d9c197-a7d1-4977-b580-508aa507f368/util/0.log" Nov 24 16:06:04 crc kubenswrapper[4790]: I1124 16:06:04.999091 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9_47d9c197-a7d1-4977-b580-508aa507f368/pull/0.log" Nov 24 16:06:05 crc kubenswrapper[4790]: I1124 16:06:05.022333 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9_47d9c197-a7d1-4977-b580-508aa507f368/util/0.log" Nov 24 16:06:05 crc kubenswrapper[4790]: I1124 16:06:05.081355 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9_47d9c197-a7d1-4977-b580-508aa507f368/pull/0.log" Nov 24 16:06:05 crc kubenswrapper[4790]: I1124 16:06:05.210656 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9_47d9c197-a7d1-4977-b580-508aa507f368/pull/0.log" Nov 24 16:06:05 crc kubenswrapper[4790]: I1124 16:06:05.230007 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9_47d9c197-a7d1-4977-b580-508aa507f368/util/0.log" Nov 24 16:06:05 crc kubenswrapper[4790]: I1124 16:06:05.242483 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abcsz9_47d9c197-a7d1-4977-b580-508aa507f368/extract/0.log" Nov 24 16:06:05 crc kubenswrapper[4790]: I1124 16:06:05.310318 4790 generic.go:334] "Generic (PLEG): container finished" podID="a49772bf-6801-49ef-931e-8c40e4a759c3" containerID="0821fc3c9e1b9acbb0dff07bc36f5a18dfb0679dff51438dff6ab3a85af9c01d" exitCode=0 Nov 24 16:06:05 crc kubenswrapper[4790]: I1124 16:06:05.310360 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qw9kh" event={"ID":"a49772bf-6801-49ef-931e-8c40e4a759c3","Type":"ContainerDied","Data":"0821fc3c9e1b9acbb0dff07bc36f5a18dfb0679dff51438dff6ab3a85af9c01d"} Nov 24 16:06:05 crc kubenswrapper[4790]: I1124 16:06:05.395187 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl_a55d437e-ba5a-4e62-8bea-a9345a428e73/util/0.log" Nov 24 16:06:05 crc kubenswrapper[4790]: I1124 16:06:05.591376 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl_a55d437e-ba5a-4e62-8bea-a9345a428e73/pull/0.log" Nov 24 16:06:05 crc kubenswrapper[4790]: I1124 16:06:05.591524 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl_a55d437e-ba5a-4e62-8bea-a9345a428e73/pull/0.log" Nov 24 16:06:05 crc kubenswrapper[4790]: I1124 16:06:05.619775 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl_a55d437e-ba5a-4e62-8bea-a9345a428e73/util/0.log" Nov 24 16:06:05 crc kubenswrapper[4790]: I1124 16:06:05.801978 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl_a55d437e-ba5a-4e62-8bea-a9345a428e73/util/0.log" Nov 24 16:06:05 crc kubenswrapper[4790]: I1124 16:06:05.812550 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl_a55d437e-ba5a-4e62-8bea-a9345a428e73/extract/0.log" Nov 24 16:06:05 crc kubenswrapper[4790]: I1124 16:06:05.896147 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eg62hl_a55d437e-ba5a-4e62-8bea-a9345a428e73/pull/0.log" Nov 24 16:06:06 crc kubenswrapper[4790]: I1124 16:06:06.070851 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8_9dd5c102-10b0-4a96-8a12-6e1ca4f63c30/util/0.log" Nov 24 16:06:06 crc kubenswrapper[4790]: I1124 16:06:06.207745 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8_9dd5c102-10b0-4a96-8a12-6e1ca4f63c30/pull/0.log" Nov 24 16:06:06 crc kubenswrapper[4790]: I1124 16:06:06.229624 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8_9dd5c102-10b0-4a96-8a12-6e1ca4f63c30/pull/0.log" Nov 24 16:06:06 crc kubenswrapper[4790]: I1124 16:06:06.326386 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qw9kh" event={"ID":"a49772bf-6801-49ef-931e-8c40e4a759c3","Type":"ContainerStarted","Data":"792e54d28308691266195e51fa9b9cef76103d0d152d87e21e3f536ee234fd0f"} Nov 24 16:06:06 crc kubenswrapper[4790]: I1124 16:06:06.345505 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qw9kh" podStartSLOduration=2.916571231 podStartE2EDuration="6.34548551s" podCreationTimestamp="2025-11-24 16:06:00 +0000 UTC" firstStartedPulling="2025-11-24 16:06:02.277826057 +0000 UTC m=+10410.657719759" lastFinishedPulling="2025-11-24 16:06:05.706740376 +0000 UTC m=+10414.086634038" observedRunningTime="2025-11-24 16:06:06.340985462 +0000 UTC m=+10414.720879124" watchObservedRunningTime="2025-11-24 16:06:06.34548551 +0000 UTC m=+10414.725379182" Nov 24 16:06:06 crc kubenswrapper[4790]: I1124 16:06:06.349621 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8_9dd5c102-10b0-4a96-8a12-6e1ca4f63c30/util/0.log" Nov 24 16:06:06 crc kubenswrapper[4790]: I1124 16:06:06.416127 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8_9dd5c102-10b0-4a96-8a12-6e1ca4f63c30/util/0.log" Nov 24 16:06:06 crc kubenswrapper[4790]: I1124 16:06:06.454352 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8_9dd5c102-10b0-4a96-8a12-6e1ca4f63c30/extract/0.log" Nov 24 16:06:06 crc kubenswrapper[4790]: I1124 16:06:06.466064 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106kpr8_9dd5c102-10b0-4a96-8a12-6e1ca4f63c30/pull/0.log" Nov 24 16:06:06 crc kubenswrapper[4790]: I1124 16:06:06.622872 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gsmqf_c67690a9-9819-45bb-9d46-24ebc69923b9/extract-utilities/0.log" Nov 24 16:06:06 crc kubenswrapper[4790]: I1124 16:06:06.873023 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gsmqf_c67690a9-9819-45bb-9d46-24ebc69923b9/extract-utilities/0.log" Nov 24 16:06:06 crc kubenswrapper[4790]: I1124 16:06:06.890951 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gsmqf_c67690a9-9819-45bb-9d46-24ebc69923b9/extract-content/0.log" Nov 24 16:06:06 crc kubenswrapper[4790]: I1124 16:06:06.907701 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gsmqf_c67690a9-9819-45bb-9d46-24ebc69923b9/extract-content/0.log" Nov 24 16:06:07 crc kubenswrapper[4790]: I1124 16:06:07.168944 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gsmqf_c67690a9-9819-45bb-9d46-24ebc69923b9/extract-utilities/0.log" Nov 24 16:06:07 crc kubenswrapper[4790]: I1124 16:06:07.191248 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gsmqf_c67690a9-9819-45bb-9d46-24ebc69923b9/extract-content/0.log" Nov 24 16:06:07 crc kubenswrapper[4790]: I1124 16:06:07.862825 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gsmqf_c67690a9-9819-45bb-9d46-24ebc69923b9/registry-server/0.log" Nov 24 16:06:07 crc kubenswrapper[4790]: I1124 16:06:07.908653 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qw9kh_a49772bf-6801-49ef-931e-8c40e4a759c3/extract-utilities/0.log" Nov 24 16:06:08 crc kubenswrapper[4790]: I1124 16:06:08.120127 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qw9kh_a49772bf-6801-49ef-931e-8c40e4a759c3/extract-utilities/0.log" Nov 24 16:06:08 crc kubenswrapper[4790]: I1124 16:06:08.132361 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qw9kh_a49772bf-6801-49ef-931e-8c40e4a759c3/extract-content/0.log" Nov 24 16:06:08 crc kubenswrapper[4790]: I1124 16:06:08.149772 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qw9kh_a49772bf-6801-49ef-931e-8c40e4a759c3/extract-content/0.log" Nov 24 16:06:08 crc kubenswrapper[4790]: I1124 16:06:08.333159 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qw9kh_a49772bf-6801-49ef-931e-8c40e4a759c3/extract-content/0.log" Nov 24 16:06:08 crc kubenswrapper[4790]: I1124 16:06:08.346144 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qw9kh_a49772bf-6801-49ef-931e-8c40e4a759c3/extract-utilities/0.log" Nov 24 16:06:08 crc kubenswrapper[4790]: I1124 16:06:08.378847 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qw9kh_a49772bf-6801-49ef-931e-8c40e4a759c3/registry-server/0.log" Nov 24 16:06:08 crc kubenswrapper[4790]: I1124 16:06:08.527114 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-87dpj_b9864022-64b2-41a9-bdc3-55beb82adb57/extract-utilities/0.log" Nov 24 16:06:08 crc kubenswrapper[4790]: I1124 16:06:08.739150 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-87dpj_b9864022-64b2-41a9-bdc3-55beb82adb57/extract-content/0.log" Nov 24 16:06:08 crc kubenswrapper[4790]: I1124 16:06:08.741011 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-87dpj_b9864022-64b2-41a9-bdc3-55beb82adb57/extract-utilities/0.log" Nov 24 16:06:08 crc kubenswrapper[4790]: I1124 16:06:08.748001 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-87dpj_b9864022-64b2-41a9-bdc3-55beb82adb57/extract-content/0.log" Nov 24 16:06:08 crc kubenswrapper[4790]: I1124 16:06:08.938261 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-87dpj_b9864022-64b2-41a9-bdc3-55beb82adb57/extract-utilities/0.log" Nov 24 16:06:08 crc kubenswrapper[4790]: I1124 16:06:08.987784 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-87dpj_b9864022-64b2-41a9-bdc3-55beb82adb57/extract-content/0.log" Nov 24 16:06:09 crc kubenswrapper[4790]: I1124 16:06:09.452414 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t_7187360e-55ef-4709-a8b5-4aa49ebdf70e/util/0.log" Nov 24 16:06:09 crc kubenswrapper[4790]: I1124 16:06:09.589129 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t_7187360e-55ef-4709-a8b5-4aa49ebdf70e/pull/0.log" Nov 24 16:06:09 crc kubenswrapper[4790]: I1124 16:06:09.650814 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t_7187360e-55ef-4709-a8b5-4aa49ebdf70e/pull/0.log" Nov 24 16:06:09 crc kubenswrapper[4790]: I1124 16:06:09.656350 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t_7187360e-55ef-4709-a8b5-4aa49ebdf70e/util/0.log" Nov 24 16:06:09 crc kubenswrapper[4790]: I1124 16:06:09.942947 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t_7187360e-55ef-4709-a8b5-4aa49ebdf70e/extract/0.log" Nov 24 16:06:10 crc kubenswrapper[4790]: I1124 16:06:10.202458 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t_7187360e-55ef-4709-a8b5-4aa49ebdf70e/pull/0.log" Nov 24 16:06:10 crc kubenswrapper[4790]: I1124 16:06:10.216627 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62kp9t_7187360e-55ef-4709-a8b5-4aa49ebdf70e/util/0.log" Nov 24 16:06:10 crc kubenswrapper[4790]: I1124 16:06:10.269099 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-vj9c9_c843c1ad-11d6-4577-b337-40cdd992ae55/marketplace-operator/0.log" Nov 24 16:06:10 crc kubenswrapper[4790]: I1124 16:06:10.411680 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-r266z_92ba1be5-af31-4941-94ef-3977f17b297e/extract-utilities/0.log" Nov 24 16:06:10 crc kubenswrapper[4790]: I1124 16:06:10.541424 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-87dpj_b9864022-64b2-41a9-bdc3-55beb82adb57/registry-server/0.log" Nov 24 16:06:10 crc kubenswrapper[4790]: I1124 16:06:10.564490 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qw9kh" Nov 24 16:06:10 crc kubenswrapper[4790]: I1124 16:06:10.565257 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qw9kh" Nov 24 16:06:10 crc kubenswrapper[4790]: I1124 16:06:10.594328 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-r266z_92ba1be5-af31-4941-94ef-3977f17b297e/extract-content/0.log" Nov 24 16:06:10 crc kubenswrapper[4790]: I1124 16:06:10.613970 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qw9kh" Nov 24 16:06:10 crc kubenswrapper[4790]: I1124 16:06:10.665286 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-r266z_92ba1be5-af31-4941-94ef-3977f17b297e/extract-content/0.log" Nov 24 16:06:10 crc kubenswrapper[4790]: I1124 16:06:10.667266 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-r266z_92ba1be5-af31-4941-94ef-3977f17b297e/extract-utilities/0.log" Nov 24 16:06:10 crc kubenswrapper[4790]: I1124 16:06:10.771831 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-r266z_92ba1be5-af31-4941-94ef-3977f17b297e/extract-content/0.log" Nov 24 16:06:10 crc kubenswrapper[4790]: I1124 16:06:10.841458 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-r266z_92ba1be5-af31-4941-94ef-3977f17b297e/extract-utilities/0.log" Nov 24 16:06:10 crc kubenswrapper[4790]: I1124 16:06:10.850135 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-m5lwm_8f4eaeea-ed9a-4cfb-a09f-e5ed26fb312c/extract-utilities/0.log" Nov 24 16:06:11 crc kubenswrapper[4790]: I1124 16:06:11.038238 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-m5lwm_8f4eaeea-ed9a-4cfb-a09f-e5ed26fb312c/extract-utilities/0.log" Nov 24 16:06:11 crc kubenswrapper[4790]: I1124 16:06:11.042864 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-m5lwm_8f4eaeea-ed9a-4cfb-a09f-e5ed26fb312c/extract-content/0.log" Nov 24 16:06:11 crc kubenswrapper[4790]: I1124 16:06:11.089838 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-m5lwm_8f4eaeea-ed9a-4cfb-a09f-e5ed26fb312c/extract-content/0.log" Nov 24 16:06:11 crc kubenswrapper[4790]: I1124 16:06:11.210324 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-r266z_92ba1be5-af31-4941-94ef-3977f17b297e/registry-server/0.log" Nov 24 16:06:11 crc kubenswrapper[4790]: I1124 16:06:11.255024 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-m5lwm_8f4eaeea-ed9a-4cfb-a09f-e5ed26fb312c/extract-content/0.log" Nov 24 16:06:11 crc kubenswrapper[4790]: I1124 16:06:11.341373 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-m5lwm_8f4eaeea-ed9a-4cfb-a09f-e5ed26fb312c/extract-utilities/0.log" Nov 24 16:06:11 crc kubenswrapper[4790]: I1124 16:06:11.444374 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qw9kh" Nov 24 16:06:11 crc kubenswrapper[4790]: I1124 16:06:11.497767 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qw9kh"] Nov 24 16:06:12 crc kubenswrapper[4790]: I1124 16:06:12.212129 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-m5lwm_8f4eaeea-ed9a-4cfb-a09f-e5ed26fb312c/registry-server/0.log" Nov 24 16:06:12 crc kubenswrapper[4790]: I1124 16:06:12.325785 4790 scope.go:117] "RemoveContainer" containerID="04f0f5e9e4cbc0de998b73d51f7a42d9e018f4e80bf0907bef135486c6f1967b" Nov 24 16:06:12 crc kubenswrapper[4790]: E1124 16:06:12.326080 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 16:06:13 crc kubenswrapper[4790]: I1124 16:06:13.399243 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qw9kh" podUID="a49772bf-6801-49ef-931e-8c40e4a759c3" containerName="registry-server" containerID="cri-o://792e54d28308691266195e51fa9b9cef76103d0d152d87e21e3f536ee234fd0f" gracePeriod=2 Nov 24 16:06:13 crc kubenswrapper[4790]: I1124 16:06:13.896503 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qw9kh" Nov 24 16:06:13 crc kubenswrapper[4790]: I1124 16:06:13.994196 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a49772bf-6801-49ef-931e-8c40e4a759c3-utilities\") pod \"a49772bf-6801-49ef-931e-8c40e4a759c3\" (UID: \"a49772bf-6801-49ef-931e-8c40e4a759c3\") " Nov 24 16:06:13 crc kubenswrapper[4790]: I1124 16:06:13.994626 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a49772bf-6801-49ef-931e-8c40e4a759c3-catalog-content\") pod \"a49772bf-6801-49ef-931e-8c40e4a759c3\" (UID: \"a49772bf-6801-49ef-931e-8c40e4a759c3\") " Nov 24 16:06:13 crc kubenswrapper[4790]: I1124 16:06:13.994742 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdbcv\" (UniqueName: \"kubernetes.io/projected/a49772bf-6801-49ef-931e-8c40e4a759c3-kube-api-access-vdbcv\") pod \"a49772bf-6801-49ef-931e-8c40e4a759c3\" (UID: \"a49772bf-6801-49ef-931e-8c40e4a759c3\") " Nov 24 16:06:14 crc kubenswrapper[4790]: I1124 16:06:13.996039 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a49772bf-6801-49ef-931e-8c40e4a759c3-utilities" (OuterVolumeSpecName: "utilities") pod "a49772bf-6801-49ef-931e-8c40e4a759c3" (UID: "a49772bf-6801-49ef-931e-8c40e4a759c3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 16:06:14 crc kubenswrapper[4790]: I1124 16:06:14.007372 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a49772bf-6801-49ef-931e-8c40e4a759c3-kube-api-access-vdbcv" (OuterVolumeSpecName: "kube-api-access-vdbcv") pod "a49772bf-6801-49ef-931e-8c40e4a759c3" (UID: "a49772bf-6801-49ef-931e-8c40e4a759c3"). InnerVolumeSpecName "kube-api-access-vdbcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 16:06:14 crc kubenswrapper[4790]: I1124 16:06:14.047681 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a49772bf-6801-49ef-931e-8c40e4a759c3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a49772bf-6801-49ef-931e-8c40e4a759c3" (UID: "a49772bf-6801-49ef-931e-8c40e4a759c3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 16:06:14 crc kubenswrapper[4790]: I1124 16:06:14.097471 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdbcv\" (UniqueName: \"kubernetes.io/projected/a49772bf-6801-49ef-931e-8c40e4a759c3-kube-api-access-vdbcv\") on node \"crc\" DevicePath \"\"" Nov 24 16:06:14 crc kubenswrapper[4790]: I1124 16:06:14.097513 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a49772bf-6801-49ef-931e-8c40e4a759c3-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 16:06:14 crc kubenswrapper[4790]: I1124 16:06:14.097526 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a49772bf-6801-49ef-931e-8c40e4a759c3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 16:06:14 crc kubenswrapper[4790]: I1124 16:06:14.411314 4790 generic.go:334] "Generic (PLEG): container finished" podID="a49772bf-6801-49ef-931e-8c40e4a759c3" containerID="792e54d28308691266195e51fa9b9cef76103d0d152d87e21e3f536ee234fd0f" exitCode=0 Nov 24 16:06:14 crc kubenswrapper[4790]: I1124 16:06:14.411359 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qw9kh" event={"ID":"a49772bf-6801-49ef-931e-8c40e4a759c3","Type":"ContainerDied","Data":"792e54d28308691266195e51fa9b9cef76103d0d152d87e21e3f536ee234fd0f"} Nov 24 16:06:14 crc kubenswrapper[4790]: I1124 16:06:14.411392 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qw9kh" event={"ID":"a49772bf-6801-49ef-931e-8c40e4a759c3","Type":"ContainerDied","Data":"2eb549ff410fb2e05509b845cc0d3e2be6183fa3a72e5b468781be172bcc6f87"} Nov 24 16:06:14 crc kubenswrapper[4790]: I1124 16:06:14.411410 4790 scope.go:117] "RemoveContainer" containerID="792e54d28308691266195e51fa9b9cef76103d0d152d87e21e3f536ee234fd0f" Nov 24 16:06:14 crc kubenswrapper[4790]: I1124 16:06:14.411417 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qw9kh" Nov 24 16:06:14 crc kubenswrapper[4790]: I1124 16:06:14.443406 4790 scope.go:117] "RemoveContainer" containerID="0821fc3c9e1b9acbb0dff07bc36f5a18dfb0679dff51438dff6ab3a85af9c01d" Nov 24 16:06:14 crc kubenswrapper[4790]: I1124 16:06:14.447103 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qw9kh"] Nov 24 16:06:14 crc kubenswrapper[4790]: I1124 16:06:14.467865 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qw9kh"] Nov 24 16:06:14 crc kubenswrapper[4790]: I1124 16:06:14.472768 4790 scope.go:117] "RemoveContainer" containerID="ada5896fbc3132da70993b7960c0dd6bae357cb225206bd5e6cb0f4ad285335f" Nov 24 16:06:14 crc kubenswrapper[4790]: I1124 16:06:14.529169 4790 scope.go:117] "RemoveContainer" containerID="792e54d28308691266195e51fa9b9cef76103d0d152d87e21e3f536ee234fd0f" Nov 24 16:06:14 crc kubenswrapper[4790]: E1124 16:06:14.529579 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"792e54d28308691266195e51fa9b9cef76103d0d152d87e21e3f536ee234fd0f\": container with ID starting with 792e54d28308691266195e51fa9b9cef76103d0d152d87e21e3f536ee234fd0f not found: ID does not exist" containerID="792e54d28308691266195e51fa9b9cef76103d0d152d87e21e3f536ee234fd0f" Nov 24 16:06:14 crc kubenswrapper[4790]: I1124 16:06:14.529623 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"792e54d28308691266195e51fa9b9cef76103d0d152d87e21e3f536ee234fd0f"} err="failed to get container status \"792e54d28308691266195e51fa9b9cef76103d0d152d87e21e3f536ee234fd0f\": rpc error: code = NotFound desc = could not find container \"792e54d28308691266195e51fa9b9cef76103d0d152d87e21e3f536ee234fd0f\": container with ID starting with 792e54d28308691266195e51fa9b9cef76103d0d152d87e21e3f536ee234fd0f not found: ID does not exist" Nov 24 16:06:14 crc kubenswrapper[4790]: I1124 16:06:14.529650 4790 scope.go:117] "RemoveContainer" containerID="0821fc3c9e1b9acbb0dff07bc36f5a18dfb0679dff51438dff6ab3a85af9c01d" Nov 24 16:06:14 crc kubenswrapper[4790]: E1124 16:06:14.529938 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0821fc3c9e1b9acbb0dff07bc36f5a18dfb0679dff51438dff6ab3a85af9c01d\": container with ID starting with 0821fc3c9e1b9acbb0dff07bc36f5a18dfb0679dff51438dff6ab3a85af9c01d not found: ID does not exist" containerID="0821fc3c9e1b9acbb0dff07bc36f5a18dfb0679dff51438dff6ab3a85af9c01d" Nov 24 16:06:14 crc kubenswrapper[4790]: I1124 16:06:14.529969 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0821fc3c9e1b9acbb0dff07bc36f5a18dfb0679dff51438dff6ab3a85af9c01d"} err="failed to get container status \"0821fc3c9e1b9acbb0dff07bc36f5a18dfb0679dff51438dff6ab3a85af9c01d\": rpc error: code = NotFound desc = could not find container \"0821fc3c9e1b9acbb0dff07bc36f5a18dfb0679dff51438dff6ab3a85af9c01d\": container with ID starting with 0821fc3c9e1b9acbb0dff07bc36f5a18dfb0679dff51438dff6ab3a85af9c01d not found: ID does not exist" Nov 24 16:06:14 crc kubenswrapper[4790]: I1124 16:06:14.529991 4790 scope.go:117] "RemoveContainer" containerID="ada5896fbc3132da70993b7960c0dd6bae357cb225206bd5e6cb0f4ad285335f" Nov 24 16:06:14 crc kubenswrapper[4790]: E1124 16:06:14.530225 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ada5896fbc3132da70993b7960c0dd6bae357cb225206bd5e6cb0f4ad285335f\": container with ID starting with ada5896fbc3132da70993b7960c0dd6bae357cb225206bd5e6cb0f4ad285335f not found: ID does not exist" containerID="ada5896fbc3132da70993b7960c0dd6bae357cb225206bd5e6cb0f4ad285335f" Nov 24 16:06:14 crc kubenswrapper[4790]: I1124 16:06:14.530267 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ada5896fbc3132da70993b7960c0dd6bae357cb225206bd5e6cb0f4ad285335f"} err="failed to get container status \"ada5896fbc3132da70993b7960c0dd6bae357cb225206bd5e6cb0f4ad285335f\": rpc error: code = NotFound desc = could not find container \"ada5896fbc3132da70993b7960c0dd6bae357cb225206bd5e6cb0f4ad285335f\": container with ID starting with ada5896fbc3132da70993b7960c0dd6bae357cb225206bd5e6cb0f4ad285335f not found: ID does not exist" Nov 24 16:06:16 crc kubenswrapper[4790]: I1124 16:06:16.329396 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a49772bf-6801-49ef-931e-8c40e4a759c3" path="/var/lib/kubelet/pods/a49772bf-6801-49ef-931e-8c40e4a759c3/volumes" Nov 24 16:06:24 crc kubenswrapper[4790]: I1124 16:06:24.314585 4790 scope.go:117] "RemoveContainer" containerID="04f0f5e9e4cbc0de998b73d51f7a42d9e018f4e80bf0907bef135486c6f1967b" Nov 24 16:06:24 crc kubenswrapper[4790]: E1124 16:06:24.315263 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 16:06:25 crc kubenswrapper[4790]: I1124 16:06:25.095835 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-gqvr8_84415f57-fc06-4b4e-9ecc-f0f1d32f8ea1/prometheus-operator/0.log" Nov 24 16:06:25 crc kubenswrapper[4790]: I1124 16:06:25.291626 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6cbcb9c64b-c4f6l_bc0609a2-7ee5-4fed-b741-b9bc0f85b6bb/prometheus-operator-admission-webhook/0.log" Nov 24 16:06:25 crc kubenswrapper[4790]: I1124 16:06:25.384732 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6cbcb9c64b-jx7cm_55eaed3b-e46e-4a6e-81b1-123e1a2f7d99/prometheus-operator-admission-webhook/0.log" Nov 24 16:06:25 crc kubenswrapper[4790]: I1124 16:06:25.519014 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-t6r5w_70de2641-0cdd-4ebe-9a2b-5aa4efa44c72/operator/0.log" Nov 24 16:06:25 crc kubenswrapper[4790]: I1124 16:06:25.577210 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-ntlws_1dd8ee5b-488a-41e3-ba58-e424899180ae/perses-operator/0.log" Nov 24 16:06:36 crc kubenswrapper[4790]: I1124 16:06:36.315343 4790 scope.go:117] "RemoveContainer" containerID="04f0f5e9e4cbc0de998b73d51f7a42d9e018f4e80bf0907bef135486c6f1967b" Nov 24 16:06:36 crc kubenswrapper[4790]: E1124 16:06:36.316265 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 16:06:50 crc kubenswrapper[4790]: I1124 16:06:50.314652 4790 scope.go:117] "RemoveContainer" containerID="04f0f5e9e4cbc0de998b73d51f7a42d9e018f4e80bf0907bef135486c6f1967b" Nov 24 16:06:50 crc kubenswrapper[4790]: E1124 16:06:50.315447 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 16:07:05 crc kubenswrapper[4790]: I1124 16:07:05.314988 4790 scope.go:117] "RemoveContainer" containerID="04f0f5e9e4cbc0de998b73d51f7a42d9e018f4e80bf0907bef135486c6f1967b" Nov 24 16:07:05 crc kubenswrapper[4790]: E1124 16:07:05.316000 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 16:07:17 crc kubenswrapper[4790]: I1124 16:07:17.316024 4790 scope.go:117] "RemoveContainer" containerID="04f0f5e9e4cbc0de998b73d51f7a42d9e018f4e80bf0907bef135486c6f1967b" Nov 24 16:07:17 crc kubenswrapper[4790]: E1124 16:07:17.317258 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 16:07:29 crc kubenswrapper[4790]: I1124 16:07:29.315428 4790 scope.go:117] "RemoveContainer" containerID="04f0f5e9e4cbc0de998b73d51f7a42d9e018f4e80bf0907bef135486c6f1967b" Nov 24 16:07:29 crc kubenswrapper[4790]: E1124 16:07:29.316725 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 16:07:42 crc kubenswrapper[4790]: I1124 16:07:42.325114 4790 scope.go:117] "RemoveContainer" containerID="04f0f5e9e4cbc0de998b73d51f7a42d9e018f4e80bf0907bef135486c6f1967b" Nov 24 16:07:42 crc kubenswrapper[4790]: E1124 16:07:42.326164 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 16:07:53 crc kubenswrapper[4790]: I1124 16:07:53.314958 4790 scope.go:117] "RemoveContainer" containerID="04f0f5e9e4cbc0de998b73d51f7a42d9e018f4e80bf0907bef135486c6f1967b" Nov 24 16:07:53 crc kubenswrapper[4790]: E1124 16:07:53.315684 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 16:08:04 crc kubenswrapper[4790]: I1124 16:08:04.315801 4790 scope.go:117] "RemoveContainer" containerID="04f0f5e9e4cbc0de998b73d51f7a42d9e018f4e80bf0907bef135486c6f1967b" Nov 24 16:08:04 crc kubenswrapper[4790]: E1124 16:08:04.316964 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 16:08:18 crc kubenswrapper[4790]: I1124 16:08:18.315213 4790 scope.go:117] "RemoveContainer" containerID="04f0f5e9e4cbc0de998b73d51f7a42d9e018f4e80bf0907bef135486c6f1967b" Nov 24 16:08:18 crc kubenswrapper[4790]: E1124 16:08:18.316186 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 16:08:31 crc kubenswrapper[4790]: I1124 16:08:31.078479 4790 generic.go:334] "Generic (PLEG): container finished" podID="7ac3f1cd-6b96-4378-949e-fc4ece0e6a09" containerID="bb6a206675c4846e0b28f4384fde2dd72e125d332d379af1ef22271b9e9ceb6a" exitCode=0 Nov 24 16:08:31 crc kubenswrapper[4790]: I1124 16:08:31.078541 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cd6sz/must-gather-99rhs" event={"ID":"7ac3f1cd-6b96-4378-949e-fc4ece0e6a09","Type":"ContainerDied","Data":"bb6a206675c4846e0b28f4384fde2dd72e125d332d379af1ef22271b9e9ceb6a"} Nov 24 16:08:31 crc kubenswrapper[4790]: I1124 16:08:31.081174 4790 scope.go:117] "RemoveContainer" containerID="bb6a206675c4846e0b28f4384fde2dd72e125d332d379af1ef22271b9e9ceb6a" Nov 24 16:08:31 crc kubenswrapper[4790]: I1124 16:08:31.758626 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-cd6sz_must-gather-99rhs_7ac3f1cd-6b96-4378-949e-fc4ece0e6a09/gather/0.log" Nov 24 16:08:33 crc kubenswrapper[4790]: I1124 16:08:33.314719 4790 scope.go:117] "RemoveContainer" containerID="04f0f5e9e4cbc0de998b73d51f7a42d9e018f4e80bf0907bef135486c6f1967b" Nov 24 16:08:33 crc kubenswrapper[4790]: E1124 16:08:33.316594 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 16:08:39 crc kubenswrapper[4790]: I1124 16:08:39.944737 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-cd6sz/must-gather-99rhs"] Nov 24 16:08:39 crc kubenswrapper[4790]: I1124 16:08:39.945516 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-cd6sz/must-gather-99rhs" podUID="7ac3f1cd-6b96-4378-949e-fc4ece0e6a09" containerName="copy" containerID="cri-o://dedf4c468ab3948b7ae8302d054d763bef0474421cc2dcb428bb0320d7418a04" gracePeriod=2 Nov 24 16:08:39 crc kubenswrapper[4790]: I1124 16:08:39.957695 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-cd6sz/must-gather-99rhs"] Nov 24 16:08:40 crc kubenswrapper[4790]: I1124 16:08:40.192283 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-cd6sz_must-gather-99rhs_7ac3f1cd-6b96-4378-949e-fc4ece0e6a09/copy/0.log" Nov 24 16:08:40 crc kubenswrapper[4790]: I1124 16:08:40.199537 4790 generic.go:334] "Generic (PLEG): container finished" podID="7ac3f1cd-6b96-4378-949e-fc4ece0e6a09" containerID="dedf4c468ab3948b7ae8302d054d763bef0474421cc2dcb428bb0320d7418a04" exitCode=143 Nov 24 16:08:40 crc kubenswrapper[4790]: I1124 16:08:40.415318 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-cd6sz_must-gather-99rhs_7ac3f1cd-6b96-4378-949e-fc4ece0e6a09/copy/0.log" Nov 24 16:08:40 crc kubenswrapper[4790]: I1124 16:08:40.416177 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cd6sz/must-gather-99rhs" Nov 24 16:08:40 crc kubenswrapper[4790]: I1124 16:08:40.431641 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7ac3f1cd-6b96-4378-949e-fc4ece0e6a09-must-gather-output\") pod \"7ac3f1cd-6b96-4378-949e-fc4ece0e6a09\" (UID: \"7ac3f1cd-6b96-4378-949e-fc4ece0e6a09\") " Nov 24 16:08:40 crc kubenswrapper[4790]: I1124 16:08:40.431785 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnnh5\" (UniqueName: \"kubernetes.io/projected/7ac3f1cd-6b96-4378-949e-fc4ece0e6a09-kube-api-access-hnnh5\") pod \"7ac3f1cd-6b96-4378-949e-fc4ece0e6a09\" (UID: \"7ac3f1cd-6b96-4378-949e-fc4ece0e6a09\") " Nov 24 16:08:40 crc kubenswrapper[4790]: I1124 16:08:40.439170 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ac3f1cd-6b96-4378-949e-fc4ece0e6a09-kube-api-access-hnnh5" (OuterVolumeSpecName: "kube-api-access-hnnh5") pod "7ac3f1cd-6b96-4378-949e-fc4ece0e6a09" (UID: "7ac3f1cd-6b96-4378-949e-fc4ece0e6a09"). InnerVolumeSpecName "kube-api-access-hnnh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 16:08:40 crc kubenswrapper[4790]: I1124 16:08:40.536712 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnnh5\" (UniqueName: \"kubernetes.io/projected/7ac3f1cd-6b96-4378-949e-fc4ece0e6a09-kube-api-access-hnnh5\") on node \"crc\" DevicePath \"\"" Nov 24 16:08:40 crc kubenswrapper[4790]: I1124 16:08:40.654227 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ac3f1cd-6b96-4378-949e-fc4ece0e6a09-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "7ac3f1cd-6b96-4378-949e-fc4ece0e6a09" (UID: "7ac3f1cd-6b96-4378-949e-fc4ece0e6a09"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 16:08:40 crc kubenswrapper[4790]: I1124 16:08:40.740080 4790 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7ac3f1cd-6b96-4378-949e-fc4ece0e6a09-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 24 16:08:41 crc kubenswrapper[4790]: I1124 16:08:41.213178 4790 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-cd6sz_must-gather-99rhs_7ac3f1cd-6b96-4378-949e-fc4ece0e6a09/copy/0.log" Nov 24 16:08:41 crc kubenswrapper[4790]: I1124 16:08:41.213756 4790 scope.go:117] "RemoveContainer" containerID="dedf4c468ab3948b7ae8302d054d763bef0474421cc2dcb428bb0320d7418a04" Nov 24 16:08:41 crc kubenswrapper[4790]: I1124 16:08:41.213902 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cd6sz/must-gather-99rhs" Nov 24 16:08:41 crc kubenswrapper[4790]: I1124 16:08:41.241077 4790 scope.go:117] "RemoveContainer" containerID="bb6a206675c4846e0b28f4384fde2dd72e125d332d379af1ef22271b9e9ceb6a" Nov 24 16:08:42 crc kubenswrapper[4790]: I1124 16:08:42.328203 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ac3f1cd-6b96-4378-949e-fc4ece0e6a09" path="/var/lib/kubelet/pods/7ac3f1cd-6b96-4378-949e-fc4ece0e6a09/volumes" Nov 24 16:08:46 crc kubenswrapper[4790]: I1124 16:08:46.314781 4790 scope.go:117] "RemoveContainer" containerID="04f0f5e9e4cbc0de998b73d51f7a42d9e018f4e80bf0907bef135486c6f1967b" Nov 24 16:08:46 crc kubenswrapper[4790]: E1124 16:08:46.315923 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 16:08:59 crc kubenswrapper[4790]: I1124 16:08:59.317711 4790 scope.go:117] "RemoveContainer" containerID="04f0f5e9e4cbc0de998b73d51f7a42d9e018f4e80bf0907bef135486c6f1967b" Nov 24 16:08:59 crc kubenswrapper[4790]: E1124 16:08:59.319081 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 16:09:13 crc kubenswrapper[4790]: I1124 16:09:13.315231 4790 scope.go:117] "RemoveContainer" containerID="04f0f5e9e4cbc0de998b73d51f7a42d9e018f4e80bf0907bef135486c6f1967b" Nov 24 16:09:13 crc kubenswrapper[4790]: E1124 16:09:13.317122 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 16:09:26 crc kubenswrapper[4790]: I1124 16:09:26.317213 4790 scope.go:117] "RemoveContainer" containerID="04f0f5e9e4cbc0de998b73d51f7a42d9e018f4e80bf0907bef135486c6f1967b" Nov 24 16:09:26 crc kubenswrapper[4790]: E1124 16:09:26.318256 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 16:09:38 crc kubenswrapper[4790]: I1124 16:09:38.314854 4790 scope.go:117] "RemoveContainer" containerID="04f0f5e9e4cbc0de998b73d51f7a42d9e018f4e80bf0907bef135486c6f1967b" Nov 24 16:09:38 crc kubenswrapper[4790]: E1124 16:09:38.315801 4790 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xz49t_openshift-machine-config-operator(0d73b133-48f0-455f-8f6a-742e633f631a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" podUID="0d73b133-48f0-455f-8f6a-742e633f631a" Nov 24 16:09:49 crc kubenswrapper[4790]: I1124 16:09:49.315527 4790 scope.go:117] "RemoveContainer" containerID="04f0f5e9e4cbc0de998b73d51f7a42d9e018f4e80bf0907bef135486c6f1967b" Nov 24 16:09:50 crc kubenswrapper[4790]: I1124 16:09:50.037770 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xz49t" event={"ID":"0d73b133-48f0-455f-8f6a-742e633f631a","Type":"ContainerStarted","Data":"de8167b03759d61a5cd3dbaeccd8cf74536108f61ac886f7679b0485f998490e"} Nov 24 16:10:43 crc kubenswrapper[4790]: I1124 16:10:43.158524 4790 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-d5hf5"] Nov 24 16:10:43 crc kubenswrapper[4790]: E1124 16:10:43.160086 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ac3f1cd-6b96-4378-949e-fc4ece0e6a09" containerName="gather" Nov 24 16:10:43 crc kubenswrapper[4790]: I1124 16:10:43.160116 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ac3f1cd-6b96-4378-949e-fc4ece0e6a09" containerName="gather" Nov 24 16:10:43 crc kubenswrapper[4790]: E1124 16:10:43.160135 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a49772bf-6801-49ef-931e-8c40e4a759c3" containerName="extract-content" Nov 24 16:10:43 crc kubenswrapper[4790]: I1124 16:10:43.160148 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="a49772bf-6801-49ef-931e-8c40e4a759c3" containerName="extract-content" Nov 24 16:10:43 crc kubenswrapper[4790]: E1124 16:10:43.160182 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ac3f1cd-6b96-4378-949e-fc4ece0e6a09" containerName="copy" Nov 24 16:10:43 crc kubenswrapper[4790]: I1124 16:10:43.160195 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ac3f1cd-6b96-4378-949e-fc4ece0e6a09" containerName="copy" Nov 24 16:10:43 crc kubenswrapper[4790]: E1124 16:10:43.160227 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a49772bf-6801-49ef-931e-8c40e4a759c3" containerName="registry-server" Nov 24 16:10:43 crc kubenswrapper[4790]: I1124 16:10:43.160239 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="a49772bf-6801-49ef-931e-8c40e4a759c3" containerName="registry-server" Nov 24 16:10:43 crc kubenswrapper[4790]: E1124 16:10:43.160277 4790 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a49772bf-6801-49ef-931e-8c40e4a759c3" containerName="extract-utilities" Nov 24 16:10:43 crc kubenswrapper[4790]: I1124 16:10:43.160290 4790 state_mem.go:107] "Deleted CPUSet assignment" podUID="a49772bf-6801-49ef-931e-8c40e4a759c3" containerName="extract-utilities" Nov 24 16:10:43 crc kubenswrapper[4790]: I1124 16:10:43.160664 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ac3f1cd-6b96-4378-949e-fc4ece0e6a09" containerName="gather" Nov 24 16:10:43 crc kubenswrapper[4790]: I1124 16:10:43.160688 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ac3f1cd-6b96-4378-949e-fc4ece0e6a09" containerName="copy" Nov 24 16:10:43 crc kubenswrapper[4790]: I1124 16:10:43.160713 4790 memory_manager.go:354] "RemoveStaleState removing state" podUID="a49772bf-6801-49ef-931e-8c40e4a759c3" containerName="registry-server" Nov 24 16:10:43 crc kubenswrapper[4790]: I1124 16:10:43.163740 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d5hf5" Nov 24 16:10:43 crc kubenswrapper[4790]: I1124 16:10:43.174215 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d5hf5"] Nov 24 16:10:43 crc kubenswrapper[4790]: I1124 16:10:43.315513 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8674d73f-1c85-420f-8ba3-e36a9fbbab93-utilities\") pod \"community-operators-d5hf5\" (UID: \"8674d73f-1c85-420f-8ba3-e36a9fbbab93\") " pod="openshift-marketplace/community-operators-d5hf5" Nov 24 16:10:43 crc kubenswrapper[4790]: I1124 16:10:43.315922 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6wkj\" (UniqueName: \"kubernetes.io/projected/8674d73f-1c85-420f-8ba3-e36a9fbbab93-kube-api-access-z6wkj\") pod \"community-operators-d5hf5\" (UID: \"8674d73f-1c85-420f-8ba3-e36a9fbbab93\") " pod="openshift-marketplace/community-operators-d5hf5" Nov 24 16:10:43 crc kubenswrapper[4790]: I1124 16:10:43.316109 4790 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8674d73f-1c85-420f-8ba3-e36a9fbbab93-catalog-content\") pod \"community-operators-d5hf5\" (UID: \"8674d73f-1c85-420f-8ba3-e36a9fbbab93\") " pod="openshift-marketplace/community-operators-d5hf5" Nov 24 16:10:43 crc kubenswrapper[4790]: I1124 16:10:43.418440 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8674d73f-1c85-420f-8ba3-e36a9fbbab93-utilities\") pod \"community-operators-d5hf5\" (UID: \"8674d73f-1c85-420f-8ba3-e36a9fbbab93\") " pod="openshift-marketplace/community-operators-d5hf5" Nov 24 16:10:43 crc kubenswrapper[4790]: I1124 16:10:43.418918 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6wkj\" (UniqueName: \"kubernetes.io/projected/8674d73f-1c85-420f-8ba3-e36a9fbbab93-kube-api-access-z6wkj\") pod \"community-operators-d5hf5\" (UID: \"8674d73f-1c85-420f-8ba3-e36a9fbbab93\") " pod="openshift-marketplace/community-operators-d5hf5" Nov 24 16:10:43 crc kubenswrapper[4790]: I1124 16:10:43.418988 4790 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8674d73f-1c85-420f-8ba3-e36a9fbbab93-catalog-content\") pod \"community-operators-d5hf5\" (UID: \"8674d73f-1c85-420f-8ba3-e36a9fbbab93\") " pod="openshift-marketplace/community-operators-d5hf5" Nov 24 16:10:43 crc kubenswrapper[4790]: I1124 16:10:43.419006 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8674d73f-1c85-420f-8ba3-e36a9fbbab93-utilities\") pod \"community-operators-d5hf5\" (UID: \"8674d73f-1c85-420f-8ba3-e36a9fbbab93\") " pod="openshift-marketplace/community-operators-d5hf5" Nov 24 16:10:43 crc kubenswrapper[4790]: I1124 16:10:43.419591 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8674d73f-1c85-420f-8ba3-e36a9fbbab93-catalog-content\") pod \"community-operators-d5hf5\" (UID: \"8674d73f-1c85-420f-8ba3-e36a9fbbab93\") " pod="openshift-marketplace/community-operators-d5hf5" Nov 24 16:10:43 crc kubenswrapper[4790]: I1124 16:10:43.438832 4790 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6wkj\" (UniqueName: \"kubernetes.io/projected/8674d73f-1c85-420f-8ba3-e36a9fbbab93-kube-api-access-z6wkj\") pod \"community-operators-d5hf5\" (UID: \"8674d73f-1c85-420f-8ba3-e36a9fbbab93\") " pod="openshift-marketplace/community-operators-d5hf5" Nov 24 16:10:43 crc kubenswrapper[4790]: I1124 16:10:43.542307 4790 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d5hf5" Nov 24 16:10:44 crc kubenswrapper[4790]: I1124 16:10:44.105594 4790 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d5hf5"] Nov 24 16:10:44 crc kubenswrapper[4790]: I1124 16:10:44.778187 4790 generic.go:334] "Generic (PLEG): container finished" podID="8674d73f-1c85-420f-8ba3-e36a9fbbab93" containerID="beb0668219a705776527a0f60d6309a8182ef4ac4f28ef4c7d030743854f4c70" exitCode=0 Nov 24 16:10:44 crc kubenswrapper[4790]: I1124 16:10:44.778256 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d5hf5" event={"ID":"8674d73f-1c85-420f-8ba3-e36a9fbbab93","Type":"ContainerDied","Data":"beb0668219a705776527a0f60d6309a8182ef4ac4f28ef4c7d030743854f4c70"} Nov 24 16:10:44 crc kubenswrapper[4790]: I1124 16:10:44.778562 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d5hf5" event={"ID":"8674d73f-1c85-420f-8ba3-e36a9fbbab93","Type":"ContainerStarted","Data":"22ade85a613d5610e2d1d2a385fcfb86bab3637fa599fc823341444746d8ebb4"} Nov 24 16:10:46 crc kubenswrapper[4790]: I1124 16:10:46.807238 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d5hf5" event={"ID":"8674d73f-1c85-420f-8ba3-e36a9fbbab93","Type":"ContainerStarted","Data":"61a47afe3f7cd4bb05145293f94b989b8ea29814d903aff17f985eacc49dee61"} Nov 24 16:10:47 crc kubenswrapper[4790]: I1124 16:10:47.828071 4790 generic.go:334] "Generic (PLEG): container finished" podID="8674d73f-1c85-420f-8ba3-e36a9fbbab93" containerID="61a47afe3f7cd4bb05145293f94b989b8ea29814d903aff17f985eacc49dee61" exitCode=0 Nov 24 16:10:47 crc kubenswrapper[4790]: I1124 16:10:47.828151 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d5hf5" event={"ID":"8674d73f-1c85-420f-8ba3-e36a9fbbab93","Type":"ContainerDied","Data":"61a47afe3f7cd4bb05145293f94b989b8ea29814d903aff17f985eacc49dee61"} Nov 24 16:10:49 crc kubenswrapper[4790]: I1124 16:10:49.858285 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d5hf5" event={"ID":"8674d73f-1c85-420f-8ba3-e36a9fbbab93","Type":"ContainerStarted","Data":"407661f48f5f8400407d3d643c7e495f9ddf1fd0a0149450d8c736a9d430ea39"} Nov 24 16:10:53 crc kubenswrapper[4790]: I1124 16:10:53.543193 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-d5hf5" Nov 24 16:10:53 crc kubenswrapper[4790]: I1124 16:10:53.543873 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-d5hf5" Nov 24 16:10:53 crc kubenswrapper[4790]: I1124 16:10:53.624908 4790 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-d5hf5" Nov 24 16:10:53 crc kubenswrapper[4790]: I1124 16:10:53.655210 4790 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-d5hf5" podStartSLOduration=7.188775328 podStartE2EDuration="10.655184149s" podCreationTimestamp="2025-11-24 16:10:43 +0000 UTC" firstStartedPulling="2025-11-24 16:10:44.780120561 +0000 UTC m=+10693.160014223" lastFinishedPulling="2025-11-24 16:10:48.246529382 +0000 UTC m=+10696.626423044" observedRunningTime="2025-11-24 16:10:49.88521774 +0000 UTC m=+10698.265111422" watchObservedRunningTime="2025-11-24 16:10:53.655184149 +0000 UTC m=+10702.035077831" Nov 24 16:10:54 crc kubenswrapper[4790]: I1124 16:10:54.982248 4790 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-d5hf5" Nov 24 16:10:55 crc kubenswrapper[4790]: I1124 16:10:55.031958 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d5hf5"] Nov 24 16:10:56 crc kubenswrapper[4790]: I1124 16:10:56.963647 4790 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-d5hf5" podUID="8674d73f-1c85-420f-8ba3-e36a9fbbab93" containerName="registry-server" containerID="cri-o://407661f48f5f8400407d3d643c7e495f9ddf1fd0a0149450d8c736a9d430ea39" gracePeriod=2 Nov 24 16:10:57 crc kubenswrapper[4790]: I1124 16:10:57.502464 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d5hf5" Nov 24 16:10:57 crc kubenswrapper[4790]: I1124 16:10:57.651524 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6wkj\" (UniqueName: \"kubernetes.io/projected/8674d73f-1c85-420f-8ba3-e36a9fbbab93-kube-api-access-z6wkj\") pod \"8674d73f-1c85-420f-8ba3-e36a9fbbab93\" (UID: \"8674d73f-1c85-420f-8ba3-e36a9fbbab93\") " Nov 24 16:10:57 crc kubenswrapper[4790]: I1124 16:10:57.653543 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8674d73f-1c85-420f-8ba3-e36a9fbbab93-utilities\") pod \"8674d73f-1c85-420f-8ba3-e36a9fbbab93\" (UID: \"8674d73f-1c85-420f-8ba3-e36a9fbbab93\") " Nov 24 16:10:57 crc kubenswrapper[4790]: I1124 16:10:57.653603 4790 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8674d73f-1c85-420f-8ba3-e36a9fbbab93-catalog-content\") pod \"8674d73f-1c85-420f-8ba3-e36a9fbbab93\" (UID: \"8674d73f-1c85-420f-8ba3-e36a9fbbab93\") " Nov 24 16:10:57 crc kubenswrapper[4790]: I1124 16:10:57.654360 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8674d73f-1c85-420f-8ba3-e36a9fbbab93-utilities" (OuterVolumeSpecName: "utilities") pod "8674d73f-1c85-420f-8ba3-e36a9fbbab93" (UID: "8674d73f-1c85-420f-8ba3-e36a9fbbab93"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 16:10:57 crc kubenswrapper[4790]: I1124 16:10:57.659222 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8674d73f-1c85-420f-8ba3-e36a9fbbab93-kube-api-access-z6wkj" (OuterVolumeSpecName: "kube-api-access-z6wkj") pod "8674d73f-1c85-420f-8ba3-e36a9fbbab93" (UID: "8674d73f-1c85-420f-8ba3-e36a9fbbab93"). InnerVolumeSpecName "kube-api-access-z6wkj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 16:10:57 crc kubenswrapper[4790]: I1124 16:10:57.756697 4790 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6wkj\" (UniqueName: \"kubernetes.io/projected/8674d73f-1c85-420f-8ba3-e36a9fbbab93-kube-api-access-z6wkj\") on node \"crc\" DevicePath \"\"" Nov 24 16:10:57 crc kubenswrapper[4790]: I1124 16:10:57.757068 4790 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8674d73f-1c85-420f-8ba3-e36a9fbbab93-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 16:10:57 crc kubenswrapper[4790]: I1124 16:10:57.912128 4790 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8674d73f-1c85-420f-8ba3-e36a9fbbab93-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8674d73f-1c85-420f-8ba3-e36a9fbbab93" (UID: "8674d73f-1c85-420f-8ba3-e36a9fbbab93"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 16:10:57 crc kubenswrapper[4790]: I1124 16:10:57.961082 4790 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8674d73f-1c85-420f-8ba3-e36a9fbbab93-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 16:10:57 crc kubenswrapper[4790]: I1124 16:10:57.980108 4790 generic.go:334] "Generic (PLEG): container finished" podID="8674d73f-1c85-420f-8ba3-e36a9fbbab93" containerID="407661f48f5f8400407d3d643c7e495f9ddf1fd0a0149450d8c736a9d430ea39" exitCode=0 Nov 24 16:10:57 crc kubenswrapper[4790]: I1124 16:10:57.980160 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d5hf5" event={"ID":"8674d73f-1c85-420f-8ba3-e36a9fbbab93","Type":"ContainerDied","Data":"407661f48f5f8400407d3d643c7e495f9ddf1fd0a0149450d8c736a9d430ea39"} Nov 24 16:10:57 crc kubenswrapper[4790]: I1124 16:10:57.980175 4790 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d5hf5" Nov 24 16:10:57 crc kubenswrapper[4790]: I1124 16:10:57.980199 4790 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d5hf5" event={"ID":"8674d73f-1c85-420f-8ba3-e36a9fbbab93","Type":"ContainerDied","Data":"22ade85a613d5610e2d1d2a385fcfb86bab3637fa599fc823341444746d8ebb4"} Nov 24 16:10:57 crc kubenswrapper[4790]: I1124 16:10:57.980222 4790 scope.go:117] "RemoveContainer" containerID="407661f48f5f8400407d3d643c7e495f9ddf1fd0a0149450d8c736a9d430ea39" Nov 24 16:10:58 crc kubenswrapper[4790]: I1124 16:10:58.014035 4790 scope.go:117] "RemoveContainer" containerID="61a47afe3f7cd4bb05145293f94b989b8ea29814d903aff17f985eacc49dee61" Nov 24 16:10:58 crc kubenswrapper[4790]: I1124 16:10:58.025750 4790 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d5hf5"] Nov 24 16:10:58 crc kubenswrapper[4790]: I1124 16:10:58.039149 4790 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-d5hf5"] Nov 24 16:10:58 crc kubenswrapper[4790]: I1124 16:10:58.039901 4790 scope.go:117] "RemoveContainer" containerID="beb0668219a705776527a0f60d6309a8182ef4ac4f28ef4c7d030743854f4c70" Nov 24 16:10:58 crc kubenswrapper[4790]: I1124 16:10:58.090094 4790 scope.go:117] "RemoveContainer" containerID="407661f48f5f8400407d3d643c7e495f9ddf1fd0a0149450d8c736a9d430ea39" Nov 24 16:10:58 crc kubenswrapper[4790]: E1124 16:10:58.090581 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"407661f48f5f8400407d3d643c7e495f9ddf1fd0a0149450d8c736a9d430ea39\": container with ID starting with 407661f48f5f8400407d3d643c7e495f9ddf1fd0a0149450d8c736a9d430ea39 not found: ID does not exist" containerID="407661f48f5f8400407d3d643c7e495f9ddf1fd0a0149450d8c736a9d430ea39" Nov 24 16:10:58 crc kubenswrapper[4790]: I1124 16:10:58.090659 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"407661f48f5f8400407d3d643c7e495f9ddf1fd0a0149450d8c736a9d430ea39"} err="failed to get container status \"407661f48f5f8400407d3d643c7e495f9ddf1fd0a0149450d8c736a9d430ea39\": rpc error: code = NotFound desc = could not find container \"407661f48f5f8400407d3d643c7e495f9ddf1fd0a0149450d8c736a9d430ea39\": container with ID starting with 407661f48f5f8400407d3d643c7e495f9ddf1fd0a0149450d8c736a9d430ea39 not found: ID does not exist" Nov 24 16:10:58 crc kubenswrapper[4790]: I1124 16:10:58.090723 4790 scope.go:117] "RemoveContainer" containerID="61a47afe3f7cd4bb05145293f94b989b8ea29814d903aff17f985eacc49dee61" Nov 24 16:10:58 crc kubenswrapper[4790]: E1124 16:10:58.091187 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61a47afe3f7cd4bb05145293f94b989b8ea29814d903aff17f985eacc49dee61\": container with ID starting with 61a47afe3f7cd4bb05145293f94b989b8ea29814d903aff17f985eacc49dee61 not found: ID does not exist" containerID="61a47afe3f7cd4bb05145293f94b989b8ea29814d903aff17f985eacc49dee61" Nov 24 16:10:58 crc kubenswrapper[4790]: I1124 16:10:58.091231 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61a47afe3f7cd4bb05145293f94b989b8ea29814d903aff17f985eacc49dee61"} err="failed to get container status \"61a47afe3f7cd4bb05145293f94b989b8ea29814d903aff17f985eacc49dee61\": rpc error: code = NotFound desc = could not find container \"61a47afe3f7cd4bb05145293f94b989b8ea29814d903aff17f985eacc49dee61\": container with ID starting with 61a47afe3f7cd4bb05145293f94b989b8ea29814d903aff17f985eacc49dee61 not found: ID does not exist" Nov 24 16:10:58 crc kubenswrapper[4790]: I1124 16:10:58.091263 4790 scope.go:117] "RemoveContainer" containerID="beb0668219a705776527a0f60d6309a8182ef4ac4f28ef4c7d030743854f4c70" Nov 24 16:10:58 crc kubenswrapper[4790]: E1124 16:10:58.091633 4790 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"beb0668219a705776527a0f60d6309a8182ef4ac4f28ef4c7d030743854f4c70\": container with ID starting with beb0668219a705776527a0f60d6309a8182ef4ac4f28ef4c7d030743854f4c70 not found: ID does not exist" containerID="beb0668219a705776527a0f60d6309a8182ef4ac4f28ef4c7d030743854f4c70" Nov 24 16:10:58 crc kubenswrapper[4790]: I1124 16:10:58.091696 4790 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"beb0668219a705776527a0f60d6309a8182ef4ac4f28ef4c7d030743854f4c70"} err="failed to get container status \"beb0668219a705776527a0f60d6309a8182ef4ac4f28ef4c7d030743854f4c70\": rpc error: code = NotFound desc = could not find container \"beb0668219a705776527a0f60d6309a8182ef4ac4f28ef4c7d030743854f4c70\": container with ID starting with beb0668219a705776527a0f60d6309a8182ef4ac4f28ef4c7d030743854f4c70 not found: ID does not exist" Nov 24 16:10:58 crc kubenswrapper[4790]: I1124 16:10:58.325709 4790 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8674d73f-1c85-420f-8ba3-e36a9fbbab93" path="/var/lib/kubelet/pods/8674d73f-1c85-420f-8ba3-e36a9fbbab93/volumes" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515111101635024437 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015111101636017355 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015111054341016500 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015111054341015450 5ustar corecore